p_start
phys_addr_t p_start, p_end;
for_each_mem_range(i, &p_start, &p_end) {
void *start = (void *)__va(p_start);
#define for_each_physmem_online_range(i, p_start, p_end) \
for (i = 0; !__get_physmem_range(i, p_start, p_end, false); i++)
#define for_each_physmem_reserved_type_range(t, range, p_start, p_end) \
for (range = &physmem_info.reserved[t], *p_start = range->start, *p_end = range->end; \
*p_start = range ? range->start : 0, *p_end = range ? range->end : 0)
#define for_each_physmem_reserved_range(t, range, p_start, p_end) \
*p_start = range ? range->start : 0, *p_end = range ? range->end : 0; \
*p_start = range ? range->start : 0, *p_end = range ? range->end : 0)
#define for_each_physmem_usable_range(i, p_start, p_end) \
for (i = 0; !__get_physmem_range(i, p_start, p_end, true); i++)
static void __meminitdata *p_start, *p_end;
if (p_start)
addr_start, addr_end-1, p_start, p_end-1, node_start);
p_start = p;
if (p_start) {
addr_start, addr_end-1, p_start, p_end-1, node_start);
p_start = NULL;
u64 p_start = vmm->managed.p.addr;
u64 p_end = p_start + vmm->managed.p.size;
if (start >= p_start && end <= p_end)
u32 *p_start,
*(p_start++) =
u32 *p_start,
u32 val = *p_start;
*p_start++ = 0;
u32 __iomem *p_start,
u32 val = in_be32(p_start);
(in_be32(p_start) &
p_start++;
u_short p_start ; /* first bit pos */
phy->r_val[plc->p_start+i] = n & 1 ;
plc->p_start = 0 ;
plc->p_start = 3 ;
plc->p_start = 4 ;
plc->p_start = 7 ;
plc->p_start = 7 ;
plc->p_start = 8 ;
long *p_start = (long *) &(sba_dev->ioc[i].res_map[idx_start]);
while (p_start < p_end)
*p_start++ = -1;
cpfile, periods[i].p_start, periods[i].p_end);
#define for_each_physmem_range(i, type, p_start, p_end) \
for (i = 0, __next_physmem_range(&i, type, p_start, p_end); \
__next_physmem_range(&i, type, p_start, p_end))
p_start, p_end, p_nid) \
p_start, p_end, p_nid); \
p_start, p_end, p_nid))
p_start, p_end, p_nid) \
p_start, p_end, p_nid); \
p_start, p_end, p_nid))
#define for_each_mem_range(i, p_start, p_end) \
p_start, p_end, NULL)
#define for_each_mem_range_rev(i, p_start, p_end) \
p_start, p_end, NULL)
#define for_each_reserved_mem_range(i, p_start, p_end) \
MEMBLOCK_NONE, p_start, p_end, NULL)
#define for_each_mem_pfn_range(i, nid, p_start, p_end, p_nid) \
for (i = -1, __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid); \
i >= 0; __next_mem_pfn_range(&i, nid, p_start, p_end, p_nid))
#define for_each_free_mem_range(i, nid, flags, p_start, p_end, p_nid) \
nid, flags, p_start, p_end, p_nid)
#define for_each_free_mem_range_reverse(i, nid, flags, p_start, p_end, \
nid, flags, p_start, p_end, p_nid)
__u64 p_start;
unsigned long long start, end, p_start, p_end;
p_start = mstart;
if (p_start > end)
if (p_start < start)
p_start = start;
if (p_start == start && p_end == end) {
} else if (p_start > start && p_end < end) {
mem->ranges[i].end = p_start - 1;
} else if (p_start != start)
mem->ranges[i].end = p_start - 1;
.start = p_start,
unsigned long p_end, p_start = mas->min;
gap = p_end - p_start + 1;
if (gap > p_end - p_start + 1) {
mas_mn(mas), i, gap, p_end, p_start,
p_end - p_start + 1);
MT_BUG_ON(mas->tree, gap > p_end - p_start + 1);
p_start = p_end + 1;
phys_addr_t p_start, p_end;
for_each_reserved_mem_range(loop, &p_start, &p_end)
kmsan_record_future_shadow_range(phys_to_virt(p_start),
const phys_addr_t *p_start, *p_size;
p_start = fdt_getprop(fdt, offset, "start", &len_start);
if (!p_start || len_start != sizeof(*p_start) || !p_size ||
if (*p_start & (align - 1)) {
name, (long)align, (long)*p_start);
reserved_mem_add(*p_start, size, name);
.start = p_start,
u32 *p_start = (u32 *)spdifrx->dmab->area;
u32 *p_end = p_start + (2 * SPDIFRX_CS_BYTES_NB) - 1;
u32 *ptr = p_start;