map_range
map_range(&ptep, fdt, (u64)_text > fdt ? min((u64)_text, efdt) : efdt,
map_range(pgd, ((u64)start + va_offset) & ~PAGE_OFFSET,
map_range(&ptep, (u64)_stext, (u64)__initdata_begin,
map_range(&ptep, (u64)__initdata_begin, (u64)_end,
map_range(pte, start, next, pa, prot, level + 1,
void map_range(phys_addr_t *pte, u64 start, u64 end, phys_addr_t pa,
static int __meminit save_mr(struct map_range *mr, int nr_range,
static void __ref adjust_range_page_size_mask(struct map_range *mr,
static const char *page_size_string(struct map_range *mr)
static int __meminit split_mem_range(struct map_range *mr, int nr_range,
(nr_range - 1 - i) * sizeof(struct map_range));
struct map_range mr[NR_RANGE_MR];
map_range(&pfn_mapped[i]);
struct xe_svm_range *range = op->map_range.range;
struct xe_svm_range *range = op->map_range.range;
xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(op->map_range.vma));
xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(op->map_range.vma));
op->map_range.vma,
op->map_range.range);
range_present_and_invalidated_tile(vm, op->map_range.range, tile->id);
op->map_range.vma = vma;
op->map_range.range = range;
struct xe_vma_op_map_range map_range;
ret = map_range(event->rb, vma);