Symbol: region
arch/arm/mm/pmsa-v7.c
146
static bool __init try_split_region(phys_addr_t base, phys_addr_t size, struct region *region)
arch/arm/mm/pmsa-v7.c
177
_set_bit(i, &region->subreg);
arch/arm/mm/pmsa-v7.c
180
_set_bit(PMSAv7_NR_SUBREGS - i, &region->subreg);
arch/arm/mm/pmsa-v7.c
183
region->base = abase;
arch/arm/mm/pmsa-v7.c
184
region->size = p2size;
arch/arm/mm/pmsa-v7.c
190
unsigned int limit, struct region *regions)
arch/arm/mm/pmsa-v7.c
25
static struct region __initdata mem[MPU_MAX_REGIONS];
arch/arm/mm/pmsa-v7.c
27
static struct region __initdata xip[MPU_MAX_REGIONS];
arch/arm/mm/pmsa-v7.c
422
int i, region = 0, err = 0;
arch/arm/mm/pmsa-v7.c
427
err |= mpu_setup_region(region++, 0, 32,
arch/arm/mm/pmsa-v7.c
441
bool need_flush = region == PMSAv7_RAM_REGION;
arch/arm/mm/pmsa-v7.c
446
err |= mpu_setup_region(region++, xip[i].base, ilog2(xip[i].size),
arch/arm/mm/pmsa-v7.c
457
err |= mpu_setup_region(region++, mem[i].base, ilog2(mem[i].size),
arch/arm/mm/pmsa-v7.c
464
err |= mpu_setup_region(region++, vectors_base, ilog2(2 * PAGE_SIZE),
arch/arm/mm/pmsa-v8.c
242
int region = PMSAv8_KERNEL_REGION;
arch/arm/mm/pmsa-v8.c
280
err |= pmsav8_setup_fixed(region++, __pa(KERNEL_START), __pa(KERNEL_END));
arch/arm/mm/pmsa-v8.c
288
err |= pmsav8_setup_io(region++, io[i].start, io[i].end);
arch/arm/mm/pmsa-v8.c
296
err |= pmsav8_setup_ram(region++, mem[i].start, mem[i].end);
arch/arm/mm/pmsa-v8.c
301
err |= pmsav8_setup_vector(region++, vectors_base, vectors_base + 2 * PAGE_SIZE);
arch/arm64/kernel/acpi.c
297
efi_memory_desc_t *md, *region = NULL;
arch/arm64/kernel/acpi.c
313
region = md;
arch/arm64/kernel/acpi.c
324
if (region) {
arch/arm64/kernel/acpi.c
325
switch (region->type) {
arch/arm64/kernel/acpi.c
372
if (region->attribute & EFI_MEMORY_WB)
arch/arm64/kernel/acpi.c
374
else if (region->attribute & EFI_MEMORY_WC)
arch/arm64/kernel/acpi.c
376
else if (region->attribute & EFI_MEMORY_WT)
arch/arm64/kernel/alternative.c
142
static int __apply_alternatives(const struct alt_region *region,
arch/arm64/kernel/alternative.c
150
for (alt = region->begin; alt < region->end; alt++) {
arch/arm64/kernel/alternative.c
205
struct alt_region region;
arch/arm64/kernel/alternative.c
219
region = (struct alt_region){
arch/arm64/kernel/alternative.c
224
__apply_alternatives(&region, false, &all_capabilities[0]);
arch/arm64/kernel/alternative.c
287
struct alt_region region = {
arch/arm64/kernel/alternative.c
295
return __apply_alternatives(&region, true, &all_capabilities[0]);
arch/arm64/kernel/setup.c
212
struct memblock_region *region;
arch/arm64/kernel/setup.c
228
for_each_mem_region(region) {
arch/arm64/kernel/setup.c
230
if (memblock_is_nomap(region)) {
arch/arm64/kernel/setup.c
233
res->start = __pfn_to_phys(memblock_region_reserved_base_pfn(region));
arch/arm64/kernel/setup.c
234
res->end = __pfn_to_phys(memblock_region_reserved_end_pfn(region)) - 1;
arch/arm64/kernel/setup.c
238
res->start = __pfn_to_phys(memblock_region_memory_base_pfn(region));
arch/arm64/kernel/setup.c
239
res->end = __pfn_to_phys(memblock_region_memory_end_pfn(region)) - 1;
arch/arm64/kvm/guest.c
418
static int sve_reg_to_region(struct sve_state_reg_region *region,
arch/arm64/kvm/guest.c
478
region->koffset = array_index_nospec(reqoffset, sve_state_size);
arch/arm64/kvm/guest.c
479
region->klen = min(maxlen, reqlen);
arch/arm64/kvm/guest.c
480
region->upad = reqlen - region->klen;
arch/arm64/kvm/guest.c
488
struct sve_state_reg_region region;
arch/arm64/kvm/guest.c
496
ret = sve_reg_to_region(&region, vcpu, reg);
arch/arm64/kvm/guest.c
503
if (copy_to_user(uptr, vcpu->arch.sve_state + region.koffset,
arch/arm64/kvm/guest.c
504
region.klen) ||
arch/arm64/kvm/guest.c
505
clear_user(uptr + region.klen, region.upad))
arch/arm64/kvm/guest.c
514
struct sve_state_reg_region region;
arch/arm64/kvm/guest.c
522
ret = sve_reg_to_region(&region, vcpu, reg);
arch/arm64/kvm/guest.c
529
if (copy_from_user(vcpu->arch.sve_state + region.koffset, uptr,
arch/arm64/kvm/guest.c
530
region.klen))
arch/arm64/kvm/vgic/vgic-its.c
1923
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-its.c
1932
region = vgic_find_mmio_region(its_registers,
arch/arm64/kvm/vgic/vgic-its.c
1935
if (!region)
arch/arm64/kvm/vgic/vgic-its.c
1945
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-its.c
1983
region = vgic_find_mmio_region(its_registers,
arch/arm64/kvm/vgic/vgic-its.c
1986
if (!region) {
arch/arm64/kvm/vgic/vgic-its.c
1993
len = region->access_flags & VGIC_ACCESS_64bit ? 8 : 4;
arch/arm64/kvm/vgic/vgic-its.c
1996
if (region->uaccess_its_write)
arch/arm64/kvm/vgic/vgic-its.c
1997
ret = region->uaccess_its_write(dev->kvm, its, addr,
arch/arm64/kvm/vgic/vgic-its.c
2000
region->its_write(dev->kvm, its, addr, len, *reg);
arch/arm64/kvm/vgic/vgic-its.c
2002
*reg = region->its_read(dev->kvm, its, addr, len);
arch/arm64/kvm/vgic/vgic-mmio-v2.c
523
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-mmio-v2.c
556
region = vgic_get_mmio_region(vcpu, &iodev, addr, sizeof(u32));
arch/arm64/kvm/vgic/vgic-mmio-v2.c
557
if (!region)
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1001
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1037
region = vgic_get_mmio_region(vcpu, &iodev, addr, sizeof(u32));
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1038
if (!region)
arch/arm64/kvm/vgic/vgic-mmio.c
1018
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-mmio.c
1021
region = vgic_get_mmio_region(vcpu, iodev, addr, len);
arch/arm64/kvm/vgic/vgic-mmio.c
1022
if (!region) {
arch/arm64/kvm/vgic/vgic-mmio.c
1029
data = region->read(vcpu, addr, len);
arch/arm64/kvm/vgic/vgic-mmio.c
1032
data = region->read(vcpu, addr, len);
arch/arm64/kvm/vgic/vgic-mmio.c
1035
data = region->read(iodev->redist_vcpu, addr, len);
arch/arm64/kvm/vgic/vgic-mmio.c
1038
data = region->its_read(vcpu->kvm, iodev->its, addr, len);
arch/arm64/kvm/vgic/vgic-mmio.c
1050
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-mmio.c
1053
region = vgic_get_mmio_region(vcpu, iodev, addr, len);
arch/arm64/kvm/vgic/vgic-mmio.c
1054
if (!region)
arch/arm64/kvm/vgic/vgic-mmio.c
1059
region->write(vcpu, addr, len, data);
arch/arm64/kvm/vgic/vgic-mmio.c
1062
region->write(vcpu, addr, len, data);
arch/arm64/kvm/vgic/vgic-mmio.c
1065
region->write(iodev->redist_vcpu, addr, len, data);
arch/arm64/kvm/vgic/vgic-mmio.c
1068
region->its_write(vcpu->kvm, iodev->its, addr, len, data);
arch/arm64/kvm/vgic/vgic-mmio.c
824
const struct vgic_register_region *region = elt;
arch/arm64/kvm/vgic/vgic-mmio.c
826
if (offset < region->reg_offset)
arch/arm64/kvm/vgic/vgic-mmio.c
829
if (offset >= region->reg_offset + region->len)
arch/arm64/kvm/vgic/vgic-mmio.c
919
const struct vgic_register_region *region,
arch/arm64/kvm/vgic/vgic-mmio.c
938
if ((region->access_flags & flags) && IS_ALIGNED(addr, len)) {
arch/arm64/kvm/vgic/vgic-mmio.c
939
if (!region->bits_per_irq)
arch/arm64/kvm/vgic/vgic-mmio.c
943
return VGIC_ADDR_TO_INTID(addr, region->bits_per_irq) < nr_irqs;
arch/arm64/kvm/vgic/vgic-mmio.c
953
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-mmio.c
955
region = vgic_find_mmio_region(iodev->regions, iodev->nr_regions,
arch/arm64/kvm/vgic/vgic-mmio.c
957
if (!region || !check_region(vcpu->kvm, region, addr, len))
arch/arm64/kvm/vgic/vgic-mmio.c
960
return region;
arch/arm64/kvm/vgic/vgic-mmio.c
966
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-mmio.c
969
region = vgic_get_mmio_region(vcpu, iodev, addr, sizeof(u32));
arch/arm64/kvm/vgic/vgic-mmio.c
970
if (!region) {
arch/arm64/kvm/vgic/vgic-mmio.c
976
if (region->uaccess_read)
arch/arm64/kvm/vgic/vgic-mmio.c
977
*val = region->uaccess_read(r_vcpu, addr, sizeof(u32));
arch/arm64/kvm/vgic/vgic-mmio.c
979
*val = region->read(r_vcpu, addr, sizeof(u32));
arch/arm64/kvm/vgic/vgic-mmio.c
987
const struct vgic_register_region *region;
arch/arm64/kvm/vgic/vgic-mmio.c
990
region = vgic_get_mmio_region(vcpu, iodev, addr, sizeof(u32));
arch/arm64/kvm/vgic/vgic-mmio.c
991
if (!region)
arch/arm64/kvm/vgic/vgic-mmio.c
995
if (region->uaccess_write)
arch/arm64/kvm/vgic/vgic-mmio.c
996
return region->uaccess_write(r_vcpu, addr, sizeof(u32), *val);
arch/arm64/kvm/vgic/vgic-mmio.c
998
region->write(r_vcpu, addr, sizeof(u32), *val);
arch/arm64/kvm/vgic/vgic.h
384
vgic_v3_redist_region_full(struct vgic_redist_region *region)
arch/arm64/kvm/vgic/vgic.h
386
if (!region->count)
arch/arm64/kvm/vgic/vgic.h
389
return (region->free_index >= region->count);
arch/loongarch/kernel/setup.c
426
struct memblock_region *region;
arch/loongarch/kernel/setup.c
439
for_each_mem_region(region) {
arch/loongarch/kernel/setup.c
441
if (!memblock_is_nomap(region)) {
arch/loongarch/kernel/setup.c
444
res->start = __pfn_to_phys(memblock_region_memory_base_pfn(region));
arch/loongarch/kernel/setup.c
445
res->end = __pfn_to_phys(memblock_region_memory_end_pfn(region)) - 1;
arch/loongarch/kernel/setup.c
449
res->start = __pfn_to_phys(memblock_region_reserved_base_pfn(region));
arch/loongarch/kernel/setup.c
450
res->end = __pfn_to_phys(memblock_region_reserved_end_pfn(region)) - 1;
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
21
# addresses, and need to have the appropriate memory region set
arch/mips/include/asm/sn/sn0/hubmd.h
449
#define MD_MEMORY_CONFIG_DIR_TYPE_GET(region) (\
arch/mips/include/asm/sn/sn0/hubmd.h
450
(REMOTE_HUB_L(region, MD_MEMORY_CONFIG) & MMC_DIR_PREMIUM_MASK) >> \
arch/mips/include/asm/sn/sn0/hubmd.h
459
#define MD_MIG_DIFF_THRESH_GET(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
460
REMOTE_HUB_L((region), MD_MIG_DIFF_THRESH) & \
arch/mips/include/asm/sn/sn0/hubmd.h
463
#define MD_MIG_DIFF_THRESH_SET(region, value) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
464
REMOTE_HUB_S((region), MD_MIG_DIFF_THRESH, \
arch/mips/include/asm/sn/sn0/hubmd.h
467
#define MD_MIG_DIFF_THRESH_DISABLE(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
468
REMOTE_HUB_S((region), MD_MIG_DIFF_THRESH, \
arch/mips/include/asm/sn/sn0/hubmd.h
469
REMOTE_HUB_L((region), MD_MIG_DIFF_THRESH) \
arch/mips/include/asm/sn/sn0/hubmd.h
472
#define MD_MIG_DIFF_THRESH_ENABLE(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
473
REMOTE_HUB_S((region), MD_MIG_DIFF_THRESH, \
arch/mips/include/asm/sn/sn0/hubmd.h
474
REMOTE_HUB_L((region), MD_MIG_DIFF_THRESH) \
arch/mips/include/asm/sn/sn0/hubmd.h
477
#define MD_MIG_DIFF_THRESH_IS_ENABLED(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
478
REMOTE_HUB_L((region), MD_MIG_DIFF_THRESH) & \
arch/mips/include/asm/sn/sn0/hubmd.h
481
#define MD_MIG_VALUE_THRESH_GET(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
482
REMOTE_HUB_L((region), MD_MIG_VALUE_THRESH) & \
arch/mips/include/asm/sn/sn0/hubmd.h
485
#define MD_MIG_VALUE_THRESH_SET(region, value) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
486
REMOTE_HUB_S((region), MD_MIG_VALUE_THRESH, \
arch/mips/include/asm/sn/sn0/hubmd.h
489
#define MD_MIG_VALUE_THRESH_DISABLE(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
490
REMOTE_HUB_S((region), MD_MIG_VALUE_THRESH, \
arch/mips/include/asm/sn/sn0/hubmd.h
491
REMOTE_HUB_L(region, MD_MIG_VALUE_THRESH) \
arch/mips/include/asm/sn/sn0/hubmd.h
494
#define MD_MIG_VALUE_THRESH_ENABLE(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
495
REMOTE_HUB_S((region), MD_MIG_VALUE_THRESH, \
arch/mips/include/asm/sn/sn0/hubmd.h
496
REMOTE_HUB_L((region), MD_MIG_VALUE_THRESH) \
arch/mips/include/asm/sn/sn0/hubmd.h
499
#define MD_MIG_VALUE_THRESH_IS_ENABLED(region) ( \
arch/mips/include/asm/sn/sn0/hubmd.h
500
REMOTE_HUB_L((region), MD_MIG_VALUE_THRESH) & \
arch/powerpc/include/asm/opal-api.h
1183
struct opal_mpipl_region region[];
arch/powerpc/kernel/pci-common.c
709
struct pci_bus_region region;
arch/powerpc/kernel/pci-common.c
712
pcibios_resource_to_bus(dev->bus, &region,
arch/powerpc/kernel/pci-common.c
714
*start = region.start;
arch/powerpc/kernel/pci-common.c
715
*end = region.end;
arch/powerpc/kernel/pci-common.c
938
struct pci_bus_region region;
arch/powerpc/kernel/pci-common.c
948
pcibios_resource_to_bus(dev->bus, &region, res);
arch/powerpc/kernel/pci-common.c
951
if (region.start != 0)
arch/powerpc/kernel/pci_of_scan.c
121
struct pci_bus_region region;
arch/powerpc/kernel/pci_of_scan.c
163
region.start = base;
arch/powerpc/kernel/pci_of_scan.c
164
region.end = base + size - 1;
arch/powerpc/kernel/pci_of_scan.c
165
pcibios_bus_to_resource(dev->bus, res, &region);
arch/powerpc/kernel/pci_of_scan.c
264
struct pci_bus_region region;
arch/powerpc/kernel/pci_of_scan.c
333
region.start = of_read_number(&ranges[1], 2);
arch/powerpc/kernel/pci_of_scan.c
334
region.end = region.start + size - 1;
arch/powerpc/kernel/pci_of_scan.c
335
pcibios_bus_to_resource(dev->bus, res, &region);
arch/powerpc/mm/pageattr.c
95
int region = get_region_id(addr);
arch/powerpc/mm/pageattr.c
97
if (WARN_ON_ONCE(region != VMALLOC_REGION_ID && region != IO_REGION_ID))
arch/powerpc/platforms/powermac/pci.c
1197
struct resource *region = NULL;
arch/powerpc/platforms/powermac/pci.c
1215
if (!region || resource_size(r) > resource_size(region))
arch/powerpc/platforms/powermac/pci.c
1216
region = r;
arch/powerpc/platforms/powermac/pci.c
1219
if (!region)
arch/powerpc/platforms/powermac/pci.c
1223
printk(KERN_INFO "PCI: Fixup U4 PCIe bridge range: %pR\n", region);
arch/powerpc/platforms/powermac/pci.c
1229
reg = ((region->start >> 16) & 0xfff0) | (region->end & 0xfff00000);
arch/powerpc/platforms/powernv/opal-core.c
520
be64_to_cpu(opalc_metadata->region[i].dest);
arch/powerpc/platforms/powernv/opal-core.c
522
be64_to_cpu(opalc_metadata->region[i].size);
arch/powerpc/platforms/powernv/opal-core.c
540
addr = be64_to_cpu(opalc_cpu_metadata->region[0].dest);
arch/powerpc/platforms/powernv/opal-core.c
548
be64_to_cpu(opalc_cpu_metadata->region[0].size);
arch/powerpc/platforms/powernv/opal-fadump.c
382
(u64)__va(be64_to_cpu(opal_cpu_metadata->region[0].dest));
arch/powerpc/platforms/powernv/opal-fadump.c
384
be64_to_cpu(opal_cpu_metadata->region[0].size);
arch/powerpc/platforms/powernv/pci-ioda.c
1918
struct pci_bus_region region;
arch/powerpc/platforms/powernv/pci-ioda.c
1927
region.start = res->start - phb->ioda.io_pci_base;
arch/powerpc/platforms/powernv/pci-ioda.c
1928
region.end = res->end - phb->ioda.io_pci_base;
arch/powerpc/platforms/powernv/pci-ioda.c
1929
index = region.start / phb->ioda.io_segsize;
arch/powerpc/platforms/powernv/pci-ioda.c
1932
region.start <= region.end) {
arch/powerpc/platforms/powernv/pci-ioda.c
1942
region.start += phb->ioda.io_segsize;
arch/powerpc/platforms/powernv/pci-ioda.c
1947
region.start = res->start -
arch/powerpc/platforms/powernv/pci-ioda.c
1950
region.end = res->end -
arch/powerpc/platforms/powernv/pci-ioda.c
1953
index = region.start / phb->ioda.m32_segsize;
arch/powerpc/platforms/powernv/pci-ioda.c
1956
region.start <= region.end) {
arch/powerpc/platforms/powernv/pci-ioda.c
1966
region.start += phb->ioda.m32_segsize;
arch/powerpc/platforms/ps3/mm.c
374
struct ps3_dma_region *region;
arch/powerpc/platforms/ps3/mm.c
387
c->region->dev->bus_id, c->region->dev->dev_id);
arch/powerpc/platforms/ps3/mm.c
388
DBG("%s:%d: r.bus_addr %lxh\n", func, line, c->region->bus_addr);
arch/powerpc/platforms/ps3/mm.c
389
DBG("%s:%d: r.page_size %u\n", func, line, c->region->page_size);
arch/powerpc/platforms/ps3/mm.c
390
DBG("%s:%d: r.len %lxh\n", func, line, c->region->len);
arch/powerpc/platforms/ps3/mm.c
391
DBG("%s:%d: r.offset %lxh\n", func, line, c->region->offset);
arch/powerpc/platforms/ps3/mm.c
462
result = lv1_unmap_device_dma_region(c->region->dev->bus_id,
arch/powerpc/platforms/ps3/mm.c
463
c->region->dev->dev_id, c->bus_addr, c->len);
arch/powerpc/platforms/ps3/mm.c
476
struct ps3_dma_region *r = c->region;
arch/powerpc/platforms/ps3/mm.c
525
c->region = r;
arch/powerpc/platforms/ps3/mm.c
531
result = lv1_map_device_dma_region(c->region->dev->bus_id,
arch/powerpc/platforms/ps3/mm.c
532
c->region->dev->dev_id, c->lpar_addr,
arch/powerpc/platforms/ps3/mm.c
570
c->region = r;
arch/powerpc/platforms/pseries/papr_scm.c
1273
p->region = nvdimm_volatile_region_create(p->bus, &ndr_desc);
arch/powerpc/platforms/pseries/papr_scm.c
1276
p->region = nvdimm_pmem_region_create(p->bus, &ndr_desc);
arch/powerpc/platforms/pseries/papr_scm.c
1278
if (!p->region) {
arch/powerpc/platforms/pseries/papr_scm.c
1298
static void papr_scm_add_badblock(struct nd_region *region,
arch/powerpc/platforms/pseries/papr_scm.c
1311
nvdimm_region_notify(region, NVDIMM_REVALIDATE_POISON);
arch/powerpc/platforms/pseries/papr_scm.c
1349
papr_scm_add_badblock(p->region, p->bus, phys_addr);
arch/powerpc/platforms/pseries/papr_scm.c
68
struct nd_region *region;
arch/riscv/kernel/acpi.c
229
efi_memory_desc_t *md, *region = NULL;
arch/riscv/kernel/acpi.c
245
region = md;
arch/riscv/kernel/acpi.c
256
if (region) {
arch/riscv/kernel/acpi.c
257
switch (region->type) {
arch/riscv/kernel/acpi.c
303
if (region->attribute & EFI_MEMORY_WB)
arch/riscv/kernel/acpi.c
305
else if ((region->attribute & EFI_MEMORY_WC) ||
arch/riscv/kernel/acpi.c
306
(region->attribute & EFI_MEMORY_WT))
arch/riscv/kernel/setup.c
144
struct memblock_region *region = NULL;
arch/riscv/kernel/setup.c
176
for_each_reserved_mem_region(region) {
arch/riscv/kernel/setup.c
181
res->start = __pfn_to_phys(memblock_region_reserved_base_pfn(region));
arch/riscv/kernel/setup.c
182
res->end = __pfn_to_phys(memblock_region_reserved_end_pfn(region)) - 1;
arch/riscv/kernel/setup.c
200
for_each_mem_region(region) {
arch/riscv/kernel/setup.c
204
if (unlikely(memblock_is_nomap(region))) {
arch/riscv/kernel/setup.c
212
res->start = __pfn_to_phys(memblock_region_memory_base_pfn(region));
arch/riscv/kernel/setup.c
213
res->end = __pfn_to_phys(memblock_region_memory_end_pfn(region)) - 1;
arch/sh/drivers/pci/fixups-dreamcast.c
32
struct pci_bus_region region;
arch/sh/drivers/pci/fixups-dreamcast.c
62
pcibios_resource_to_bus(dev->bus, &region, &res);
arch/sh/drivers/pci/fixups-dreamcast.c
65
region.start,
arch/sparc/kernel/pci.c
187
struct pci_bus_region region;
arch/sparc/kernel/pci.c
195
pcibios_resource_to_bus(dev->bus, &region, res);
arch/sparc/kernel/pci.c
196
if (region.end <= ~((u32)0))
arch/sparc/kernel/pci.c
419
struct pci_bus_region region;
arch/sparc/kernel/pci.c
428
region.start = (first << 21);
arch/sparc/kernel/pci.c
429
region.end = (last << 21) + ((1 << 21) - 1);
arch/sparc/kernel/pci.c
430
pcibios_bus_to_resource(dev->bus, res, &region);
arch/sparc/kernel/pci.c
436
region.start = (first << 29);
arch/sparc/kernel/pci.c
437
region.end = (last << 29) + ((1 << 29) - 1);
arch/sparc/kernel/pci.c
438
pcibios_bus_to_resource(dev->bus, res, &region);
arch/sparc/kernel/pci.c
454
struct pci_bus_region region;
arch/sparc/kernel/pci.c
555
region.start = start = GET_64BIT(ranges, 1);
arch/sparc/kernel/pci.c
556
region.end = region.start + size - 1;
arch/sparc/kernel/pci.c
562
pcibios_bus_to_resource(dev->bus, res, &region);
arch/sparc/kernel/pci.c
654
struct pci_bus_region region;
arch/sparc/kernel/pci.c
667
region.start = 0xa0000UL;
arch/sparc/kernel/pci.c
668
region.end = region.start + 0x1ffffUL;
arch/sparc/kernel/pci.c
669
pcibios_bus_to_resource(dev->bus, p, &region);
arch/sparc/kernel/pci.c
854
struct pci_bus_region region;
arch/sparc/kernel/pci.c
863
pcibios_resource_to_bus(pdev->bus, &region, (struct resource *) rp);
arch/sparc/kernel/pci.c
864
*start = region.start;
arch/sparc/kernel/pci.c
865
*end = region.end;
arch/um/drivers/vfio_user.c
197
dev->region = uml_kmalloc(sizeof(*dev->region) * dev->num_regions,
arch/um/drivers/vfio_user.c
199
if (!dev->region) {
arch/um/drivers/vfio_user.c
205
struct vfio_region_info region = {
arch/um/drivers/vfio_user.c
206
.argsz = sizeof(region),
arch/um/drivers/vfio_user.c
209
if (ioctl(dev->device, VFIO_DEVICE_GET_REGION_INFO, &region) < 0) {
arch/um/drivers/vfio_user.c
213
dev->region[i].size = region.size;
arch/um/drivers/vfio_user.c
214
dev->region[i].offset = region.offset;
arch/um/drivers/vfio_user.c
243
kfree(dev->region);
arch/um/drivers/vfio_user.c
252
kfree(dev->region);
arch/um/drivers/vfio_user.c
282
if (index >= dev->num_regions || offset + size > dev->region[index].size)
arch/um/drivers/vfio_user.c
285
if (pread(dev->device, buf, size, dev->region[index].offset + offset) < 0)
arch/um/drivers/vfio_user.c
294
if (index >= dev->num_regions || offset + size > dev->region[index].size)
arch/um/drivers/vfio_user.c
297
if (pwrite(dev->device, buf, size, dev->region[index].offset + offset) < 0)
arch/um/drivers/vfio_user.h
11
} *region;
arch/x86/boot/compressed/kaslr.c
463
static void store_slot_info(struct mem_vector *region, unsigned long image_size)
arch/x86/boot/compressed/kaslr.c
470
slot_area.addr = region->start;
arch/x86/boot/compressed/kaslr.c
471
slot_area.num = 1 + (region->size - image_size) / CONFIG_PHYSICAL_ALIGN;
arch/x86/boot/compressed/kaslr.c
482
process_gb_huge_pages(struct mem_vector *region, unsigned long image_size)
arch/x86/boot/compressed/kaslr.c
489
store_slot_info(region, image_size);
arch/x86/boot/compressed/kaslr.c
494
pud_start = ALIGN(region->start, PUD_SIZE);
arch/x86/boot/compressed/kaslr.c
495
pud_end = ALIGN_DOWN(region->start + region->size, PUD_SIZE);
arch/x86/boot/compressed/kaslr.c
499
store_slot_info(region, image_size);
arch/x86/boot/compressed/kaslr.c
504
if (pud_start >= region->start + image_size) {
arch/x86/boot/compressed/kaslr.c
505
tmp.start = region->start;
arch/x86/boot/compressed/kaslr.c
506
tmp.size = pud_start - region->start;
arch/x86/boot/compressed/kaslr.c
520
if (region->start + region->size >= pud_end + image_size) {
arch/x86/boot/compressed/kaslr.c
522
tmp.size = region->start + region->size - pud_end;
arch/x86/boot/compressed/kaslr.c
555
struct mem_vector region, overlap;
arch/x86/boot/compressed/kaslr.c
559
region.start = max_t(u64, entry->start, minimum);
arch/x86/boot/compressed/kaslr.c
565
region.start = ALIGN(region.start, CONFIG_PHYSICAL_ALIGN);
arch/x86/boot/compressed/kaslr.c
568
if (region.start > region_end)
arch/x86/boot/compressed/kaslr.c
572
region.size = region_end - region.start;
arch/x86/boot/compressed/kaslr.c
575
if (region.size < image_size)
arch/x86/boot/compressed/kaslr.c
579
if (!mem_avoid_overlap(&region, &overlap)) {
arch/x86/boot/compressed/kaslr.c
580
process_gb_huge_pages(&region, image_size);
arch/x86/boot/compressed/kaslr.c
585
if (overlap.start >= region.start + image_size) {
arch/x86/boot/compressed/kaslr.c
586
region.size = overlap.start - region.start;
arch/x86/boot/compressed/kaslr.c
587
process_gb_huge_pages(&region, image_size);
arch/x86/boot/compressed/kaslr.c
591
region.start = overlap.start + overlap.size;
arch/x86/boot/compressed/kaslr.c
595
static bool process_mem_region(struct mem_vector *region,
arch/x86/boot/compressed/kaslr.c
605
__process_mem_region(region, minimum, image_size);
arch/x86/boot/compressed/kaslr.c
623
if (!mem_overlaps(region, &immovable_mem[i]))
arch/x86/boot/compressed/kaslr.c
628
region_end = region->start + region->size;
arch/x86/boot/compressed/kaslr.c
630
entry.start = clamp(region->start, start, end);
arch/x86/boot/compressed/kaslr.c
682
struct mem_vector region;
arch/x86/boot/compressed/kaslr.c
728
region.start = md->phys_addr;
arch/x86/boot/compressed/kaslr.c
729
region.size = md->num_pages << EFI_PAGE_SHIFT;
arch/x86/boot/compressed/kaslr.c
730
if (process_mem_region(&region, minimum, image_size))
arch/x86/boot/compressed/kaslr.c
747
struct mem_vector region;
arch/x86/boot/compressed/kaslr.c
756
region.start = entry->addr;
arch/x86/boot/compressed/kaslr.c
757
region.size = entry->size;
arch/x86/boot/compressed/kaslr.c
758
if (process_mem_region(&region, minimum, image_size))
arch/x86/boot/compressed/kaslr.c
794
struct mem_vector region = {
arch/x86/boot/compressed/kaslr.c
799
if (process_mem_region(&region, minimum, image_size))
arch/x86/include/uapi/asm/kvm.h
400
__u32 region[1024];
arch/x86/kvm/svm/sev.c
2701
struct enc_region *region;
arch/x86/kvm/svm/sev.c
2714
region = kzalloc_obj(*region, GFP_KERNEL_ACCOUNT);
arch/x86/kvm/svm/sev.c
2715
if (!region)
arch/x86/kvm/svm/sev.c
2719
region->pages = sev_pin_memory(kvm, range->addr, range->size, &region->npages,
arch/x86/kvm/svm/sev.c
2721
if (IS_ERR(region->pages)) {
arch/x86/kvm/svm/sev.c
2722
ret = PTR_ERR(region->pages);
arch/x86/kvm/svm/sev.c
2735
sev_clflush_pages(region->pages, region->npages);
arch/x86/kvm/svm/sev.c
2737
region->uaddr = range->addr;
arch/x86/kvm/svm/sev.c
2738
region->size = range->size;
arch/x86/kvm/svm/sev.c
2740
list_add_tail(&region->list, &sev->regions_list);
arch/x86/kvm/svm/sev.c
2746
kfree(region);
arch/x86/kvm/svm/sev.c
2767
struct enc_region *region)
arch/x86/kvm/svm/sev.c
2769
sev_unpin_memory(kvm, region->pages, region->npages);
arch/x86/kvm/svm/sev.c
2770
list_del(&region->list);
arch/x86/kvm/svm/sev.c
2771
kfree(region);
arch/x86/kvm/svm/sev.c
2777
struct enc_region *region;
arch/x86/kvm/svm/sev.c
2791
region = find_enc_region(kvm, range);
arch/x86/kvm/svm/sev.c
2792
if (!region) {
arch/x86/kvm/svm/sev.c
2799
__unregister_enc_region_locked(kvm, region);
arch/x86/kvm/vmx/tdx.c
3162
struct kvm_tdx_init_mem_region region;
arch/x86/kvm/vmx/tdx.c
3177
if (copy_from_user(&region, u64_to_user_ptr(cmd->data), sizeof(region)))
arch/x86/kvm/vmx/tdx.c
3180
if (!PAGE_ALIGNED(region.source_addr) || !PAGE_ALIGNED(region.gpa) ||
arch/x86/kvm/vmx/tdx.c
3181
!region.nr_pages ||
arch/x86/kvm/vmx/tdx.c
3182
region.gpa + (region.nr_pages << PAGE_SHIFT) <= region.gpa ||
arch/x86/kvm/vmx/tdx.c
3183
!vt_is_tdx_private_gpa(kvm, region.gpa) ||
arch/x86/kvm/vmx/tdx.c
3184
!vt_is_tdx_private_gpa(kvm, region.gpa + (region.nr_pages << PAGE_SHIFT) - 1))
arch/x86/kvm/vmx/tdx.c
3188
while (region.nr_pages) {
arch/x86/kvm/vmx/tdx.c
3198
gmem_ret = kvm_gmem_populate(kvm, gpa_to_gfn(region.gpa),
arch/x86/kvm/vmx/tdx.c
3199
u64_to_user_ptr(region.source_addr),
arch/x86/kvm/vmx/tdx.c
3211
region.source_addr += PAGE_SIZE;
arch/x86/kvm/vmx/tdx.c
3212
region.gpa += PAGE_SIZE;
arch/x86/kvm/vmx/tdx.c
3213
region.nr_pages--;
arch/x86/kvm/vmx/tdx.c
3218
if (copy_to_user(u64_to_user_ptr(cmd->data), &region, sizeof(region)))
arch/x86/kvm/x86.c
5818
return kvm_vcpu_ioctl_x86_get_xsave2(vcpu, (void *)guest_xsave->region,
arch/x86/kvm/x86.c
5819
sizeof(guest_xsave->region));
arch/x86/kvm/x86.c
5825
union fpregs_state *xstate = (union fpregs_state *)guest_xsave->region;
arch/x86/kvm/x86.c
5838
guest_xsave->region,
arch/x86/kvm/x86.c
7561
struct kvm_enc_region region;
arch/x86/kvm/x86.c
7564
if (copy_from_user(&region, argp, sizeof(region)))
arch/x86/kvm/x86.c
7571
r = kvm_x86_call(mem_enc_register_region)(kvm, &region);
arch/x86/kvm/x86.c
7575
struct kvm_enc_region region;
arch/x86/kvm/x86.c
7578
if (copy_from_user(&region, argp, sizeof(region)))
arch/x86/kvm/x86.c
7585
r = kvm_x86_call(mem_enc_unregister_region)(kvm, &region);
arch/x86/mm/kaslr.c
73
static inline unsigned long get_padding(struct kaslr_memory_region *region)
arch/x86/mm/kaslr.c
75
return (region->size_tb << TB_SHIFT);
arch/x86/platform/uv/bios_uv.c
100
if (region)
arch/x86/platform/uv/bios_uv.c
101
*region = part.region_size;
arch/x86/platform/uv/bios_uv.c
82
long *region, long *ssn)
drivers/accel/ethosu/ethosu_gem.c
101
s8 region;
drivers/accel/ethosu/ethosu_gem.c
118
s8 region;
drivers/accel/ethosu/ethosu_gem.c
126
s8 region;
drivers/accel/ethosu/ethosu_gem.c
174
if (dma->region >= 0)
drivers/accel/ethosu/ethosu_gem.c
175
info->region_size[dma->region] = max(info->region_size[dma->region],
drivers/accel/ethosu/ethosu_gem.c
189
if (fm->region < 0)
drivers/accel/ethosu/ethosu_gem.c
229
info->region_size[fm->region] = max(info->region_size[fm->region], addr + 1);
drivers/accel/ethosu/ethosu_gem.c
259
op, st->ifm.region, st->ifm.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
268
op, st->ifm2.region, st->ifm2.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
275
op, st->weight[0].region, st->weight[0].base,
drivers/accel/ethosu/ethosu_gem.c
277
if (st->weight[0].region < 0 || st->weight[0].base == U64_MAX ||
drivers/accel/ethosu/ethosu_gem.c
280
info->region_size[st->weight[0].region] =
drivers/accel/ethosu/ethosu_gem.c
281
max(info->region_size[st->weight[0].region],
drivers/accel/ethosu/ethosu_gem.c
287
op, st->scale[0].region, st->scale[0].base,
drivers/accel/ethosu/ethosu_gem.c
289
if (st->scale[0].region < 0 || st->scale[0].base == U64_MAX ||
drivers/accel/ethosu/ethosu_gem.c
292
info->region_size[st->scale[0].region] =
drivers/accel/ethosu/ethosu_gem.c
293
max(info->region_size[st->scale[0].region],
drivers/accel/ethosu/ethosu_gem.c
300
op, st->ofm.region, st->ofm.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
303
info->output_region[st->ofm.region] = true;
drivers/accel/ethosu/ethosu_gem.c
324
op, st->ifm.region, st->ifm.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
337
op, st->ifm2.region, st->ifm2.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
345
op, st->ofm.region, st->ofm.base[0], len);
drivers/accel/ethosu/ethosu_gem.c
348
info->output_region[st->ofm.region] = true;
drivers/accel/ethosu/ethosu_gem.c
399
if (st.dma.dst.region >= 0)
drivers/accel/ethosu/ethosu_gem.c
400
info->output_region[st.dma.dst.region] = true;
drivers/accel/ethosu/ethosu_gem.c
402
st.dma.src.region, st.dma.src.offset, srclen,
drivers/accel/ethosu/ethosu_gem.c
403
st.dma.dst.region, st.dma.dst.offset, dstlen);
drivers/accel/ethosu/ethosu_gem.c
467
st.ifm.region = param & 0x7f;
drivers/accel/ethosu/ethosu_gem.c
507
st.ofm.region = param & 0x7;
drivers/accel/ethosu/ethosu_gem.c
541
st.ifm2.region = param & 0x7;
drivers/accel/ethosu/ethosu_gem.c
569
st.weight[0].region = param & 0x7;
drivers/accel/ethosu/ethosu_gem.c
572
st.scale[0].region = param & 0x7;
drivers/accel/ethosu/ethosu_gem.c
613
st.dma.src.region = -1;
drivers/accel/ethosu/ethosu_gem.c
615
st.dma.src.region = param & 0x7;
drivers/accel/ethosu/ethosu_gem.c
620
st.dma.dst.region = -1;
drivers/accel/ethosu/ethosu_gem.c
622
st.dma.dst.region = param & 0x7;
drivers/accel/ethosu/ethosu_job.c
52
int region = job->region_bo_num[i];
drivers/accel/ethosu/ethosu_job.c
55
writel_relaxed(lower_32_bits(bo->dma_addr), dev->regs + NPU_REG_BASEP(region));
drivers/accel/ethosu/ethosu_job.c
56
writel_relaxed(upper_32_bits(bo->dma_addr), dev->regs + NPU_REG_BASEP_HI(region));
drivers/accel/ethosu/ethosu_job.c
57
dev_dbg(dev->base.dev, "Region %d base addr = %pad\n", region, &bo->dma_addr);
drivers/accel/habanalabs/common/device.c
111
rc = hl_set_dram_bar(hdev, old_base, region, NULL);
drivers/accel/habanalabs/common/device.c
51
static u64 hl_set_dram_bar(struct hl_device *hdev, u64 addr, struct pci_mem_region *region,
drivers/accel/habanalabs/common/device.c
60
bar_base_addr = region->region_base +
drivers/accel/habanalabs/common/device.c
61
div64_u64((addr - region->region_base), prop->dram_pci_bar_size) *
drivers/accel/habanalabs/common/device.c
76
struct pci_mem_region *region = &hdev->pci_mem_region[region_type];
drivers/accel/habanalabs/common/device.c
77
u64 old_base = 0, rc, bar_region_base = region->region_base;
drivers/accel/habanalabs/common/device.c
81
old_base = hl_set_dram_bar(hdev, addr, region, &bar_region_base);
drivers/accel/habanalabs/common/device.c
86
acc_addr = hdev->pcie_bar[region->bar_id] + region->offset_in_bar +
drivers/accel/habanalabs/common/firmware_if.c
2040
struct pci_mem_region *region)
drivers/accel/habanalabs/common/firmware_if.c
2046
if (end_addr >= region->region_base + region->region_size) {
drivers/accel/habanalabs/common/firmware_if.c
2058
if (end_addr >= region->region_base - region->offset_in_bar +
drivers/accel/habanalabs/common/firmware_if.c
2059
region->bar_size) {
drivers/accel/habanalabs/common/firmware_if.c
2081
struct pci_mem_region *region;
drivers/accel/habanalabs/common/firmware_if.c
2122
region = &hdev->pci_mem_region[region_id];
drivers/accel/habanalabs/common/firmware_if.c
2125
fw_loader->dynamic_loader.image_region = region;
drivers/accel/habanalabs/common/firmware_if.c
2133
region);
drivers/accel/habanalabs/common/firmware_if.c
2147
struct pci_mem_region *region)
drivers/accel/habanalabs/common/firmware_if.c
2152
device_addr = region->region_base + response->ram_offset;
drivers/accel/habanalabs/common/firmware_if.c
2161
region);
drivers/accel/habanalabs/common/firmware_if.c
2214
struct pci_mem_region *region;
drivers/accel/habanalabs/common/firmware_if.c
2228
region = &hdev->pci_mem_region[region_id];
drivers/accel/habanalabs/common/firmware_if.c
2230
rc = hl_fw_dynamic_validate_response(hdev, response, region);
drivers/accel/habanalabs/common/firmware_if.c
2244
src = hdev->pcie_bar[region->bar_id] + region->offset_in_bar +
drivers/accel/habanalabs/common/firmware_if.c
2368
struct pci_mem_region *region;
drivers/accel/habanalabs/common/firmware_if.c
2377
region = fw_loader->dynamic_loader.image_region;
drivers/accel/habanalabs/common/firmware_if.c
2379
dest = hdev->pcie_bar[region->bar_id] + region->offset_in_bar +
drivers/accel/habanalabs/common/firmware_if.c
2380
(addr - region->region_base);
drivers/accel/habanalabs/common/firmware_if.c
2400
struct pci_mem_region *region;
drivers/accel/habanalabs/common/firmware_if.c
2409
region = fw_loader->dynamic_loader.image_region;
drivers/accel/habanalabs/common/firmware_if.c
2411
dest = hdev->pcie_bar[region->bar_id] + region->offset_in_bar +
drivers/accel/habanalabs/common/firmware_if.c
2412
(addr - region->region_base);
drivers/accel/habanalabs/common/habanalabs.h
4046
int hl_pci_set_inbound_region(struct hl_device *hdev, u8 region,
drivers/accel/habanalabs/common/pci/pci.c
245
int hl_pci_set_inbound_region(struct hl_device *hdev, u8 region,
drivers/accel/habanalabs/common/pci/pci.c
254
offset = (0x200 * region) + 0x100;
drivers/accel/habanalabs/common/pci/pci.c
357
struct pci_mem_region *region = &hdev->pci_mem_region[i];
drivers/accel/habanalabs/common/pci/pci.c
359
if (!region->used)
drivers/accel/habanalabs/common/pci/pci.c
362
if ((addr >= region->region_base) &&
drivers/accel/habanalabs/common/pci/pci.c
363
(addr < region->region_base + region->region_size))
drivers/accel/habanalabs/gaudi/gaudi.c
1809
struct pci_mem_region *region;
drivers/accel/habanalabs/gaudi/gaudi.c
1812
region = &hdev->pci_mem_region[PCI_REGION_CFG];
drivers/accel/habanalabs/gaudi/gaudi.c
1813
region->region_base = CFG_BASE;
drivers/accel/habanalabs/gaudi/gaudi.c
1814
region->region_size = CFG_SIZE;
drivers/accel/habanalabs/gaudi/gaudi.c
1815
region->offset_in_bar = CFG_BASE - SPI_FLASH_BASE_ADDR;
drivers/accel/habanalabs/gaudi/gaudi.c
1816
region->bar_size = CFG_BAR_SIZE;
drivers/accel/habanalabs/gaudi/gaudi.c
1817
region->bar_id = CFG_BAR_ID;
drivers/accel/habanalabs/gaudi/gaudi.c
1818
region->used = 1;
drivers/accel/habanalabs/gaudi/gaudi.c
1821
region = &hdev->pci_mem_region[PCI_REGION_SRAM];
drivers/accel/habanalabs/gaudi/gaudi.c
1822
region->region_base = SRAM_BASE_ADDR;
drivers/accel/habanalabs/gaudi/gaudi.c
1823
region->region_size = SRAM_SIZE;
drivers/accel/habanalabs/gaudi/gaudi.c
1824
region->offset_in_bar = 0;
drivers/accel/habanalabs/gaudi/gaudi.c
1825
region->bar_size = SRAM_BAR_SIZE;
drivers/accel/habanalabs/gaudi/gaudi.c
1826
region->bar_id = SRAM_BAR_ID;
drivers/accel/habanalabs/gaudi/gaudi.c
1827
region->used = 1;
drivers/accel/habanalabs/gaudi/gaudi.c
1830
region = &hdev->pci_mem_region[PCI_REGION_DRAM];
drivers/accel/habanalabs/gaudi/gaudi.c
1831
region->region_base = DRAM_PHYS_BASE;
drivers/accel/habanalabs/gaudi/gaudi.c
1832
region->region_size = hdev->asic_prop.dram_size;
drivers/accel/habanalabs/gaudi/gaudi.c
1833
region->offset_in_bar = 0;
drivers/accel/habanalabs/gaudi/gaudi.c
1834
region->bar_size = prop->dram_pci_bar_size;
drivers/accel/habanalabs/gaudi/gaudi.c
1835
region->bar_id = HBM_BAR_ID;
drivers/accel/habanalabs/gaudi/gaudi.c
1836
region->used = 1;
drivers/accel/habanalabs/gaudi/gaudi.c
1839
region = &hdev->pci_mem_region[PCI_REGION_SP_SRAM];
drivers/accel/habanalabs/gaudi/gaudi.c
1840
region->region_base = PSOC_SCRATCHPAD_ADDR;
drivers/accel/habanalabs/gaudi/gaudi.c
1841
region->region_size = PSOC_SCRATCHPAD_SIZE;
drivers/accel/habanalabs/gaudi/gaudi.c
1842
region->offset_in_bar = PSOC_SCRATCHPAD_ADDR - SPI_FLASH_BASE_ADDR;
drivers/accel/habanalabs/gaudi/gaudi.c
1843
region->bar_size = CFG_BAR_SIZE;
drivers/accel/habanalabs/gaudi/gaudi.c
1844
region->bar_id = CFG_BAR_ID;
drivers/accel/habanalabs/gaudi/gaudi.c
1845
region->used = 1;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3840
struct pci_mem_region *region;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3843
region = &hdev->pci_mem_region[PCI_REGION_CFG];
drivers/accel/habanalabs/gaudi2/gaudi2.c
3844
region->region_base = CFG_BASE;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3845
region->region_size = CFG_SIZE;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3846
region->offset_in_bar = CFG_BASE - STM_FLASH_BASE_ADDR;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3847
region->bar_size = CFG_BAR_SIZE;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3848
region->bar_id = SRAM_CFG_BAR_ID;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3849
region->used = 1;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3852
region = &hdev->pci_mem_region[PCI_REGION_SRAM];
drivers/accel/habanalabs/gaudi2/gaudi2.c
3853
region->region_base = SRAM_BASE_ADDR;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3854
region->region_size = SRAM_SIZE;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3855
region->offset_in_bar = CFG_REGION_SIZE + BAR0_RSRVD_SIZE;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3856
region->bar_size = CFG_BAR_SIZE;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3857
region->bar_id = SRAM_CFG_BAR_ID;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3858
region->used = 1;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3861
region = &hdev->pci_mem_region[PCI_REGION_DRAM];
drivers/accel/habanalabs/gaudi2/gaudi2.c
3862
region->region_base = DRAM_PHYS_BASE;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3863
region->region_size = hdev->asic_prop.dram_size;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3864
region->offset_in_bar = 0;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3865
region->bar_size = prop->dram_pci_bar_size;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3866
region->bar_id = DRAM_BAR_ID;
drivers/accel/habanalabs/gaudi2/gaudi2.c
3867
region->used = 1;
drivers/accel/habanalabs/goya/goya.c
930
struct pci_mem_region *region;
drivers/accel/habanalabs/goya/goya.c
933
region = &hdev->pci_mem_region[PCI_REGION_CFG];
drivers/accel/habanalabs/goya/goya.c
934
region->region_base = CFG_BASE;
drivers/accel/habanalabs/goya/goya.c
935
region->region_size = CFG_SIZE;
drivers/accel/habanalabs/goya/goya.c
936
region->offset_in_bar = CFG_BASE - SRAM_BASE_ADDR;
drivers/accel/habanalabs/goya/goya.c
937
region->bar_size = CFG_BAR_SIZE;
drivers/accel/habanalabs/goya/goya.c
938
region->bar_id = SRAM_CFG_BAR_ID;
drivers/accel/habanalabs/goya/goya.c
939
region->used = 1;
drivers/accel/habanalabs/goya/goya.c
942
region = &hdev->pci_mem_region[PCI_REGION_SRAM];
drivers/accel/habanalabs/goya/goya.c
943
region->region_base = SRAM_BASE_ADDR;
drivers/accel/habanalabs/goya/goya.c
944
region->region_size = SRAM_SIZE;
drivers/accel/habanalabs/goya/goya.c
945
region->offset_in_bar = 0;
drivers/accel/habanalabs/goya/goya.c
946
region->bar_size = CFG_BAR_SIZE;
drivers/accel/habanalabs/goya/goya.c
947
region->bar_id = SRAM_CFG_BAR_ID;
drivers/accel/habanalabs/goya/goya.c
948
region->used = 1;
drivers/accel/habanalabs/goya/goya.c
951
region = &hdev->pci_mem_region[PCI_REGION_DRAM];
drivers/accel/habanalabs/goya/goya.c
952
region->region_base = DRAM_PHYS_BASE;
drivers/accel/habanalabs/goya/goya.c
953
region->region_size = hdev->asic_prop.dram_size;
drivers/accel/habanalabs/goya/goya.c
954
region->offset_in_bar = 0;
drivers/accel/habanalabs/goya/goya.c
955
region->bar_size = prop->dram_pci_bar_size;
drivers/accel/habanalabs/goya/goya.c
956
region->bar_id = DDR_BAR_ID;
drivers/accel/habanalabs/goya/goya.c
957
region->used = 1;
drivers/accel/habanalabs/include/gaudi2/asic_reg/gaudi2_regs.h
263
#define ARC_REGION_CFG_OFFSET(region) \
drivers/accel/habanalabs/include/gaudi2/asic_reg/gaudi2_regs.h
264
(mmARC_FARM_ARC0_AUX_ARC_REGION_CFG_0 + (region * 4) - mmARC_FARM_ARC0_AUX_BASE)
drivers/acpi/acpica/acobject.h
413
struct acpi_object_region region;
drivers/acpi/acpica/dbnames.c
532
if (node->object->field.region_obj->region.space_id !=
drivers/acpi/acpica/dbstats.c
169
acpi_db_enumerate_object(obj_desc->region.handler);
drivers/acpi/acpica/dbtest.c
747
switch (region_obj->region.space_id) {
drivers/acpi/acpica/dbtest.c
776
acpi_ut_get_region_name(region_obj->region.
drivers/acpi/acpica/dbtest.c
784
acpi_ut_get_region_name(region_obj->region.space_id),
drivers/acpi/acpica/dbtest.c
785
region_obj->region.node->name.ascii);
drivers/acpi/acpica/dsargs.c
223
status = acpi_ut_add_address_range(obj_desc->region.space_id,
drivers/acpi/acpica/dsargs.c
224
obj_desc->region.address,
drivers/acpi/acpica/dsargs.c
225
obj_desc->region.length, node);
drivers/acpi/acpica/dsargs.c
340
if (obj_desc->region.flags & AOPOBJ_DATA_VALID) {
drivers/acpi/acpica/dsargs.c
351
node = obj_desc->region.node;
drivers/acpi/acpica/dsargs.c
370
status = acpi_ut_add_address_range(obj_desc->region.space_id,
drivers/acpi/acpica/dsargs.c
371
obj_desc->region.address,
drivers/acpi/acpica/dsargs.c
372
obj_desc->region.length, node);
drivers/acpi/acpica/dsfield.c
524
if (info.region_node->object->region.space_id ==
drivers/acpi/acpica/dsfield.c
527
ACPI_ALLOCATE_ZEROED(info.region_node->object->region.
drivers/acpi/acpica/dsfield.c
94
obj_desc->region.node = *node;
drivers/acpi/acpica/dsopcode.c
407
obj_desc->region.length = (u32) operand_desc->integer.value;
drivers/acpi/acpica/dsopcode.c
412
if (!obj_desc->region.length
drivers/acpi/acpica/dsopcode.c
425
obj_desc->region.address = (acpi_physical_address)
drivers/acpi/acpica/dsopcode.c
431
ACPI_FORMAT_UINT64(obj_desc->region.address),
drivers/acpi/acpica/dsopcode.c
432
obj_desc->region.length));
drivers/acpi/acpica/dsopcode.c
434
status = acpi_ut_add_address_range(obj_desc->region.space_id,
drivers/acpi/acpica/dsopcode.c
435
obj_desc->region.address,
drivers/acpi/acpica/dsopcode.c
436
obj_desc->region.length, node);
drivers/acpi/acpica/dsopcode.c
440
obj_desc->region.flags |= AOPOBJ_DATA_VALID;
drivers/acpi/acpica/dsopcode.c
532
obj_desc->region.address = ACPI_PTR_TO_PHYSADDR(table);
drivers/acpi/acpica/dsopcode.c
533
obj_desc->region.length = table->length;
drivers/acpi/acpica/dsopcode.c
534
obj_desc->region.pointer = table;
drivers/acpi/acpica/dsopcode.c
538
ACPI_FORMAT_UINT64(obj_desc->region.address),
drivers/acpi/acpica/dsopcode.c
539
obj_desc->region.length));
drivers/acpi/acpica/dsopcode.c
543
obj_desc->region.flags |= AOPOBJ_DATA_VALID;
drivers/acpi/acpica/evhandler.c
252
if (obj_desc->region.space_id != handler_obj->address_space.space_id) {
drivers/acpi/acpica/evregion.c
129
handler_desc = region_obj->region.handler;
drivers/acpi/acpica/evregion.c
133
acpi_ut_get_node_name(region_obj->region.node),
drivers/acpi/acpica/evregion.c
135
acpi_ut_get_region_name(region_obj->region.
drivers/acpi/acpica/evregion.c
149
if (!(region_obj->region.flags & AOPOBJ_SETUP_COMPLETE)) {
drivers/acpi/acpica/evregion.c
161
acpi_ut_get_region_name(region_obj->region.
drivers/acpi/acpica/evregion.c
167
&& region_obj->region.space_id ==
drivers/acpi/acpica/evregion.c
174
ctx->length = (u16)region_obj->region.length;
drivers/acpi/acpica/evregion.c
175
ctx->subspace_id = (u8)region_obj->region.address;
drivers/acpi/acpica/evregion.c
178
if (region_obj->region.space_id ==
drivers/acpi/acpica/evregion.c
183
ctx->length = region_obj->region.length;
drivers/acpi/acpica/evregion.c
184
ctx->offset = region_obj->region.address;
drivers/acpi/acpica/evregion.c
207
region.
drivers/acpi/acpica/evregion.c
214
if (!(region_obj->region.flags & AOPOBJ_SETUP_COMPLETE)) {
drivers/acpi/acpica/evregion.c
215
region_obj->region.flags |= AOPOBJ_SETUP_COMPLETE;
drivers/acpi/acpica/evregion.c
231
address = (region_obj->region.address + region_offset);
drivers/acpi/acpica/evregion.c
235
&region_obj->region.handler->address_space, handler,
drivers/acpi/acpica/evregion.c
237
acpi_ut_get_region_name(region_obj->region.
drivers/acpi/acpica/evregion.c
269
if ((region_obj->region.space_id == ACPI_ADR_SPACE_GSBUS ||
drivers/acpi/acpica/evregion.c
270
region_obj->region.space_id == ACPI_ADR_SPACE_GPIO) &&
drivers/acpi/acpica/evregion.c
287
if (region_obj->region.space_id == ACPI_ADR_SPACE_GPIO) {
drivers/acpi/acpica/evregion.c
304
acpi_ut_get_region_name(region_obj->region.
drivers/acpi/acpica/evregion.c
311
if ((region_obj->region.space_id == ACPI_ADR_SPACE_EC) &&
drivers/acpi/acpica/evregion.c
368
handler_obj = region_obj->region.handler;
drivers/acpi/acpica/evregion.c
393
*last_obj_ptr = obj_desc->region.next;
drivers/acpi/acpica/evregion.c
394
obj_desc->region.next = NULL; /* Must clear field */
drivers/acpi/acpica/evregion.c
413
(region_obj->region.space_id)));
drivers/acpi/acpica/evregion.c
428
if (region_obj->region.flags & AOPOBJ_SETUP_COMPLETE) {
drivers/acpi/acpica/evregion.c
450
(region_obj->region.
drivers/acpi/acpica/evregion.c
454
region_obj->region.flags &=
drivers/acpi/acpica/evregion.c
467
region_obj->region.handler = NULL;
drivers/acpi/acpica/evregion.c
475
last_obj_ptr = &obj_desc->region.next;
drivers/acpi/acpica/evregion.c
476
obj_desc = obj_desc->region.next;
drivers/acpi/acpica/evregion.c
522
if (region_obj->region.handler) {
drivers/acpi/acpica/evregion.c
528
acpi_ut_get_node_name(region_obj->region.node),
drivers/acpi/acpica/evregion.c
530
acpi_ut_get_region_name(region_obj->region.
drivers/acpi/acpica/evregion.c
535
region_obj->region.next = handler_obj->address_space.region_list;
drivers/acpi/acpica/evregion.c
537
region_obj->region.handler = handler_obj;
drivers/acpi/acpica/evregion.c
571
region_obj->region.handler == NULL) {
drivers/acpi/acpica/evregion.c
585
node = region_obj->region.node->parent;
drivers/acpi/acpica/evregion.c
633
acpi_ut_create_integer_object((u64)region_obj->region.space_id);
drivers/acpi/acpica/evregion.c
792
if (obj_desc->region.space_id != info->space_id) {
drivers/acpi/acpica/evregion.c
863
(next_node->object->region.space_id == space_id)) {
drivers/acpi/acpica/evrgnini.c
147
handler_obj = region_obj->region.handler;
drivers/acpi/acpica/evrgnini.c
167
parent_node = region_obj->region.node->parent;
drivers/acpi/acpica/evrgnini.c
226
if (region_obj->region.flags & AOPOBJ_SETUP_COMPLETE) {
drivers/acpi/acpica/evrgnini.c
244
pci_device_node = region_obj->region.node;
drivers/acpi/acpica/evrgnini.c
290
region_obj->region.node);
drivers/acpi/acpica/evrgnini.c
455
local_region_context->pointer = region_desc->region.pointer;
drivers/acpi/acpica/evrgnini.c
547
node = region_obj->region.node->parent;
drivers/acpi/acpica/evrgnini.c
548
space_id = region_obj->region.space_id;
drivers/acpi/acpica/evrgnini.c
75
local_region_context->length = region_desc->region.length;
drivers/acpi/acpica/evrgnini.c
76
local_region_context->address = region_desc->region.address;
drivers/acpi/acpica/exconfig.c
315
if (obj_desc->region.space_id != ACPI_ADR_SPACE_SYSTEM_MEMORY) {
drivers/acpi/acpica/excreate.c
279
obj_desc->region.space_id = space_id;
drivers/acpi/acpica/excreate.c
280
obj_desc->region.address = 0;
drivers/acpi/acpica/excreate.c
281
obj_desc->region.length = 0;
drivers/acpi/acpica/excreate.c
282
obj_desc->region.pointer = NULL;
drivers/acpi/acpica/excreate.c
283
obj_desc->region.node = node;
drivers/acpi/acpica/excreate.c
284
obj_desc->region.handler = NULL;
drivers/acpi/acpica/exdump.c
116
{ACPI_EXD_UINT8, ACPI_EXD_OFFSET(region.space_id), "Space Id"},
drivers/acpi/acpica/exdump.c
117
{ACPI_EXD_UINT8, ACPI_EXD_OFFSET(region.flags), "Flags"},
drivers/acpi/acpica/exdump.c
118
{ACPI_EXD_NODE, ACPI_EXD_OFFSET(region.node), "Parent Node"},
drivers/acpi/acpica/exdump.c
119
{ACPI_EXD_ADDRESS, ACPI_EXD_OFFSET(region.address), "Address"},
drivers/acpi/acpica/exdump.c
120
{ACPI_EXD_UINT32, ACPI_EXD_OFFSET(region.length), "Length"},
drivers/acpi/acpica/exdump.c
121
{ACPI_EXD_HDLR_LIST, ACPI_EXD_OFFSET(region.handler), "Handler"},
drivers/acpi/acpica/exdump.c
122
{ACPI_EXD_POINTER, ACPI_EXD_OFFSET(region.next), "Next"}
drivers/acpi/acpica/exdump.c
520
while (next->region.next) {
drivers/acpi/acpica/exdump.c
526
next = next->region.next;
drivers/acpi/acpica/exdump.c
739
acpi_ut_get_region_name(obj_desc->region.
drivers/acpi/acpica/exdump.c
741
obj_desc->region.space_id);
drivers/acpi/acpica/exdump.c
747
if (!(obj_desc->region.flags & AOPOBJ_DATA_VALID)) {
drivers/acpi/acpica/exdump.c
751
ACPI_FORMAT_UINT64(obj_desc->region.
drivers/acpi/acpica/exdump.c
753
obj_desc->region.length);
drivers/acpi/acpica/exfield.c
137
(obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
139
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
141
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
143
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
145
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
194
(obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
202
(obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
295
(obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
303
(obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
305
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
307
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
309
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
311
|| obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfield.c
321
(obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exfldio.c
121
if (rgn_desc->region.length <
drivers/acpi/acpica/exfldio.c
130
if (ACPI_ROUND_UP(rgn_desc->region.length,
drivers/acpi/acpica/exfldio.c
141
if (rgn_desc->region.length <
drivers/acpi/acpica/exfldio.c
154
acpi_ut_get_node_name(rgn_desc->region.
drivers/acpi/acpica/exfldio.c
156
rgn_desc->region.length));
drivers/acpi/acpica/exfldio.c
170
acpi_ut_get_node_name(rgn_desc->region.node),
drivers/acpi/acpica/exfldio.c
171
rgn_desc->region.length));
drivers/acpi/acpica/exfldio.c
235
acpi_ut_get_region_name(rgn_desc->region.
drivers/acpi/acpica/exfldio.c
237
rgn_desc->region.space_id,
drivers/acpi/acpica/exfldio.c
241
ACPI_FORMAT_UINT64(rgn_desc->region.address +
drivers/acpi/acpica/exfldio.c
257
acpi_ut_get_region_name(rgn_desc->region.
drivers/acpi/acpica/exfldio.c
259
rgn_desc->region.space_id));
drivers/acpi/acpica/exfldio.c
263
acpi_ut_get_region_name(rgn_desc->region.
drivers/acpi/acpica/exfldio.c
265
rgn_desc->region.space_id));
drivers/acpi/acpica/exfldio.c
70
space_id = rgn_desc->region.space_id;
drivers/acpi/acpica/exprep.c
465
if ((obj_desc->field.region_obj->region.space_id ==
drivers/acpi/acpica/exserial.c
158
switch (obj_desc->field.region_obj->region.space_id) {
drivers/acpi/acpica/exserial.c
293
switch (obj_desc->field.region_obj->region.space_id) {
drivers/acpi/acpica/nsdump.c
315
acpi_ut_get_region_name(obj_desc->region.
drivers/acpi/acpica/nsdump.c
317
if (obj_desc->region.flags & AOPOBJ_DATA_VALID) {
drivers/acpi/acpica/nsdump.c
320
region.
drivers/acpi/acpica/nsdump.c
322
obj_desc->region.length);
drivers/acpi/acpica/nsdump.c
353
region_obj->region.
drivers/acpi/acpica/nsdump.c
362
region_obj->region.
drivers/acpi/acpica/nsobject.c
190
acpi_ut_remove_address_range(obj_desc->region.space_id, node);
drivers/acpi/acpica/utcopy.c
741
if (dest_desc->region.handler) {
drivers/acpi/acpica/utcopy.c
742
acpi_ut_add_reference(dest_desc->region.handler);
drivers/acpi/acpica/utdelete.c
191
if (!(object->region.node->flags & ANOBJ_TEMPORARY)) {
drivers/acpi/acpica/utdelete.c
192
acpi_ut_remove_address_range(object->region.space_id,
drivers/acpi/acpica/utdelete.c
193
object->region.node);
drivers/acpi/acpica/utdelete.c
203
handler_desc = object->region.handler;
drivers/acpi/acpica/utdelete.c
216
next_desc->region.next;
drivers/acpi/acpica/utdelete.c
222
last_obj_ptr = &next_desc->region.next;
drivers/acpi/acpica/utdelete.c
223
next_desc = next_desc->region.next;
drivers/acpi/apei/bert.c
46
static void __init bert_print_all(struct acpi_bert_region *region,
drivers/acpi/apei/bert.c
50
(struct acpi_hest_generic_status *)region;
drivers/acpi/arm64/iort.c
1271
struct iommu_resv_region *region;
drivers/acpi/arm64/iort.c
1273
region = iommu_alloc_resv_region(base + SZ_64K, SZ_64K,
drivers/acpi/arm64/iort.c
1276
if (region)
drivers/acpi/arm64/iort.c
1277
list_add_tail(&region->list, head);
drivers/acpi/arm64/iort.c
922
struct iommu_resv_region *region)
drivers/acpi/arm64/iort.c
926
rmr_data = container_of(region, struct iommu_iort_rmr_data, rr);
drivers/acpi/arm64/iort.c
937
struct iommu_resv_region *region;
drivers/acpi/arm64/iort.c
965
region = &rmr_data->rr;
drivers/acpi/arm64/iort.c
966
INIT_LIST_HEAD(&region->list);
drivers/acpi/arm64/iort.c
967
region->start = addr;
drivers/acpi/arm64/iort.c
968
region->length = size;
drivers/acpi/arm64/iort.c
969
region->prot = prot;
drivers/acpi/arm64/iort.c
970
region->type = type;
drivers/acpi/arm64/iort.c
971
region->free = iort_rmr_free;
drivers/acpi/nvs.c
40
struct nvs_region *region;
drivers/acpi/nvs.c
42
region = kmalloc_obj(*region);
drivers/acpi/nvs.c
43
if (!region)
drivers/acpi/nvs.c
45
region->phys_start = start;
drivers/acpi/nvs.c
46
region->size = size;
drivers/acpi/nvs.c
47
list_add_tail(&region->node, &nvs_region_list);
drivers/acpi/nvs.c
56
struct nvs_region *region;
drivers/acpi/nvs.c
58
list_for_each_entry(region, &nvs_region_list, node) {
drivers/acpi/nvs.c
59
rc = func(region->phys_start, region->size, data);
drivers/crypto/intel/qat/qat_common/icp_qat_uclo.h
150
struct icp_qat_uclo_region *region;
drivers/crypto/intel/qat/qat_common/icp_qat_uclo.h
160
struct icp_qat_uclo_region *region;
drivers/crypto/intel/qat/qat_common/icp_qat_uclo.h
293
char region;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
344
switch (init_mem->region) {
drivers/crypto/intel/qat/qat_common/qat_uclo.c
354
pr_err("initmem region error. region type=0x%x\n", init_mem->region);
drivers/crypto/intel/qat/qat_common/qat_uclo.c
45
ae_slice->region = kzalloc_obj(*ae_slice->region);
drivers/crypto/intel/qat/qat_common/qat_uclo.c
46
if (!ae_slice->region)
drivers/crypto/intel/qat/qat_common/qat_uclo.c
53
ae_slice->page->region = ae_slice->region;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
57
kfree(ae_slice->region);
drivers/crypto/intel/qat/qat_common/qat_uclo.c
58
ae_slice->region = NULL;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
72
kfree(ae_data->ae_slices[i].region);
drivers/crypto/intel/qat/qat_common/qat_uclo.c
73
ae_data->ae_slices[i].region = NULL;
drivers/cxl/core/cdat.c
339
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/core.h
126
struct rw_semaphore region;
drivers/cxl/core/edac.c
118
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/edac.c
1343
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/edac.c
160
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/edac.c
1789
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/hdm.c
1026
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/hdm.c
20
.region = __RWSEM_INITIALIZER(cxl_rwsem.region),
drivers/cxl/core/hdm.c
550
if (cxled->cxld.region) {
drivers/cxl/core/hdm.c
552
dev_name(&cxled->cxld.region->dev));
drivers/cxl/core/hdm.c
610
if (cxled->cxld.region) {
drivers/cxl/core/hdm.c
612
dev_name(&cxled->cxld.region->dev));
drivers/cxl/core/hdm.c
875
lockdep_assert_held_write(&cxl_rwsem.region);
drivers/cxl/core/mbox.c
1299
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/mbox.c
923
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/memdev.c
242
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/memdev.c
299
lockdep_assert_held(&cxl_rwsem.region);
drivers/cxl/core/memdev.c
334
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/memdev.c
359
lockdep_assert_held(&cxl_rwsem.region);
drivers/cxl/core/memdev.c
403
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/port.c
1611
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/port.c
182
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/port.c
1918
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/port.c
387
SET_CXL_REGION_ATTR(region)
drivers/cxl/core/port.c
407
SET_CXL_REGION_ATTR(region)
drivers/cxl/core/port.c
47
lockdep_assert_held(&cxl_rwsem.region);
drivers/cxl/core/port.c
569
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/region.c
1005
cxl_rr->region = cxlr;
drivers/cxl/core/region.c
1023
struct cxl_region *cxlr = cxl_rr->region;
drivers/cxl/core/region.c
1029
dev_WARN_ONCE(&cxlr->dev, cxld->region != cxlr, "region mismatch\n");
drivers/cxl/core/region.c
1030
if (cxld->region == cxlr) {
drivers/cxl/core/region.c
1031
cxld->region = NULL;
drivers/cxl/core/region.c
1039
struct cxl_region *cxlr = cxl_rr->region;
drivers/cxl/core/region.c
1052
struct cxl_region *cxlr = cxl_rr->region;
drivers/cxl/core/region.c
1064
if (!cxld->region) {
drivers/cxl/core/region.c
1065
cxld->region = cxlr;
drivers/cxl/core/region.c
1077
if (cxld->region) {
drivers/cxl/core/region.c
1080
dev_name(&cxld->region->dev));
drivers/cxl/core/region.c
1148
lockdep_assert_held_write(&cxl_rwsem.region);
drivers/cxl/core/region.c
1260
lockdep_assert_held_write(&cxl_rwsem.region);
drivers/cxl/core/region.c
1299
struct cxl_region *cxlr = cxl_rr->region;
drivers/cxl/core/region.c
152
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
169
lockdep_assert_held(&cxl_rwsem.region);
drivers/cxl/core/region.c
199
ACQUIRE(rwsem_write_kill, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
2163
lockdep_assert_held_write(&cxl_rwsem.region);
drivers/cxl/core/region.c
2178
cxlr = cxled->cxld.region;
drivers/cxl/core/region.c
2237
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/region.c
2242
ACQUIRE(rwsem_write_kill, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
2262
ACQUIRE(rwsem_write_kill, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
2268
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/region.c
2718
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
2722
if (cxld->region)
drivers/cxl/core/region.c
2723
return sysfs_emit(buf, "%s\n", dev_name(&cxld->region->dev));
drivers/cxl/core/region.c
2726
DEVICE_ATTR_RO(region);
drivers/cxl/core/region.c
2876
rc = cxl_mem_get_poison(cxlmd, offset, length, cxled->cxld.region);
drivers/cxl/core/region.c
2937
cxlr = cxled->cxld.region;
drivers/cxl/core/region.c
3399
lockdep_assert_held(&cxl_rwsem.region);
drivers/cxl/core/region.c
3462
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/region.c
3551
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/region.c
362
ACQUIRE(rwsem_write_kill, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
3772
guard(rwsem_read)(&cxl_rwsem.region);
drivers/cxl/core/region.c
380
ACQUIRE(rwsem_write_kill, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
3825
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/region.c
3979
scoped_guard(rwsem_read, &cxl_rwsem.region) {
drivers/cxl/core/region.c
4007
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/region.c
4010
struct cxl_region_params *p = &iter->region->params;
drivers/cxl/core/region.c
4085
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
4118
ACQUIRE(rwsem_read_intr, region_rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
4183
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
445
guard(rwsem_write)(&cxl_rwsem.region);
drivers/cxl/core/region.c
466
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
480
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
518
ACQUIRE(rwsem_write_kill, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
545
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
581
ACQUIRE(rwsem_write_kill, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
602
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
636
lockdep_assert_held_write(&cxl_rwsem.region);
drivers/cxl/core/region.c
680
lockdep_assert_held_write(&cxl_rwsem.region);
drivers/cxl/core/region.c
697
lockdep_assert_held_write(&cxl_rwsem.region);
drivers/cxl/core/region.c
723
ACQUIRE(rwsem_write_kill, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
746
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
763
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
817
ACQUIRE(rwsem_read_intr, rwsem)(&cxl_rwsem.region);
drivers/cxl/core/region.c
861
if (cxld->region) {
drivers/cxl/core/region.c
864
dev_name(dev), dev_name(&cxld->region->dev));
drivers/cxl/core/region.c
985
struct cxl_region_params *ip = &iter->region->params;
drivers/cxl/core/region.c
991
if (auto_order_ok(port, iter->region, cxld))
drivers/cxl/core/region.c
996
dev_name(&iter->region->dev), ip->res, p->res);
drivers/cxl/core/trace.h
1044
__string(region, cxlr ? dev_name(&cxlr->dev) : "")
drivers/cxl/core/trace.h
1066
__assign_str(region);
drivers/cxl/core/trace.h
1076
__assign_str(region);
drivers/cxl/core/trace.h
1090
__get_str(region),
drivers/cxl/cxl.h
373
struct cxl_region *region;
drivers/cxl/cxl.h
752
struct cxl_region *region;
drivers/dax/bus.c
1100
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
1169
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
1231
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
1274
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
1292
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
1368
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
1434
dev_dax->region = dax_region;
drivers/dax/bus.c
183
return is_static(dev_dax->region);
drivers/dax/bus.c
204
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
445
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
502
dax_region = dev_dax->region;
drivers/dax/bus.c
523
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
624
static void dax_region_unregister(void *region)
drivers/dax/bus.c
626
struct dax_region *dax_region = region;
drivers/dax/bus.c
73
if (dev_dax->region->res.flags & IORESOURCE_DAX_KMEM)
drivers/dax/bus.c
799
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
842
struct dax_region *dax_region = dev_dax->region;
drivers/dax/bus.c
958
struct dax_region *dax_region = dev_dax->region;
drivers/dax/dax-private.h
84
struct dax_region *region;
drivers/dma/altera-msgdma.c
771
struct resource *region;
drivers/dma/altera-msgdma.c
786
region = devm_request_mem_region(device, (*res)->start,
drivers/dma/altera-msgdma.c
788
if (region == NULL) {
drivers/dma/altera-msgdma.c
793
*ptr = devm_ioremap(device, region->start,
drivers/dma/altera-msgdma.c
794
resource_size(region));
drivers/dma/dw-edma/dw-edma-pcie.c
101
struct pci_bus_region region;
drivers/dma/dw-edma/dw-edma-pcie.c
108
pcibios_resource_to_bus(pdev->bus, &region, &res);
drivers/dma/dw-edma/dw-edma-pcie.c
109
return region.start;
drivers/dma/ppc4xx/adma.c
3459
int region = test_bit(PPC440SPE_DESC_RXOR12,
drivers/dma/ppc4xx/adma.c
3462
if (index < region) {
drivers/dma/ppc4xx/adma.c
3475
index - region +
drivers/edac/pnd2_edac.c
318
static void mk_region(char *name, struct region *rp, u64 base, u64 limit)
drivers/edac/pnd2_edac.c
326
static void mk_region_mask(char *name, struct region *rp, u64 base, u64 mask)
drivers/edac/pnd2_edac.c
346
static bool in_region(struct region *rp, u64 addr)
drivers/edac/pnd2_edac.c
425
static void apl_mk_region(char *name, struct region *rp, void *asym)
drivers/edac/pnd2_edac.c
435
static void dnv_mk_region(char *name, struct region *rp, void *asym)
drivers/edac/pnd2_edac.c
94
void (*mk_region)(char *name, struct region *rp, void *asym);
drivers/firewire/core-cdev.c
837
struct fw_address_region region;
drivers/firewire/core-cdev.c
844
region.start = a->offset;
drivers/firewire/core-cdev.c
846
region.end = a->offset + a->length;
drivers/firewire/core-cdev.c
848
region.end = a->region_end;
drivers/firewire/core-cdev.c
856
ret = fw_core_add_address_handler(&r->handler, &region);
drivers/firewire/core-transaction.c
650
const struct fw_address_region *region)
drivers/firewire/core-transaction.c
655
if (region->start & 0xffff000000000003ULL ||
drivers/firewire/core-transaction.c
656
region->start >= region->end ||
drivers/firewire/core-transaction.c
657
region->end > 0x0001000000000000ULL ||
drivers/firewire/core-transaction.c
664
handler->offset = region->start;
drivers/firewire/core-transaction.c
665
while (handler->offset + handler->length <= region->end) {
drivers/firmware/cirrus/cs_dsp.c
1185
const struct wmfw_region *region,
drivers/firmware/cirrus/cs_dsp.c
1189
unsigned int data_len = le32_to_cpu(region->len);
drivers/firmware/cirrus/cs_dsp.c
1193
raw = (const struct wmfw_adsp_alg_data *)region->data;
drivers/firmware/cirrus/cs_dsp.c
1212
tmp = region->data;
drivers/firmware/cirrus/cs_dsp.c
1214
pos = tmp - region->data;
drivers/firmware/cirrus/cs_dsp.c
1216
tmp = &region->data[pos];
drivers/firmware/cirrus/cs_dsp.c
1222
pos = tmp - region->data;
drivers/firmware/cirrus/cs_dsp.c
1227
pos = tmp - region->data;
drivers/firmware/cirrus/cs_dsp.c
1247
const struct wmfw_region *region,
drivers/firmware/cirrus/cs_dsp.c
1252
unsigned int data_len = le32_to_cpu(region->len);
drivers/firmware/cirrus/cs_dsp.c
1256
raw = (const struct wmfw_adsp_coeff_data *)&region->data[pos];
drivers/firmware/cirrus/cs_dsp.c
1286
tmp = &region->data[pos];
drivers/firmware/cirrus/cs_dsp.c
1292
pos = tmp - region->data;
drivers/firmware/cirrus/cs_dsp.c
1297
pos = tmp - region->data;
drivers/firmware/cirrus/cs_dsp.c
1302
pos = tmp - region->data;
drivers/firmware/cirrus/cs_dsp.c
1341
const struct wmfw_region *region)
drivers/firmware/cirrus/cs_dsp.c
1348
pos = cs_dsp_coeff_parse_alg(dsp, region, &alg_blk);
drivers/firmware/cirrus/cs_dsp.c
1353
pos = cs_dsp_coeff_parse_coeff(dsp, region, pos, &coeff_blk);
drivers/firmware/cirrus/cs_dsp.c
1488
const struct wmfw_region *region;
drivers/firmware/cirrus/cs_dsp.c
1549
if (sizeof(*region) > firmware->size - pos) {
drivers/firmware/cirrus/cs_dsp.c
1554
region = (void *)&(firmware->data[pos]);
drivers/firmware/cirrus/cs_dsp.c
1556
if (le32_to_cpu(region->len) > firmware->size - pos - sizeof(*region)) {
drivers/firmware/cirrus/cs_dsp.c
1563
offset = le32_to_cpu(region->offset) & 0xffffff;
drivers/firmware/cirrus/cs_dsp.c
1564
type = be32_to_cpu(region->type) & 0xff;
drivers/firmware/cirrus/cs_dsp.c
1571
min(le32_to_cpu(region->len), 100), region->data);
drivers/firmware/cirrus/cs_dsp.c
1575
ret = cs_dsp_parse_coeff(dsp, region);
drivers/firmware/cirrus/cs_dsp.c
1609
regions, le32_to_cpu(region->len), offset,
drivers/firmware/cirrus/cs_dsp.c
1619
region_len = le32_to_cpu(region->len);
drivers/firmware/cirrus/cs_dsp.c
1630
memcpy(buf, region->data, region_len);
drivers/firmware/cirrus/cs_dsp.c
1640
pos += le32_to_cpu(region->len) + sizeof(*region);
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
126
const struct cs_dsp_region *region = priv->dsp->mem;
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
130
if (region[i].type == mem_type)
drivers/firmware/cirrus/test/cs_dsp_mock_mem_maps.c
131
return region[i].base;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
341
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
357
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
360
KUNIT_ASSERT_EQ(test, le32_to_cpu(region->offset) >> 24, param->block_type);
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
361
KUNIT_ASSERT_EQ(test, le32_to_cpu(region->len), sizeof(payload));
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
363
region->len = cpu_to_le32(0x8000);
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
368
region->len = cpu_to_le32(0xffff);
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
373
region->len = cpu_to_le32(0x7fffffff);
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
378
region->len = cpu_to_le32(0x80000000);
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
383
region->len = cpu_to_le32(0xffffffff);
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
429
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
443
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
444
alg_data = (struct wmfw_adsp_alg_data *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
484
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
501
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
502
alg_data = (__force __le32 *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
541
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
558
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
559
alg_data = (__force __le32 *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
608
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
627
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
628
alg_data = (struct wmfw_adsp_alg_data *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
663
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
682
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
683
alg_data = (__force __le32 *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
721
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
740
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
741
alg_data = (__force __le32 *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
780
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
795
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
796
alg_data = (struct wmfw_adsp_alg_data *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
840
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
859
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
860
alg_data = (__force __le32 *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
889
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
909
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
910
alg_data = (__force __le32 *)region->data;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
943
struct wmfw_region *region;
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
963
region = (struct wmfw_region *)&wmfw->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_wmfw_error.c
964
alg_data = (__force __le32 *)region->data;
drivers/fpga/dfl-afu-dma-region.c
102
static bool afu_dma_check_continuous_pages(struct dfl_afu_dma_region *region)
drivers/fpga/dfl-afu-dma-region.c
104
int npages = region->length >> PAGE_SHIFT;
drivers/fpga/dfl-afu-dma-region.c
108
if (page_to_pfn(region->pages[i]) + 1 !=
drivers/fpga/dfl-afu-dma-region.c
109
page_to_pfn(region->pages[i + 1]))
drivers/fpga/dfl-afu-dma-region.c
124
static bool dma_region_check_iova(struct dfl_afu_dma_region *region,
drivers/fpga/dfl-afu-dma-region.c
127
if (!size && region->iova != iova)
drivers/fpga/dfl-afu-dma-region.c
130
return (region->iova <= iova) &&
drivers/fpga/dfl-afu-dma-region.c
131
(region->length + region->iova >= iova + size);
drivers/fpga/dfl-afu-dma-region.c
144
struct dfl_afu_dma_region *region)
drivers/fpga/dfl-afu-dma-region.c
150
(unsigned long long)region->iova);
drivers/fpga/dfl-afu-dma-region.c
161
if (dma_region_check_iova(this, region->iova, region->length))
drivers/fpga/dfl-afu-dma-region.c
164
if (region->iova < this->iova)
drivers/fpga/dfl-afu-dma-region.c
166
else if (region->iova > this->iova)
drivers/fpga/dfl-afu-dma-region.c
172
rb_link_node(&region->node, parent, new);
drivers/fpga/dfl-afu-dma-region.c
173
rb_insert_color(&region->node, &afu->dma_regions);
drivers/fpga/dfl-afu-dma-region.c
186
struct dfl_afu_dma_region *region)
drivers/fpga/dfl-afu-dma-region.c
191
(unsigned long long)region->iova);
drivers/fpga/dfl-afu-dma-region.c
194
rb_erase(&region->node, &afu->dma_regions);
drivers/fpga/dfl-afu-dma-region.c
207
struct dfl_afu_dma_region *region;
drivers/fpga/dfl-afu-dma-region.c
210
region = container_of(node, struct dfl_afu_dma_region, node);
drivers/fpga/dfl-afu-dma-region.c
213
(unsigned long long)region->iova);
drivers/fpga/dfl-afu-dma-region.c
217
if (region->iova)
drivers/fpga/dfl-afu-dma-region.c
219
region->iova, region->length,
drivers/fpga/dfl-afu-dma-region.c
222
if (region->pages)
drivers/fpga/dfl-afu-dma-region.c
223
afu_dma_unpin_pages(fdata, region);
drivers/fpga/dfl-afu-dma-region.c
226
kfree(region);
drivers/fpga/dfl-afu-dma-region.c
252
struct dfl_afu_dma_region *region;
drivers/fpga/dfl-afu-dma-region.c
254
region = container_of(node, struct dfl_afu_dma_region, node);
drivers/fpga/dfl-afu-dma-region.c
256
if (dma_region_check_iova(region, iova, size)) {
drivers/fpga/dfl-afu-dma-region.c
258
(unsigned long long)region->iova);
drivers/fpga/dfl-afu-dma-region.c
259
return region;
drivers/fpga/dfl-afu-dma-region.c
262
if (iova < region->iova)
drivers/fpga/dfl-afu-dma-region.c
264
else if (iova > region->iova)
drivers/fpga/dfl-afu-dma-region.c
305
struct dfl_afu_dma_region *region;
drivers/fpga/dfl-afu-dma-region.c
319
region = kzalloc(sizeof(*region), GFP_KERNEL);
drivers/fpga/dfl-afu-dma-region.c
320
if (!region)
drivers/fpga/dfl-afu-dma-region.c
323
region->user_addr = user_addr;
drivers/fpga/dfl-afu-dma-region.c
324
region->length = length;
drivers/fpga/dfl-afu-dma-region.c
327
ret = afu_dma_pin_pages(fdata, region);
drivers/fpga/dfl-afu-dma-region.c
334
if (!afu_dma_check_continuous_pages(region)) {
drivers/fpga/dfl-afu-dma-region.c
341
region->iova = dma_map_page(dfl_fpga_fdata_to_parent(fdata),
drivers/fpga/dfl-afu-dma-region.c
342
region->pages[0], 0,
drivers/fpga/dfl-afu-dma-region.c
343
region->length,
drivers/fpga/dfl-afu-dma-region.c
345
if (dma_mapping_error(dfl_fpga_fdata_to_parent(fdata), region->iova)) {
drivers/fpga/dfl-afu-dma-region.c
35
struct dfl_afu_dma_region *region)
drivers/fpga/dfl-afu-dma-region.c
351
*iova = region->iova;
drivers/fpga/dfl-afu-dma-region.c
354
ret = afu_dma_region_add(fdata, region);
drivers/fpga/dfl-afu-dma-region.c
365
region->iova, region->length, DMA_BIDIRECTIONAL);
drivers/fpga/dfl-afu-dma-region.c
367
afu_dma_unpin_pages(fdata, region);
drivers/fpga/dfl-afu-dma-region.c
369
kfree(region);
drivers/fpga/dfl-afu-dma-region.c
37
int npages = region->length >> PAGE_SHIFT;
drivers/fpga/dfl-afu-dma-region.c
383
struct dfl_afu_dma_region *region;
drivers/fpga/dfl-afu-dma-region.c
386
region = afu_dma_region_find_iova(fdata, iova);
drivers/fpga/dfl-afu-dma-region.c
387
if (!region) {
drivers/fpga/dfl-afu-dma-region.c
392
if (region->in_use) {
drivers/fpga/dfl-afu-dma-region.c
397
afu_dma_region_remove(fdata, region);
drivers/fpga/dfl-afu-dma-region.c
401
region->iova, region->length, DMA_BIDIRECTIONAL);
drivers/fpga/dfl-afu-dma-region.c
402
afu_dma_unpin_pages(fdata, region);
drivers/fpga/dfl-afu-dma-region.c
403
kfree(region);
drivers/fpga/dfl-afu-dma-region.c
45
region->pages = kzalloc_objs(struct page *, npages);
drivers/fpga/dfl-afu-dma-region.c
46
if (!region->pages) {
drivers/fpga/dfl-afu-dma-region.c
51
pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE,
drivers/fpga/dfl-afu-dma-region.c
52
region->pages);
drivers/fpga/dfl-afu-dma-region.c
66
unpin_user_pages(region->pages, pinned);
drivers/fpga/dfl-afu-dma-region.c
68
kfree(region->pages);
drivers/fpga/dfl-afu-dma-region.c
83
struct dfl_afu_dma_region *region)
drivers/fpga/dfl-afu-dma-region.c
85
long npages = region->length >> PAGE_SHIFT;
drivers/fpga/dfl-afu-dma-region.c
88
unpin_user_pages(region->pages, npages);
drivers/fpga/dfl-afu-dma-region.c
89
kfree(region->pages);
drivers/fpga/dfl-afu-main.c
685
struct dfl_afu_mmio_region region;
drivers/fpga/dfl-afu-main.c
697
ret = afu_mmio_region_get_by_index(fdata, rinfo.index, &region);
drivers/fpga/dfl-afu-main.c
701
rinfo.flags = region.flags;
drivers/fpga/dfl-afu-main.c
702
rinfo.size = region.size;
drivers/fpga/dfl-afu-main.c
703
rinfo.offset = region.offset;
drivers/fpga/dfl-afu-main.c
813
struct dfl_afu_mmio_region region;
drivers/fpga/dfl-afu-main.c
823
ret = afu_mmio_region_get_by_offset(fdata, offset, size, &region);
drivers/fpga/dfl-afu-main.c
827
if (!(region.flags & DFL_PORT_REGION_MMAP))
drivers/fpga/dfl-afu-main.c
830
if ((vma->vm_flags & VM_READ) && !(region.flags & DFL_PORT_REGION_READ))
drivers/fpga/dfl-afu-main.c
834
!(region.flags & DFL_PORT_REGION_WRITE))
drivers/fpga/dfl-afu-main.c
843
(region.phys + (offset - region.offset)) >> PAGE_SHIFT,
drivers/fpga/dfl-afu-region.c
100
struct dfl_afu_mmio_region *tmp, *region;
drivers/fpga/dfl-afu-region.c
102
list_for_each_entry_safe(region, tmp, &afu->regions, node)
drivers/fpga/dfl-afu-region.c
103
devm_kfree(&fdata->dev->dev, region);
drivers/fpga/dfl-afu-region.c
118
struct dfl_afu_mmio_region *region;
drivers/fpga/dfl-afu-region.c
124
region = get_region_by_index(afu, region_index);
drivers/fpga/dfl-afu-region.c
125
if (!region) {
drivers/fpga/dfl-afu-region.c
129
*pregion = *region;
drivers/fpga/dfl-afu-region.c
152
struct dfl_afu_mmio_region *region;
drivers/fpga/dfl-afu-region.c
158
for_each_region(region, afu)
drivers/fpga/dfl-afu-region.c
159
if (region->offset <= offset &&
drivers/fpga/dfl-afu-region.c
160
region->offset + region->size >= offset + size) {
drivers/fpga/dfl-afu-region.c
161
*pregion = *region;
drivers/fpga/dfl-afu-region.c
24
#define for_each_region(region, afu) \
drivers/fpga/dfl-afu-region.c
25
list_for_each_entry((region), &(afu)->regions, node)
drivers/fpga/dfl-afu-region.c
30
struct dfl_afu_mmio_region *region;
drivers/fpga/dfl-afu-region.c
32
for_each_region(region, afu)
drivers/fpga/dfl-afu-region.c
33
if (region->index == region_index)
drivers/fpga/dfl-afu-region.c
34
return region;
drivers/fpga/dfl-afu-region.c
54
struct dfl_afu_mmio_region *region;
drivers/fpga/dfl-afu-region.c
58
region = devm_kzalloc(dev, sizeof(*region), GFP_KERNEL);
drivers/fpga/dfl-afu-region.c
59
if (!region)
drivers/fpga/dfl-afu-region.c
62
region->index = region_index;
drivers/fpga/dfl-afu-region.c
63
region->size = region_size;
drivers/fpga/dfl-afu-region.c
64
region->phys = phys;
drivers/fpga/dfl-afu-region.c
65
region->flags = flags;
drivers/fpga/dfl-afu-region.c
79
region->offset = afu->region_cur_offset;
drivers/fpga/dfl-afu-region.c
80
list_add(&region->node, &afu->regions);
drivers/fpga/dfl-afu-region.c
89
devm_kfree(dev, region);
drivers/fpga/dfl-fme-pr.c
133
region = dfl_fme_region_find(fme, port_pr.port_id);
drivers/fpga/dfl-fme-pr.c
134
if (!region) {
drivers/fpga/dfl-fme-pr.c
139
fpga_image_info_free(region->info);
drivers/fpga/dfl-fme-pr.c
144
region->info = info;
drivers/fpga/dfl-fme-pr.c
146
ret = fpga_region_program_fpga(region);
drivers/fpga/dfl-fme-pr.c
153
if (region->get_bridges)
drivers/fpga/dfl-fme-pr.c
154
fpga_bridges_put(&region->bridge_list);
drivers/fpga/dfl-fme-pr.c
156
put_device(&region->dev);
drivers/fpga/dfl-fme-pr.c
320
fme_region->region = platform_device_alloc(DFL_FPGA_FME_REGION, br->id);
drivers/fpga/dfl-fme-pr.c
321
if (!fme_region->region)
drivers/fpga/dfl-fme-pr.c
324
fme_region->region->dev.parent = dev;
drivers/fpga/dfl-fme-pr.c
326
ret = platform_device_add_data(fme_region->region, &region_pdata,
drivers/fpga/dfl-fme-pr.c
331
ret = platform_device_add(fme_region->region);
drivers/fpga/dfl-fme-pr.c
340
platform_device_put(fme_region->region);
drivers/fpga/dfl-fme-pr.c
350
platform_device_unregister(fme_region->region);
drivers/fpga/dfl-fme-pr.c
52
struct fpga_region *region;
drivers/fpga/dfl-fme-pr.c
58
region = fpga_region_class_find(NULL, &fme_region->region->dev,
drivers/fpga/dfl-fme-pr.c
60
if (!region)
drivers/fpga/dfl-fme-pr.c
63
return region;
drivers/fpga/dfl-fme-pr.c
72
struct fpga_region *region;
drivers/fpga/dfl-fme-pr.h
31
struct platform_device *region;
drivers/fpga/dfl-fme-region.c
22
static int fme_region_get_bridges(struct fpga_region *region)
drivers/fpga/dfl-fme-region.c
24
struct dfl_fme_region_pdata *pdata = region->priv;
drivers/fpga/dfl-fme-region.c
27
return fpga_bridge_get_to_list(dev, region->info, &region->bridge_list);
drivers/fpga/dfl-fme-region.c
35
struct fpga_region *region;
drivers/fpga/dfl-fme-region.c
47
region = fpga_region_register_full(dev, &info);
drivers/fpga/dfl-fme-region.c
48
if (IS_ERR(region)) {
drivers/fpga/dfl-fme-region.c
49
ret = PTR_ERR(region);
drivers/fpga/dfl-fme-region.c
53
platform_set_drvdata(pdev, region);
drivers/fpga/dfl-fme-region.c
66
struct fpga_region *region = platform_get_drvdata(pdev);
drivers/fpga/dfl-fme-region.c
67
struct fpga_manager *mgr = region->mgr;
drivers/fpga/dfl-fme-region.c
69
fpga_region_unregister(region);
drivers/fpga/dfl.c
1545
device_for_each_child(&cdev->region->dev, NULL, remove_feature_dev);
drivers/fpga/dfl.c
1577
cdev->region = fpga_region_register(info->dev, NULL, NULL);
drivers/fpga/dfl.c
1578
if (IS_ERR(cdev->region)) {
drivers/fpga/dfl.c
1579
ret = PTR_ERR(cdev->region);
drivers/fpga/dfl.c
1616
fpga_region_unregister(cdev->region);
drivers/fpga/dfl.c
1640
fpga_region_unregister(cdev->region);
drivers/fpga/dfl.c
877
fdev->dev.parent = &fdata->dfl_cdev->region->dev;
drivers/fpga/dfl.h
543
struct fpga_region *region;
drivers/fpga/fpga-region.c
100
struct fpga_image_info *info = region->info;
drivers/fpga/fpga-region.c
103
region = fpga_region_get(region);
drivers/fpga/fpga-region.c
104
if (IS_ERR(region)) {
drivers/fpga/fpga-region.c
106
return PTR_ERR(region);
drivers/fpga/fpga-region.c
109
ret = fpga_mgr_lock(region->mgr);
drivers/fpga/fpga-region.c
119
if (region->get_bridges) {
drivers/fpga/fpga-region.c
120
ret = region->get_bridges(region);
drivers/fpga/fpga-region.c
127
ret = fpga_bridges_disable(&region->bridge_list);
drivers/fpga/fpga-region.c
133
ret = fpga_mgr_load(region->mgr, info);
drivers/fpga/fpga-region.c
139
ret = fpga_bridges_enable(&region->bridge_list);
drivers/fpga/fpga-region.c
145
fpga_mgr_unlock(region->mgr);
drivers/fpga/fpga-region.c
146
fpga_region_put(region);
drivers/fpga/fpga-region.c
151
if (region->get_bridges)
drivers/fpga/fpga-region.c
152
fpga_bridges_put(&region->bridge_list);
drivers/fpga/fpga-region.c
154
fpga_mgr_unlock(region->mgr);
drivers/fpga/fpga-region.c
156
fpga_region_put(region);
drivers/fpga/fpga-region.c
165
struct fpga_region *region = to_fpga_region(dev);
drivers/fpga/fpga-region.c
167
if (!region->compat_id)
drivers/fpga/fpga-region.c
171
(unsigned long long)region->compat_id->id_h,
drivers/fpga/fpga-region.c
172
(unsigned long long)region->compat_id->id_l);
drivers/fpga/fpga-region.c
195
struct fpga_region *region;
drivers/fpga/fpga-region.c
204
region = kzalloc_obj(*region);
drivers/fpga/fpga-region.c
205
if (!region)
drivers/fpga/fpga-region.c
214
region->mgr = info->mgr;
drivers/fpga/fpga-region.c
215
region->compat_id = info->compat_id;
drivers/fpga/fpga-region.c
216
region->priv = info->priv;
drivers/fpga/fpga-region.c
217
region->get_bridges = info->get_bridges;
drivers/fpga/fpga-region.c
218
region->ops_owner = owner;
drivers/fpga/fpga-region.c
220
mutex_init(&region->mutex);
drivers/fpga/fpga-region.c
221
INIT_LIST_HEAD(&region->bridge_list);
drivers/fpga/fpga-region.c
223
region->dev.class = &fpga_region_class;
drivers/fpga/fpga-region.c
224
region->dev.parent = parent;
drivers/fpga/fpga-region.c
225
region->dev.of_node = parent->of_node;
drivers/fpga/fpga-region.c
226
region->dev.id = id;
drivers/fpga/fpga-region.c
228
ret = dev_set_name(&region->dev, "region%d", id);
drivers/fpga/fpga-region.c
232
ret = device_register(&region->dev);
drivers/fpga/fpga-region.c
234
put_device(&region->dev);
drivers/fpga/fpga-region.c
238
return region;
drivers/fpga/fpga-region.c
243
kfree(region);
drivers/fpga/fpga-region.c
281
void fpga_region_unregister(struct fpga_region *region)
drivers/fpga/fpga-region.c
283
device_unregister(&region->dev);
drivers/fpga/fpga-region.c
289
struct fpga_region *region = to_fpga_region(dev);
drivers/fpga/fpga-region.c
291
ida_free(&fpga_region_ida, region->dev.id);
drivers/fpga/fpga-region.c
292
kfree(region);
drivers/fpga/fpga-region.c
46
static struct fpga_region *fpga_region_get(struct fpga_region *region)
drivers/fpga/fpga-region.c
48
struct device *dev = &region->dev;
drivers/fpga/fpga-region.c
50
if (!mutex_trylock(&region->mutex)) {
drivers/fpga/fpga-region.c
56
if (!try_module_get(region->ops_owner)) {
drivers/fpga/fpga-region.c
58
mutex_unlock(&region->mutex);
drivers/fpga/fpga-region.c
64
return region;
drivers/fpga/fpga-region.c
72
static void fpga_region_put(struct fpga_region *region)
drivers/fpga/fpga-region.c
74
struct device *dev = &region->dev;
drivers/fpga/fpga-region.c
78
module_put(region->ops_owner);
drivers/fpga/fpga-region.c
80
mutex_unlock(&region->mutex);
drivers/fpga/fpga-region.c
97
int fpga_region_program_fpga(struct fpga_region *region)
drivers/fpga/fpga-region.c
99
struct device *dev = &region->dev;
drivers/fpga/of-fpga-region.c
130
&region->bridge_list);
drivers/fpga/of-fpga-region.c
135
fpga_bridges_put(&region->bridge_list);
drivers/fpga/of-fpga-region.c
195
of_fpga_region_parse_ov(struct fpga_region *region,
drivers/fpga/of-fpga-region.c
198
struct device *dev = &region->dev;
drivers/fpga/of-fpga-region.c
203
if (region->info) {
drivers/fpga/of-fpga-region.c
286
static int of_fpga_region_notify_pre_apply(struct fpga_region *region,
drivers/fpga/of-fpga-region.c
289
struct device *dev = &region->dev;
drivers/fpga/of-fpga-region.c
293
info = of_fpga_region_parse_ov(region, nd->overlay);
drivers/fpga/of-fpga-region.c
301
if (region->info) {
drivers/fpga/of-fpga-region.c
306
region->info = info;
drivers/fpga/of-fpga-region.c
307
ret = fpga_region_program_fpga(region);
drivers/fpga/of-fpga-region.c
311
region->info = NULL;
drivers/fpga/of-fpga-region.c
326
static void of_fpga_region_notify_post_remove(struct fpga_region *region,
drivers/fpga/of-fpga-region.c
329
fpga_bridges_disable(&region->bridge_list);
drivers/fpga/of-fpga-region.c
330
fpga_bridges_put(&region->bridge_list);
drivers/fpga/of-fpga-region.c
331
fpga_image_info_free(region->info);
drivers/fpga/of-fpga-region.c
332
region->info = NULL;
drivers/fpga/of-fpga-region.c
350
struct fpga_region *region;
drivers/fpga/of-fpga-region.c
370
region = of_fpga_region_find(nd->target);
drivers/fpga/of-fpga-region.c
371
if (!region)
drivers/fpga/of-fpga-region.c
377
ret = of_fpga_region_notify_pre_apply(region, nd);
drivers/fpga/of-fpga-region.c
381
of_fpga_region_notify_post_remove(region, nd);
drivers/fpga/of-fpga-region.c
385
put_device(&region->dev);
drivers/fpga/of-fpga-region.c
401
struct fpga_region *region;
drivers/fpga/of-fpga-region.c
410
region = fpga_region_register(dev, mgr, of_fpga_region_get_bridges);
drivers/fpga/of-fpga-region.c
411
if (IS_ERR(region)) {
drivers/fpga/of-fpga-region.c
412
ret = PTR_ERR(region);
drivers/fpga/of-fpga-region.c
416
of_platform_populate(np, fpga_region_of_match, NULL, &region->dev);
drivers/fpga/of-fpga-region.c
417
platform_set_drvdata(pdev, region);
drivers/fpga/of-fpga-region.c
430
struct fpga_region *region = platform_get_drvdata(pdev);
drivers/fpga/of-fpga-region.c
431
struct fpga_manager *mgr = region->mgr;
drivers/fpga/of-fpga-region.c
433
fpga_region_unregister(region);
drivers/fpga/of-fpga-region.c
88
static int of_fpga_region_get_bridges(struct fpga_region *region)
drivers/fpga/of-fpga-region.c
90
struct device *dev = &region->dev;
drivers/fpga/of-fpga-region.c
92
struct fpga_image_info *info = region->info;
drivers/fpga/of-fpga-region.c
98
&region->bridge_list);
drivers/fpga/tests/fpga-region-test.c
105
struct fpga_region *region;
drivers/fpga/tests/fpga-region-test.c
107
region = fpga_region_class_find(NULL, ctx->region_dev, fake_region_match);
drivers/fpga/tests/fpga-region-test.c
108
KUNIT_EXPECT_PTR_EQ(test, region, ctx->region);
drivers/fpga/tests/fpga-region-test.c
110
put_device(&region->dev);
drivers/fpga/tests/fpga-region-test.c
133
ctx->region->info = img_info;
drivers/fpga/tests/fpga-region-test.c
134
ret = fpga_region_program_fpga(ctx->region);
drivers/fpga/tests/fpga-region-test.c
140
fpga_bridges_put(&ctx->region->bridge_list);
drivers/fpga/tests/fpga-region-test.c
142
ret = fpga_region_program_fpga(ctx->region);
drivers/fpga/tests/fpga-region-test.c
148
fpga_bridges_put(&ctx->region->bridge_list);
drivers/fpga/tests/fpga-region-test.c
192
ctx->region = fpga_region_register_full(ctx->region_dev, &region_info);
drivers/fpga/tests/fpga-region-test.c
193
KUNIT_ASSERT_FALSE(test, IS_ERR_OR_NULL(ctx->region));
drivers/fpga/tests/fpga-region-test.c
195
ret = kunit_add_action_or_reset(test, fpga_region_unregister_wrapper, ctx->region);
drivers/fpga/tests/fpga-region-test.c
32
struct fpga_region *region;
drivers/fpga/tests/fpga-region-test.c
90
static int fake_region_get_bridges(struct fpga_region *region)
drivers/fpga/tests/fpga-region-test.c
92
struct fpga_bridge *bridge = region->priv;
drivers/fpga/tests/fpga-region-test.c
94
return fpga_bridge_get_to_list(bridge->dev.parent, region->info, &region->bridge_list);
drivers/gpib/include/amcc5920.h
23
static inline uint32_t amcc_wait_state_bits(unsigned int region, unsigned int num_wait_states)
drivers/gpib/include/amcc5920.h
25
return (num_wait_states & 0x7) << (--region * bits_per_region);
drivers/gpib/include/amcc5920.h
35
static inline uint32_t amcc_prefetch_bits(unsigned int region, enum amcc_prefetch_bits prefetch)
drivers/gpib/include/amcc5920.h
37
return prefetch << (--region * bits_per_region);
drivers/gpib/include/amcc5920.h
40
static inline uint32_t amcc_PTADR_mode_bit(unsigned int region)
drivers/gpib/include/amcc5920.h
42
return 0x80 << (--region * bits_per_region);
drivers/gpib/include/amcc5920.h
45
static inline uint32_t amcc_disable_write_fifo_bit(unsigned int region)
drivers/gpib/include/amcc5920.h
47
return 0x20 << (--region * bits_per_region);
drivers/gpib/ines/ines_gpib.c
832
static const int region = 1;
drivers/gpib/ines/ines_gpib.c
836
bits = amcc_prefetch_bits(region, PREFETCH_DISABLED);
drivers/gpib/ines/ines_gpib.c
837
bits |= amcc_PTADR_mode_bit(region);
drivers/gpib/ines/ines_gpib.c
838
bits |= amcc_disable_write_fifo_bit(region);
drivers/gpib/ines/ines_gpib.c
839
bits |= amcc_wait_state_bits(region, num_wait_states);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5667
struct ras_critical_region *region;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5679
list_for_each_entry(region, &con->critical_region_head, node)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5680
if (region->bo == bo)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5685
region = kzalloc_obj(*region);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5686
if (!region) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5690
region->bo = bo;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5691
region->start = amdgpu_vram_mgr_block_start(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5692
region->size = amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5693
list_add_tail(&region->node, &con->critical_region_head);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5710
struct ras_critical_region *region, *tmp;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5713
list_for_each_entry_safe(region, tmp, &con->critical_region_head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5714
list_del(&region->node);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5715
kfree(region);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5723
struct ras_critical_region *region;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5727
list_for_each_entry(region, &con->critical_region_head, node) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5728
if ((region->start <= addr) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5729
(addr < (region->start + region->size))) {
drivers/gpu/drm/amd/display/dmub/dmub_srv.h
204
struct dmub_region region;
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
171
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
173
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
180
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
182
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
202
if (cw2->region.base != cw2->region.top) {
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
208
REG_WRITE(DMCUB_REGION3_CW2_BASE_ADDRESS, cw2->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
210
DMCUB_REGION3_CW2_TOP_ADDRESS, cw2->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
223
REG_WRITE(DMCUB_REGION3_CW3_BASE_ADDRESS, cw3->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
225
DMCUB_REGION3_CW3_TOP_ADDRESS, cw3->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
235
REG_WRITE(DMCUB_REGION3_CW4_BASE_ADDRESS, cw4->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
237
DMCUB_REGION3_CW4_TOP_ADDRESS, cw4->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
244
cw4->region.top - cw4->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
252
REG_WRITE(DMCUB_REGION3_CW5_BASE_ADDRESS, cw5->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
254
DMCUB_REGION3_CW5_TOP_ADDRESS, cw5->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
261
cw5->region.top - cw5->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
268
REG_WRITE(DMCUB_REGION3_CW6_BASE_ADDRESS, cw6->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
270
DMCUB_REGION3_CW6_TOP_ADDRESS, cw6->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
104
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
106
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
113
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
115
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
136
if (cw2->region.base != cw2->region.top) {
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
139
REG_WRITE(DMCUB_REGION3_CW2_BASE_ADDRESS, cw2->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
141
DMCUB_REGION3_CW2_TOP_ADDRESS, cw2->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
154
REG_WRITE(DMCUB_REGION3_CW3_BASE_ADDRESS, cw3->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
156
DMCUB_REGION3_CW3_TOP_ADDRESS, cw3->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
165
REG_WRITE(DMCUB_REGION3_CW4_BASE_ADDRESS, cw4->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
167
DMCUB_REGION3_CW4_TOP_ADDRESS, cw4->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
174
cw4->region.top - cw4->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
182
REG_WRITE(DMCUB_REGION3_CW5_BASE_ADDRESS, cw5->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
184
DMCUB_REGION3_CW5_TOP_ADDRESS, cw5->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
191
cw5->region.top - cw5->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
198
REG_WRITE(DMCUB_REGION3_CW6_BASE_ADDRESS, cw6->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
200
DMCUB_REGION3_CW6_TOP_ADDRESS, cw6->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
172
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
174
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
181
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
183
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
204
REG_WRITE(DMCUB_REGION3_CW3_BASE_ADDRESS, cw3->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
206
DMCUB_REGION3_CW3_TOP_ADDRESS, cw3->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
213
REG_WRITE(DMCUB_REGION3_CW4_BASE_ADDRESS, cw4->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
215
DMCUB_REGION3_CW4_TOP_ADDRESS, cw4->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
222
REG_WRITE(DMCUB_REGION3_CW5_BASE_ADDRESS, cw5->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
224
DMCUB_REGION3_CW5_TOP_ADDRESS, cw5->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
231
cw5->region.top - cw5->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
238
REG_WRITE(DMCUB_REGION3_CW6_BASE_ADDRESS, cw6->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
240
DMCUB_REGION3_CW6_TOP_ADDRESS, cw6->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
182
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
184
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
191
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
193
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
214
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
216
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
223
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
225
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
246
REG_WRITE(DMCUB_REGION3_CW3_BASE_ADDRESS, cw3->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
248
DMCUB_REGION3_CW3_TOP_ADDRESS, cw3->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
255
REG_WRITE(DMCUB_REGION3_CW4_BASE_ADDRESS, cw4->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
257
DMCUB_REGION3_CW4_TOP_ADDRESS, cw4->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
264
REG_WRITE(DMCUB_REGION3_CW5_BASE_ADDRESS, cw5->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
266
DMCUB_REGION3_CW5_TOP_ADDRESS, cw5->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
273
cw5->region.top - cw5->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
280
REG_WRITE(DMCUB_REGION3_CW6_BASE_ADDRESS, cw6->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
282
DMCUB_REGION3_CW6_TOP_ADDRESS, cw6->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
175
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
177
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
184
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
186
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
203
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
205
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
210
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
212
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
231
REG_WRITE(DMCUB_REGION3_CW3_BASE_ADDRESS, cw3->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
233
DMCUB_REGION3_CW3_TOP_ADDRESS, cw3->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
240
REG_WRITE(DMCUB_REGION3_CW4_BASE_ADDRESS, cw4->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
242
DMCUB_REGION3_CW4_TOP_ADDRESS, cw4->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
249
REG_WRITE(DMCUB_REGION3_CW5_BASE_ADDRESS, cw5->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
251
DMCUB_REGION3_CW5_TOP_ADDRESS, cw5->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
258
cw5->region.top - cw5->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
265
REG_WRITE(DMCUB_REGION3_CW6_BASE_ADDRESS, cw6->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
267
DMCUB_REGION3_CW6_TOP_ADDRESS, cw6->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
276
region6->region.top - region6->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
156
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
158
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
165
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
167
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
189
REG_WRITE(DMCUB_REGION3_CW0_BASE_ADDRESS, cw0->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
191
DMCUB_REGION3_CW0_TOP_ADDRESS, cw0->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
198
REG_WRITE(DMCUB_REGION3_CW1_BASE_ADDRESS, cw1->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
200
DMCUB_REGION3_CW1_TOP_ADDRESS, cw1->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
222
REG_WRITE(DMCUB_REGION3_CW3_BASE_ADDRESS, cw3->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
224
DMCUB_REGION3_CW3_TOP_ADDRESS, cw3->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
231
REG_WRITE(DMCUB_REGION3_CW4_BASE_ADDRESS, cw4->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
233
DMCUB_REGION3_CW4_TOP_ADDRESS, cw4->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
240
REG_WRITE(DMCUB_REGION3_CW5_BASE_ADDRESS, cw5->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
242
DMCUB_REGION3_CW5_TOP_ADDRESS, cw5->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
249
cw5->region.top - cw5->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
256
REG_WRITE(DMCUB_REGION3_CW6_BASE_ADDRESS, cw6->region.base);
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
258
DMCUB_REGION3_CW6_TOP_ADDRESS, cw6->region.top,
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
267
region6->region.top - region6->region.base - 1,
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
551
struct dmub_region *region = &out->regions[i];
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
553
region->base = dmub_align(top, 256);
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
554
region->top = region->base + dmub_align(window_sizes[i], 64);
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
555
top = region->top;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
722
cw0.region.base = DMUB_CW0_BASE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
723
cw0.region.top = cw0.region.base + inst_fb->size - 1;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
726
cw1.region.base = DMUB_CW1_BASE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
727
cw1.region.top = cw1.region.base + stack_fb->size - 1;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
748
cw2.region.base = DMUB_CW0_BASE + inst_fb->size;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
749
cw2.region.top = cw2.region.base + data_fb->size;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
752
cw3.region.base = DMUB_CW3_BASE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
753
cw3.region.top = cw3.region.base + bios_fb->size;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
756
cw4.region.base = DMUB_CW4_BASE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
757
cw4.region.top = cw4.region.base + mail_fb->size;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
766
inbox1.base = cw4.region.base;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
767
inbox1.top = cw4.region.base + DMUB_RB_SIZE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
772
cw5.region.base = DMUB_CW5_BASE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
773
cw5.region.top = cw5.region.base + tracebuff_fb->size;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
779
cw6.region.base = DMUB_CW6_BASE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
780
cw6.region.top = cw6.region.base + fw_state_fb->size;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
785
region6.region.base = DMUB_CW6_BASE;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
786
region6.region.top = region6.region.base + shared_state_fb->size;
drivers/gpu/drm/drm_drv.c
965
struct dmem_cgroup_region *region;
drivers/gpu/drm/drm_drv.c
968
region = dmem_cgroup_register_region(size, "drm/%s/%s", dev->unique, region_name);
drivers/gpu/drm/drm_drv.c
969
if (IS_ERR_OR_NULL(region))
drivers/gpu/drm/drm_drv.c
970
return region;
drivers/gpu/drm/drm_drv.c
972
ret = drmm_add_action_or_reset(dev, drmm_cg_unregister_region, region);
drivers/gpu/drm/drm_drv.c
976
return region;
drivers/gpu/drm/drm_file.c
848
const char *region,
drivers/gpu/drm/drm_file.c
861
prefix, stat, region, sz, units[u]);
drivers/gpu/drm/drm_file.c
886
const char *region)
drivers/gpu/drm/drm_file.c
890
drm_fdinfo_print_size(p, prefix, "total", region,
drivers/gpu/drm/drm_file.c
892
drm_fdinfo_print_size(p, prefix, "shared", region, stats->shared);
drivers/gpu/drm/drm_file.c
895
drm_fdinfo_print_size(p, prefix, "active", region, stats->active);
drivers/gpu/drm/drm_file.c
898
drm_fdinfo_print_size(p, prefix, "resident", region,
drivers/gpu/drm/drm_file.c
902
drm_fdinfo_print_size(p, prefix, "purgeable", region,
drivers/gpu/drm/i915/display/intel_fbdev_fb.c
67
struct intel_memory_region *mem = obj->mm.region;
drivers/gpu/drm/i915/display/intel_fbdev_fb.c
73
mem->region.start);
drivers/gpu/drm/i915/gem/i915_gem_create.c
304
struct drm_i915_gem_memory_class_instance region;
drivers/gpu/drm/i915/gem/i915_gem_create.c
307
if (copy_from_user(&region, uregions, sizeof(region)))
drivers/gpu/drm/i915/gem/i915_gem_create.c
311
region.memory_class,
drivers/gpu/drm/i915/gem/i915_gem_create.c
312
region.memory_instance);
drivers/gpu/drm/i915/gem/i915_gem_create.c
315
region.memory_class, region.memory_instance, i);
drivers/gpu/drm/i915/gem/i915_gem_create.c
322
mr->name, region.memory_class,
drivers/gpu/drm/i915/gem/i915_gem_create.c
323
region.memory_instance, i);
drivers/gpu/drm/i915/gem/i915_gem_lmem.c
23
offset -= obj->mm.region->region.start;
drivers/gpu/drm/i915/gem/i915_gem_lmem.c
25
return io_mapping_map_wc(&obj->mm.region->iomap, offset, size);
drivers/gpu/drm/i915/gem/i915_gem_lmem.c
44
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
275
iomap = obj->mm.region->iomap.base;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
276
iomap -= obj->mm.region->region.start;
drivers/gpu/drm/i915/gem/i915_gem_object.c
524
src_map = io_mapping_map_wc(&obj->mm.region->iomap,
drivers/gpu/drm/i915/gem/i915_gem_object.c
525
dma - obj->mm.region->region.start,
drivers/gpu/drm/i915/gem/i915_gem_object.c
622
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
drivers/gpu/drm/i915/gem/i915_gem_object.c
703
if (obj->mm.region == mr)
drivers/gpu/drm/i915/gem/i915_gem_object.c
804
if (GEM_WARN_ON(obj->mm.region != mr))
drivers/gpu/drm/i915/gem/i915_gem_object_types.h
650
struct intel_memory_region *region;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
335
resource_size_t iomap = obj->mm.region->iomap.base -
drivers/gpu/drm/i915/gem/i915_gem_pages.c
336
obj->mm.region->region.start;
drivers/gpu/drm/i915/gem/i915_gem_region.c
128
if (range_overflows(offset, size, resource_size(&mem->region)))
drivers/gpu/drm/i915/gem/i915_gem_region.c
16
obj->mm.region = mem;
drivers/gpu/drm/i915/gem/i915_gem_region.c
194
if (obj->mm.region == mr)
drivers/gpu/drm/i915/gem/i915_gem_region.c
25
struct intel_memory_region *mem = obj->mm.region;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
216
struct intel_memory_region *mem = obj->mm.region;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
85
if (size > resource_size(&mr->region))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
556
if (adjust_stolen(i915, &mem->region))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
559
if (request_smem_stolen(i915, &mem->region))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
562
i915->dsm.stolen = mem->region;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
568
mem->region.end = i915->dsm.reserved.start - 1;
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
571
resource_size(&mem->region)));
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
573
i915->dsm.usable_size = resource_size(&mem->region);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
858
if (GEM_WARN_ON(resource_size(&mem->region) == 0))
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
1306
obj->mm.region = mem;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
168
obj->mm.region, &places[0], obj->bo_offset,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
576
page_alignment = obj->mm.region->min_page_size;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
587
rsgt = intel_region_ttm_resource_to_rsgt(obj->mm.region,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
598
return intel_region_ttm_resource_to_rsgt(obj->mm.region, res,
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
697
base = obj->mm.region->iomap.base - obj->mm.region->region.start;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
708
resource_size_t iomap = obj->mm.region->iomap.base -
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
709
obj->mm.region->region.start;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
892
if (obj->mm.region != mr) {
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
134
if (intel_region_to_ttm_type(obj->mm.region) != mem_type) {
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
139
mr != obj->mm.region) {
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
327
&dst_rsgt->table, dst_reg->region.start);
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
333
src_reg->region.start);
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1718
addr = round_up(addr, obj->mm.region->min_page_size);
drivers/gpu/drm/i915/gem/selftests/i915_gem_dmabuf.c
278
if (obj->mm.region != i915->mm.regions[INTEL_REGION_SMEM]) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1104
obj->mm.region->name, I915_MMAP_TYPE_FIXED, addr);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1113
obj->mm.region->name, i * sizeof(x));
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1122
obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1128
obj->mm.region->name,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1136
obj->mm.region->name, i * sizeof(x));
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1198
obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1257
if (!err && obj->mm.region != expected_mr) {
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1458
obj->mm.region->name, repr_mmap_type(type));
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1467
obj->mm.region->name, repr_mmap_type(type));
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1474
obj->mm.region->name, repr_mmap_type(type));
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1483
obj->mm.region->name, repr_mmap_type(type));
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1489
obj->mm.region->name, repr_mmap_type(type),
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1575
pr_err("%s: Unable to write to mmap\n", obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1619
__func__, engine->name, obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1761
pr_err("%s: was not present\n", obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
1791
pr_err("%s: was not absent\n", obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
829
obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
865
obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
923
pr_debug("igt_mmap(%s, %d) @ %lx\n", obj->mm.region->name, type, addr);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
930
obj->mm.region->name);
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
941
obj->mm.region->name, i * sizeof(x));
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
948
obj->mm.region->name,
drivers/gpu/drm/i915/gem/selftests/i915_gem_mman.c
957
obj->mm.region->name, i * sizeof(x));
drivers/gpu/drm/i915/gt/intel_gtt.h
452
struct intel_memory_region *mr = READ_ONCE(obj->mm.region);
drivers/gpu/drm/i915/gt/selftest_tlb.c
151
ce->engine->name, va->obj->mm.region->name ?: "smem",
drivers/gpu/drm/i915/gt/selftest_tlb.c
95
ce->engine->name, va->obj->mm.region->name ?: "smem",
drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c
1279
struct intel_memory_region *mr = uc_fw->obj->mm.region;
drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c
1324
addr - mr->region.start);
drivers/gpu/drm/i915/gvt/gvt.h
222
struct vfio_region *region;
drivers/gpu/drm/i915/gvt/kvmgt.c
1229
info->size = vgpu->region[i].size;
drivers/gpu/drm/i915/gvt/kvmgt.c
1230
info->flags = vgpu->region[i].flags;
drivers/gpu/drm/i915/gvt/kvmgt.c
1232
cap_type.type = vgpu->region[i].type;
drivers/gpu/drm/i915/gvt/kvmgt.c
1233
cap_type.subtype = vgpu->region[i].subtype;
drivers/gpu/drm/i915/gvt/kvmgt.c
1597
if (!vgpu->region)
drivers/gpu/drm/i915/gvt/kvmgt.c
1601
if (vgpu->region[i].ops->release)
drivers/gpu/drm/i915/gvt/kvmgt.c
1602
vgpu->region[i].ops->release(vgpu,
drivers/gpu/drm/i915/gvt/kvmgt.c
1603
&vgpu->region[i]);
drivers/gpu/drm/i915/gvt/kvmgt.c
1605
kfree(vgpu->region);
drivers/gpu/drm/i915/gvt/kvmgt.c
1606
vgpu->region = NULL;
drivers/gpu/drm/i915/gvt/kvmgt.c
405
void *base = vgpu->region[i].data;
drivers/gpu/drm/i915/gvt/kvmgt.c
409
if (pos >= vgpu->region[i].size || iswrite) {
drivers/gpu/drm/i915/gvt/kvmgt.c
413
count = min(count, (size_t)(vgpu->region[i].size - pos));
drivers/gpu/drm/i915/gvt/kvmgt.c
420
struct vfio_region *region)
drivers/gpu/drm/i915/gvt/kvmgt.c
442
struct vfio_edid_region *region, char *buf,
drivers/gpu/drm/i915/gvt/kvmgt.c
445
struct vfio_region_gfx_edid *regs = &region->vfio_edid_regs;
drivers/gpu/drm/i915/gvt/kvmgt.c
459
if (!edid_valid(region->edid_blob, EDID_SIZE)) {
drivers/gpu/drm/i915/gvt/kvmgt.c
494
static int handle_edid_blob(struct vfio_edid_region *region, char *buf,
drivers/gpu/drm/i915/gvt/kvmgt.c
497
if (offset + count > region->vfio_edid_regs.edid_size)
drivers/gpu/drm/i915/gvt/kvmgt.c
501
memcpy(region->edid_blob + offset, buf, count);
drivers/gpu/drm/i915/gvt/kvmgt.c
503
memcpy(buf, region->edid_blob + offset, count);
drivers/gpu/drm/i915/gvt/kvmgt.c
514
struct vfio_edid_region *region = vgpu->region[i].data;
drivers/gpu/drm/i915/gvt/kvmgt.c
517
if (pos < region->vfio_edid_regs.edid_offset) {
drivers/gpu/drm/i915/gvt/kvmgt.c
518
ret = handle_edid_regs(vgpu, region, buf, count, pos, iswrite);
drivers/gpu/drm/i915/gvt/kvmgt.c
521
ret = handle_edid_blob(region, buf, count, pos, iswrite);
drivers/gpu/drm/i915/gvt/kvmgt.c
531
struct vfio_region *region)
drivers/gpu/drm/i915/gvt/kvmgt.c
533
kfree(region->data);
drivers/gpu/drm/i915/gvt/kvmgt.c
546
struct vfio_region *region;
drivers/gpu/drm/i915/gvt/kvmgt.c
548
region = krealloc(vgpu->region,
drivers/gpu/drm/i915/gvt/kvmgt.c
549
(vgpu->num_regions + 1) * sizeof(*region),
drivers/gpu/drm/i915/gvt/kvmgt.c
551
if (!region)
drivers/gpu/drm/i915/gvt/kvmgt.c
554
vgpu->region = region;
drivers/gpu/drm/i915/gvt/kvmgt.c
555
vgpu->region[vgpu->num_regions].type = type;
drivers/gpu/drm/i915/gvt/kvmgt.c
556
vgpu->region[vgpu->num_regions].subtype = subtype;
drivers/gpu/drm/i915/gvt/kvmgt.c
557
vgpu->region[vgpu->num_regions].ops = ops;
drivers/gpu/drm/i915/gvt/kvmgt.c
558
vgpu->region[vgpu->num_regions].size = size;
drivers/gpu/drm/i915/gvt/kvmgt.c
559
vgpu->region[vgpu->num_regions].flags = flags;
drivers/gpu/drm/i915/gvt/kvmgt.c
560
vgpu->region[vgpu->num_regions].data = data;
drivers/gpu/drm/i915/gvt/kvmgt.c
75
struct vfio_region *region);
drivers/gpu/drm/i915/gvt/kvmgt.c
841
return vgpu->region[index].ops->rw(vgpu, buf, count,
drivers/gpu/drm/i915/i915_drm_client.c
52
const enum intel_region_id id = obj->mm.region ?
drivers/gpu/drm/i915/i915_drm_client.c
53
obj->mm.region->id : INTEL_REGION_SMEM;
drivers/gpu/drm/i915/i915_gpu_error.c
1212
dma_addr_t offset = dma - mem->region.start;
drivers/gpu/drm/i915/i915_initial_plane.c
129
mem->region.name);
drivers/gpu/drm/i915/i915_initial_plane.c
65
mem->region.name);
drivers/gpu/drm/i915/i915_initial_plane.c
69
if (dma_addr < mem->region.start || dma_addr > mem->region.end) {
drivers/gpu/drm/i915/i915_initial_plane.c
72
&dma_addr, mem->region.name, &mem->region.start, &mem->region.end);
drivers/gpu/drm/i915/i915_initial_plane.c
79
plane_config->phys_base = dma_addr - mem->region.start;
drivers/gpu/drm/i915/i915_query.c
502
info.region.memory_class = mr->type;
drivers/gpu/drm/i915/i915_query.c
503
info.region.memory_instance = mr->instance;
drivers/gpu/drm/i915/i915_vma.c
456
i915_gem_object_is_lmem(obj), obj->mm.region,
drivers/gpu/drm/i915/intel_memory_region.c
266
mem->region = DEFINE_RES_MEM(start, size);
drivers/gpu/drm/i915/intel_memory_region.c
396
region_size = resource_size(&mem->region) >> 20;
drivers/gpu/drm/i915/intel_memory_region.c
401
mem->id, mem->name, region_size, &mem->region, io_size, &mem->io);
drivers/gpu/drm/i915/intel_memory_region.c
404
mem->id, mem->name, region_size, &mem->region);
drivers/gpu/drm/i915/intel_memory_region.c
419
struct intel_memory_region *region =
drivers/gpu/drm/i915/intel_memory_region.c
422
if (region)
drivers/gpu/drm/i915/intel_memory_region.c
423
intel_memory_region_destroy(region);
drivers/gpu/drm/i915/intel_memory_region.c
54
&mem->region, &mem->io.start, &offset, caller,
drivers/gpu/drm/i915/intel_memory_region.h
65
struct resource region;
drivers/gpu/drm/i915/intel_region_ttm.c
172
mem->region.start,
drivers/gpu/drm/i915/intel_region_ttm.c
175
return i915_rsgt_from_buddy_resource(res, mem->region.start,
drivers/gpu/drm/i915/intel_region_ttm.c
89
resource_size(&mem->region),
drivers/gpu/drm/i915/selftests/i915_perf_selftests.h
21
selftest(region, intel_memory_region_perf_selftests)
drivers/gpu/drm/i915/selftests/intel_memory_region.c
160
resource_size_t avail = resource_size(&mem->region);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
249
total = resource_size(&mem->region);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
527
struct intel_memory_region *mr = obj->mm.region;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
57
resource_size_t total = resource_size(&mem->region);
drivers/gpu/drm/i915/selftests/mock_region.c
20
intel_region_ttm_resource_free(obj->mm.region, obj->mm.res);
drivers/gpu/drm/i915/selftests/mock_region.c
28
obj->mm.res = intel_region_ttm_resource_alloc(obj->mm.region,
drivers/gpu/drm/i915/selftests/mock_region.c
35
obj->mm.rsgt = intel_region_ttm_resource_to_rsgt(obj->mm.region,
drivers/gpu/drm/i915/selftests/mock_region.c
37
obj->mm.region->min_page_size);
drivers/gpu/drm/i915/selftests/mock_region.c
49
intel_region_ttm_resource_free(obj->mm.region, obj->mm.res);
drivers/gpu/drm/i915/selftests/mock_region.c
70
if (size > resource_size(&mem->region))
drivers/gpu/drm/nouveau/include/nvkm/core/engine.h
27
void (*tile)(struct nvkm_engine *, int region, struct nvkm_fb_tile *);
drivers/gpu/drm/nouveau/include/nvkm/core/engine.h
53
void nvkm_engine_tile(struct nvkm_engine *, int region);
drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h
52
struct nvkm_fb_tile region[16];
drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h
65
void nvkm_fb_tile_init(struct nvkm_fb *, int region, u32 addr, u32 size,
drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h
67
void nvkm_fb_tile_fini(struct nvkm_fb *, int region, struct nvkm_fb_tile *);
drivers/gpu/drm/nouveau/include/nvkm/subdev/fb.h
68
void nvkm_fb_tile_prog(struct nvkm_fb *, int region, struct nvkm_fb_tile *);
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
117
} region[16];
drivers/gpu/drm/nouveau/nouveau_bo.c
123
} else if (tile && fb->tile.region[i].pitch) {
drivers/gpu/drm/nouveau/nouveau_bo.c
62
struct nvkm_fb_tile *tile = &fb->tile.region[i];
drivers/gpu/drm/nouveau/nouveau_uvmm.c
210
bool sparse = !!uvma->region;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
656
uvma->region = args->region;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
715
.region = uvma_from_va(va)->region,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
802
struct nouveau_uvma_region *region,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
807
.region = region,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
84
struct nouveau_uvma_region *region;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
869
bool sparse = !!uvma->region;
drivers/gpu/drm/nouveau/nouveau_uvmm.h
34
struct nouveau_uvma_region *region;
drivers/gpu/drm/nouveau/nvkm/core/engine.c
74
nvkm_engine_tile(struct nvkm_engine *engine, int region)
drivers/gpu/drm/nouveau/nvkm/core/engine.c
78
engine->func->tile(engine, region, &fb->tile.region[region]);
drivers/gpu/drm/nouveau/nvkm/engine/gr/base.c
65
nvkm_gr_tile(struct nvkm_engine *engine, int region, struct nvkm_fb_tile *tile)
drivers/gpu/drm/nouveau/nvkm/engine/gr/base.c
69
gr->func->tile(gr, region, tile);
drivers/gpu/drm/nouveau/nvkm/engine/gr/priv.h
23
void (*tile)(struct nvkm_gr *, int region, struct nvkm_fb_tile *);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
202
fb->func->tile.prog(fb, i, &fb->tile.region[i]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
241
fb->func->tile.fini(fb, i, &fb->tile.region[i]);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
35
nvkm_fb_tile_fini(struct nvkm_fb *fb, int region, struct nvkm_fb_tile *tile)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
37
fb->func->tile.fini(fb, region, tile);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
41
nvkm_fb_tile_init(struct nvkm_fb *fb, int region, u32 addr, u32 size,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
44
fb->func->tile.init(fb, region, addr, size, pitch, flags, tile);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
48
nvkm_fb_tile_prog(struct nvkm_fb *fb, int region, struct nvkm_fb_tile *tile)
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
52
fb->func->tile.prog(fb, region, tile);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
54
nvkm_engine_tile(&device->gr->engine, region);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/base.c
56
nvkm_engine_tile(device->mpeg, region);
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
50
gsp->fb.region[i].addr >> NVKM_RAM_MM_SHIFT,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/r535.c
51
gsp->fb.region[i].size >> NVKM_RAM_MM_SHIFT,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
193
!WARN_ON_ONCE(gsp->fb.region_nr >= ARRAY_SIZE(gsp->fb.region))) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
196
gsp->fb.region[gsp->fb.region_nr].addr = reg->base;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
197
gsp->fb.region[gsp->fb.region_nr].size = size;
drivers/gpu/drm/panfrost/panfrost_mmu.c
107
u64 region;
drivers/gpu/drm/panfrost/panfrost_mmu.c
130
region = region_width | region_start;
drivers/gpu/drm/panfrost/panfrost_mmu.c
133
mmu_write(pfdev, AS_LOCKADDR_LO(as_nr), lower_32_bits(region));
drivers/gpu/drm/panfrost/panfrost_mmu.c
134
mmu_write(pfdev, AS_LOCKADDR_HI(as_nr), upper_32_bits(region));
drivers/gpu/drm/xe/tests/xe_migrate.c
173
u32 region;
drivers/gpu/drm/xe/tests/xe_migrate.c
179
region = XE_BO_FLAG_VRAM1;
drivers/gpu/drm/xe/tests/xe_migrate.c
181
region = XE_BO_FLAG_VRAM0;
drivers/gpu/drm/xe/tests/xe_migrate.c
182
test_copy(m, bo, test, region, exec);
drivers/gpu/drm/xe/tests/xe_migrate.c
73
struct kunit *test, u32 region, struct drm_exec *exec)
drivers/gpu/drm/xe/tests/xe_migrate.c
85
region |
drivers/gpu/drm/xe/xe_vm.c
2349
op->prefetch.region = prefetch_region;
drivers/gpu/drm/xe/xe_vm.c
3026
u32 region;
drivers/gpu/drm/xe/xe_vm.c
3029
region = op->prefetch.region;
drivers/gpu/drm/xe/xe_vm.c
3030
xe_assert(vm->xe, region == DRM_XE_CONSULT_MEM_ADVISE_PREF_LOC ||
drivers/gpu/drm/xe/xe_vm.c
3031
region <= ARRAY_SIZE(region_to_mem_type));
drivers/gpu/drm/xe/xe_vm.c
3039
region_to_mem_type[region],
drivers/gpu/drm/xe/xe_vm_types.h
391
u32 region;
drivers/hv/mshv_regions.c
101
page = region->mreg_pages[page_offset + count];
drivers/hv/mshv_regions.c
113
ret = handler(region, flags, page_offset, count, stride > 1);
drivers/hv/mshv_regions.c
139
static int mshv_region_process_range(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
142
int (*handler)(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
150
if (page_offset + page_count > region->nr_pages)
drivers/hv/mshv_regions.c
155
if (!region->mreg_pages[page_offset]) {
drivers/hv/mshv_regions.c
161
ret = mshv_region_process_chunk(region, flags,
drivers/hv/mshv_regions.c
178
struct mshv_mem_region *region;
drivers/hv/mshv_regions.c
180
region = vzalloc(sizeof(*region) + sizeof(struct page *) * nr_pages);
drivers/hv/mshv_regions.c
181
if (!region)
drivers/hv/mshv_regions.c
184
region->nr_pages = nr_pages;
drivers/hv/mshv_regions.c
185
region->start_gfn = guest_pfn;
drivers/hv/mshv_regions.c
186
region->start_uaddr = uaddr;
drivers/hv/mshv_regions.c
187
region->hv_map_flags = HV_MAP_GPA_READABLE | HV_MAP_GPA_ADJUSTABLE;
drivers/hv/mshv_regions.c
189
region->hv_map_flags |= HV_MAP_GPA_WRITABLE;
drivers/hv/mshv_regions.c
191
region->hv_map_flags |= HV_MAP_GPA_EXECUTABLE;
drivers/hv/mshv_regions.c
193
kref_init(&region->mreg_refcount);
drivers/hv/mshv_regions.c
195
return region;
drivers/hv/mshv_regions.c
198
static int mshv_region_chunk_share(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
206
return hv_call_modify_spa_host_access(region->partition->pt_id,
drivers/hv/mshv_regions.c
207
region->mreg_pages + page_offset,
drivers/hv/mshv_regions.c
214
int mshv_region_share(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
218
return mshv_region_process_range(region, flags,
drivers/hv/mshv_regions.c
219
0, region->nr_pages,
drivers/hv/mshv_regions.c
223
static int mshv_region_chunk_unshare(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
231
return hv_call_modify_spa_host_access(region->partition->pt_id,
drivers/hv/mshv_regions.c
232
region->mreg_pages + page_offset,
drivers/hv/mshv_regions.c
237
int mshv_region_unshare(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
241
return mshv_region_process_range(region, flags,
drivers/hv/mshv_regions.c
242
0, region->nr_pages,
drivers/hv/mshv_regions.c
246
static int mshv_region_chunk_remap(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
254
return hv_call_map_gpa_pages(region->partition->pt_id,
drivers/hv/mshv_regions.c
255
region->start_gfn + page_offset,
drivers/hv/mshv_regions.c
257
region->mreg_pages + page_offset);
drivers/hv/mshv_regions.c
260
static int mshv_region_remap_pages(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
264
return mshv_region_process_range(region, map_flags,
drivers/hv/mshv_regions.c
269
int mshv_region_map(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
271
u32 map_flags = region->hv_map_flags;
drivers/hv/mshv_regions.c
273
return mshv_region_remap_pages(region, map_flags,
drivers/hv/mshv_regions.c
274
0, region->nr_pages);
drivers/hv/mshv_regions.c
277
static void mshv_region_invalidate_pages(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
280
if (region->mreg_type == MSHV_REGION_TYPE_MEM_PINNED)
drivers/hv/mshv_regions.c
281
unpin_user_pages(region->mreg_pages + page_offset, page_count);
drivers/hv/mshv_regions.c
283
memset(region->mreg_pages + page_offset, 0,
drivers/hv/mshv_regions.c
287
void mshv_region_invalidate(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
289
mshv_region_invalidate_pages(region, 0, region->nr_pages);
drivers/hv/mshv_regions.c
292
int mshv_region_pin(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
299
for (done_count = 0; done_count < region->nr_pages; done_count += ret) {
drivers/hv/mshv_regions.c
300
pages = region->mreg_pages + done_count;
drivers/hv/mshv_regions.c
301
userspace_addr = region->start_uaddr +
drivers/hv/mshv_regions.c
303
nr_pages = min(region->nr_pages - done_count,
drivers/hv/mshv_regions.c
326
mshv_region_invalidate_pages(region, 0, done_count);
drivers/hv/mshv_regions.c
330
static int mshv_region_chunk_unmap(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
338
return hv_call_unmap_gpa_pages(region->partition->pt_id,
drivers/hv/mshv_regions.c
339
region->start_gfn + page_offset,
drivers/hv/mshv_regions.c
343
static int mshv_region_unmap(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
345
return mshv_region_process_range(region, 0,
drivers/hv/mshv_regions.c
346
0, region->nr_pages,
drivers/hv/mshv_regions.c
352
struct mshv_mem_region *region =
drivers/hv/mshv_regions.c
354
struct mshv_partition *partition = region->partition;
drivers/hv/mshv_regions.c
357
if (region->mreg_type == MSHV_REGION_TYPE_MEM_MOVABLE)
drivers/hv/mshv_regions.c
358
mshv_region_movable_fini(region);
drivers/hv/mshv_regions.c
361
ret = mshv_region_share(region);
drivers/hv/mshv_regions.c
370
mshv_region_unmap(region);
drivers/hv/mshv_regions.c
372
mshv_region_invalidate(region);
drivers/hv/mshv_regions.c
374
vfree(region);
drivers/hv/mshv_regions.c
377
void mshv_region_put(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
379
kref_put(&region->mreg_refcount, mshv_region_destroy);
drivers/hv/mshv_regions.c
382
int mshv_region_get(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
384
return kref_get_unless_zero(&region->mreg_refcount);
drivers/hv/mshv_regions.c
404
static int mshv_region_hmm_fault_and_lock(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
410
mmap_read_lock(region->mreg_mni.mm);
drivers/hv/mshv_regions.c
412
mmap_read_unlock(region->mreg_mni.mm);
drivers/hv/mshv_regions.c
416
mutex_lock(&region->mreg_mutex);
drivers/hv/mshv_regions.c
419
mutex_unlock(&region->mreg_mutex);
drivers/hv/mshv_regions.c
439
static int mshv_region_range_fault(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
443
.notifier = &region->mreg_mni,
drivers/hv/mshv_regions.c
455
range.start = region->start_uaddr + page_offset * HV_HYP_PAGE_SIZE;
drivers/hv/mshv_regions.c
459
ret = mshv_region_hmm_fault_and_lock(region, &range);
drivers/hv/mshv_regions.c
466
region->mreg_pages[page_offset + i] = hmm_pfn_to_page(pfns[i]);
drivers/hv/mshv_regions.c
468
ret = mshv_region_remap_pages(region, region->hv_map_flags,
drivers/hv/mshv_regions.c
471
mutex_unlock(&region->mreg_mutex);
drivers/hv/mshv_regions.c
477
bool mshv_region_handle_gfn_fault(struct mshv_mem_region *region, u64 gfn)
drivers/hv/mshv_regions.c
483
page_offset = ALIGN_DOWN(gfn - region->start_gfn,
drivers/hv/mshv_regions.c
487
page_count = min(region->nr_pages - page_offset,
drivers/hv/mshv_regions.c
490
ret = mshv_region_range_fault(region, page_offset, page_count);
drivers/hv/mshv_regions.c
494
region->partition->pt_id, region->start_uaddr,
drivers/hv/mshv_regions.c
495
region->start_uaddr + (region->nr_pages << HV_HYP_PAGE_SHIFT),
drivers/hv/mshv_regions.c
523
struct mshv_mem_region *region = container_of(mni,
drivers/hv/mshv_regions.c
530
mstart = max(range->start, region->start_uaddr);
drivers/hv/mshv_regions.c
531
mend = min(range->end, region->start_uaddr +
drivers/hv/mshv_regions.c
532
(region->nr_pages << HV_HYP_PAGE_SHIFT));
drivers/hv/mshv_regions.c
534
page_offset = HVPFN_DOWN(mstart - region->start_uaddr);
drivers/hv/mshv_regions.c
538
mutex_lock(&region->mreg_mutex);
drivers/hv/mshv_regions.c
539
else if (!mutex_trylock(&region->mreg_mutex))
drivers/hv/mshv_regions.c
544
ret = mshv_region_remap_pages(region, HV_MAP_GPA_NO_ACCESS,
drivers/hv/mshv_regions.c
549
mshv_region_invalidate_pages(region, page_offset, page_count);
drivers/hv/mshv_regions.c
551
mutex_unlock(&region->mreg_mutex);
drivers/hv/mshv_regions.c
556
mutex_unlock(&region->mreg_mutex);
drivers/hv/mshv_regions.c
560
region->start_uaddr,
drivers/hv/mshv_regions.c
561
region->start_uaddr + (region->nr_pages << HV_HYP_PAGE_SHIFT),
drivers/hv/mshv_regions.c
571
void mshv_region_movable_fini(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
573
mmu_interval_notifier_remove(&region->mreg_mni);
drivers/hv/mshv_regions.c
576
bool mshv_region_movable_init(struct mshv_mem_region *region)
drivers/hv/mshv_regions.c
580
ret = mmu_interval_notifier_insert(&region->mreg_mni, current->mm,
drivers/hv/mshv_regions.c
581
region->start_uaddr,
drivers/hv/mshv_regions.c
582
region->nr_pages << HV_HYP_PAGE_SHIFT,
drivers/hv/mshv_regions.c
587
mutex_init(&region->mreg_mutex);
drivers/hv/mshv_regions.c
77
static long mshv_region_process_chunk(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
80
int (*handler)(struct mshv_mem_region *region,
drivers/hv/mshv_regions.c
86
u64 gfn = region->start_gfn + page_offset;
drivers/hv/mshv_regions.c
91
page = region->mreg_pages[page_offset];
drivers/hv/mshv_root.h
369
int mshv_region_share(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
370
int mshv_region_unshare(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
371
int mshv_region_map(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
372
void mshv_region_invalidate(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
373
int mshv_region_pin(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
374
void mshv_region_put(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
375
int mshv_region_get(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
376
bool mshv_region_handle_gfn_fault(struct mshv_mem_region *region, u64 gfn);
drivers/hv/mshv_root.h
377
void mshv_region_movable_fini(struct mshv_mem_region *region);
drivers/hv/mshv_root.h
378
bool mshv_region_movable_init(struct mshv_mem_region *region);
drivers/hv/mshv_root_main.c
1231
static int mshv_prepare_pinned_region(struct mshv_mem_region *region)
drivers/hv/mshv_root_main.c
1233
struct mshv_partition *partition = region->partition;
drivers/hv/mshv_root_main.c
1236
ret = mshv_region_pin(region);
drivers/hv/mshv_root_main.c
1251
ret = mshv_region_unshare(region);
drivers/hv/mshv_root_main.c
1255
region->start_gfn, ret);
drivers/hv/mshv_root_main.c
1260
ret = mshv_region_map(region);
drivers/hv/mshv_root_main.c
1264
shrc = mshv_region_share(region);
drivers/hv/mshv_root_main.c
1270
region->start_gfn, shrc);
drivers/hv/mshv_root_main.c
1281
mshv_region_invalidate(region);
drivers/hv/mshv_root_main.c
1302
struct mshv_mem_region *region;
drivers/hv/mshv_root_main.c
1321
ret = mshv_partition_create_region(partition, mem, &region,
drivers/hv/mshv_root_main.c
1326
switch (region->mreg_type) {
drivers/hv/mshv_root_main.c
1328
ret = mshv_prepare_pinned_region(region);
drivers/hv/mshv_root_main.c
1337
region->start_gfn,
drivers/hv/mshv_root_main.c
1338
region->nr_pages,
drivers/hv/mshv_root_main.c
1343
region->start_gfn,
drivers/hv/mshv_root_main.c
1345
region->nr_pages);
drivers/hv/mshv_root_main.c
1353
hlist_add_head(&region->hnode, &partition->pt_mem_regions);
drivers/hv/mshv_root_main.c
1359
mshv_region_put(region);
drivers/hv/mshv_root_main.c
1368
struct mshv_mem_region *region;
drivers/hv/mshv_root_main.c
1375
region = mshv_partition_region_by_gfn(partition, mem->guest_pfn);
drivers/hv/mshv_root_main.c
1376
if (!region) {
drivers/hv/mshv_root_main.c
1382
if (region->start_uaddr != mem->userspace_addr ||
drivers/hv/mshv_root_main.c
1383
region->start_gfn != mem->guest_pfn ||
drivers/hv/mshv_root_main.c
1384
region->nr_pages != HVPFN_DOWN(mem->size)) {
drivers/hv/mshv_root_main.c
1389
hlist_del(&region->hnode);
drivers/hv/mshv_root_main.c
1393
mshv_region_put(region);
drivers/hv/mshv_root_main.c
1735
struct mshv_mem_region *region;
drivers/hv/mshv_root_main.c
1803
hlist_for_each_entry_safe(region, n, &partition->pt_mem_regions,
drivers/hv/mshv_root_main.c
1805
hlist_del(&region->hnode);
drivers/hv/mshv_root_main.c
1806
mshv_region_put(region);
drivers/hv/mshv_root_main.c
591
struct mshv_mem_region *region;
drivers/hv/mshv_root_main.c
593
hlist_for_each_entry(region, &partition->pt_mem_regions, hnode) {
drivers/hv/mshv_root_main.c
594
if (gfn >= region->start_gfn &&
drivers/hv/mshv_root_main.c
595
gfn < region->start_gfn + region->nr_pages)
drivers/hv/mshv_root_main.c
596
return region;
drivers/hv/mshv_root_main.c
605
struct mshv_mem_region *region;
drivers/hv/mshv_root_main.c
608
region = mshv_partition_region_by_gfn(p, gfn);
drivers/hv/mshv_root_main.c
609
if (!region || !mshv_region_get(region)) {
drivers/hv/mshv_root_main.c
615
return region;
drivers/hv/mshv_root_main.c
632
struct mshv_mem_region *region;
drivers/hv/mshv_root_main.c
649
region = mshv_partition_region_by_gfn_get(p, gfn);
drivers/hv/mshv_root_main.c
650
if (!region)
drivers/hv/mshv_root_main.c
654
!(region->hv_map_flags & HV_MAP_GPA_WRITABLE))
drivers/hv/mshv_root_main.c
658
!(region->hv_map_flags & HV_MAP_GPA_EXECUTABLE))
drivers/hv/mshv_root_main.c
662
if (region->mreg_type == MSHV_REGION_TYPE_MEM_MOVABLE)
drivers/hv/mshv_root_main.c
663
ret = mshv_region_handle_gfn_fault(region, gfn);
drivers/hv/mshv_root_main.c
666
mshv_region_put(region);
drivers/hwmon/gpd-fan.c
604
const struct resource *region;
drivers/hwmon/gpd-fan.c
613
region = devm_request_region(dev, res->start,
drivers/hwmon/gpd-fan.c
615
if (!region)
drivers/infiniband/hw/hns/hns_roce_cq.c
261
buf_attr.region[0].size = hr_cq->cq_depth * hr_cq->cqe_size;
drivers/infiniband/hw/hns/hns_roce_cq.c
262
buf_attr.region[0].hopnum = hr_dev->caps.cqe_hop_num;
drivers/infiniband/hw/hns/hns_roce_device.h
300
} region[HNS_ROCE_MAX_BT_REGION];
drivers/infiniband/hw/hns/hns_roce_device.h
316
struct hns_roce_buf_region region[HNS_ROCE_MAX_BT_REGION];
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
6909
buf_attr.region[0].size = eq->entries * eq->eqe_size;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
6910
buf_attr.region[0].hopnum = eq->hop_num;
drivers/infiniband/hw/hns/hns_roce_mr.c
1007
cfg->region, cfg->region_count,
drivers/infiniband/hw/hns/hns_roce_mr.c
102
buf_attr.region[0].size = mr->size;
drivers/infiniband/hw/hns/hns_roce_mr.c
103
buf_attr.region[0].hopnum = mr->pbl_hop_num;
drivers/infiniband/hw/hns/hns_roce_mr.c
122
mr->pbl_hop_num = buf_attr.region[0].hopnum;
drivers/infiniband/hw/hns/hns_roce_mr.c
473
mtr->hem_cfg.region[0].offset = 0;
drivers/infiniband/hw/hns/hns_roce_mr.c
474
mtr->hem_cfg.region[0].count = mr->npages;
drivers/infiniband/hw/hns/hns_roce_mr.c
475
mtr->hem_cfg.region[0].hopnum = mr->pbl_hop_num;
drivers/infiniband/hw/hns/hns_roce_mr.c
493
struct hns_roce_buf_region *region, dma_addr_t *pages,
drivers/infiniband/hw/hns/hns_roce_mr.c
502
offset = region->offset;
drivers/infiniband/hw/hns/hns_roce_mr.c
503
end = offset + region->count;
drivers/infiniband/hw/hns/hns_roce_mr.c
529
if (attr->region[i].hopnum != HNS_ROCE_HOP_NUM_0 &&
drivers/infiniband/hw/hns/hns_roce_mr.c
530
attr->region[i].hopnum > 0)
drivers/infiniband/hw/hns/hns_roce_mr.c
546
size += attr->region[i].size;
drivers/infiniband/hw/hns/hns_roce_mr.c
619
struct hns_roce_buf_region *region;
drivers/infiniband/hw/hns/hns_roce_mr.c
624
region = &mtr->hem_cfg.region[i];
drivers/infiniband/hw/hns/hns_roce_mr.c
625
page_cnt += region->count;
drivers/infiniband/hw/hns/hns_roce_mr.c
710
r = &mtr->hem_cfg.region[i];
drivers/infiniband/hw/hns/hns_roce_mr.c
755
npage = cfg->region[mtt_count].offset;
drivers/infiniband/hw/hns/hns_roce_mr.c
865
ba_cnt = DIV_ROUND_UP(buf_attr->region[0].size, buf_pg_sz);
drivers/infiniband/hw/hns/hns_roce_mr.c
882
buf_attr->region[0].hopnum = hop_num;
drivers/infiniband/hw/hns/hns_roce_mr.c
892
if (attr->region_count > ARRAY_SIZE(attr->region) ||
drivers/infiniband/hw/hns/hns_roce_mr.c
940
r = &cfg->region[i];
drivers/infiniband/hw/hns/hns_roce_mr.c
942
buf_size = hr_hw_page_align(attr->region[i].size + pgoff);
drivers/infiniband/hw/hns/hns_roce_mr.c
950
r->hopnum = to_hr_hem_hopnum(attr->region[i].hopnum, r->count);
drivers/infiniband/hw/hns/hns_roce_mr.c
979
re = &mtr->hem_cfg.region[i];
drivers/infiniband/hw/hns/hns_roce_qp.c
710
if (buf_size > 0 && idx < ARRAY_SIZE(buf_attr->region)) {
drivers/infiniband/hw/hns/hns_roce_qp.c
711
buf_attr->region[idx].size = buf_size;
drivers/infiniband/hw/hns/hns_roce_qp.c
712
buf_attr->region[idx].hopnum = hr_dev->caps.wqe_sq_hop_num;
drivers/infiniband/hw/hns/hns_roce_qp.c
721
if (buf_size > 0 && idx < ARRAY_SIZE(buf_attr->region)) {
drivers/infiniband/hw/hns/hns_roce_qp.c
722
buf_attr->region[idx].size = buf_size;
drivers/infiniband/hw/hns/hns_roce_qp.c
723
buf_attr->region[idx].hopnum = hr_dev->caps.wqe_sge_hop_num;
drivers/infiniband/hw/hns/hns_roce_qp.c
732
if (buf_size > 0 && idx < ARRAY_SIZE(buf_attr->region)) {
drivers/infiniband/hw/hns/hns_roce_qp.c
733
buf_attr->region[idx].size = buf_size;
drivers/infiniband/hw/hns/hns_roce_qp.c
734
buf_attr->region[idx].hopnum = hr_dev->caps.wqe_rq_hop_num;
drivers/infiniband/hw/hns/hns_roce_srq.c
176
buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt,
drivers/infiniband/hw/hns/hns_roce_srq.c
178
buf_attr.region[0].hopnum = hr_dev->caps.idx_hop_num;
drivers/infiniband/hw/hns/hns_roce_srq.c
231
buf_attr.region[0].size = to_hr_hem_entries_size(srq->wqe_cnt,
drivers/infiniband/hw/hns/hns_roce_srq.c
233
buf_attr.region[0].hopnum = hr_dev->caps.srqwqe_hop_num;
drivers/infiniband/hw/hns/hns_roce_trace.h
157
__entry->region0_size = attr->region[0].size;
drivers/infiniband/hw/hns/hns_roce_trace.h
158
__entry->region0_hopnum = attr->region[0].hopnum;
drivers/infiniband/hw/hns/hns_roce_trace.h
159
__entry->region1_size = attr->region[1].size;
drivers/infiniband/hw/hns/hns_roce_trace.h
160
__entry->region1_hopnum = attr->region[1].hopnum;
drivers/infiniband/hw/hns/hns_roce_trace.h
161
__entry->region2_size = attr->region[2].size;
drivers/infiniband/hw/hns/hns_roce_trace.h
162
__entry->region2_hopnum = attr->region[2].hopnum;
drivers/infiniband/hw/irdma/ig3rdma_hw.c
140
static void __iomem *__ig3rdma_get_reg_addr(struct irdma_mmio_region *region, u64 reg_offset)
drivers/infiniband/hw/irdma/ig3rdma_hw.c
142
if (reg_offset >= region->offset &&
drivers/infiniband/hw/irdma/ig3rdma_hw.c
143
reg_offset < (region->offset + region->len)) {
drivers/infiniband/hw/irdma/ig3rdma_hw.c
144
reg_offset -= region->offset;
drivers/infiniband/hw/irdma/ig3rdma_hw.c
146
return region->addr + reg_offset;
drivers/infiniband/hw/irdma/verbs.c
2787
struct ib_umem *region = iwmr->region;
drivers/infiniband/hw/irdma/verbs.c
2798
iwpbl->qp_mr.sq_page = sg_page(region->sgt_append.sgt.sgl);
drivers/infiniband/hw/irdma/verbs.c
2800
rdma_umem_for_each_dma_block(region, &biter, iwmr->page_size) {
drivers/infiniband/hw/irdma/verbs.c
3363
static struct irdma_mr *irdma_alloc_iwmr(struct ib_umem *region,
drivers/infiniband/hw/irdma/verbs.c
3378
iwmr->region = region;
drivers/infiniband/hw/irdma/verbs.c
3387
iwmr->page_size = ib_umem_find_best_pgsz(region, pgsz_bitmap, virt);
drivers/infiniband/hw/irdma/verbs.c
3393
iwmr->len = region->length;
drivers/infiniband/hw/irdma/verbs.c
3395
iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size);
drivers/infiniband/hw/irdma/verbs.c
3419
ib_umem_offset(iwmr->region))
drivers/infiniband/hw/irdma/verbs.c
3525
struct ib_umem *region = NULL;
drivers/infiniband/hw/irdma/verbs.c
3538
region = ib_umem_get(pd->device, start, len, access);
drivers/infiniband/hw/irdma/verbs.c
3540
if (IS_ERR(region)) {
drivers/infiniband/hw/irdma/verbs.c
3543
return (struct ib_mr *)region;
drivers/infiniband/hw/irdma/verbs.c
3547
ib_umem_release(region);
drivers/infiniband/hw/irdma/verbs.c
3551
iwmr = irdma_alloc_iwmr(region, pd, virt, req.reg_type);
drivers/infiniband/hw/irdma/verbs.c
3553
ib_umem_release(region);
drivers/infiniband/hw/irdma/verbs.c
3588
ib_umem_release(region);
drivers/infiniband/hw/irdma/verbs.c
3697
struct ib_umem *region;
drivers/infiniband/hw/irdma/verbs.c
3700
region = ib_umem_get(pd->device, start, len, iwmr->access);
drivers/infiniband/hw/irdma/verbs.c
3701
if (IS_ERR(region))
drivers/infiniband/hw/irdma/verbs.c
3702
return PTR_ERR(region);
drivers/infiniband/hw/irdma/verbs.c
3704
iwmr->region = region;
drivers/infiniband/hw/irdma/verbs.c
3707
iwmr->page_size = ib_umem_find_best_pgsz(region,
drivers/infiniband/hw/irdma/verbs.c
3715
iwmr->len = region->length;
drivers/infiniband/hw/irdma/verbs.c
3717
iwmr->page_cnt = ib_umem_num_dma_blocks(region, iwmr->page_size);
drivers/infiniband/hw/irdma/verbs.c
3726
ib_umem_release(region);
drivers/infiniband/hw/irdma/verbs.c
3727
iwmr->region = NULL;
drivers/infiniband/hw/irdma/verbs.c
3780
if (iwmr->region) {
drivers/infiniband/hw/irdma/verbs.c
3781
ib_umem_release(iwmr->region);
drivers/infiniband/hw/irdma/verbs.c
3782
iwmr->region = NULL;
drivers/infiniband/hw/irdma/verbs.c
3917
if (iwmr->region) {
drivers/infiniband/hw/irdma/verbs.c
3937
if (iwmr->region)
drivers/infiniband/hw/irdma/verbs.c
3938
ib_umem_release(iwmr->region);
drivers/infiniband/hw/irdma/verbs.h
112
struct ib_umem *region;
drivers/iommu/amd/iommu.c
3052
struct iommu_resv_region *region;
drivers/iommu/amd/iommu.c
3083
region = iommu_alloc_resv_region(entry->address_start,
drivers/iommu/amd/iommu.c
3086
if (!region) {
drivers/iommu/amd/iommu.c
3090
list_add_tail(&region->list, head);
drivers/iommu/amd/iommu.c
3093
region = iommu_alloc_resv_region(MSI_RANGE_START,
drivers/iommu/amd/iommu.c
3096
if (!region)
drivers/iommu/amd/iommu.c
3098
list_add_tail(&region->list, head);
drivers/iommu/amd/iommu.c
3103
region = iommu_alloc_resv_region(HT_RANGE_START,
drivers/iommu/amd/iommu.c
3106
if (!region)
drivers/iommu/amd/iommu.c
3108
list_add_tail(&region->list, head);
drivers/iommu/apple-dart.c
988
struct iommu_resv_region *region;
drivers/iommu/apple-dart.c
991
region = iommu_alloc_resv_region(DOORBELL_ADDR,
drivers/iommu/apple-dart.c
994
if (!region)
drivers/iommu/apple-dart.c
997
list_add_tail(&region->list, head);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3705
struct iommu_resv_region *region;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3708
region = iommu_alloc_resv_region(MSI_IOVA_BASE, MSI_IOVA_LENGTH,
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3710
if (!region)
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3713
list_add_tail(&region->list, head);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1609
struct iommu_resv_region *region;
drivers/iommu/arm/arm-smmu/arm-smmu.c
1612
region = iommu_alloc_resv_region(MSI_IOVA_BASE, MSI_IOVA_LENGTH,
drivers/iommu/arm/arm-smmu/arm-smmu.c
1614
if (!region)
drivers/iommu/arm/arm-smmu/arm-smmu.c
1617
list_add_tail(&region->list, head);
drivers/iommu/dma-iommu.c
557
struct iommu_resv_region *region;
drivers/iommu/dma-iommu.c
568
list_for_each_entry(region, &resv_regions, list) {
drivers/iommu/dma-iommu.c
572
if (region->type == IOMMU_RESV_SW_MSI)
drivers/iommu/dma-iommu.c
575
lo = iova_pfn(iovad, region->start);
drivers/iommu/dma-iommu.c
576
hi = iova_pfn(iovad, region->start + region->length - 1);
drivers/iommu/dma-iommu.c
579
if (region->type == IOMMU_RESV_MSI)
drivers/iommu/dma-iommu.c
580
ret = cookie_init_hw_msi_region(cookie, region->start,
drivers/iommu/dma-iommu.c
581
region->start + region->length);
drivers/iommu/iommu.c
2950
struct iommu_resv_region *region;
drivers/iommu/iommu.c
2952
region = kzalloc_obj(*region, gfp);
drivers/iommu/iommu.c
2953
if (!region)
drivers/iommu/iommu.c
2956
INIT_LIST_HEAD(&region->list);
drivers/iommu/iommu.c
2957
region->start = start;
drivers/iommu/iommu.c
2958
region->length = length;
drivers/iommu/iommu.c
2959
region->prot = prot;
drivers/iommu/iommu.c
2960
region->type = type;
drivers/iommu/iommu.c
2961
return region;
drivers/iommu/iommu.c
959
struct iommu_resv_region *region, *next;
drivers/iommu/iommu.c
966
list_for_each_entry_safe(region, next, &group_resv_regions, list) {
drivers/iommu/iommu.c
968
(long long)region->start,
drivers/iommu/iommu.c
969
(long long)(region->start +
drivers/iommu/iommu.c
970
region->length - 1),
drivers/iommu/iommu.c
971
iommu_group_resv_type_string[region->type]);
drivers/iommu/iommu.c
972
kfree(region);
drivers/iommu/mtk_iommu.c
1033
struct iommu_resv_region *region;
drivers/iommu/mtk_iommu.c
1047
region = iommu_alloc_resv_region(resv->iova_base, resv->size,
drivers/iommu/mtk_iommu.c
1050
if (!region)
drivers/iommu/mtk_iommu.c
1053
list_add_tail(&region->list, head);
drivers/iommu/mtk_iommu.c
600
const struct mtk_iommu_iova_region *region;
drivers/iommu/mtk_iommu.c
614
region = data->plat_data->iova_region + regionid;
drivers/iommu/mtk_iommu.c
617
larb_mmu->bank[portid] = upper_32_bits(region->iova_base);
drivers/iommu/mtk_iommu.c
621
portid_msk, regionid, upper_32_bits(region->iova_base));
drivers/iommu/mtk_iommu.c
659
const struct mtk_iommu_iova_region *region;
drivers/iommu/mtk_iommu.c
696
region = data->plat_data->iova_region + region_id;
drivers/iommu/mtk_iommu.c
697
dom->domain.geometry.aperture_start = region->iova_base;
drivers/iommu/mtk_iommu.c
698
dom->domain.geometry.aperture_end = region->iova_base + region->size - 1;
drivers/iommu/of_iommu.c
248
struct iommu_resv_region *region;
drivers/iommu/of_iommu.c
263
region = iommu_alloc_resv_region(iova, length, prot, type,
drivers/iommu/of_iommu.c
265
if (region)
drivers/iommu/of_iommu.c
266
list_add_tail(&region->list, list);
drivers/iommu/s390-iommu.c
734
struct iommu_resv_region *region;
drivers/iommu/s390-iommu.c
739
region = iommu_alloc_resv_region(0, zdev->start_dma, 0,
drivers/iommu/s390-iommu.c
741
if (!region)
drivers/iommu/s390-iommu.c
743
list_add_tail(&region->list, list);
drivers/iommu/s390-iommu.c
758
region = iommu_alloc_resv_region(zdev->end_dma + 1, end_resv,
drivers/iommu/s390-iommu.c
761
if (!region)
drivers/iommu/s390-iommu.c
763
list_add_tail(&region->list, list);
drivers/iommu/virtio-iommu.c
489
struct iommu_resv_region *region = NULL, *next;
drivers/iommu/virtio-iommu.c
509
region = iommu_alloc_resv_region(start, size, 0,
drivers/iommu/virtio-iommu.c
514
region = iommu_alloc_resv_region(start, size, prot,
drivers/iommu/virtio-iommu.c
519
if (!region)
drivers/iommu/virtio-iommu.c
524
if (next->start > region->start)
drivers/iommu/virtio-iommu.c
527
list_add_tail(&region->list, &next->list);
drivers/ipack/carriers/tpci200.c
497
dev->region[space].start =
drivers/ipack/carriers/tpci200.c
500
dev->region[space].size = tpci200_space_size[space];
drivers/ipack/devices/ipoctal.c
279
struct ipack_region *region;
drivers/ipack/devices/ipoctal.c
286
region = &ipoctal->dev->region[IPACK_IO_SPACE];
drivers/ipack/devices/ipoctal.c
288
region->start, region->size);
drivers/ipack/devices/ipoctal.c
301
region = &ipoctal->dev->region[IPACK_INT_SPACE];
drivers/ipack/devices/ipoctal.c
304
region->start, region->size);
drivers/ipack/devices/ipoctal.c
312
region = &ipoctal->dev->region[IPACK_MEM8_SPACE];
drivers/ipack/devices/ipoctal.c
315
region->start, 0x8000);
drivers/ipack/ipack.c
351
idmem = ioremap(dev->region[IPACK_ID_SPACE].start,
drivers/ipack/ipack.c
352
dev->region[IPACK_ID_SPACE].size);
drivers/irqchip/irq-gic-v3.c
1023
static int __gic_populate_rdist(struct redist_region *region, void __iomem *ptr)
drivers/irqchip/irq-gic-v3.c
1042
u64 offset = ptr - region->redist_base;
drivers/irqchip/irq-gic-v3.c
1045
gic_data_rdist()->phys_base = region->phys_base + offset;
drivers/irqchip/irq-gic-v3.c
1049
(int)(region - gic_data.redist_regions),
drivers/irqchip/irq-gic-v3.c
1070
static int __gic_update_rdist_properties(struct redist_region *region,
drivers/irqchip/qcom-pdc.c
269
struct pdc_pin_region *region;
drivers/irqchip/qcom-pdc.c
286
region = get_pin_region(hwirq);
drivers/irqchip/qcom-pdc.c
287
if (!region)
drivers/irqchip/qcom-pdc.c
299
parent_fwspec.param[1] = pin_to_hwirq(region, hwirq);
drivers/md/dm-bufio.c
1308
struct dm_io_region region = {
drivers/md/dm-bufio.c
1322
r = dm_io(&io_req, 1, &region, NULL, ioprio);
drivers/md/dm-io.c
101
bio->bi_private = (void *)((unsigned long)io | region);
drivers/md/dm-io.c
105
unsigned int *region)
drivers/md/dm-io.c
110
*region = val & (DM_IO_MAX_REGIONS - 1);
drivers/md/dm-io.c
133
static void dec_count(struct io *io, unsigned int region, blk_status_t error)
drivers/md/dm-io.c
136
set_bit(region, &io->error_bits);
drivers/md/dm-io.c
145
unsigned int region;
drivers/md/dm-io.c
154
retrieve_io_and_region_from_bio(bio, &io, &region);
drivers/md/dm-io.c
159
dec_count(io, region, error);
drivers/md/dm-io.c
306
static void do_region(const blk_opf_t opf, unsigned int region,
drivers/md/dm-io.c
331
dec_count(io, region, BLK_STS_NOTSUPP);
drivers/md/dm-io.c
358
store_io_and_region_in_bio(bio, io, region);
drivers/md/dm-io.c
94
unsigned int region)
drivers/md/dm-log-userspace-base.c
25
region_t region;
drivers/md/dm-log-userspace-base.c
407
static int userspace_is_clean(struct dm_dirty_log *log, region_t region)
drivers/md/dm-log-userspace-base.c
410
uint64_t region64 = (uint64_t)region;
drivers/md/dm-log-userspace-base.c
434
static int userspace_in_sync(struct dm_dirty_log *log, region_t region,
drivers/md/dm-log-userspace-base.c
438
uint64_t region64 = region;
drivers/md/dm-log-userspace-base.c
473
(char *)&fe->region,
drivers/md/dm-log-userspace-base.c
474
sizeof(fe->region),
drivers/md/dm-log-userspace-base.c
500
group[count] = fe->region;
drivers/md/dm-log-userspace-base.c
648
static void userspace_mark_region(struct dm_dirty_log *log, region_t region)
drivers/md/dm-log-userspace-base.c
660
fe->region = region;
drivers/md/dm-log-userspace-base.c
675
static void userspace_clear_region(struct dm_dirty_log *log, region_t region)
drivers/md/dm-log-userspace-base.c
695
fe->region = region;
drivers/md/dm-log-userspace-base.c
708
static int userspace_get_resync_work(struct dm_dirty_log *log, region_t *region)
drivers/md/dm-log-userspace-base.c
725
*region = pkg.r;
drivers/md/dm-log-userspace-base.c
736
region_t region, int in_sync)
drivers/md/dm-log-userspace-base.c
744
pkg.r = region;
drivers/md/dm-log-userspace-base.c
831
region_t region)
drivers/md/dm-log-userspace-base.c
834
uint64_t region64 = region;
drivers/md/dm-log-userspace-base.c
850
if (region < lc->in_sync_hint)
drivers/md/dm-log.c
668
static int core_is_clean(struct dm_dirty_log *log, region_t region)
drivers/md/dm-log.c
672
return log_test_bit(lc->clean_bits, region);
drivers/md/dm-log.c
675
static int core_in_sync(struct dm_dirty_log *log, region_t region, int block)
drivers/md/dm-log.c
679
return log_test_bit(lc->sync_bits, region);
drivers/md/dm-log.c
728
static void core_mark_region(struct dm_dirty_log *log, region_t region)
drivers/md/dm-log.c
732
log_clear_bit(lc, lc->clean_bits, region);
drivers/md/dm-log.c
735
static void core_clear_region(struct dm_dirty_log *log, region_t region)
drivers/md/dm-log.c
740
log_set_bit(lc, lc->clean_bits, region);
drivers/md/dm-log.c
743
static int core_get_resync_work(struct dm_dirty_log *log, region_t *region)
drivers/md/dm-log.c
751
*region = find_next_zero_bit_le(lc->sync_bits,
drivers/md/dm-log.c
754
lc->sync_search = *region + 1;
drivers/md/dm-log.c
756
if (*region >= lc->region_count)
drivers/md/dm-log.c
759
} while (log_test_bit(lc->recovering_bits, *region));
drivers/md/dm-log.c
761
log_set_bit(lc, lc->recovering_bits, *region);
drivers/md/dm-log.c
765
static void core_set_region_sync(struct dm_dirty_log *log, region_t region,
drivers/md/dm-log.c
770
log_clear_bit(lc, lc->recovering_bits, region);
drivers/md/dm-log.c
772
log_set_bit(lc, lc->sync_bits, region);
drivers/md/dm-log.c
774
} else if (log_test_bit(lc->sync_bits, region)) {
drivers/md/dm-log.c
776
log_clear_bit(lc, lc->sync_bits, region);
drivers/md/dm-raid1.c
447
region_t region = dm_rh_bio_to_region(ms->rh, bio);
drivers/md/dm-raid1.c
449
if (log->type->in_sync(log, region, 0))
drivers/md/dm-raid1.c
559
static inline int region_in_sync(struct mirror_set *ms, region_t region,
drivers/md/dm-raid1.c
562
int state = dm_rh_get_state(ms->rh, region, may_block);
drivers/md/dm-raid1.c
568
region_t region;
drivers/md/dm-raid1.c
573
region = dm_rh_bio_to_region(ms->rh, bio);
drivers/md/dm-raid1.c
579
if (likely(region_in_sync(ms, region, 1)))
drivers/md/dm-raid1.c
693
region_t region;
drivers/md/dm-raid1.c
713
region = dm_rh_bio_to_region(ms->rh, bio);
drivers/md/dm-raid1.c
716
log->type->is_remote_recovering(log, region)) {
drivers/md/dm-raid1.c
721
state = dm_rh_get_state(ms->rh, region, 1);
drivers/md/dm-region-hash.c
125
sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
127
return region << rh->region_shift;
drivers/md/dm-region-hash.c
269
static unsigned int rh_hash(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
271
return (unsigned int) ((region * rh->prime) >> rh->shift) & rh->mask;
drivers/md/dm-region-hash.c
274
static struct dm_region *__rh_lookup(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
277
struct list_head *bucket = rh->buckets + rh_hash(rh, region);
drivers/md/dm-region-hash.c
280
if (reg->key == region)
drivers/md/dm-region-hash.c
291
static struct dm_region *__rh_alloc(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
299
nreg->state = rh->log->type->in_sync(rh->log, region, 1) ?
drivers/md/dm-region-hash.c
302
nreg->key = region;
drivers/md/dm-region-hash.c
308
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
327
static struct dm_region *__rh_find(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
331
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
334
reg = __rh_alloc(rh, region);
drivers/md/dm-region-hash.c
341
int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block)
drivers/md/dm-region-hash.c
347
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
357
r = rh->log->type->in_sync(rh->log, region, may_block);
drivers/md/dm-region-hash.c
403
region_t region = dm_rh_bio_to_region(rh, bio);
drivers/md/dm-region-hash.c
415
log->type->set_region_sync(log, region, 0);
drivers/md/dm-region-hash.c
418
reg = __rh_find(rh, region);
drivers/md/dm-region-hash.c
506
static void rh_inc(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
511
reg = __rh_find(rh, region);
drivers/md/dm-region-hash.c
541
void dm_rh_dec(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
548
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
593
region_t region;
drivers/md/dm-region-hash.c
599
r = rh->log->type->get_resync_work(rh->log, &region);
drivers/md/dm-region-hash.c
608
reg = __rh_find(rh, region);
drivers/md/dm-vdo/dm-vdo-target.c
2543
struct dm_io_region *region)
drivers/md/dm-vdo/dm-vdo-target.c
2547
*region = (struct dm_io_region) {
drivers/md/dm-vdo/indexer/index-layout.c
304
static int open_region_reader(struct index_layout *layout, struct layout_region *region,
drivers/md/dm-vdo/indexer/index-layout.c
307
return open_layout_reader(layout, region, -layout->super.start_offset,
drivers/md/dm-vdo/indexer/index-layout.c
319
static int open_region_writer(struct index_layout *layout, struct layout_region *region,
drivers/md/dm-vdo/indexer/index-layout.c
322
return open_layout_writer(layout, region, -layout->super.start_offset,
drivers/md/dm-writecache.c
496
struct dm_io_region region;
drivers/md/dm-writecache.c
514
region.bdev = wc->ssd_dev->bdev;
drivers/md/dm-writecache.c
515
region.sector = (sector_t)i * (BITMAP_GRANULARITY >> SECTOR_SHIFT);
drivers/md/dm-writecache.c
516
region.count = (sector_t)(j - i) * (BITMAP_GRANULARITY >> SECTOR_SHIFT);
drivers/md/dm-writecache.c
518
if (unlikely(region.sector >= wc->metadata_sectors))
drivers/md/dm-writecache.c
520
if (unlikely(region.sector + region.count > wc->metadata_sectors))
drivers/md/dm-writecache.c
521
region.count = wc->metadata_sectors - region.sector;
drivers/md/dm-writecache.c
523
region.sector += wc->start_sector;
drivers/md/dm-writecache.c
533
(void) dm_io(&req, 1, &region, NULL, IOPRIO_DEFAULT);
drivers/md/dm-writecache.c
551
struct dm_io_region region;
drivers/md/dm-writecache.c
554
region.bdev = wc->ssd_dev->bdev;
drivers/md/dm-writecache.c
555
region.sector = 0;
drivers/md/dm-writecache.c
556
region.count = max(4096U, wc->block_size) >> SECTOR_SHIFT;
drivers/md/dm-writecache.c
558
if (unlikely(region.sector + region.count > wc->metadata_sectors))
drivers/md/dm-writecache.c
559
region.count = wc->metadata_sectors - region.sector;
drivers/md/dm-writecache.c
561
region.sector += wc->start_sector;
drivers/md/dm-writecache.c
570
r = dm_io(&req, 1, &region, NULL, IOPRIO_DEFAULT);
drivers/md/dm-writecache.c
586
struct dm_io_region region;
drivers/md/dm-writecache.c
589
region.bdev = dev->bdev;
drivers/md/dm-writecache.c
590
region.sector = 0;
drivers/md/dm-writecache.c
591
region.count = 0;
drivers/md/dm-writecache.c
598
r = dm_io(&req, 1, &region, NULL, IOPRIO_DEFAULT);
drivers/md/dm-writecache.c
980
struct dm_io_region region;
drivers/md/dm-writecache.c
983
region.bdev = wc->ssd_dev->bdev;
drivers/md/dm-writecache.c
984
region.sector = wc->start_sector;
drivers/md/dm-writecache.c
985
region.count = n_sectors;
drivers/md/dm-writecache.c
992
return dm_io(&req, 1, &region, NULL, IOPRIO_DEFAULT);
drivers/media/i2c/adv748x/adv748x-core.c
59
static int adv748x_configure_regmap(struct adv748x_state *state, int region)
drivers/media/i2c/adv748x/adv748x-core.c
63
if (!state->i2c_clients[region])
drivers/media/i2c/adv748x/adv748x-core.c
66
state->regmap[region] =
drivers/media/i2c/adv748x/adv748x-core.c
67
devm_regmap_init_i2c(state->i2c_clients[region],
drivers/media/i2c/adv748x/adv748x-core.c
68
&adv748x_regmap_cnf[region]);
drivers/media/i2c/adv748x/adv748x-core.c
70
if (IS_ERR(state->regmap[region])) {
drivers/media/i2c/adv748x/adv748x-core.c
71
err = PTR_ERR(state->regmap[region]);
drivers/media/i2c/adv748x/adv748x-core.c
74
region, err);
drivers/media/i2c/adv7604.c
3419
static int configure_regmap(struct adv76xx_state *state, int region)
drivers/media/i2c/adv7604.c
3423
if (!state->i2c_clients[region])
drivers/media/i2c/adv7604.c
3426
state->regmap[region] =
drivers/media/i2c/adv7604.c
3427
devm_regmap_init_i2c(state->i2c_clients[region],
drivers/media/i2c/adv7604.c
3428
&adv76xx_regmap_cnf[region]);
drivers/media/i2c/adv7604.c
3430
if (IS_ERR(state->regmap[region])) {
drivers/media/i2c/adv7604.c
3431
err = PTR_ERR(state->regmap[region]);
drivers/media/i2c/adv7604.c
3432
v4l_err(state->i2c_clients[region],
drivers/media/i2c/adv7604.c
3434
region, err);
drivers/media/platform/amphion/vpu_imx8q.c
183
const struct vpu_rpc_region_t *region = &imx8q_regions[i];
drivers/media/platform/amphion/vpu_imx8q.c
185
if (addr >= region->start && addr + size < region->end)
drivers/media/platform/amphion/vpu_imx8q.c
186
return region->type;
drivers/media/platform/ti/omap3isp/isphist.c
321
if (user_cfg->region[c].h_start & ~ISPHIST_REG_START_END_MASK)
drivers/media/platform/ti/omap3isp/isphist.c
323
if (user_cfg->region[c].h_end & ~ISPHIST_REG_START_END_MASK)
drivers/media/platform/ti/omap3isp/isphist.c
325
if (user_cfg->region[c].v_start & ~ISPHIST_REG_START_END_MASK)
drivers/media/platform/ti/omap3isp/isphist.c
327
if (user_cfg->region[c].v_end & ~ISPHIST_REG_START_END_MASK)
drivers/media/platform/ti/omap3isp/isphist.c
329
if (user_cfg->region[c].h_start > user_cfg->region[c].h_end)
drivers/media/platform/ti/omap3isp/isphist.c
331
if (user_cfg->region[c].v_start > user_cfg->region[c].v_end)
drivers/media/platform/ti/omap3isp/isphist.c
387
if (cur_cfg->region[c].h_start != user_cfg->region[c].h_start)
drivers/media/platform/ti/omap3isp/isphist.c
389
if (cur_cfg->region[c].h_end != user_cfg->region[c].h_end)
drivers/media/platform/ti/omap3isp/isphist.c
391
if (cur_cfg->region[c].v_start != user_cfg->region[c].v_start)
drivers/media/platform/ti/omap3isp/isphist.c
393
if (cur_cfg->region[c].v_end != user_cfg->region[c].v_end)
drivers/media/platform/ti/omap3isp/isphist.c
87
reg_hor[c] = (conf->region[c].h_start <<
drivers/media/platform/ti/omap3isp/isphist.c
89
| (conf->region[c].h_end <<
drivers/media/platform/ti/omap3isp/isphist.c
91
reg_ver[c] = (conf->region[c].v_start <<
drivers/media/platform/ti/omap3isp/isphist.c
93
| (conf->region[c].v_end <<
drivers/misc/sgi-xp/xpc_partition.c
420
int region;
drivers/misc/sgi-xp/xpc_partition.c
465
for (region = 0; region < max_regions; region++) {
drivers/misc/sgi-xp/xpc_partition.c
470
dev_dbg(xpc_part, "searching region %d\n", region);
drivers/misc/sgi-xp/xpc_partition.c
472
for (nasid = (region * region_size * 2);
drivers/misc/sgi-xp/xpc_partition.c
473
nasid < ((region + 1) * region_size * 2); nasid += 2) {
drivers/mtd/chips/cfi_cmdset_0001.c
2497
struct mtd_erase_region_info *region;
drivers/mtd/chips/cfi_cmdset_0001.c
2503
region = &mtd->eraseregions[i];
drivers/mtd/chips/cfi_cmdset_0001.c
2504
if (!region->lockmap)
drivers/mtd/chips/cfi_cmdset_0001.c
2507
for (block = 0; block < region->numblocks; block++){
drivers/mtd/chips/cfi_cmdset_0001.c
2508
len = region->erasesize;
drivers/mtd/chips/cfi_cmdset_0001.c
2509
adr = region->offset + block * len;
drivers/mtd/chips/cfi_cmdset_0001.c
2514
set_bit(block, region->lockmap);
drivers/mtd/chips/cfi_cmdset_0001.c
2516
clear_bit(block, region->lockmap);
drivers/mtd/chips/cfi_cmdset_0001.c
2598
struct mtd_erase_region_info *region;
drivers/mtd/chips/cfi_cmdset_0001.c
2604
region = &mtd->eraseregions[i];
drivers/mtd/chips/cfi_cmdset_0001.c
2605
if (!region->lockmap)
drivers/mtd/chips/cfi_cmdset_0001.c
2608
for_each_clear_bit(block, region->lockmap, region->numblocks) {
drivers/mtd/chips/cfi_cmdset_0001.c
2609
len = region->erasesize;
drivers/mtd/chips/cfi_cmdset_0001.c
2610
adr = region->offset + block * len;
drivers/mtd/chips/cfi_cmdset_0001.c
2686
struct mtd_erase_region_info *region;
drivers/mtd/chips/cfi_cmdset_0001.c
2695
region = &mtd->eraseregions[i];
drivers/mtd/chips/cfi_cmdset_0001.c
2696
kfree(region->lockmap);
drivers/mtd/devices/mtd_intel_dg.c
172
static bool idg_nvm_region_readable(u32 access_map, u8 region)
drivers/mtd/devices/mtd_intel_dg.c
174
if (region < 12)
drivers/mtd/devices/mtd_intel_dg.c
175
return access_map & BIT(region + 8); /* [19:8] */
drivers/mtd/devices/mtd_intel_dg.c
177
return access_map & BIT(region - 12); /* [3:0] */
drivers/mtd/devices/mtd_intel_dg.c
180
static bool idg_nvm_region_writable(u32 access_map, u8 region)
drivers/mtd/devices/mtd_intel_dg.c
182
if (region < 12)
drivers/mtd/devices/mtd_intel_dg.c
183
return access_map & BIT(region + 20); /* [31:20] */
drivers/mtd/devices/mtd_intel_dg.c
185
return access_map & BIT(region - 8); /* [7:4] */
drivers/mtd/devices/mtd_intel_dg.c
235
static ssize_t idg_write(struct intel_dg_nvm *nvm, u8 region,
drivers/mtd/devices/mtd_intel_dg.c
246
idg_nvm_set_region_id(nvm, region);
drivers/mtd/devices/mtd_intel_dg.c
309
static ssize_t idg_read(struct intel_dg_nvm *nvm, u8 region,
drivers/mtd/devices/mtd_intel_dg.c
319
idg_nvm_set_region_id(nvm, region);
drivers/mtd/devices/mtd_intel_dg.c
385
idg_erase(struct intel_dg_nvm *nvm, u8 region, loff_t from, u64 len, u64 *fail_addr)
drivers/mtd/devices/mtd_intel_dg.c
396
iowrite32(region << 24 | block, base + NVM_ERASE_REG);
drivers/mtd/devices/mtd_intel_dg.c
443
u32 address, base, limit, region;
drivers/mtd/devices/mtd_intel_dg.c
447
region = idg_nvm_read32(nvm, address);
drivers/mtd/devices/mtd_intel_dg.c
449
base = FIELD_GET(NVM_FREG_BASE_MASK, region) << NVM_FREG_ADDR_SHIFT;
drivers/mtd/devices/mtd_intel_dg.c
450
limit = (FIELD_GET(NVM_FREG_ADDR_MASK, region) << NVM_FREG_ADDR_SHIFT) |
drivers/mtd/devices/mtd_intel_dg.c
454
id, nvm->regions[i].name, region, base, limit);
drivers/mtd/devices/mtd_intel_dg.c
505
u8 region;
drivers/mtd/devices/mtd_intel_dg.c
548
region = nvm->regions[idx].id;
drivers/mtd/devices/mtd_intel_dg.c
554
region, nvm->regions[idx].name, from, len);
drivers/mtd/devices/mtd_intel_dg.c
556
bytes = idg_erase(nvm, region, from, len, &info->fail_addr);
drivers/mtd/devices/mtd_intel_dg.c
578
u8 region;
drivers/mtd/devices/mtd_intel_dg.c
594
region = nvm->regions[idx].id;
drivers/mtd/devices/mtd_intel_dg.c
606
ret = idg_read(nvm, region, from, len, buf);
drivers/mtd/devices/mtd_intel_dg.c
624
u8 region;
drivers/mtd/devices/mtd_intel_dg.c
640
region = nvm->regions[idx].id;
drivers/mtd/devices/mtd_intel_dg.c
652
ret = idg_write(nvm, region, to, len, buf);
drivers/mtd/devices/mtd_intel_dg.c
86
static inline void idg_nvm_set_region_id(struct intel_dg_nvm *nvm, u8 region)
drivers/mtd/devices/mtd_intel_dg.c
88
iowrite32((u32)region, nvm->base + NVM_REGION_ID_REG);
drivers/mtd/maps/pismo.c
101
res.end = base + region->size - 1;
drivers/mtd/maps/pismo.c
131
struct pismo_mem *region)
drivers/mtd/maps/pismo.c
134
.width = region->width,
drivers/mtd/maps/pismo.c
140
return pismo_add_device(pismo, i, region, "physmap-flash",
drivers/mtd/maps/pismo.c
145
struct pismo_mem *region)
drivers/mtd/maps/pismo.c
148
.bankwidth = region->width,
drivers/mtd/maps/pismo.c
151
return pismo_add_device(pismo, i, region, "mtd-ram",
drivers/mtd/maps/pismo.c
159
struct pismo_mem region;
drivers/mtd/maps/pismo.c
161
region.base = base;
drivers/mtd/maps/pismo.c
162
region.type = cs->type;
drivers/mtd/maps/pismo.c
163
region.width = pismo_width_to_bytes(cs->width);
drivers/mtd/maps/pismo.c
164
region.access = le16_to_cpu(cs->access);
drivers/mtd/maps/pismo.c
165
region.size = le32_to_cpu(cs->size);
drivers/mtd/maps/pismo.c
167
if (region.width == 0) {
drivers/mtd/maps/pismo.c
179
i, cs->device, region.type, region.access, region.size / 1024);
drivers/mtd/maps/pismo.c
181
switch (region.type) {
drivers/mtd/maps/pismo.c
189
pismo_add_nor(pismo, i, &region);
drivers/mtd/maps/pismo.c
193
pismo_add_sram(pismo, i, &region);
drivers/mtd/maps/pismo.c
89
struct pismo_mem *region, const char *name,
drivers/mtd/maps/pismo.c
94
phys_addr_t base = region->base;
drivers/mtd/maps/scb2_flash.c
106
struct mtd_erase_region_info *region = &mtd->eraseregions[i];
drivers/mtd/maps/scb2_flash.c
107
region->erasesize /= 2;
drivers/mtd/maps/scb2_flash.c
117
struct mtd_erase_region_info *region = &mtd->eraseregions[i];
drivers/mtd/maps/scb2_flash.c
119
if (region->numblocks * region->erasesize > mtd->size) {
drivers/mtd/maps/scb2_flash.c
120
region->numblocks = ((unsigned long)mtd->size /
drivers/mtd/maps/scb2_flash.c
121
region->erasesize);
drivers/mtd/maps/scb2_flash.c
124
region->numblocks = 0;
drivers/mtd/maps/scb2_flash.c
126
region->offset = 0;
drivers/mtd/mtdcore.c
389
struct mtd_oob_region *region))
drivers/mtd/mtdcore.c
395
struct mtd_oob_region region;
drivers/mtd/mtdcore.c
398
err = iter(mtd, section, &region);
drivers/mtd/mtdcore.c
406
seq_printf(s, "%-3d %4u %4u\n", section, region.offset,
drivers/mtd/mtdcore.c
407
region.length);
drivers/mtd/mtdcore.c
428
struct mtd_oob_region region;
drivers/mtd/mtdcore.c
438
if (mtd_ooblayout_ecc(mtd, 0, &region) == 0)
drivers/mtd/mtdcore.c
442
if (mtd_ooblayout_free(mtd, 0, &region) == 0)
drivers/mtd/nand/onenand/onenand_base.c
2275
struct mtd_erase_region_info *region,
drivers/mtd/nand/onenand/onenand_base.c
2284
if (region) {
drivers/mtd/nand/onenand/onenand_base.c
2286
region_end = region->offset + region->erasesize * region->numblocks;
drivers/mtd/nand/onenand/onenand_base.c
2317
if (region && addr == region_end) {
drivers/mtd/nand/onenand/onenand_base.c
2320
region++;
drivers/mtd/nand/onenand/onenand_base.c
2322
block_size = region->erasesize;
drivers/mtd/nand/onenand/onenand_base.c
2323
region_end = region->offset + region->erasesize * region->numblocks;
drivers/mtd/nand/onenand/onenand_base.c
2350
struct mtd_erase_region_info *region = NULL;
drivers/mtd/nand/onenand/onenand_base.c
2361
region = &mtd->eraseregions[i];
drivers/mtd/nand/onenand/onenand_base.c
2362
block_size = region->erasesize;
drivers/mtd/nand/onenand/onenand_base.c
2367
region_offset = region->offset;
drivers/mtd/nand/onenand/onenand_base.c
2386
if (ONENAND_IS_4KB_PAGE(this) || region ||
drivers/mtd/nand/onenand/onenand_base.c
2390
region, block_size);
drivers/mtd/nand/raw/nand_base.c
297
const struct nand_secure_region *region = &chip->secure_regions[i];
drivers/mtd/nand/raw/nand_base.c
299
if (offset + size <= region->offset ||
drivers/mtd/nand/raw/nand_base.c
300
offset >= region->offset + region->size)
drivers/mtd/nand/spi/alliancememory.c
48
struct mtd_oob_region *region)
drivers/mtd/nand/spi/alliancememory.c
56
region->offset = mtd->oobsize - ecc_bytes;
drivers/mtd/nand/spi/alliancememory.c
57
region->length = ecc_bytes;
drivers/mtd/nand/spi/alliancememory.c
63
struct mtd_oob_region *region)
drivers/mtd/nand/spi/alliancememory.c
84
region->offset = 2;
drivers/mtd/nand/spi/alliancememory.c
85
region->length = mtd->oobsize - 2 - ecc_bytes;
drivers/mtd/nand/spi/ato.c
31
struct mtd_oob_region *region)
drivers/mtd/nand/spi/ato.c
36
region->offset = (16 * section) + 8;
drivers/mtd/nand/spi/ato.c
37
region->length = 8;
drivers/mtd/nand/spi/ato.c
42
struct mtd_oob_region *region)
drivers/mtd/nand/spi/ato.c
48
region->offset = (16 * section);
drivers/mtd/nand/spi/ato.c
49
region->length = 8;
drivers/mtd/nand/spi/ato.c
52
region->offset = 1;
drivers/mtd/nand/spi/ato.c
53
region->length = 7;
drivers/mtd/nand/spi/core.c
320
struct mtd_oob_region *region)
drivers/mtd/nand/spi/core.c
326
struct mtd_oob_region *region)
drivers/mtd/nand/spi/core.c
332
region->offset = 2;
drivers/mtd/nand/spi/core.c
333
region->length = 62;
drivers/mtd/nand/spi/dosilicon.c
27
struct mtd_oob_region *region)
drivers/mtd/nand/spi/dosilicon.c
32
region->offset = 8 + (section * 16);
drivers/mtd/nand/spi/dosilicon.c
33
region->length = 8;
drivers/mtd/nand/spi/dosilicon.c
39
struct mtd_oob_region *region)
drivers/mtd/nand/spi/dosilicon.c
46
region->offset = 2;
drivers/mtd/nand/spi/dosilicon.c
47
region->length = 6;
drivers/mtd/nand/spi/dosilicon.c
49
region->offset = section * 16;
drivers/mtd/nand/spi/dosilicon.c
50
region->length = 8;
drivers/mtd/nand/spi/esmt.c
101
region->length = 2;
drivers/mtd/nand/spi/esmt.c
72
struct mtd_oob_region *region)
drivers/mtd/nand/spi/esmt.c
79
region->offset = section * ESMT_OOB_SECTION_SIZE(nand) +
drivers/mtd/nand/spi/esmt.c
81
region->length = ESMT_OOB_ECC_SIZE(nand);
drivers/mtd/nand/spi/esmt.c
87
struct mtd_oob_region *region)
drivers/mtd/nand/spi/esmt.c
98
region->offset = section * ESMT_OOB_SECTION_SIZE(nand) + 2;
drivers/mtd/nand/spi/fmsh.c
100
region->length = 12;
drivers/mtd/nand/spi/fmsh.c
38
struct mtd_oob_region *region)
drivers/mtd/nand/spi/fmsh.c
44
struct mtd_oob_region *region)
drivers/mtd/nand/spi/fmsh.c
49
region->offset = 2;
drivers/mtd/nand/spi/fmsh.c
50
region->length = 62;
drivers/mtd/nand/spi/fmsh.c
82
struct mtd_oob_region *region)
drivers/mtd/nand/spi/fmsh.c
87
region->offset = 64;
drivers/mtd/nand/spi/fmsh.c
88
region->length = 64;
drivers/mtd/nand/spi/fmsh.c
94
struct mtd_oob_region *region)
drivers/mtd/nand/spi/fmsh.c
99
region->offset = (16 * section) + 4;
drivers/mtd/nand/spi/foresee.c
34
struct mtd_oob_region *region)
drivers/mtd/nand/spi/foresee.c
40
struct mtd_oob_region *region)
drivers/mtd/nand/spi/foresee.c
46
region->offset = 2;
drivers/mtd/nand/spi/foresee.c
47
region->length = 62;
drivers/mtd/nand/spi/gigadevice.c
148
struct mtd_oob_region *region)
drivers/mtd/nand/spi/gigadevice.c
153
region->offset = (16 * section) + 8;
drivers/mtd/nand/spi/gigadevice.c
154
region->length = 8;
drivers/mtd/nand/spi/gigadevice.c
160
struct mtd_oob_region *region)
drivers/mtd/nand/spi/gigadevice.c
166
region->offset = 16 * section;
drivers/mtd/nand/spi/gigadevice.c
167
region->length = 8;
drivers/mtd/nand/spi/gigadevice.c
170
region->offset = 1;
drivers/mtd/nand/spi/gigadevice.c
171
region->length = 7;
drivers/mtd/nand/spi/gigadevice.c
206
struct mtd_oob_region *region)
drivers/mtd/nand/spi/gigadevice.c
211
region->offset = 64;
drivers/mtd/nand/spi/gigadevice.c
212
region->length = 64;
drivers/mtd/nand/spi/gigadevice.c
218
struct mtd_oob_region *region)
drivers/mtd/nand/spi/gigadevice.c
224
region->offset = 1;
drivers/mtd/nand/spi/gigadevice.c
225
region->length = 63;
drivers/mtd/nand/spi/macronix.c
62
struct mtd_oob_region *region)
drivers/mtd/nand/spi/macronix.c
68
struct mtd_oob_region *region)
drivers/mtd/nand/spi/macronix.c
73
region->offset = 2;
drivers/mtd/nand/spi/macronix.c
74
region->length = mtd->oobsize - 2;
drivers/mtd/nand/spi/micron.c
105
region->offset = (section * 16) + 8;
drivers/mtd/nand/spi/micron.c
106
region->length = 8;
drivers/mtd/nand/spi/micron.c
112
struct mtd_oob_region *region)
drivers/mtd/nand/spi/micron.c
121
region->offset = 16 * section;
drivers/mtd/nand/spi/micron.c
122
region->length = 8;
drivers/mtd/nand/spi/micron.c
125
region->offset = 2;
drivers/mtd/nand/spi/micron.c
126
region->length = 6;
drivers/mtd/nand/spi/micron.c
67
struct mtd_oob_region *region)
drivers/mtd/nand/spi/micron.c
72
region->offset = mtd->oobsize / 2;
drivers/mtd/nand/spi/micron.c
73
region->length = mtd->oobsize / 2;
drivers/mtd/nand/spi/micron.c
79
struct mtd_oob_region *region)
drivers/mtd/nand/spi/micron.c
85
region->offset = 2;
drivers/mtd/nand/spi/micron.c
86
region->length = (mtd->oobsize / 2) - 2;
drivers/mtd/nand/spi/micron.c
97
struct mtd_oob_region *region)
drivers/mtd/nand/spi/paragon.c
42
struct mtd_oob_region *region)
drivers/mtd/nand/spi/paragon.c
47
region->offset = 6 + (15 * section); /* 4 BBM + 2 user bytes */
drivers/mtd/nand/spi/paragon.c
48
region->length = 13;
drivers/mtd/nand/spi/paragon.c
54
struct mtd_oob_region *region)
drivers/mtd/nand/spi/paragon.c
60
region->offset = 64;
drivers/mtd/nand/spi/paragon.c
61
region->length = 64;
drivers/mtd/nand/spi/paragon.c
63
region->offset = 4 + (15 * section);
drivers/mtd/nand/spi/paragon.c
64
region->length = 2;
drivers/mtd/nand/spi/skyhigh.c
36
struct mtd_oob_region *region)
drivers/mtd/nand/spi/skyhigh.c
43
struct mtd_oob_region *region)
drivers/mtd/nand/spi/skyhigh.c
49
region->offset = 2;
drivers/mtd/nand/spi/skyhigh.c
50
region->length = mtd->oobsize - 2;
drivers/mtd/nand/spi/toshiba.c
42
struct mtd_oob_region *region)
drivers/mtd/nand/spi/toshiba.c
47
region->offset = mtd->oobsize / 2;
drivers/mtd/nand/spi/toshiba.c
48
region->length = mtd->oobsize / 2;
drivers/mtd/nand/spi/toshiba.c
54
struct mtd_oob_region *region)
drivers/mtd/nand/spi/toshiba.c
60
region->offset = 2;
drivers/mtd/nand/spi/toshiba.c
61
region->length = (mtd->oobsize / 2) - 2;
drivers/mtd/nand/spi/winbond.c
136
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
141
region->offset = (16 * section) + 8;
drivers/mtd/nand/spi/winbond.c
142
region->length = 8;
drivers/mtd/nand/spi/winbond.c
148
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
153
region->offset = (16 * section) + 2;
drivers/mtd/nand/spi/winbond.c
154
region->length = 6;
drivers/mtd/nand/spi/winbond.c
175
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
180
region->offset = 64 + (8 * section);
drivers/mtd/nand/spi/winbond.c
181
region->length = 7;
drivers/mtd/nand/spi/winbond.c
187
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
192
region->offset = 64 + (16 * section);
drivers/mtd/nand/spi/winbond.c
193
region->length = 13;
drivers/mtd/nand/spi/winbond.c
199
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
204
region->offset = (16 * section) + 2;
drivers/mtd/nand/spi/winbond.c
205
region->length = 14;
drivers/mtd/nand/spi/winbond.c
221
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
226
region->offset = (16 * section) + 12;
drivers/mtd/nand/spi/winbond.c
227
region->length = 4;
drivers/mtd/nand/spi/winbond.c
233
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
238
region->offset = (16 * section);
drivers/mtd/nand/spi/winbond.c
239
region->length = 12;
drivers/mtd/nand/spi/winbond.c
243
region->offset += 2;
drivers/mtd/nand/spi/winbond.c
244
region->length -= 2;
drivers/mtd/nand/spi/winbond.c
251
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
256
region->offset = (16 * section) + 12;
drivers/mtd/nand/spi/winbond.c
257
region->length = 4;
drivers/mtd/nand/spi/winbond.c
263
struct mtd_oob_region *region)
drivers/mtd/nand/spi/winbond.c
268
region->offset = 16 * section;
drivers/mtd/nand/spi/winbond.c
269
region->length = 12;
drivers/mtd/nand/spi/winbond.c
273
region->offset += 2;
drivers/mtd/nand/spi/winbond.c
274
region->length -= 2;
drivers/mtd/nand/spi/xtx.c
100
region->offset = mtd->oobsize / 2;
drivers/mtd/nand/spi/xtx.c
101
region->length = mtd->oobsize / 2;
drivers/mtd/nand/spi/xtx.c
107
struct mtd_oob_region *region)
drivers/mtd/nand/spi/xtx.c
112
region->offset = 2;
drivers/mtd/nand/spi/xtx.c
113
region->length = mtd->oobsize / 2 - 2;
drivers/mtd/nand/spi/xtx.c
42
struct mtd_oob_region *region)
drivers/mtd/nand/spi/xtx.c
47
region->offset = 48;
drivers/mtd/nand/spi/xtx.c
48
region->length = 16;
drivers/mtd/nand/spi/xtx.c
54
struct mtd_oob_region *region)
drivers/mtd/nand/spi/xtx.c
59
region->offset = 1;
drivers/mtd/nand/spi/xtx.c
60
region->length = 47;
drivers/mtd/nand/spi/xtx.c
95
struct mtd_oob_region *region)
drivers/mtd/spi-nor/core.c
1506
const struct spi_nor_erase_region *region,
drivers/mtd/spi-nor/core.c
1519
if (!(region->erase_mask & BIT(i)))
drivers/mtd/spi-nor/core.c
1527
if (region->overlaid && region->size <= len)
drivers/mtd/spi-nor/core.c
1551
spi_nor_init_erase_cmd(const struct spi_nor_erase_region *region,
drivers/mtd/spi-nor/core.c
1564
if (region->overlaid)
drivers/mtd/spi-nor/core.c
1565
cmd->size = region->size;
drivers/mtd/spi-nor/core.c
1605
struct spi_nor_erase_region *region;
drivers/mtd/spi-nor/core.c
1612
region = &map->regions[i];
drivers/mtd/spi-nor/core.c
1613
region_end = region->offset + region->size;
drivers/mtd/spi-nor/core.c
1615
while (len && addr >= region->offset && addr < region_end) {
drivers/mtd/spi-nor/core.c
1616
erase = spi_nor_find_best_erase_type(map, region, addr,
drivers/mtd/spi-nor/core.c
1622
region->overlaid) {
drivers/mtd/spi-nor/core.c
1623
cmd = spi_nor_init_erase_cmd(region, erase);
drivers/mtd/spi-nor/core.c
3458
spi_nor_get_region_erasesize(const struct spi_nor_erase_region *region,
drivers/mtd/spi-nor/core.c
3463
if (region->overlaid)
drivers/mtd/spi-nor/core.c
3464
return region->size;
drivers/mtd/spi-nor/core.c
3467
if (region->erase_mask & BIT(i))
drivers/mtd/spi-nor/core.c
3477
const struct spi_nor_erase_region *region = map->regions;
drivers/mtd/spi-nor/core.c
3488
erasesize = spi_nor_get_region_erasesize(&region[i],
drivers/mtd/spi-nor/core.c
3494
mtd_region[i].numblocks = div_u64(region[i].size, erasesize);
drivers/mtd/spi-nor/core.c
3495
mtd_region[i].offset = region[i].offset;
drivers/mtd/spi-nor/core.h
319
int (*lock)(struct spi_nor *nor, unsigned int region);
drivers/mtd/spi-nor/core.h
321
int (*is_locked)(struct spi_nor *nor, unsigned int region);
drivers/mtd/spi-nor/core.h
650
int spi_nor_otp_lock_sr2(struct spi_nor *nor, unsigned int region);
drivers/mtd/spi-nor/core.h
651
int spi_nor_otp_is_locked_sr2(struct spi_nor *nor, unsigned int region);
drivers/mtd/spi-nor/debugfs.c
148
u64 start = region[i].offset;
drivers/mtd/spi-nor/debugfs.c
149
u64 end = start + region[i].size - 1;
drivers/mtd/spi-nor/debugfs.c
150
u8 erase_mask = region[i].erase_mask;
drivers/mtd/spi-nor/debugfs.c
158
region[i].overlaid ? "yes" : "no");
drivers/mtd/spi-nor/debugfs.c
81
struct spi_nor_erase_region *region = erase_map->regions;
drivers/mtd/spi-nor/otp.c
156
static int spi_nor_otp_lock_bit_cr(unsigned int region)
drivers/mtd/spi-nor/otp.c
160
if (region >= ARRAY_SIZE(lock_bits))
drivers/mtd/spi-nor/otp.c
163
return lock_bits[region];
drivers/mtd/spi-nor/otp.c
176
int spi_nor_otp_lock_sr2(struct spi_nor *nor, unsigned int region)
drivers/mtd/spi-nor/otp.c
181
lock_bit = spi_nor_otp_lock_bit_cr(region);
drivers/mtd/spi-nor/otp.c
208
int spi_nor_otp_is_locked_sr2(struct spi_nor *nor, unsigned int region)
drivers/mtd/spi-nor/otp.c
213
lock_bit = spi_nor_otp_lock_bit_cr(region);
drivers/mtd/spi-nor/otp.c
224
static loff_t spi_nor_otp_region_start(const struct spi_nor *nor, unsigned int region)
drivers/mtd/spi-nor/otp.c
228
return org->base + region * org->offset;
drivers/mtd/spi-nor/otp.c
237
static loff_t spi_nor_otp_region_to_offset(struct spi_nor *nor, unsigned int region)
drivers/mtd/spi-nor/otp.c
239
return region * spi_nor_otp_region_len(nor);
drivers/mtd/spi-nor/otp.c
289
unsigned int region;
drivers/mtd/spi-nor/otp.c
296
for (region = spi_nor_otp_offset_to_region(nor, ofs);
drivers/mtd/spi-nor/otp.c
297
region <= spi_nor_otp_offset_to_region(nor, ofs + len - 1);
drivers/mtd/spi-nor/otp.c
298
region++) {
drivers/mtd/spi-nor/otp.c
299
locked = ops->is_locked(nor, region);
drivers/mtd/spi-nor/otp.c
316
unsigned int region;
drivers/mtd/spi-nor/otp.c
350
region = spi_nor_otp_offset_to_region(nor, ofs);
drivers/mtd/spi-nor/otp.c
351
rstart = spi_nor_otp_region_start(nor, region);
drivers/mtd/spi-nor/otp.c
401
unsigned int region;
drivers/mtd/spi-nor/otp.c
432
region = spi_nor_otp_offset_to_region(nor, from);
drivers/mtd/spi-nor/otp.c
433
rstart = spi_nor_otp_region_start(nor, region);
drivers/mtd/spi-nor/otp.c
454
unsigned int region;
drivers/mtd/spi-nor/otp.c
469
region = spi_nor_otp_offset_to_region(nor, from);
drivers/mtd/spi-nor/otp.c
470
ret = ops->lock(nor, region);
drivers/mtd/spi-nor/sfdp.c
391
struct spi_nor_erase_region *region = map->regions;
drivers/mtd/spi-nor/sfdp.c
397
spi_nor_sort_erase_mask(map, region[i].erase_mask);
drivers/mtd/spi-nor/sfdp.c
400
region[i].erase_mask = sorted_erase_mask;
drivers/mtd/spi-nor/sfdp.c
840
spi_nor_region_check_overlay(struct spi_nor_erase_region *region,
drivers/mtd/spi-nor/sfdp.c
849
if (region->size & erase[i].size_mask) {
drivers/mtd/spi-nor/sfdp.c
850
region->overlaid = true;
drivers/mtd/spi-nor/sfdp.c
868
struct spi_nor_erase_region *region;
drivers/mtd/spi-nor/sfdp.c
880
region = devm_kcalloc(nor->dev, region_count, sizeof(*region),
drivers/mtd/spi-nor/sfdp.c
882
if (!region)
drivers/mtd/spi-nor/sfdp.c
884
map->regions = region;
drivers/mtd/spi-nor/sfdp.c
893
region[i].offset = offset;
drivers/mtd/spi-nor/sfdp.c
894
region[i].size = SMPT_MAP_REGION_SIZE(smpt[j]);
drivers/mtd/spi-nor/sfdp.c
896
region[i].erase_mask = erase_type;
drivers/mtd/spi-nor/sfdp.c
898
spi_nor_region_check_overlay(&region[i], erase, erase_type);
drivers/mtd/spi-nor/sfdp.c
912
offset = region[i].offset + region[i].size;
drivers/net/dsa/hirschmann/hellcreek.c
1339
struct devlink_region *region;
drivers/net/dsa/hirschmann/hellcreek.c
1347
region = dsa_devlink_region_create(ds, ops, 1, size);
drivers/net/dsa/hirschmann/hellcreek.c
1348
if (IS_ERR(region))
drivers/net/dsa/hirschmann/hellcreek.c
1349
return PTR_ERR(region);
drivers/net/dsa/hirschmann/hellcreek.c
1351
hellcreek->vlan_region = region;
drivers/net/dsa/hirschmann/hellcreek.c
1357
region = dsa_devlink_region_create(ds, ops, 1, size);
drivers/net/dsa/hirschmann/hellcreek.c
1358
if (IS_ERR(region)) {
drivers/net/dsa/hirschmann/hellcreek.c
1359
ret = PTR_ERR(region);
drivers/net/dsa/hirschmann/hellcreek.c
1363
hellcreek->fdb_region = region;
drivers/net/dsa/mv88e6xxx/chip.h
294
struct devlink_region *region;
drivers/net/dsa/mv88e6xxx/devlink.c
739
dsa_devlink_region_destroy(chip->ports[port].region);
drivers/net/dsa/mv88e6xxx/devlink.c
745
struct devlink_region *region;
drivers/net/dsa/mv88e6xxx/devlink.c
747
region = dsa_devlink_port_region_create(ds,
drivers/net/dsa/mv88e6xxx/devlink.c
751
if (IS_ERR(region))
drivers/net/dsa/mv88e6xxx/devlink.c
752
return PTR_ERR(region);
drivers/net/dsa/mv88e6xxx/devlink.c
754
chip->ports[port].region = region;
drivers/net/dsa/mv88e6xxx/devlink.c
764
struct devlink_region *region;
drivers/net/dsa/mv88e6xxx/devlink.c
791
region = dsa_devlink_region_create(ds, ops, 1, size);
drivers/net/dsa/mv88e6xxx/devlink.c
792
if (IS_ERR(region))
drivers/net/dsa/mv88e6xxx/devlink.c
794
chip->regions[i] = region;
drivers/net/dsa/mv88e6xxx/devlink.c
802
return PTR_ERR(region);
drivers/net/dsa/sja1105/sja1105_devlink.c
102
priv->regions[i] = region;
drivers/net/dsa/sja1105/sja1105_devlink.c
82
struct devlink_region *region;
drivers/net/dsa/sja1105/sja1105_devlink.c
93
region = dsa_devlink_region_create(ds, ops, 1, size);
drivers/net/dsa/sja1105/sja1105_devlink.c
94
if (IS_ERR(region)) {
drivers/net/dsa/sja1105/sja1105_devlink.c
99
return PTR_ERR(region);
drivers/net/ethernet/altera/altera_tse_main.c
1101
struct resource *region;
drivers/net/ethernet/altera/altera_tse_main.c
1109
region = devm_request_mem_region(device, (*res)->start,
drivers/net/ethernet/altera/altera_tse_main.c
1111
if (region == NULL) {
drivers/net/ethernet/altera/altera_tse_main.c
1116
*ptr = devm_ioremap(device, region->start,
drivers/net/ethernet/altera/altera_tse_main.c
1117
resource_size(region));
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4123
if (flash->region == ETHTOOL_FLASH_ALL_REGIONS ||
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4124
flash->region > 0xffff)
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4126
flash->region, NULL);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
4128
return bnxt_flash_firmware_from_file(dev, flash->region, flash->data);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1395
const u8 *data, u32 size, u32 region)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1400
switch (region) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1421
flash_region_strings[region]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1505
int region;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1526
if (ef->region == ETHTOOL_FLASH_ALL_REGIONS) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1529
region = cxgb4_ethtool_get_flash_region(fw_data, &size);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1530
if (region < 0 || !size) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1531
ret = region;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1536
region);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1545
ef->region);
drivers/net/ethernet/intel/i40e/i40e_ddp.c
443
if (flash->region != I40_DDP_FLASH_REGION) {
drivers/net/ethernet/intel/ice/virt/fdir.c
1186
input->q_region = action->act_conf.queue.region;
drivers/net/ethernet/intel/idpf/idpf.h
916
struct idpf_mmio_reg *region = &hw->lan_regs[i];
drivers/net/ethernet/intel/idpf/idpf.h
918
if (reg_offset >= region->addr_start &&
drivers/net/ethernet/intel/idpf/idpf.h
919
reg_offset < (region->addr_start + region->addr_len)) {
drivers/net/ethernet/intel/idpf/idpf.h
924
reg_offset -= region->addr_start;
drivers/net/ethernet/intel/idpf/idpf.h
926
return region->vaddr + reg_offset;
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
235
int region;
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
239
for (region = 0; region < num; region++) {
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
240
if (!test_bit(region, pf_bmap))
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
244
bar += region * MBOX_SIZE;
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
245
mbox_addr[region] = (void *)bar;
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
247
if (!mbox_addr[region])
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
253
for (region = 0; region < num; region++) {
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
254
if (!test_bit(region, pf_bmap))
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
258
bar += region * MBOX_SIZE;
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
260
mbox_addr[region] = (void *)bar;
drivers/net/ethernet/marvell/octeontx2/af/cn20k/mbox_init.c
262
if (!mbox_addr[region])
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2432
int region;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2448
for (region = 0; region < num; region++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2449
if (!test_bit(region, pf_bmap))
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2456
bar4 += region * MBOX_SIZE;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2459
bar4 += region * MBOX_SIZE;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2461
mbox_addr[region] = ioremap_wc(bar4, MBOX_SIZE);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2462
if (!mbox_addr[region])
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2472
for (region = 0; region < num; region++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2473
if (!test_bit(region, pf_bmap))
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2478
RVU_AF_PFX_BAR4_ADDR(region));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2482
bar4 += region * MBOX_SIZE;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2484
mbox_addr[region] = ioremap_wc(bar4, MBOX_SIZE);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2485
if (!mbox_addr[region])
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2491
while (region--)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2492
iounmap(mbox_addr[region]);
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
238
struct mtk_wed_wo_memory_region *region)
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
247
region->phy_addr = res.start;
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
248
region->size = resource_size(&res);
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
249
region->addr = devm_ioremap_resource(hw->dev, &res);
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
250
if (IS_ERR(region->addr))
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
251
return PTR_ERR(region->addr);
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
278
struct mtk_wed_wo_memory_region *region;
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
280
region = &mem_region[i];
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
281
if (region->phy_addr != le32_to_cpu(fw_region->addr))
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
284
if (region->size < length)
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
287
if (region->shared && region->consumed)
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
290
if (!region->shared || !region->consumed) {
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
291
memcpy_toio(region->addr, ptr, length);
drivers/net/ethernet/mediatek/mtk_wed_mcu.c
292
region->consumed = true;
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
2159
if (flash->region != ETHTOOL_FLASH_ALL_REGIONS)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1118
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1122
struct mlxsw_sp_acl_tcam_region *region);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
101
struct mlxsw_sp1_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
103
struct mlxsw_sp_acl_rule_info *rulei = region->catchall.rulei;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
105
mlxsw_sp_acl_ctcam_entry_del(mlxsw_sp, &region->cregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
106
&region->catchall.cchunk,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
107
&region->catchall.centry);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
109
mlxsw_sp_acl_ctcam_chunk_fini(&region->catchall.cchunk);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
118
struct mlxsw_sp1_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
121
err = mlxsw_sp_acl_ctcam_region_init(mlxsw_sp, &region->cregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
126
err = mlxsw_sp1_acl_ctcam_region_catchall_add(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
129
region->region = _region;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
133
mlxsw_sp_acl_ctcam_region_fini(&region->cregion);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
14
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
140
struct mlxsw_sp1_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
142
mlxsw_sp1_acl_ctcam_region_catchall_del(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
143
mlxsw_sp_acl_ctcam_region_fini(&region->cregion);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
148
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
156
struct mlxsw_sp1_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
159
mlxsw_sp_acl_ctcam_chunk_init(&region->cregion, &chunk->cchunk,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
175
struct mlxsw_sp1_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
179
return mlxsw_sp_acl_ctcam_entry_add(mlxsw_sp, &region->cregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
188
struct mlxsw_sp1_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
192
mlxsw_sp_acl_ctcam_entry_del(mlxsw_sp, &region->cregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
227
struct mlxsw_sp1_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
233
region->region,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
62
struct mlxsw_sp1_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
67
mlxsw_sp_acl_ctcam_chunk_init(&region->cregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
68
&region->catchall.cchunk,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
81
err = mlxsw_sp_acl_ctcam_entry_add(mlxsw_sp, &region->cregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
82
&region->catchall.cchunk,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
83
&region->catchall.centry,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
87
region->catchall.rulei = rulei;
drivers/net/ethernet/mellanox/mlxsw/spectrum1_acl_tcam.c
95
mlxsw_sp_acl_ctcam_chunk_fini(&region->catchall.cchunk);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
297
struct mlxsw_sp1_mr_tcam_region *region = &mr_tcam->tcam_regions[0];
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
306
&region[MLXSW_SP_L3_PROTO_IPV4],
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
313
&region[MLXSW_SP_L3_PROTO_IPV6],
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
321
mlxsw_sp1_mr_tcam_region_fini(&region[MLXSW_SP_L3_PROTO_IPV4]);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
328
struct mlxsw_sp1_mr_tcam_region *region = &mr_tcam->tcam_regions[0];
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
330
mlxsw_sp1_mr_tcam_region_fini(&region[MLXSW_SP_L3_PROTO_IPV6]);
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
331
mlxsw_sp1_mr_tcam_region_fini(&region[MLXSW_SP_L3_PROTO_IPV4]);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
155
struct mlxsw_sp2_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
158
region->region = _region;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
161
&region->aregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
169
struct mlxsw_sp2_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
171
mlxsw_sp_acl_atcam_region_fini(&region->aregion);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
176
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
178
return mlxsw_sp_acl_atcam_region_associate(mlxsw_sp, region->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
18
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
183
struct mlxsw_sp2_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
185
return mlxsw_sp_acl_atcam_rehash_hints_get(&region->aregion);
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
196
struct mlxsw_sp2_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
199
mlxsw_sp_acl_atcam_chunk_init(&region->aregion, &chunk->achunk,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
215
struct mlxsw_sp2_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
220
return mlxsw_sp_acl_atcam_entry_add(mlxsw_sp, &region->aregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
229
struct mlxsw_sp2_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
233
mlxsw_sp_acl_atcam_entry_del(mlxsw_sp, &region->aregion, &chunk->achunk,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
242
struct mlxsw_sp2_acl_tcam_region *region = region_priv;
drivers/net/ethernet/mellanox/mlxsw/spectrum2_acl_tcam.c
247
&region->aregion,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
120
struct mlxsw_sp *mlxsw_sp = aregion->region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
225
struct mlxsw_sp_acl_tcam_region *region = aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
227
struct mlxsw_sp *mlxsw_sp = region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
293
struct mlxsw_sp_acl_tcam_region *region = aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
300
blocks_count = mlxsw_afk_key_info_blocks_count_get(region->key_info);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
318
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
324
aregion->region = region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
340
region, ops);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
383
struct mlxsw_sp_acl_tcam_region *region = aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
402
priority, region->tcam_region_info,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
426
struct mlxsw_sp_acl_tcam_region *region = aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
431
region->tcam_region_info,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
450
struct mlxsw_sp_acl_tcam_region *region = aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
460
priority, region->tcam_region_info,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
476
struct mlxsw_sp_acl_tcam_region *region = aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_atcam.c
483
mlxsw_afk_encode(afk, region->key_info, &rulei->values,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_bloom_filter.c
252
struct mlxsw_afk_key_info *key_info = aregion->region->key_info;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_bloom_filter.c
261
(aregion->region->id << 4));
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_bloom_filter.c
394
struct mlxsw_afk_key_info *key_info = aregion->region->key_info;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
111
cregion->region->tcam_region_info,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
125
struct mlxsw_sp_acl_tcam_region *region = cregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
126
struct mlxsw_sp *mlxsw_sp = region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
132
return mlxsw_sp_acl_ctcam_region_resize(mlxsw_sp, region, new_count);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
141
struct mlxsw_sp_acl_tcam_region *region = cregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
142
struct mlxsw_sp *mlxsw_sp = region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
144
mlxsw_sp_acl_ctcam_region_move(mlxsw_sp, region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
15
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
159
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
162
cregion->region = region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
21
region->key_type, new_size, region->id,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
22
region->tcam_region_info);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
28
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
34
region->tcam_region_info, src_offset,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
35
region->tcam_region_info, dst_offset, size);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
46
struct mlxsw_sp_acl_tcam_region *region = cregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
61
region->tcam_region_info,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
65
mlxsw_afk_encode(afk, region->key_info, &rulei->values, key, mask);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_ctcam.c
94
cregion->region->tcam_region_info,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
1348
struct mlxsw_sp *mlxsw_sp = aregion->region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
1351
mlxsw_reg_percr_pack(percr_pl, aregion->region->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
1358
struct mlxsw_sp *mlxsw_sp = aregion->region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
1361
mlxsw_reg_pererp_pack(pererp_pl, aregion->region->id, false, false, 0,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
1409
struct mlxsw_sp *mlxsw_sp = aregion->region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
183
struct mlxsw_sp_acl_tcam_region *region = erp_table->aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
184
struct mlxsw_sp *mlxsw_sp = region->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
188
mlxsw_reg_percr_pack(percr_pl, region->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
427
struct mlxsw_sp_acl_tcam_region *region = erp_table->aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
431
mlxsw_reg_pererp_pack(pererp_pl, region->id, ctcam_le, true, 0,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
442
struct mlxsw_sp_acl_tcam_region *region = erp_table->aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
452
mlxsw_reg_pererp_pack(pererp_pl, region->id, false, false, 0, 0,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
534
err = mlxsw_sp_acl_bf_entry_add(aregion->region->mlxsw_sp,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
546
mlxsw_sp_acl_bf_entry_del(aregion->region->mlxsw_sp,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
561
mlxsw_sp_acl_bf_entry_del(aregion->region->mlxsw_sp,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
649
struct mlxsw_sp_acl_tcam_region *region = erp_table->aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
654
mlxsw_reg_pererp_pack(pererp_pl, region->id, ctcam_le, true, 0,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
666
struct mlxsw_sp_acl_tcam_region *region = erp_table->aregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_erp.c
671
mlxsw_reg_pererp_pack(pererp_pl, region->id, ctcam_le, true, 0,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1063
err = ops->entry_add(mlxsw_sp, chunk->region->priv, chunk->priv,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1080
ops->entry_del(mlxsw_sp, entry->chunk->region->priv,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1087
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1093
return ops->entry_action_replace(mlxsw_sp, region->priv,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1104
return ops->entry_activity_get(mlxsw_sp, entry->chunk->region->priv,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1168
vchunk->vregion->region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1213
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1220
new_chunk = mlxsw_sp_acl_tcam_chunk_create(mlxsw_sp, vchunk, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1244
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1251
if (vchunk->chunk->region != region) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1253
region, ctx);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1343
vregion->region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1369
swap(vregion->region, vregion->region2);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1404
hints_priv = ops->region_rehash_hints_get(vregion->region->priv);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1418
vregion->region2 = vregion->region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1419
vregion->region = new_region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1434
vregion->region = vregion->region2;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
149
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
169
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
210
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
215
list_for_each_entry(region, &group->region_list, list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
219
if (region->list.next != &group->region_list &&
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
220
list_next_entry(region, list)->vregion == region->vregion)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
223
region->id, multi);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
385
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
413
list_add_tail(&region->list, pos);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
414
region->group = group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
425
list_del(&region->list);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
433
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
435
struct mlxsw_sp_acl_tcam_group *group = region->group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
438
list_del(&region->list);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
463
vregion->region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
483
mlxsw_sp_acl_tcam_group_region_detach(mlxsw_sp, vregion->region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
571
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
573
struct mlxsw_afk_key_info *key_info = region->key_info;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
580
region->key_type,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
582
region->id, region->tcam_region_info);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
593
mlxsw_reg_ptar_unpack(ptar_pl, region->tcam_region_info);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
599
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
604
region->key_type, 0, region->id,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
605
region->tcam_region_info);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
611
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
615
mlxsw_reg_pacl_pack(pacl_pl, region->id, true,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
616
region->tcam_region_info);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
622
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
626
mlxsw_reg_pacl_pack(pacl_pl, region->id, false,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
627
region->tcam_region_info);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
638
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
641
region = kzalloc(sizeof(*region) + ops->region_priv_size, GFP_KERNEL);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
642
if (!region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
644
region->mlxsw_sp = mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
645
region->vregion = vregion;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
646
region->key_info = vregion->key_info;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
648
err = mlxsw_sp_acl_tcam_region_id_get(tcam, &region->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
652
err = ops->region_associate(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
656
region->key_type = ops->key_type;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
657
err = mlxsw_sp_acl_tcam_region_alloc(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
661
err = mlxsw_sp_acl_tcam_region_enable(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
665
err = ops->region_init(mlxsw_sp, region->priv, tcam->priv,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
666
region, hints_priv);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
670
return region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
673
mlxsw_sp_acl_tcam_region_disable(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
675
mlxsw_sp_acl_tcam_region_free(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
678
mlxsw_sp_acl_tcam_region_id_put(tcam, region->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
680
kfree(region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
686
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
691
ops->region_fini(mlxsw_sp, region->priv);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
692
mlxsw_sp_acl_tcam_region_disable(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
693
mlxsw_sp_acl_tcam_region_free(mlxsw_sp, region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
694
mlxsw_sp_acl_tcam_region_id_put(tcam, region->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
695
kfree(region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
797
vregion->region = mlxsw_sp_acl_tcam_region_create(mlxsw_sp, tcam,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
799
if (IS_ERR(vregion->region)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
800
err = PTR_ERR(vregion->region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
822
mlxsw_sp_acl_tcam_region_destroy(mlxsw_sp, vregion->region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
851
mlxsw_sp_acl_tcam_region_destroy(mlxsw_sp, vregion->region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
903
struct mlxsw_sp_acl_tcam_region *region)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
912
chunk->region = region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
914
ops->chunk_init(region->priv, chunk->priv, vchunk->priority);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
966
vchunk->vregion->region);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.h
115
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.h
162
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.h
212
struct mlxsw_sp_acl_tcam_region *region,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.h
93
struct mlxsw_sp_acl_tcam_region *region;
drivers/net/ethernet/mscc/ocelot_stats.c
325
struct ocelot_stats_region *region;
drivers/net/ethernet/mscc/ocelot_stats.c
331
list_for_each_entry(region, &ocelot->stats_regions, node) {
drivers/net/ethernet/mscc/ocelot_stats.c
332
err = ocelot_bulk_read(ocelot, region->base, region->buf,
drivers/net/ethernet/mscc/ocelot_stats.c
333
region->count);
drivers/net/ethernet/mscc/ocelot_stats.c
346
struct ocelot_stats_region *region;
drivers/net/ethernet/mscc/ocelot_stats.c
349
list_for_each_entry(region, &ocelot->stats_regions, node) {
drivers/net/ethernet/mscc/ocelot_stats.c
350
unsigned int idx = port * OCELOT_NUM_STATS + region->first_stat;
drivers/net/ethernet/mscc/ocelot_stats.c
352
for (j = 0; j < region->count; j++) {
drivers/net/ethernet/mscc/ocelot_stats.c
354
u64 val = region->buf[j];
drivers/net/ethernet/mscc/ocelot_stats.c
911
struct ocelot_stats_region *region = NULL;
drivers/net/ethernet/mscc/ocelot_stats.c
935
if (region && ocelot->map[SYS][layout[i].reg & REG_MASK] ==
drivers/net/ethernet/mscc/ocelot_stats.c
937
region->count++;
drivers/net/ethernet/mscc/ocelot_stats.c
939
region = devm_kzalloc(ocelot->dev, sizeof(*region),
drivers/net/ethernet/mscc/ocelot_stats.c
941
if (!region)
drivers/net/ethernet/mscc/ocelot_stats.c
944
region->base = layout[i].reg;
drivers/net/ethernet/mscc/ocelot_stats.c
945
region->first_stat = i;
drivers/net/ethernet/mscc/ocelot_stats.c
946
region->count = 1;
drivers/net/ethernet/mscc/ocelot_stats.c
947
list_add_tail(&region->node, &ocelot->stats_regions);
drivers/net/ethernet/mscc/ocelot_stats.c
953
list_for_each_entry(region, &ocelot->stats_regions, node) {
drivers/net/ethernet/mscc/ocelot_stats.c
957
ocelot_reg_to_target_addr(ocelot, region->base, &target,
drivers/net/ethernet/mscc/ocelot_stats.c
962
region->count, addr / 4);
drivers/net/ethernet/mscc/ocelot_stats.c
963
region->buf = devm_kcalloc(ocelot->dev, region->count,
drivers/net/ethernet/mscc/ocelot_stats.c
964
sizeof(*region->buf), GFP_KERNEL);
drivers/net/ethernet/mscc/ocelot_stats.c
965
if (!region->buf)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c
100
entry.region.cpp_action,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c
101
entry.region.cpp_token);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c
102
res->addr = (u64)entry.region.page_offset << 8;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c
103
res->size = (u64)entry.region.page_size << 8;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c
54
} region;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_resource.c
99
res->cpp_id = NFP_CPP_ID(entry.region.cpp_target,
drivers/net/ethernet/pensando/ionic/ionic_dev.c
242
offset = le32_to_cpu(ident->cmb_layout.region[i].offset);
drivers/net/ethernet/pensando/ionic/ionic_dev.c
243
length = le32_to_cpu(ident->cmb_layout.region[i].length);
drivers/net/ethernet/pensando/ionic/ionic_dev.c
247
i, ident->cmb_layout.region[i].bar_num,
drivers/net/ethernet/pensando/ionic/ionic_dev.c
248
ident->cmb_layout.region[i].cmb_type,
drivers/net/ethernet/pensando/ionic/ionic_dev.c
259
if (ident->cmb_layout.region[0].bar_num == bar->res_index &&
drivers/net/ethernet/pensando/ionic/ionic_dev.c
260
le32_to_cpu(ident->cmb_layout.region[0].length) == bar->len &&
drivers/net/ethernet/pensando/ionic/ionic_dev.c
261
!ident->cmb_layout.region[0].offset) {
drivers/net/ethernet/pensando/ionic/ionic_dev.c
267
length_reg0 = le32_to_cpu(ident->cmb_layout.region[0].length);
drivers/net/ethernet/pensando/ionic/ionic_dev.c
298
if (le32_to_cpu(ident->cmb_layout.region[i].length) !=
drivers/net/ethernet/pensando/ionic/ionic_dev.c
302
offset = le32_to_cpu(ident->cmb_layout.region[i].offset);
drivers/net/ethernet/pensando/ionic/ionic_dev.c
304
switch (ident->cmb_layout.region[i].cmb_type) {
drivers/net/ethernet/pensando/ionic/ionic_dev.c
345
i, ident->cmb_layout.region[i].cmb_type);
drivers/net/ethernet/pensando/ionic/ionic_if.h
2292
union ionic_cmb_region region[IONIC_MAX_CMB_REGIONS];
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
251
u8 region;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
641
static int qlcnic_get_flt_entry(struct qlcnic_adapter *adapter, u8 region,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
674
if (flt_entry[i].region == region)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
680
"region=%x not found in %d regions\n", region, i);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
1020
struct __mem *rom = &entry->region.mem;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
241
} region;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
389
struct __crb *crb = &entry->region.crb;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
406
struct __ctrl *ctr = &entry->region.ctrl;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
505
struct __mux *mux = &entry->region.mux;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
523
struct __queue *que = &entry->region.que;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
547
struct __ocm *ocm = &entry->region.ocm;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
563
struct __mem *rom = &entry->region.mem;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
593
struct __cache *l1 = &entry->region.cache;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
619
struct __cache *l2 = &entry->region.cache;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
858
struct __mem *mem = &entry->region.mem;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
909
struct __pollrdmwr *poll = &entry->region.pollrdmwr;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
953
struct __pollrd *pollrd = &entry->region.pollrd;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
988
struct __mux2 *mux2 = &entry->region.mux2;
drivers/net/ethernet/realtek/r8169_main.c
5600
int jumbo_max, region, rc;
drivers/net/ethernet/realtek/r8169_main.c
5634
region = ffs(pci_select_bars(pdev, IORESOURCE_MEM)) - 1;
drivers/net/ethernet/realtek/r8169_main.c
5635
if (region < 0)
drivers/net/ethernet/realtek/r8169_main.c
5638
tp->mmio_addr = pcim_iomap_region(pdev, region, KBUILD_MODNAME);
drivers/net/ethernet/sfc/falcon/falcon.c
1786
void *region;
drivers/net/ethernet/sfc/falcon/falcon.c
1798
region = kmalloc(FALCON_NVCONFIG_END, GFP_KERNEL);
drivers/net/ethernet/sfc/falcon/falcon.c
1799
if (!region)
drivers/net/ethernet/sfc/falcon/falcon.c
1801
nvconfig = region + FALCON_NVCONFIG_OFFSET;
drivers/net/ethernet/sfc/falcon/falcon.c
1804
rc = falcon_spi_read(efx, spi, 0, FALCON_NVCONFIG_END, NULL, region);
drivers/net/ethernet/sfc/falcon/falcon.c
1831
word = region;
drivers/net/ethernet/sfc/falcon/falcon.c
1832
limit = region + FALCON_NVCONFIG_END;
drivers/net/ethernet/sfc/falcon/falcon.c
1848
kfree(region);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
283
static void icssm_prueth_clearmem(struct prueth *prueth, enum prueth_mem region)
drivers/net/ethernet/ti/icssm/icssm_prueth.c
285
memset_io(prueth->mem[region].va, 0, prueth->mem[region].size);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
384
enum prueth_mem region;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
413
region = emac->dram;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
419
icssm_prueth_write_reg(prueth, region, PHY_SPEED_OFFSET,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
434
writeb(port_status, prueth->mem[region].va +
drivers/net/ethernet/ti/icssm/icssm_prueth.c
44
enum prueth_mem region,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
47
writel_relaxed(val, prueth->mem[region].va + reg);
drivers/net/fddi/defxx.c
541
struct resource *region;
drivers/net/fddi/defxx.c
582
region = request_mem_region(bar_start[0], bar_len[0],
drivers/net/fddi/defxx.c
584
if (!region && (dfx_bus_eisa || dfx_bus_pci)) {
drivers/net/fddi/defxx.c
590
region = request_region(bar_start[0], bar_len[0],
drivers/net/fddi/defxx.c
592
if (!region) {
drivers/net/fddi/defxx.c
599
region = request_region(bar_start[1], bar_len[1],
drivers/net/fddi/defxx.c
601
if (!region) {
drivers/net/fddi/defxx.c
609
region = request_region(bar_start[2], bar_len[2],
drivers/net/fddi/defxx.c
611
if (!region) {
drivers/net/wireless/ath/ath.h
82
enum nl80211_dfs_regions region;
drivers/net/wireless/ath/ath10k/ahb.c
686
u32 val = 0, region = addr & 0xfffff;
drivers/net/wireless/ath/ath10k/ahb.c
690
if (region >= QCA4019_SRAM_ADDR && region <=
drivers/net/wireless/ath/ath10k/ahb.c
695
val |= region;
drivers/net/wireless/ath/ath10k/ahb.c
697
val |= 0x100000 | region;
drivers/net/wireless/ath/ath10k/mac.c
3526
nl_dfs_reg = ar->dfs_detector->region;
drivers/net/wireless/ath/ath10k/pci.c
1586
const struct ath10k_mem_region *region,
drivers/net/wireless/ath/ath10k/pci.c
1593
base_addr += region->start;
drivers/net/wireless/ath/ath10k/pci.c
1595
for (i = 0; i < region->len; i += 4) {
drivers/net/wireless/ath/ath10k/pci.c
1600
return region->len;
drivers/net/wireless/ath/ath10k/pci.c
1605
const struct ath10k_mem_region *region,
drivers/net/wireless/ath/ath10k/pci.c
1619
for (i = 0; i < region->len; i += 4)
drivers/net/wireless/ath/ath10k/pci.c
1620
*(u32 *)(buf + i) = ioread32(ar_pci->mem + region->start + i);
drivers/net/wireless/ath/ath10k/pci.c
1622
ret = region->len;
drivers/net/wireless/ath/ath10k/pci.c
857
u32 val = 0, region = addr & 0xfffff;
drivers/net/wireless/ath/ath10k/pci.c
861
val |= 0x100000 | region;
drivers/net/wireless/ath/ath10k/pci.c
872
u32 val = 0, region = addr & 0xfffff;
drivers/net/wireless/ath/ath10k/pci.c
876
val |= ((addr >= 0x100000) ? 0x100000 : 0) | region;
drivers/net/wireless/ath/ath10k/pci.c
882
u32 val = 0, region = addr & 0xfffff;
drivers/net/wireless/ath/ath10k/pci.c
885
val |= 0x100000 | region;
drivers/net/wireless/ath/ath10k/wmi.c
4135
ar->dfs_detector->region == NL80211_DFS_FCC) {
drivers/net/wireless/ath/ath9k/dfs_debug.c
73
sc->dfs_detector->region);
drivers/net/wireless/ath/dfs_pattern_detector.c
123
.region = NL80211_DFS_JP,
drivers/net/wireless/ath/dfs_pattern_detector.c
141
get_dfs_domain_radar_types(enum nl80211_dfs_regions region)
drivers/net/wireless/ath/dfs_pattern_detector.c
145
if (dfs_domains[i]->region == region)
drivers/net/wireless/ath/dfs_pattern_detector.c
278
if (dpd->region == NL80211_DFS_UNSET)
drivers/net/wireless/ath/dfs_pattern_detector.c
31
enum nl80211_dfs_regions region;
drivers/net/wireless/ath/dfs_pattern_detector.c
316
enum nl80211_dfs_regions region)
drivers/net/wireless/ath/dfs_pattern_detector.c
321
if (dpd->region == region)
drivers/net/wireless/ath/dfs_pattern_detector.c
324
dpd->region = NL80211_DFS_UNSET;
drivers/net/wireless/ath/dfs_pattern_detector.c
326
rt = get_dfs_domain_radar_types(region);
drivers/net/wireless/ath/dfs_pattern_detector.c
336
dpd->region = region;
drivers/net/wireless/ath/dfs_pattern_detector.c
345
.region = NL80211_DFS_UNSET,
drivers/net/wireless/ath/dfs_pattern_detector.c
350
enum nl80211_dfs_regions region)
drivers/net/wireless/ath/dfs_pattern_detector.c
365
if (dpd->set_dfs_domain(dpd, region))
drivers/net/wireless/ath/dfs_pattern_detector.c
368
ath_dbg(common, DFS,"Could not set DFS domain to %d", region);
drivers/net/wireless/ath/dfs_pattern_detector.c
66
.region = NL80211_DFS_ETSI,
drivers/net/wireless/ath/dfs_pattern_detector.c
98
.region = NL80211_DFS_FCC,
drivers/net/wireless/ath/dfs_pattern_detector.h
104
enum nl80211_dfs_regions region;
drivers/net/wireless/ath/dfs_pattern_detector.h
121
enum nl80211_dfs_regions region);
drivers/net/wireless/ath/dfs_pattern_detector.h
98
enum nl80211_dfs_regions region);
drivers/net/wireless/ath/regd.c
525
reg->region = request->dfs_region;
drivers/net/wireless/ath/regd.c
791
switch (reg->region) {
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
731
const struct iwl_fw_ini_region_tlv *region;
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
735
if (length < offsetof(typeof(*region), special_mem) +
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
736
sizeof(region->special_mem))
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
739
region = (const void *)tlv->data;
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
740
addr = le32_to_cpu(region->special_mem.base_addr);
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
741
addr += le32_to_cpu(region->special_mem.offset);
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
744
if (region->type != IWL_FW_INI_REGION_SPECIAL_DEVICE_MEMORY)
drivers/net/wireless/intel/iwlwifi/iwl-drv.c
747
switch (region->sub_type) {
drivers/net/wireless/marvell/mwifiex/cfp.c
153
u8 region[IEEE80211_COUNTRY_STRING_LEN] __nonstring;
drivers/net/wireless/marvell/mwifiex/cfp.c
175
return region_code_mapping_t[i].region;
drivers/net/wireless/mediatek/mt76/eeprom.c
215
if (dev->region < ARRAY_SIZE(region_names))
drivers/net/wireless/mediatek/mt76/eeprom.c
216
region_name = region_names[dev->region];
drivers/net/wireless/mediatek/mt76/mac80211.c
2042
if (dev->region == NL80211_DFS_UNSET ||
drivers/net/wireless/mediatek/mt76/mt76.h
994
enum nl80211_dfs_regions region;
drivers/net/wireless/mediatek/mt76/mt7603/debugfs.c
52
dev->mt76.region == NL80211_DFS_ETSI;
drivers/net/wireless/mediatek/mt76/mt7603/init.c
431
dev->mt76.region = request->dfs_region;
drivers/net/wireless/mediatek/mt76/mt7603/init.c
433
dev->mt76.region == NL80211_DFS_ETSI;
drivers/net/wireless/mediatek/mt76/mt7615/init.c
341
dev->mt76.region = request->dfs_region;
drivers/net/wireless/mediatek/mt76/mt7615/mac.c
2184
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2962
const struct mt76_connac2_fw_region *region;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2966
region = (const void *)((const u8 *)hdr -
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2967
(hdr->n_region - i) * sizeof(*region));
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2968
mode = mt76_connac_mcu_gen_dl_mode(dev, region->feature_set,
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2970
len = le32_to_cpu(region->len);
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2971
addr = le32_to_cpu(region->addr);
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2973
if (region->feature_set & FW_FEATURE_NON_DL)
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
2976
if (region->feature_set & FW_FEATURE_OVERRIDE_ADDR)
drivers/net/wireless/mediatek/mt76/mt76x02_debugfs.c
91
enum nl80211_dfs_regions region = dev->mt76.region;
drivers/net/wireless/mediatek/mt76/mt76x02_debugfs.c
97
region == NL80211_DFS_ETSI;
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
275
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
449
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
678
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
718
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
859
dev->mt76.region = NL80211_DFS_UNSET;
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
866
enum nl80211_dfs_regions region)
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
871
if (dev->mt76.region != region) {
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
875
region == NL80211_DFS_ETSI;
drivers/net/wireless/mediatek/mt76/mt76x02_dfs.c
878
dev->mt76.region = region;
drivers/net/wireless/mediatek/mt76/mt7915/init.c
346
dev->mt76.region = request->dfs_region;
drivers/net/wireless/mediatek/mt76/mt7915/init.c
348
if (dev->mt76.region == NL80211_DFS_UNSET)
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
2042
int err, region;
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
2044
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
2046
region = 0;
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
2049
region = 2;
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
2053
region = 1;
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
2057
err = mt76_connac_mcu_rdd_cmd(&dev->mt76, RDD_START, rdd_idx, 0, region);
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
2099
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt7915/main.c
1677
if (dev->mt76.region == NL80211_DFS_UNSET)
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2728
int err, region, rdd_idx;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2748
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2750
region = 0;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2753
region = 2;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2757
region = 1;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2761
return mt76_connac_mcu_rdd_cmd(&dev->mt76, RDD_START, rdd_idx, 0, region);
drivers/net/wireless/mediatek/mt76/mt7921/init.c
137
dev->mt76.region = request->dfs_region;
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
416
const struct mt76_connac2_fw_region *region;
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
448
region = (const void *)((const u8 *)hdr -
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
449
(hdr->n_region - i) * sizeof(*region));
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
450
len = le32_to_cpu(region->len);
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
459
if ((region->feature_set & FW_FEATURE_NON_DL) &&
drivers/net/wireless/mediatek/mt76/mt7921/mcu.c
460
region->type == FW_TYPE_CLC) {
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
732
const struct mt76_connac2_fw_region *region;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
764
region = (const void *)((const u8 *)hdr -
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
765
(hdr->n_region - i) * sizeof(*region));
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
766
len = le32_to_cpu(region->len);
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
775
if ((region->feature_set & FW_FEATURE_NON_DL) &&
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
776
region->type == FW_TYPE_CLC) {
drivers/net/wireless/mediatek/mt76/mt7925/regd.c
191
mdev->region = req->dfs_region;
drivers/net/wireless/mediatek/mt76/mt792x_acpi_sar.c
245
switch (phy->mt76->dev->region) {
drivers/net/wireless/mediatek/mt76/mt792x_core.c
730
const struct mt76_connac2_fw_region *region;
drivers/net/wireless/mediatek/mt76/mt792x_core.c
732
region = (const void *)((const u8 *)hdr -
drivers/net/wireless/mediatek/mt76/mt792x_core.c
733
(hdr->n_region - i) * sizeof(*region));
drivers/net/wireless/mediatek/mt76/mt792x_core.c
734
offset += le32_to_cpu(region->len);
drivers/net/wireless/mediatek/mt76/mt7996/init.c
390
dev->mt76.region = request->dfs_region;
drivers/net/wireless/mediatek/mt76/mt7996/init.c
393
if (dev->mt76.region == NL80211_DFS_UNSET)
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2977
int err, region;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2979
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2981
region = 0;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2984
region = 2;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2988
region = 1;
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
2992
err = mt7996_mcu_rdd_cmd(dev, RDD_START, rdd_idx, region);
drivers/net/wireless/mediatek/mt76/mt7996/mac.c
3025
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt7996/main.c
2129
if (dev->mt76.region == NL80211_DFS_UNSET)
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3041
const struct mt7996_fw_region *region;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3045
region = (const struct mt7996_fw_region *)((const u8 *)hdr -
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3046
(hdr->n_region - i) * sizeof(*region));
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3049
region->feature_set,
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3051
len = le32_to_cpu(region->len);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3052
addr = le32_to_cpu(region->addr);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3054
if (region->feature_set & FW_FEATURE_OVERRIDE_ADDR)
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3664
int err, region, rdd_idx = mt7996_get_rdd_idx(phy, true);
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3680
switch (dev->mt76.region) {
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3682
region = 0;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3685
region = 2;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3689
region = 1;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3693
return mt7996_mcu_rdd_cmd(dev, RDD_START, rdd_idx, region);
drivers/nvdimm/namespace_devs.c
2059
static void deactivate_labels(void *region)
drivers/nvdimm/namespace_devs.c
2061
struct nd_region *nd_region = region;
drivers/nvdimm/of_pmem.c
56
struct nd_region *region;
drivers/nvdimm/of_pmem.c
70
region = nvdimm_volatile_region_create(bus, &ndr_desc);
drivers/nvdimm/of_pmem.c
73
region = nvdimm_pmem_region_create(bus, &ndr_desc);
drivers/nvdimm/of_pmem.c
76
if (!region)
drivers/nvmem/imx-ocotp-scu.c
100
if ((index >= data->region[i].start) &&
drivers/nvmem/imx-ocotp-scu.c
101
(index <= data->region[i].end))
drivers/nvmem/imx-ocotp-scu.c
38
struct ocotp_region region[];
drivers/nvmem/imx-ocotp-scu.c
58
.region = {
drivers/nvmem/imx-ocotp-scu.c
69
.region = {
drivers/nvmem/imx-ocotp-scu.c
82
if (data->region[i].flag & HOLE_REGION) {
drivers/nvmem/imx-ocotp-scu.c
83
if ((index >= data->region[i].start) &&
drivers/nvmem/imx-ocotp-scu.c
84
(index <= data->region[i].end))
drivers/nvmem/imx-ocotp-scu.c
99
if (data->region[i].flag & ECC_REGION) {
drivers/pci/bus.c
174
struct pci_bus_region *region)
drivers/pci/bus.c
179
if (r.start < region->start)
drivers/pci/bus.c
180
r.start = region->start;
drivers/pci/bus.c
181
if (r.end > region->end)
drivers/pci/bus.c
182
r.end = region->end;
drivers/pci/bus.c
195
struct pci_bus_region *region)
drivers/pci/bus.c
223
pci_clip_resource_to_region(bus, &avail, region);
drivers/pci/controller/pci-xgene.c
490
int region;
drivers/pci/controller/pci-xgene.c
492
region = xgene_pcie_select_ib_reg(ib_reg_mask, range->size);
drivers/pci/controller/pci-xgene.c
493
if (region < 0) {
drivers/pci/controller/pci-xgene.c
502
switch (region) {
drivers/pci/controller/pcie-rockchip-ep.c
66
u32 region)
drivers/pci/controller/pcie-rockchip-ep.c
69
ROCKCHIP_PCIE_AT_OB_REGION_PCI_ADDR0(region));
drivers/pci/controller/pcie-rockchip-ep.c
71
ROCKCHIP_PCIE_AT_OB_REGION_PCI_ADDR1(region));
drivers/pci/controller/pcie-rockchip-ep.c
73
ROCKCHIP_PCIE_AT_OB_REGION_DESC0(region));
drivers/pci/controller/pcie-rockchip-ep.c
75
ROCKCHIP_PCIE_AT_OB_REGION_DESC1(region));
drivers/pci/host-bridge.c
100
res->end = region->end + offset;
drivers/pci/host-bridge.c
52
void pcibios_resource_to_bus(struct pci_bus *bus, struct pci_bus_region *region,
drivers/pci/host-bridge.c
66
region->start = res->start - offset;
drivers/pci/host-bridge.c
67
region->end = res->end - offset;
drivers/pci/host-bridge.c
78
struct pci_bus_region *region)
drivers/pci/host-bridge.c
93
if (region_contains(&bus_region, region)) {
drivers/pci/host-bridge.c
99
res->start = region->start + offset;
drivers/pci/ide.c
538
static void mem_assoc_to_regs(struct pci_bus_region *region,
drivers/pci/ide.c
542
struct range r = { region->start, region->end };
drivers/pci/iov.c
1041
struct pci_bus_region region;
drivers/pci/iov.c
1075
pcibios_resource_to_bus(dev->bus, &region, res);
drivers/pci/iov.c
1076
new = region.start;
drivers/pci/iov.c
1082
new = region.start >> 16 >> 16;
drivers/pci/probe.c
2022
struct pci_bus_region region;
drivers/pci/probe.c
2126
region.start = 0x1F0;
drivers/pci/probe.c
2127
region.end = 0x1F7;
drivers/pci/probe.c
2130
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/probe.c
2133
region.start = 0x3F6;
drivers/pci/probe.c
2134
region.end = 0x3F6;
drivers/pci/probe.c
2137
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/probe.c
2142
region.start = 0x170;
drivers/pci/probe.c
2143
region.end = 0x177;
drivers/pci/probe.c
2146
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/probe.c
2149
region.start = 0x376;
drivers/pci/probe.c
2150
region.end = 0x376;
drivers/pci/probe.c
2153
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/probe.c
223
struct pci_bus_region region, inverted_region;
drivers/pci/probe.c
305
region.start = l64;
drivers/pci/probe.c
306
region.end = l64 + sz64 - 1;
drivers/pci/probe.c
308
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/probe.c
322
if (inverted_region.start != region.start) {
drivers/pci/probe.c
325
res->end = region.end - region.start;
drivers/pci/probe.c
327
res_name, (unsigned long long)region.start);
drivers/pci/probe.c
396
struct pci_bus_region region;
drivers/pci/probe.c
423
region.start = base;
drivers/pci/probe.c
424
region.end = limit + io_granularity - 1;
drivers/pci/probe.c
425
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/probe.c
439
struct pci_bus_region region;
drivers/pci/probe.c
449
region.start = base;
drivers/pci/probe.c
450
region.end = limit + 0xfffff;
drivers/pci/probe.c
451
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/probe.c
466
struct pci_bus_region region;
drivers/pci/probe.c
505
region.start = base;
drivers/pci/probe.c
506
region.end = limit + 0xfffff;
drivers/pci/probe.c
507
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/pci/quirks.c
621
u32 region;
drivers/pci/quirks.c
626
pci_read_config_dword(dev, PCI_BASE_ADDRESS_0 + (pos << 2), &region);
drivers/pci/quirks.c
628
if (!region)
drivers/pci/quirks.c
632
res->flags = region & ~PCI_BASE_ADDRESS_IO_MASK;
drivers/pci/quirks.c
635
region &= ~(size - 1);
drivers/pci/quirks.c
638
bus_region.start = region;
drivers/pci/quirks.c
639
bus_region.end = region + size - 1;
drivers/pci/quirks.c
671
u16 region;
drivers/pci/quirks.c
675
pci_read_config_word(dev, port, &region);
drivers/pci/quirks.c
676
region &= ~(size - 1);
drivers/pci/quirks.c
678
if (!region)
drivers/pci/quirks.c
685
bus_region.start = region;
drivers/pci/quirks.c
686
bus_region.end = region + size - 1;
drivers/pci/rom.c
27
struct pci_bus_region region;
drivers/pci/rom.c
42
pcibios_resource_to_bus(pdev->bus, &region, res);
drivers/pci/rom.c
45
rom_addr |= region.start | PCI_ROM_ADDRESS_ENABLE;
drivers/pci/setup-bus.c
1792
struct pci_bus_region region;
drivers/pci/setup-bus.c
1798
pcibios_resource_to_bus(dev->bus, &region, r);
drivers/pci/setup-bus.c
1799
if (!region.start) {
drivers/pci/setup-bus.c
799
struct pci_bus_region region;
drivers/pci/setup-bus.c
812
pcibios_resource_to_bus(bridge->bus, &region, res);
drivers/pci/setup-bus.c
815
io_base_lo = (region.start >> 8) & io_mask;
drivers/pci/setup-bus.c
816
io_limit_lo = (region.end >> 8) & io_mask;
drivers/pci/setup-bus.c
819
io_upper16 = (region.end & 0xffff0000) | (region.start >> 16);
drivers/pci/setup-bus.c
838
struct pci_bus_region region;
drivers/pci/setup-bus.c
844
pcibios_resource_to_bus(bridge->bus, &region, res);
drivers/pci/setup-bus.c
846
l = (region.start >> 16) & 0xfff0;
drivers/pci/setup-bus.c
847
l |= region.end & 0xfff00000;
drivers/pci/setup-bus.c
859
struct pci_bus_region region;
drivers/pci/setup-bus.c
873
pcibios_resource_to_bus(bridge->bus, &region, res);
drivers/pci/setup-bus.c
875
l = (region.start >> 16) & 0xfff0;
drivers/pci/setup-bus.c
876
l |= region.end & 0xfff00000;
drivers/pci/setup-bus.c
878
bu = upper_32_bits(region.start);
drivers/pci/setup-bus.c
879
lu = upper_32_bits(region.end);
drivers/pci/setup-cardbus.c
131
struct pci_bus_region region;
drivers/pci/setup-cardbus.c
137
pcibios_resource_to_bus(bridge->bus, &region, res);
drivers/pci/setup-cardbus.c
145
region.start);
drivers/pci/setup-cardbus.c
147
region.end);
drivers/pci/setup-cardbus.c
151
pcibios_resource_to_bus(bridge->bus, &region, res);
drivers/pci/setup-cardbus.c
155
region.start);
drivers/pci/setup-cardbus.c
157
region.end);
drivers/pci/setup-cardbus.c
161
pcibios_resource_to_bus(bridge->bus, &region, res);
drivers/pci/setup-cardbus.c
165
region.start);
drivers/pci/setup-cardbus.c
167
region.end);
drivers/pci/setup-cardbus.c
171
pcibios_resource_to_bus(bridge->bus, &region, res);
drivers/pci/setup-cardbus.c
175
region.start);
drivers/pci/setup-cardbus.c
177
region.end);
drivers/pci/setup-res.c
113
new = region.start >> 16 >> 16;
drivers/pci/setup-res.c
27
struct pci_bus_region region;
drivers/pci/setup-res.c
57
pcibios_resource_to_bus(dev->bus, &region, res);
drivers/pci/setup-res.c
58
new = region.start;
drivers/pcmcia/i82092.c
596
struct pci_bus_region region;
drivers/pcmcia/i82092.c
600
pcibios_resource_to_bus(sock_info->dev->bus, &region, mem->res);
drivers/pcmcia/i82092.c
606
if ((mem->card_start > 0x3ffffff) || (region.start > region.end) ||
drivers/pcmcia/i82092.c
611
(unsigned long long)region.start,
drivers/pcmcia/i82092.c
612
(unsigned long long)region.end,
drivers/pcmcia/i82092.c
623
i = (region.start >> 12) & 0x0fff;
drivers/pcmcia/i82092.c
632
i = (region.end >> 12) & 0x0fff;
drivers/pcmcia/i82092.c
651
i = ((mem->card_start - region.start) >> 12) & 0x3fff;
drivers/pcmcia/yenta_socket.c
1048
struct pci_bus_region region;
drivers/pcmcia/yenta_socket.c
1050
pcibios_resource_to_bus(socket->dev->bus, &region, &dev->resource[0]);
drivers/pcmcia/yenta_socket.c
1053
config_writel(socket, PCI_BASE_ADDRESS_0, region.start);
drivers/pcmcia/yenta_socket.c
445
struct pci_bus_region region;
drivers/pcmcia/yenta_socket.c
451
pcibios_resource_to_bus(socket->dev->bus, &region, mem->res);
drivers/pcmcia/yenta_socket.c
454
start = region.start;
drivers/pcmcia/yenta_socket.c
455
stop = region.end;
drivers/pcmcia/yenta_socket.c
697
struct pci_bus_region region;
drivers/pcmcia/yenta_socket.c
713
region.start = config_readl(socket, addr_start) & mask;
drivers/pcmcia/yenta_socket.c
714
region.end = config_readl(socket, addr_end) | ~mask;
drivers/pcmcia/yenta_socket.c
715
if (region.start && region.end > region.start && !override_bios) {
drivers/pcmcia/yenta_socket.c
716
pcibios_bus_to_resource(dev->bus, res, &region);
drivers/perf/arm-ccn.c
1327
int (*callback)(struct arm_ccn *ccn, int region,
drivers/perf/arm-ccn.c
1330
int region;
drivers/perf/arm-ccn.c
1332
for (region = 0; region < CCN_NUM_REGIONS; region++) {
drivers/perf/arm-ccn.c
1338
4 * (region / 32));
drivers/perf/arm-ccn.c
1339
if (!(val & (1 << (region % 32))))
drivers/perf/arm-ccn.c
1342
base = ccn->base + region * CCN_REGION_SIZE;
drivers/perf/arm-ccn.c
1349
err = callback(ccn, region, base, type, id);
drivers/perf/arm-ccn.c
1357
static int arm_ccn_get_nodes_num(struct arm_ccn *ccn, int region,
drivers/perf/arm-ccn.c
1369
static int arm_ccn_init_nodes(struct arm_ccn *ccn, int region,
drivers/perf/arm-ccn.c
1374
dev_dbg(ccn->dev, "Region %d: id=%u, type=0x%02x\n", region, id, type);
drivers/perf/arm_v7_pmu.c
1339
static u32 krait_get_pmresrn_event(unsigned int region)
drivers/perf/arm_v7_pmu.c
1344
return pmresrn_table[region];
drivers/perf/arm_v7_pmu.c
1352
unsigned int region = EVENT_REGION(config_base);
drivers/perf/arm_v7_pmu.c
1365
val = krait_get_pmresrn_event(region);
drivers/perf/arm_v7_pmu.c
1380
val = krait_read_pmresrn(region);
drivers/perf/arm_v7_pmu.c
1384
krait_write_pmresrn(region, val);
drivers/perf/arm_v7_pmu.c
1408
unsigned int region = EVENT_REGION(config_base);
drivers/perf/arm_v7_pmu.c
1419
val = krait_read_pmresrn(region);
drivers/perf/arm_v7_pmu.c
1421
krait_write_pmresrn(region, val);
drivers/perf/arm_v7_pmu.c
1489
static int krait_event_to_bit(struct perf_event *event, unsigned int region,
drivers/perf/arm_v7_pmu.c
1499
bit = krait_get_pmresrn_event(region);
drivers/perf/arm_v7_pmu.c
1521
unsigned int region = EVENT_REGION(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1529
if (group > 3 || region > 2)
drivers/perf/arm_v7_pmu.c
1534
bit = krait_event_to_bit(event, region, group);
drivers/perf/arm_v7_pmu.c
1551
unsigned int region = EVENT_REGION(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1558
bit = krait_event_to_bit(event, region, group);
drivers/perf/arm_v7_pmu.c
1662
static u32 scorpion_get_pmresrn_event(unsigned int region)
drivers/perf/arm_v7_pmu.c
1668
return pmresrn_table[region];
drivers/perf/arm_v7_pmu.c
1676
unsigned int region = EVENT_REGION(config_base);
drivers/perf/arm_v7_pmu.c
1689
val = scorpion_get_pmresrn_event(region);
drivers/perf/arm_v7_pmu.c
1706
val = scorpion_read_pmresrn(region);
drivers/perf/arm_v7_pmu.c
1710
scorpion_write_pmresrn(region, val);
drivers/perf/arm_v7_pmu.c
1718
unsigned int region = EVENT_REGION(config_base);
drivers/perf/arm_v7_pmu.c
1729
val = scorpion_read_pmresrn(region);
drivers/perf/arm_v7_pmu.c
1731
scorpion_write_pmresrn(region, val);
drivers/perf/arm_v7_pmu.c
1799
static int scorpion_event_to_bit(struct perf_event *event, unsigned int region,
drivers/perf/arm_v7_pmu.c
1809
bit = scorpion_get_pmresrn_event(region);
drivers/perf/arm_v7_pmu.c
1831
unsigned int region = EVENT_REGION(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1838
if (group > 3 || region > 3)
drivers/perf/arm_v7_pmu.c
1841
bit = scorpion_event_to_bit(event, region, group);
drivers/perf/arm_v7_pmu.c
1858
unsigned int region = EVENT_REGION(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1865
bit = scorpion_event_to_bit(event, region, group);
drivers/phy/samsung/phy-exynos5-usbdrd.c
420
.region = PTR_PHY \
drivers/phy/samsung/phy-exynos5-usbdrd.c
427
.region = PTR_PCS \
drivers/phy/samsung/phy-exynos5-usbdrd.c
434
.region = PTR_PMA, \
drivers/phy/samsung/phy-exynos5-usbdrd.c
437
#define PHY_TUNING_ENTRY_LAST { .region = PTR_INVALID }
drivers/phy/samsung/phy-exynos5-usbdrd.c
440
for (; (tune)->region != PTR_INVALID; ++(tune))
drivers/phy/samsung/phy-exynos5-usbdrd.c
446
char region;
drivers/phy/samsung/phy-exynos5-usbdrd.c
686
switch (tune->region) {
drivers/phy/samsung/phy-exynos5-usbdrd.c
698
"unknown phy region %d\n", tune->region);
drivers/pinctrl/nxp/pinctrl-s32cc.c
135
struct s32_pinctrl_mem_region *region;
drivers/pinctrl/nxp/pinctrl-s32cc.c
138
region = s32_get_region(pctldev, pin);
drivers/pinctrl/nxp/pinctrl-s32cc.c
139
if (!region)
drivers/pinctrl/nxp/pinctrl-s32cc.c
142
offset = (pin - region->pin_range->start) *
drivers/pinctrl/nxp/pinctrl-s32cc.c
143
regmap_get_reg_stride(region->map);
drivers/pinctrl/nxp/pinctrl-s32cc.c
145
return regmap_read(region->map, offset, val);
drivers/pinctrl/nxp/pinctrl-s32cc.c
152
struct s32_pinctrl_mem_region *region;
drivers/pinctrl/nxp/pinctrl-s32cc.c
155
region = s32_get_region(pctldev, pin);
drivers/pinctrl/nxp/pinctrl-s32cc.c
156
if (!region)
drivers/pinctrl/nxp/pinctrl-s32cc.c
159
offset = (pin - region->pin_range->start) *
drivers/pinctrl/nxp/pinctrl-s32cc.c
160
regmap_get_reg_stride(region->map);
drivers/pinctrl/nxp/pinctrl-s32cc.c
162
return regmap_write(region->map, offset, val);
drivers/pinctrl/nxp/pinctrl-s32cc.c
169
struct s32_pinctrl_mem_region *region;
drivers/pinctrl/nxp/pinctrl-s32cc.c
172
region = s32_get_region(pctldev, pin);
drivers/pinctrl/nxp/pinctrl-s32cc.c
173
if (!region)
drivers/pinctrl/nxp/pinctrl-s32cc.c
176
offset = (pin - region->pin_range->start) *
drivers/pinctrl/nxp/pinctrl-s32cc.c
177
regmap_get_reg_stride(region->map);
drivers/pinctrl/nxp/pinctrl-s32cc.c
179
return regmap_update_bits(region->map, offset, mask, val);
drivers/pinctrl/uniphier/pinctrl-uniphier-core.c
746
struct uniphier_pinctrl_reg_region *region;
drivers/pinctrl/uniphier/pinctrl-uniphier-core.c
754
region = devm_kzalloc(dev, struct_size(region, vals, nregs),
drivers/pinctrl/uniphier/pinctrl-uniphier-core.c
756
if (!region)
drivers/pinctrl/uniphier/pinctrl-uniphier-core.c
759
region->base = base;
drivers/pinctrl/uniphier/pinctrl-uniphier-core.c
760
region->nregs = nregs;
drivers/pinctrl/uniphier/pinctrl-uniphier-core.c
762
list_add_tail(&region->node, &priv->reg_regions);
drivers/platform/x86/intel/pmt/discovery-kunit.c
27
struct telemetry_region *region = &feature_group->regions[i];
drivers/platform/x86/intel/pmt/discovery-kunit.c
30
i, region->plat_info.cdie_mask, region->plat_info.package_id,
drivers/platform/x86/intel/pmt/discovery-kunit.c
31
region->plat_info.partition, region->plat_info.segment);
drivers/platform/x86/intel/pmt/discovery-kunit.c
33
region->plat_info.bus_number, region->plat_info.device_number,
drivers/platform/x86/intel/pmt/discovery-kunit.c
34
region->plat_info.function_number, region->guid);
drivers/platform/x86/intel/pmt/discovery-kunit.c
35
kunit_info(test, "\t\taddr=%p, size=%zu, num_rmids=%u", region->addr, region->size,
drivers/platform/x86/intel/pmt/discovery-kunit.c
36
region->num_rmids);
drivers/platform/x86/intel/pmt/discovery-kunit.c
39
KUNIT_ASSERT_GE(test, region->plat_info.cdie_mask, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
40
KUNIT_ASSERT_GE(test, region->plat_info.package_id, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
41
KUNIT_ASSERT_GE(test, region->plat_info.partition, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
42
KUNIT_ASSERT_GE(test, region->plat_info.segment, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
43
KUNIT_ASSERT_GE(test, region->plat_info.bus_number, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
44
KUNIT_ASSERT_GE(test, region->plat_info.device_number, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
45
KUNIT_ASSERT_GE(test, region->plat_info.function_number, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
47
KUNIT_ASSERT_NE(test, region->guid, 0);
drivers/platform/x86/intel/pmt/discovery-kunit.c
49
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, (__force const void *)region->addr);
drivers/platform/x86/intel/pmt/telemetry.c
217
static int pmt_copy_region(struct telemetry_region *region,
drivers/platform/x86/intel/pmt/telemetry.c
227
region->plat_info = *plat_info;
drivers/platform/x86/intel/pmt/telemetry.c
228
region->guid = entry->guid;
drivers/platform/x86/intel/pmt/telemetry.c
229
region->addr = entry->ep->base;
drivers/platform/x86/intel/pmt/telemetry.c
230
region->size = entry->size;
drivers/platform/x86/intel/pmt/telemetry.c
231
region->num_rmids = entry->num_rmids;
drivers/platform/x86/intel/pmt/telemetry.c
247
struct telemetry_region *region;
drivers/platform/x86/intel/pmt/telemetry.c
272
region = feature_group->regions;
drivers/platform/x86/intel/pmt/telemetry.c
279
ret = pmt_copy_region(region, entry);
drivers/platform/x86/intel/pmt/telemetry.c
283
region++;
drivers/pnp/quirks.c
356
struct pci_bus_region region;
drivers/pnp/quirks.c
379
region.start = addr_lo & ~0x7fff;
drivers/pnp/quirks.c
381
region.start |= (u64) addr_hi << 32;
drivers/pnp/quirks.c
382
region.end = region.start + 32*1024 - 1;
drivers/pnp/quirks.c
386
pcibios_bus_to_resource(host->bus, &mch, &region);
drivers/ps3/ps3av.c
39
int region;
drivers/ps3/ps3av.c
649
if (ps3av->region & PS3AV_REGION_60)
drivers/ps3/ps3av.c
782
if (ps3av->region & PS3AV_REGION_60)
drivers/ps3/ps3av.c
786
if (ps3av->region & PS3AV_REGION_RGB)
drivers/ps3/ps3av.c
951
ps3av->region = PS3AV_REGION_60;
drivers/ps3/ps3av.c
955
ps3av->region = PS3AV_REGION_50;
drivers/ps3/ps3av.c
958
ps3av->region = PS3AV_REGION_50 | PS3AV_REGION_RGB;
drivers/ps3/ps3av.c
961
ps3av->region = PS3AV_REGION_60;
drivers/remoteproc/qcom_common.c
111
struct minidump_region region;
drivers/remoteproc/qcom_common.c
129
memcpy_fromio(&region, ptr + i, sizeof(region));
drivers/remoteproc/qcom_common.c
130
if (le32_to_cpu(region.valid) == MINIDUMP_REGION_VALID) {
drivers/remoteproc/qcom_common.c
131
name = kstrndup(region.name, MAX_REGION_NAME_LENGTH - 1, GFP_KERNEL);
drivers/remoteproc/qcom_common.c
136
da = le64_to_cpu(region.address);
drivers/remoteproc/qcom_common.c
137
size = le64_to_cpu(region.size);
drivers/s390/cio/vfio_ccw_async.c
20
struct ccw_cmd_region *region;
drivers/s390/cio/vfio_ccw_async.c
23
if (pos + count > sizeof(*region))
drivers/s390/cio/vfio_ccw_async.c
27
region = private->region[i].data;
drivers/s390/cio/vfio_ccw_async.c
28
if (copy_to_user(buf, (void *)region + pos, count))
drivers/s390/cio/vfio_ccw_async.c
42
struct ccw_cmd_region *region;
drivers/s390/cio/vfio_ccw_async.c
45
if (pos + count > sizeof(*region))
drivers/s390/cio/vfio_ccw_async.c
51
region = private->region[i].data;
drivers/s390/cio/vfio_ccw_async.c
52
if (copy_from_user((void *)region + pos, buf, count)) {
drivers/s390/cio/vfio_ccw_async.c
59
ret = region->ret_code ? region->ret_code : count;
drivers/s390/cio/vfio_ccw_async.c
67
struct vfio_ccw_region *region)
drivers/s390/cio/vfio_ccw_chp.c
100
region = private->region[i].data;
drivers/s390/cio/vfio_ccw_chp.c
103
memcpy(&region->crw, &crw->crw, sizeof(region->crw));
drivers/s390/cio/vfio_ccw_chp.c
105
if (copy_to_user(buf, (void *)region + pos, count))
drivers/s390/cio/vfio_ccw_chp.c
110
region->crw = 0;
drivers/s390/cio/vfio_ccw_chp.c
131
struct vfio_ccw_region *region)
drivers/s390/cio/vfio_ccw_chp.c
22
struct ccw_schib_region *region;
drivers/s390/cio/vfio_ccw_chp.c
25
if (pos + count > sizeof(*region))
drivers/s390/cio/vfio_ccw_chp.c
29
region = private->region[i].data;
drivers/s390/cio/vfio_ccw_chp.c
36
memcpy(region, &sch->schib, sizeof(*region));
drivers/s390/cio/vfio_ccw_chp.c
38
if (copy_to_user(buf, (void *)region + pos, count)) {
drivers/s390/cio/vfio_ccw_chp.c
59
struct vfio_ccw_region *region)
drivers/s390/cio/vfio_ccw_chp.c
86
struct ccw_crw_region *region;
drivers/s390/cio/vfio_ccw_chp.c
90
if (pos + count > sizeof(*region))
drivers/s390/cio/vfio_ccw_ops.c
213
struct ccw_io_region *region;
drivers/s390/cio/vfio_ccw_ops.c
216
if (pos + count > sizeof(*region))
drivers/s390/cio/vfio_ccw_ops.c
220
region = private->io_region;
drivers/s390/cio/vfio_ccw_ops.c
221
if (copy_to_user(buf, (void *)region + pos, count))
drivers/s390/cio/vfio_ccw_ops.c
246
return private->region[index].ops->read(private, buf, count,
drivers/s390/cio/vfio_ccw_ops.c
258
struct ccw_io_region *region;
drivers/s390/cio/vfio_ccw_ops.c
261
if (pos + count > sizeof(*region))
drivers/s390/cio/vfio_ccw_ops.c
267
region = private->io_region;
drivers/s390/cio/vfio_ccw_ops.c
268
if (copy_from_user((void *)region + pos, buf, count)) {
drivers/s390/cio/vfio_ccw_ops.c
274
ret = (region->ret_code != 0) ? region->ret_code : count;
drivers/s390/cio/vfio_ccw_ops.c
298
return private->region[index].ops->write(private, buf, count,
drivers/s390/cio/vfio_ccw_ops.c
348
info->size = private->region[i].size;
drivers/s390/cio/vfio_ccw_ops.c
349
info->flags = private->region[i].flags;
drivers/s390/cio/vfio_ccw_ops.c
351
cap_type.type = private->region[i].type;
drivers/s390/cio/vfio_ccw_ops.c
352
cap_type.subtype = private->region[i].subtype;
drivers/s390/cio/vfio_ccw_ops.c
458
struct vfio_ccw_region *region;
drivers/s390/cio/vfio_ccw_ops.c
460
region = krealloc(private->region,
drivers/s390/cio/vfio_ccw_ops.c
461
(private->num_regions + 1) * sizeof(*region),
drivers/s390/cio/vfio_ccw_ops.c
463
if (!region)
drivers/s390/cio/vfio_ccw_ops.c
466
private->region = region;
drivers/s390/cio/vfio_ccw_ops.c
467
private->region[private->num_regions].type = VFIO_REGION_TYPE_CCW;
drivers/s390/cio/vfio_ccw_ops.c
468
private->region[private->num_regions].subtype = subtype;
drivers/s390/cio/vfio_ccw_ops.c
469
private->region[private->num_regions].ops = ops;
drivers/s390/cio/vfio_ccw_ops.c
470
private->region[private->num_regions].size = size;
drivers/s390/cio/vfio_ccw_ops.c
471
private->region[private->num_regions].flags = flags;
drivers/s390/cio/vfio_ccw_ops.c
472
private->region[private->num_regions].data = data;
drivers/s390/cio/vfio_ccw_ops.c
484
private->region[i].ops->release(private, &private->region[i]);
drivers/s390/cio/vfio_ccw_ops.c
486
kfree(private->region);
drivers/s390/cio/vfio_ccw_ops.c
487
private->region = NULL;
drivers/s390/cio/vfio_ccw_private.h
112
struct vfio_ccw_region *region;
drivers/s390/cio/vfio_ccw_private.h
43
struct vfio_ccw_region *region);
drivers/scsi/be2iscsi/be_mgmt.c
43
unsigned short region, sector_size, sector, offset;
drivers/scsi/be2iscsi/be_mgmt.c
47
region = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
drivers/scsi/be2iscsi/be_mgmt.c
51
req->region = region;
drivers/scsi/be2iscsi/be_mgmt.h
134
unsigned short region;
drivers/scsi/pm8001/pm8001_ctl.c
369
u8 *ptr = (u8 *)pm8001_ha->memoryMap.region[AAP1].virt_ptr;
drivers/scsi/pm8001/pm8001_ctl.c
411
memoryMap.region[ib_offset].virt_ptr + \
drivers/scsi/pm8001/pm8001_ctl.c
448
memoryMap.region[ob_offset].virt_ptr + \
drivers/scsi/pm8001/pm8001_ctl.c
539
u32 *temp = (u32 *)pm8001_ha->memoryMap.region[IOP].virt_ptr;
drivers/scsi/pm8001/pm8001_hwi.c
213
pm8001_ha->memoryMap.region[AAP1].phys_addr_hi;
drivers/scsi/pm8001/pm8001_hwi.c
215
pm8001_ha->memoryMap.region[AAP1].phys_addr_lo;
drivers/scsi/pm8001/pm8001_hwi.c
220
pm8001_ha->memoryMap.region[IOP].phys_addr_hi;
drivers/scsi/pm8001/pm8001_hwi.c
222
pm8001_ha->memoryMap.region[IOP].phys_addr_lo;
drivers/scsi/pm8001/pm8001_hwi.c
231
pm8001_ha->memoryMap.region[ib_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm8001_hwi.c
233
pm8001_ha->memoryMap.region[ib_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm8001_hwi.c
235
(u8 *)pm8001_ha->memoryMap.region[ib_offset + i].virt_ptr;
drivers/scsi/pm8001/pm8001_hwi.c
237
pm8001_ha->memoryMap.region[ib_offset + i].total_len;
drivers/scsi/pm8001/pm8001_hwi.c
239
pm8001_ha->memoryMap.region[ci_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm8001_hwi.c
241
pm8001_ha->memoryMap.region[ci_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm8001_hwi.c
243
pm8001_ha->memoryMap.region[ci_offset + i].virt_ptr;
drivers/scsi/pm8001/pm8001_hwi.c
258
pm8001_ha->memoryMap.region[ob_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm8001_hwi.c
260
pm8001_ha->memoryMap.region[ob_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm8001_hwi.c
262
(u8 *)pm8001_ha->memoryMap.region[ob_offset + i].virt_ptr;
drivers/scsi/pm8001/pm8001_hwi.c
264
pm8001_ha->memoryMap.region[ob_offset + i].total_len;
drivers/scsi/pm8001/pm8001_hwi.c
266
pm8001_ha->memoryMap.region[pi_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm8001_hwi.c
268
pm8001_ha->memoryMap.region[pi_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm8001_hwi.c
272
pm8001_ha->memoryMap.region[pi_offset + i].virt_ptr;
drivers/scsi/pm8001/pm8001_hwi.c
2897
void *virt_addr = pm8001_ha->memoryMap.region[NVMD].virt_ptr;
drivers/scsi/pm8001/pm8001_hwi.c
2943
pm8001_ha->memoryMap.region[NVMD].virt_ptr,
drivers/scsi/pm8001/pm8001_hwi.c
4400
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4402
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4409
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4411
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4418
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4420
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4427
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4429
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4437
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4439
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4471
memcpy(pm8001_ha->memoryMap.region[NVMD].virt_ptr,
drivers/scsi/pm8001/pm8001_hwi.c
4494
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4496
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4504
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4506
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4513
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4515
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4522
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi);
drivers/scsi/pm8001/pm8001_hwi.c
4524
cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo);
drivers/scsi/pm8001/pm8001_hwi.c
4578
void *buffer = pm8001_ha->memoryMap.region[FW_FLASH].virt_ptr;
drivers/scsi/pm8001/pm8001_hwi.c
4579
dma_addr_t phys_addr = pm8001_ha->memoryMap.region[FW_FLASH].phys_addr;
drivers/scsi/pm8001/pm8001_init.c
191
if (pm8001_ha->memoryMap.region[i].virt_ptr != NULL) {
drivers/scsi/pm8001/pm8001_init.c
193
(pm8001_ha->memoryMap.region[i].total_len +
drivers/scsi/pm8001/pm8001_init.c
194
pm8001_ha->memoryMap.region[i].alignment),
drivers/scsi/pm8001/pm8001_init.c
195
pm8001_ha->memoryMap.region[i].virt_ptr,
drivers/scsi/pm8001/pm8001_init.c
196
pm8001_ha->memoryMap.region[i].phys_addr);
drivers/scsi/pm8001/pm8001_init.c
347
pm8001_ha->memoryMap.region[AAP1].num_elements = 1;
drivers/scsi/pm8001/pm8001_init.c
348
pm8001_ha->memoryMap.region[AAP1].element_size = PM8001_EVENT_LOG_SIZE;
drivers/scsi/pm8001/pm8001_init.c
349
pm8001_ha->memoryMap.region[AAP1].total_len = PM8001_EVENT_LOG_SIZE;
drivers/scsi/pm8001/pm8001_init.c
350
pm8001_ha->memoryMap.region[AAP1].alignment = 32;
drivers/scsi/pm8001/pm8001_init.c
353
pm8001_ha->memoryMap.region[IOP].num_elements = 1;
drivers/scsi/pm8001/pm8001_init.c
354
pm8001_ha->memoryMap.region[IOP].element_size = PM8001_EVENT_LOG_SIZE;
drivers/scsi/pm8001/pm8001_init.c
355
pm8001_ha->memoryMap.region[IOP].total_len = PM8001_EVENT_LOG_SIZE;
drivers/scsi/pm8001/pm8001_init.c
356
pm8001_ha->memoryMap.region[IOP].alignment = 32;
drivers/scsi/pm8001/pm8001_init.c
362
pm8001_ha->memoryMap.region[ci_offset+i].num_elements = 1;
drivers/scsi/pm8001/pm8001_init.c
363
pm8001_ha->memoryMap.region[ci_offset+i].element_size = 4;
drivers/scsi/pm8001/pm8001_init.c
364
pm8001_ha->memoryMap.region[ci_offset+i].total_len = 4;
drivers/scsi/pm8001/pm8001_init.c
365
pm8001_ha->memoryMap.region[ci_offset+i].alignment = 4;
drivers/scsi/pm8001/pm8001_init.c
369
pm8001_ha->memoryMap.region[ib_offset+i].num_elements =
drivers/scsi/pm8001/pm8001_init.c
371
pm8001_ha->memoryMap.region[ib_offset+i].element_size
drivers/scsi/pm8001/pm8001_init.c
373
pm8001_ha->memoryMap.region[ib_offset+i].total_len =
drivers/scsi/pm8001/pm8001_init.c
375
pm8001_ha->memoryMap.region[ib_offset+i].alignment
drivers/scsi/pm8001/pm8001_init.c
378
pm8001_ha->memoryMap.region[ib_offset+i].num_elements =
drivers/scsi/pm8001/pm8001_init.c
380
pm8001_ha->memoryMap.region[ib_offset+i].element_size
drivers/scsi/pm8001/pm8001_init.c
382
pm8001_ha->memoryMap.region[ib_offset+i].total_len =
drivers/scsi/pm8001/pm8001_init.c
384
pm8001_ha->memoryMap.region[ib_offset+i].alignment = 64;
drivers/scsi/pm8001/pm8001_init.c
392
pm8001_ha->memoryMap.region[pi_offset+i].num_elements = 1;
drivers/scsi/pm8001/pm8001_init.c
393
pm8001_ha->memoryMap.region[pi_offset+i].element_size = 4;
drivers/scsi/pm8001/pm8001_init.c
394
pm8001_ha->memoryMap.region[pi_offset+i].total_len = 4;
drivers/scsi/pm8001/pm8001_init.c
395
pm8001_ha->memoryMap.region[pi_offset+i].alignment = 4;
drivers/scsi/pm8001/pm8001_init.c
399
pm8001_ha->memoryMap.region[ob_offset+i].num_elements =
drivers/scsi/pm8001/pm8001_init.c
401
pm8001_ha->memoryMap.region[ob_offset+i].element_size
drivers/scsi/pm8001/pm8001_init.c
403
pm8001_ha->memoryMap.region[ob_offset+i].total_len =
drivers/scsi/pm8001/pm8001_init.c
405
pm8001_ha->memoryMap.region[ob_offset+i].alignment
drivers/scsi/pm8001/pm8001_init.c
409
pm8001_ha->memoryMap.region[ob_offset+i].num_elements =
drivers/scsi/pm8001/pm8001_init.c
411
pm8001_ha->memoryMap.region[ob_offset+i].element_size
drivers/scsi/pm8001/pm8001_init.c
413
pm8001_ha->memoryMap.region[ob_offset+i].total_len =
drivers/scsi/pm8001/pm8001_init.c
415
pm8001_ha->memoryMap.region[ob_offset+i].alignment = 64;
drivers/scsi/pm8001/pm8001_init.c
420
pm8001_ha->memoryMap.region[NVMD].num_elements = 1;
drivers/scsi/pm8001/pm8001_init.c
421
pm8001_ha->memoryMap.region[NVMD].element_size = 4096;
drivers/scsi/pm8001/pm8001_init.c
422
pm8001_ha->memoryMap.region[NVMD].total_len = 4096;
drivers/scsi/pm8001/pm8001_init.c
425
pm8001_ha->memoryMap.region[FW_FLASH].total_len = 4096;
drivers/scsi/pm8001/pm8001_init.c
427
pm8001_ha->memoryMap.region[FORENSIC_MEM].num_elements = 1;
drivers/scsi/pm8001/pm8001_init.c
428
pm8001_ha->memoryMap.region[FORENSIC_MEM].total_len = 0x10000;
drivers/scsi/pm8001/pm8001_init.c
429
pm8001_ha->memoryMap.region[FORENSIC_MEM].element_size = 0x10000;
drivers/scsi/pm8001/pm8001_init.c
430
pm8001_ha->memoryMap.region[FORENSIC_MEM].alignment = 0x10000;
drivers/scsi/pm8001/pm8001_init.c
432
struct mpi_mem *region = &pm8001_ha->memoryMap.region[i];
drivers/scsi/pm8001/pm8001_init.c
435
&region->virt_ptr,
drivers/scsi/pm8001/pm8001_init.c
436
&region->phys_addr,
drivers/scsi/pm8001/pm8001_init.c
437
&region->phys_addr_hi,
drivers/scsi/pm8001/pm8001_init.c
438
&region->phys_addr_lo,
drivers/scsi/pm8001/pm8001_init.c
439
region->total_len,
drivers/scsi/pm8001/pm8001_init.c
440
region->alignment) != 0) {
drivers/scsi/pm8001/pm8001_init.c
461
if (pm8001_ha->memoryMap.region[i].virt_ptr != NULL) {
drivers/scsi/pm8001/pm8001_init.c
463
(pm8001_ha->memoryMap.region[i].total_len +
drivers/scsi/pm8001/pm8001_init.c
464
pm8001_ha->memoryMap.region[i].alignment),
drivers/scsi/pm8001/pm8001_init.c
465
pm8001_ha->memoryMap.region[i].virt_ptr,
drivers/scsi/pm8001/pm8001_init.c
466
pm8001_ha->memoryMap.region[i].phys_addr);
drivers/scsi/pm8001/pm8001_sas.h
318
struct mpi_mem region[USI_MAX_MEMCNT];
drivers/scsi/pm8001/pm80xx_hwi.c
179
temp = (u32 *)pm8001_ha->memoryMap.region[FORENSIC_MEM].virt_ptr;
drivers/scsi/pm8001/pm80xx_hwi.c
198
pm8001_ha->memoryMap.region[FORENSIC_MEM].virt_ptr,
drivers/scsi/pm8001/pm80xx_hwi.c
403
temp = (u32 *)pm8001_ha->memoryMap.region[FORENSIC_MEM].virt_ptr;
drivers/scsi/pm8001/pm80xx_hwi.c
418
pm8001_ha->memoryMap.region[FORENSIC_MEM].phys_addr_lo);
drivers/scsi/pm8001/pm80xx_hwi.c
422
pm8001_ha->memoryMap.region[FORENSIC_MEM].phys_addr_hi);
drivers/scsi/pm8001/pm80xx_hwi.c
754
pm8001_ha->memoryMap.region[AAP1].phys_addr_hi;
drivers/scsi/pm8001/pm80xx_hwi.c
756
pm8001_ha->memoryMap.region[AAP1].phys_addr_lo;
drivers/scsi/pm8001/pm80xx_hwi.c
761
pm8001_ha->memoryMap.region[IOP].phys_addr_hi;
drivers/scsi/pm8001/pm80xx_hwi.c
763
pm8001_ha->memoryMap.region[IOP].phys_addr_lo;
drivers/scsi/pm8001/pm80xx_hwi.c
781
pm8001_ha->memoryMap.region[ib_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm80xx_hwi.c
783
pm8001_ha->memoryMap.region[ib_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm80xx_hwi.c
785
(u8 *)pm8001_ha->memoryMap.region[ib_offset + i].virt_ptr;
drivers/scsi/pm8001/pm80xx_hwi.c
787
pm8001_ha->memoryMap.region[ib_offset + i].total_len;
drivers/scsi/pm8001/pm80xx_hwi.c
789
pm8001_ha->memoryMap.region[ci_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm80xx_hwi.c
791
pm8001_ha->memoryMap.region[ci_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm80xx_hwi.c
793
pm8001_ha->memoryMap.region[ci_offset + i].virt_ptr;
drivers/scsi/pm8001/pm80xx_hwi.c
813
pm8001_ha->memoryMap.region[ob_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm80xx_hwi.c
815
pm8001_ha->memoryMap.region[ob_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm80xx_hwi.c
817
(u8 *)pm8001_ha->memoryMap.region[ob_offset + i].virt_ptr;
drivers/scsi/pm8001/pm80xx_hwi.c
819
pm8001_ha->memoryMap.region[ob_offset + i].total_len;
drivers/scsi/pm8001/pm80xx_hwi.c
821
pm8001_ha->memoryMap.region[pi_offset + i].phys_addr_hi;
drivers/scsi/pm8001/pm80xx_hwi.c
823
pm8001_ha->memoryMap.region[pi_offset + i].phys_addr_lo;
drivers/scsi/pm8001/pm80xx_hwi.c
827
pm8001_ha->memoryMap.region[pi_offset + i].virt_ptr;
drivers/scsi/qla2xxx/qla_fw.h
1712
struct qla_flt_region region[];
drivers/scsi/qla2xxx/qla_mbx.c
6995
uint16_t region, uint32_t len, dma_addr_t sfub_dma_addr,
drivers/scsi/qla2xxx/qla_mbx.c
7004
mcp->mb[2] = region;
drivers/scsi/qla2xxx/qla_sup.c
2704
struct qla_flt_region *region)
drivers/scsi/qla2xxx/qla_sup.c
2708
struct qla_flt_region *flt_reg = &flt->region[0];
drivers/scsi/qla2xxx/qla_sup.c
2718
memcpy((uint8_t *)region, flt_reg,
drivers/scsi/qla2xxx/qla_sup.c
2743
struct qla_flt_region region;
drivers/scsi/qla2xxx/qla_sup.c
2749
rval = qla28xx_get_flash_region(vha, offset, &region);
drivers/scsi/qla2xxx/qla_sup.c
2772
if (ha->flags.secure_adapter && region.attribute) {
drivers/scsi/qla2xxx/qla_sup.c
2775
"Region %x is secure\n", le16_to_cpu(region.code));
drivers/scsi/qla2xxx/qla_sup.c
2777
switch (le16_to_cpu(region.code)) {
drivers/scsi/qla2xxx/qla_sup.c
2829
le16_to_cpu(region.code));
drivers/scsi/qla2xxx/qla_sup.c
2944
if (region.attribute && buf_size_without_sfub) {
drivers/scsi/qla2xxx/qla_sup.c
2948
le16_to_cpu(region.code),
drivers/scsi/qla2xxx/qla_sup.c
689
struct qla_flt_region *region = &flt->region[0];
drivers/scsi/qla2xxx/qla_sup.c
726
cnt = le16_to_cpu(flt->length) / sizeof(*region);
drivers/scsi/qla2xxx/qla_sup.c
727
for ( ; cnt; cnt--, region++) {
drivers/scsi/qla2xxx/qla_sup.c
729
start = le32_to_cpu(region->start) >> 2;
drivers/scsi/qla2xxx/qla_sup.c
732
le16_to_cpu(region->code), start,
drivers/scsi/qla2xxx/qla_sup.c
733
le32_to_cpu(region->end) >> 2,
drivers/scsi/qla2xxx/qla_sup.c
734
le32_to_cpu(region->size) >> 2);
drivers/scsi/qla2xxx/qla_sup.c
735
if (region->attribute)
drivers/scsi/qla2xxx/qla_sup.c
737
"Region %x is secure\n", region->code);
drivers/scsi/qla2xxx/qla_sup.c
739
switch (le16_to_cpu(region->code)) {
drivers/scsi/qla4xxx/ql4_bsg.c
346
uint32_t region = 0;
drivers/scsi/qla4xxx/ql4_bsg.c
365
region = bsg_req->rqst_data.h_vendor.vendor_cmd[1];
drivers/scsi/qla4xxx/ql4_bsg.c
369
rval = qla4xxx_restore_factory_defaults(ha, region, field0, field1);
drivers/scsi/qla4xxx/ql4_glbl.h
174
uint32_t region, uint32_t field0,
drivers/scsi/qla4xxx/ql4_mbx.c
2249
uint32_t region, uint32_t field0,
drivers/scsi/qla4xxx/ql4_mbx.c
2260
mbox_cmd[3] = region;
drivers/scsi/qla4xxx/ql4_nx.c
3700
struct qla_flt_region *region;
drivers/scsi/qla4xxx/ql4_nx.c
3706
region = (struct qla_flt_region *)&flt[1];
drivers/scsi/qla4xxx/ql4_nx.c
3742
for ( ; cnt; cnt--, region++) {
drivers/scsi/qla4xxx/ql4_nx.c
3744
start = le32_to_cpu(region->start) >> 2;
drivers/scsi/qla4xxx/ql4_nx.c
3747
"end=0x%x size=0x%x.\n", le32_to_cpu(region->code), start,
drivers/scsi/qla4xxx/ql4_nx.c
3748
le32_to_cpu(region->end) >> 2, le32_to_cpu(region->size)));
drivers/scsi/qla4xxx/ql4_nx.c
3750
switch (le32_to_cpu(region->code) & 0xff) {
drivers/scsi/qla4xxx/ql4_nx.c
3769
hw->flt_chap_size = le32_to_cpu(region->size);
drivers/scsi/qla4xxx/ql4_nx.c
3773
hw->flt_ddb_size = le32_to_cpu(region->size);
drivers/soc/aspeed/aspeed-p2a-ctrl.c
140
const struct region *curr = &ctrl->config->regions[i];
drivers/soc/aspeed/aspeed-p2a-ctrl.c
52
struct region regions[P2A_REGION_COUNT];
drivers/soc/qcom/ocmem.c
119
struct ocmem_region *region = &ocmem->regions[i];
drivers/soc/qcom/ocmem.c
121
if (region->mode == THIN_MODE)
drivers/soc/qcom/ocmem.c
131
struct ocmem_region *region = &ocmem->regions[i];
drivers/soc/qcom/ocmem.c
134
data = OCMEM_PSGSC_CTL_MACRO0_MODE(region->macro_state[0]) |
drivers/soc/qcom/ocmem.c
135
OCMEM_PSGSC_CTL_MACRO1_MODE(region->macro_state[1]) |
drivers/soc/qcom/ocmem.c
136
OCMEM_PSGSC_CTL_MACRO2_MODE(region->macro_state[2]) |
drivers/soc/qcom/ocmem.c
137
OCMEM_PSGSC_CTL_MACRO3_MODE(region->macro_state[3]);
drivers/soc/qcom/ocmem.c
169
struct ocmem_region *region = &ocmem->regions[i];
drivers/soc/qcom/ocmem.c
172
region->mode = rmode;
drivers/soc/qcom/ocmem.c
174
for (j = 0; j < region->num_macros; j++) {
drivers/soc/qcom/ocmem.c
177
region->macro_state[j] = mstate;
drivers/soc/qcom/ocmem.c
179
offset += region->macro_size;
drivers/soc/qcom/ocmem.c
381
struct ocmem_region *region = &ocmem->regions[i];
drivers/soc/qcom/ocmem.c
383
if (WARN_ON(num_banks > ARRAY_SIZE(region->macro_state))) {
drivers/soc/qcom/ocmem.c
388
region->mode = MODE_DEFAULT;
drivers/soc/qcom/ocmem.c
389
region->num_macros = num_banks;
drivers/soc/qcom/ocmem.c
393
region->macro_size = ocmem->config->macro_size / 2;
drivers/soc/qcom/ocmem.c
394
region->region_size = region_size / 2;
drivers/soc/qcom/ocmem.c
396
region->macro_size = ocmem->config->macro_size;
drivers/soc/qcom/ocmem.c
397
region->region_size = region_size;
drivers/soc/qcom/ocmem.c
400
for (j = 0; j < ARRAY_SIZE(region->macro_state); j++)
drivers/soc/qcom/ocmem.c
401
region->macro_state[j] = CLK_OFF;
drivers/soc/qcom/smem.c
1069
static int qcom_smem_map_toc(struct qcom_smem *smem, struct smem_region *region)
drivers/soc/qcom/smem.c
1074
region->virt_base = devm_ioremap_wc(smem->dev, region->aux_base, SZ_4K);
drivers/soc/qcom/smem.c
1075
ptable_start = region->aux_base + region->size - SZ_4K;
drivers/soc/qcom/smem.c
1079
if (!region->virt_base || !smem->ptable)
drivers/soc/qcom/smem.c
1101
struct smem_region *region)
drivers/soc/qcom/smem.c
1119
region->aux_base = r.start;
drivers/soc/qcom/smem.c
1120
region->size = resource_size(&r);
drivers/soc/qcom/smem.c
554
struct smem_region *region;
drivers/soc/qcom/smem.c
569
region = &smem->regions[i];
drivers/soc/qcom/smem.c
571
if ((u32)region->aux_base == aux_base || !aux_base) {
drivers/soc/qcom/smem.c
575
if (WARN_ON(e_size + entry_offset > region->size))
drivers/soc/qcom/smem.c
581
return region->virt_base + entry_offset;
drivers/soc/ti/knav_qmss.h
222
struct knav_region *region;
drivers/soc/ti/knav_qmss.h
348
#define for_each_region(kdev, region) \
drivers/soc/ti/knav_qmss.h
349
list_for_each_entry(region, &kdev->regions, list)
drivers/soc/ti/knav_qmss_queue.c
1000
region->num_desc = 0;
drivers/soc/ti/knav_qmss_queue.c
1002
region->name);
drivers/soc/ti/knav_qmss_queue.c
1006
size = region->num_desc * region->desc_size;
drivers/soc/ti/knav_qmss_queue.c
1007
region->virt_start = alloc_pages_exact(size, GFP_KERNEL | GFP_DMA |
drivers/soc/ti/knav_qmss_queue.c
1009
if (!region->virt_start) {
drivers/soc/ti/knav_qmss_queue.c
1010
region->num_desc = 0;
drivers/soc/ti/knav_qmss_queue.c
1012
region->name);
drivers/soc/ti/knav_qmss_queue.c
1015
region->virt_end = region->virt_start + size;
drivers/soc/ti/knav_qmss_queue.c
1016
page = virt_to_page(region->virt_start);
drivers/soc/ti/knav_qmss_queue.c
1018
region->dma_start = dma_map_page(kdev->dev, page, 0, size,
drivers/soc/ti/knav_qmss_queue.c
1020
if (dma_mapping_error(kdev->dev, region->dma_start)) {
drivers/soc/ti/knav_qmss_queue.c
1022
region->name);
drivers/soc/ti/knav_qmss_queue.c
1025
region->dma_end = region->dma_start + size;
drivers/soc/ti/knav_qmss_queue.c
1033
pool->region_offset = region->num_desc;
drivers/soc/ti/knav_qmss_queue.c
1034
list_add(&pool->region_inst, &region->pools);
drivers/soc/ti/knav_qmss_queue.c
1038
region->name, id, region->desc_size, region->num_desc,
drivers/soc/ti/knav_qmss_queue.c
1039
region->link_index, &region->dma_start, &region->dma_end,
drivers/soc/ti/knav_qmss_queue.c
1040
region->virt_start, region->virt_end);
drivers/soc/ti/knav_qmss_queue.c
1042
hw_desc_size = (region->desc_size / 16) - 1;
drivers/soc/ti/knav_qmss_queue.c
1047
writel_relaxed((u32)region->dma_start, &regs->base);
drivers/soc/ti/knav_qmss_queue.c
1048
writel_relaxed(region->link_index, &regs->start_index);
drivers/soc/ti/knav_qmss_queue.c
1055
if (region->dma_start)
drivers/soc/ti/knav_qmss_queue.c
1056
dma_unmap_page(kdev->dev, region->dma_start, size,
drivers/soc/ti/knav_qmss_queue.c
1058
if (region->virt_start)
drivers/soc/ti/knav_qmss_queue.c
1059
free_pages_exact(region->virt_start, size);
drivers/soc/ti/knav_qmss_queue.c
1060
region->num_desc = 0;
drivers/soc/ti/knav_qmss_queue.c
1081
struct knav_region *region;
drivers/soc/ti/knav_qmss_queue.c
1090
region = devm_kzalloc(dev, sizeof(*region), GFP_KERNEL);
drivers/soc/ti/knav_qmss_queue.c
1091
if (!region)
drivers/soc/ti/knav_qmss_queue.c
1094
region->name = knav_queue_find_name(child);
drivers/soc/ti/knav_qmss_queue.c
1095
of_property_read_u32(child, "id", &region->id);
drivers/soc/ti/knav_qmss_queue.c
1098
region->num_desc = temp[0];
drivers/soc/ti/knav_qmss_queue.c
1099
region->desc_size = temp[1];
drivers/soc/ti/knav_qmss_queue.c
1101
dev_err(dev, "invalid region info %s\n", region->name);
drivers/soc/ti/knav_qmss_queue.c
1102
devm_kfree(dev, region);
drivers/soc/ti/knav_qmss_queue.c
1107
&region->link_index);
drivers/soc/ti/knav_qmss_queue.c
1110
region->name);
drivers/soc/ti/knav_qmss_queue.c
1111
devm_kfree(dev, region);
drivers/soc/ti/knav_qmss_queue.c
1115
INIT_LIST_HEAD(&region->pools);
drivers/soc/ti/knav_qmss_queue.c
1116
list_add_tail(&region->list, &kdev->regions);
drivers/soc/ti/knav_qmss_queue.c
1123
for_each_region(kdev, region)
drivers/soc/ti/knav_qmss_queue.c
1124
knav_queue_setup_region(kdev, region);
drivers/soc/ti/knav_qmss_queue.c
1349
struct knav_region *region;
drivers/soc/ti/knav_qmss_queue.c
1354
region = first_region(kdev);
drivers/soc/ti/knav_qmss_queue.c
1355
if (!region)
drivers/soc/ti/knav_qmss_queue.c
1357
list_for_each_entry_safe(pool, tmp, &region->pools, region_inst)
drivers/soc/ti/knav_qmss_queue.c
1360
size = region->virt_end - region->virt_start;
drivers/soc/ti/knav_qmss_queue.c
1362
free_pages_exact(region->virt_start, size);
drivers/soc/ti/knav_qmss_queue.c
1363
list_del(&region->list);
drivers/soc/ti/knav_qmss_queue.c
1364
devm_kfree(kdev->dev, region);
drivers/soc/ti/knav_qmss_queue.c
688
struct knav_region *region;
drivers/soc/ti/knav_qmss_queue.c
691
region = pool->region;
drivers/soc/ti/knav_qmss_queue.c
692
pool->desc_size = region->desc_size;
drivers/soc/ti/knav_qmss_queue.c
697
dma_addr = region->dma_start + (region->desc_size * index);
drivers/soc/ti/knav_qmss_queue.c
735
return pool->region->dma_start + (virt - pool->region->virt_start);
drivers/soc/ti/knav_qmss_queue.c
742
return pool->region->virt_start + (dma - pool->region->dma_start);
drivers/soc/ti/knav_qmss_queue.c
759
struct knav_region *reg_itr, *region = NULL;
drivers/soc/ti/knav_qmss_queue.c
780
region = reg_itr;
drivers/soc/ti/knav_qmss_queue.c
784
if (!region) {
drivers/soc/ti/knav_qmss_queue.c
805
if (num_desc > (region->num_desc - region->used_desc)) {
drivers/soc/ti/knav_qmss_queue.c
817
node = &region->pools;
drivers/soc/ti/knav_qmss_queue.c
818
list_for_each_entry(iter, &region->pools, region_inst) {
drivers/soc/ti/knav_qmss_queue.c
828
pool->region = region;
drivers/soc/ti/knav_qmss_queue.c
831
region->used_desc += num_desc;
drivers/soc/ti/knav_qmss_queue.c
865
if (!pool->region)
drivers/soc/ti/knav_qmss_queue.c
871
pool->region->used_desc -= pool->num_desc;
drivers/soc/ti/knav_qmss_queue.c
913
knav_queue_push(pool->queue, dma, pool->region->desc_size, 0);
drivers/soc/ti/knav_qmss_queue.c
932
size = min(size, pool->region->desc_size);
drivers/soc/ti/knav_qmss_queue.c
959
desc_sz = min(dma_sz, pool->region->desc_size);
drivers/soc/ti/knav_qmss_queue.c
980
struct knav_region *region)
drivers/soc/ti/knav_qmss_queue.c
986
int id = region->id;
drivers/soc/ti/knav_qmss_queue.c
990
if (!region->num_desc) {
drivers/soc/ti/knav_qmss_queue.c
991
dev_warn(kdev->dev, "unused region %s\n", region->name);
drivers/soc/ti/knav_qmss_queue.c
996
hw_num_desc = ilog2(region->num_desc - 1) + 1;
drivers/soc/ti/knav_qmss_queue.c
999
if (region->num_desc < 32) {
drivers/soc/ti/pruss.c
113
struct pruss_mem_region *region)
drivers/soc/ti/pruss.c
115
if (!pruss || !region || mem_id >= PRUSS_MEM_MAX)
drivers/soc/ti/pruss.c
125
*region = pruss->mem_regions[mem_id];
drivers/soc/ti/pruss.c
126
pruss->mem_in_use[mem_id] = region;
drivers/soc/ti/pruss.c
146
struct pruss_mem_region *region)
drivers/soc/ti/pruss.c
150
if (!pruss || !region)
drivers/soc/ti/pruss.c
157
if (pruss->mem_in_use[id] == region)
drivers/spi/spi-intel.c
1247
u32 region, base, limit;
drivers/spi/spi-intel.c
1249
region = readl(ispi->base + FREG(i));
drivers/spi/spi-intel.c
1250
base = region & FREG_BASE_MASK;
drivers/spi/spi-intel.c
1251
limit = (region & FREG_LIMIT_MASK) >> FREG_LIMIT_SHIFT;
drivers/spi/spi-intel.c
259
u32 region, base, limit;
drivers/spi/spi-intel.c
261
region = readl(ispi->base + FREG(i));
drivers/spi/spi-intel.c
262
base = region & FREG_BASE_MASK;
drivers/spi/spi-intel.c
263
limit = (region & FREG_LIMIT_MASK) >> FREG_LIMIT_SHIFT;
drivers/staging/sm750fb/sm750.c
159
const struct fb_fillrect *region)
drivers/staging/sm750fb/sm750.c
180
color = (bpp == 1) ? region->color :
drivers/staging/sm750fb/sm750.c
181
((u32 *)info->pseudo_palette)[region->color];
drivers/staging/sm750fb/sm750.c
182
rop = (region->rop != ROP_COPY) ? HW_ROP2_XOR : HW_ROP2_COPY;
drivers/staging/sm750fb/sm750.c
194
region->dx, region->dy,
drivers/staging/sm750fb/sm750.c
195
region->width, region->height,
drivers/staging/sm750fb/sm750.c
201
const struct fb_copyarea *region)
drivers/staging/sm750fb/sm750.c
227
base, pitch, region->sx, region->sy,
drivers/staging/sm750fb/sm750.c
228
base, pitch, bpp, region->dx, region->dy,
drivers/staging/sm750fb/sm750.c
229
region->width, region->height,
drivers/staging/vme_user/vme_tsi148.c
810
struct pci_bus_region region;
drivers/staging/vme_user/vme_tsi148.c
850
pcibios_resource_to_bus(pdev->bus, &region,
drivers/staging/vme_user/vme_tsi148.c
852
pci_base = region.start;
drivers/tty/serial/8250/8250_pci.c
783
struct pci_bus_region region;
drivers/tty/serial/8250/8250_pci.c
801
pcibios_resource_to_bus(dev->bus, &region, &dev->resource[bar]);
drivers/tty/serial/8250/8250_pci.c
802
device_window = ((region.start + MITE_IOWBSR1_WIN_OFFSET) & 0xffffff00)
drivers/ufs/core/ufs-rpmb.c
140
int region;
drivers/ufs/core/ufs-rpmb.c
163
for (region = 0; region < ARRAY_SIZE(hba->dev_info.rpmb_region_size); region++) {
drivers/ufs/core/ufs-rpmb.c
164
cap = hba->dev_info.rpmb_region_size[region];
drivers/ufs/core/ufs-rpmb.c
178
dev_set_name(&ufs_rpmb->dev, "ufs_rpmb%d", region);
drivers/ufs/core/ufs-rpmb.c
185
dev_err(hba->dev, "Failed to register UFS RPMB device %d\n", region);
drivers/ufs/core/ufs-rpmb.c
191
cid = kasprintf(GFP_KERNEL, "%s-R%d", hba->dev_info.device_id, region);
drivers/ufs/core/ufs-rpmb.c
215
ufs_rpmb->region_id = region;
drivers/ufs/core/ufs-rpmb.c
219
dev_info(hba->dev, "UFS RPMB region %d registered (capacity=%u)\n", region, cap);
drivers/usb/core/hcd-pci.c
232
int region;
drivers/usb/core/hcd-pci.c
234
for (region = 0; region < PCI_STD_NUM_BARS; region++) {
drivers/usb/core/hcd-pci.c
235
if (!(pci_resource_flags(dev, region) &
drivers/usb/core/hcd-pci.c
239
hcd->rsrc_start = pci_resource_start(dev, region);
drivers/usb/core/hcd-pci.c
240
hcd->rsrc_len = pci_resource_len(dev, region);
drivers/usb/core/hcd-pci.c
245
if (region == PCI_STD_NUM_BARS) {
drivers/usb/gadget/udc/net2280.c
3605
if (dev->region)
drivers/usb/gadget/udc/net2280.c
3664
dev->region = 1;
drivers/usb/gadget/udc/net2280.h
158
region:1,
drivers/vfio/cdx/main.c
227
static int vfio_cdx_mmap_mmio(struct vfio_cdx_region region,
drivers/vfio/cdx/main.c
237
if (base + size > region.size)
drivers/vfio/cdx/main.c
240
vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff;
drivers/vfio/fsl-mc/vfio_fsl_mc.c
242
struct vfio_fsl_mc_region *region;
drivers/vfio/fsl-mc/vfio_fsl_mc.c
249
region = &vdev->regions[index];
drivers/vfio/fsl-mc/vfio_fsl_mc.c
251
if (!(region->flags & VFIO_REGION_INFO_FLAG_READ))
drivers/vfio/fsl-mc/vfio_fsl_mc.c
254
if (!region->ioaddr) {
drivers/vfio/fsl-mc/vfio_fsl_mc.c
255
region->ioaddr = ioremap(region->addr, region->size);
drivers/vfio/fsl-mc/vfio_fsl_mc.c
256
if (!region->ioaddr)
drivers/vfio/fsl-mc/vfio_fsl_mc.c
264
data[i] = readq(region->ioaddr + i * sizeof(uint64_t));
drivers/vfio/fsl-mc/vfio_fsl_mc.c
321
struct vfio_fsl_mc_region *region;
drivers/vfio/fsl-mc/vfio_fsl_mc.c
328
region = &vdev->regions[index];
drivers/vfio/fsl-mc/vfio_fsl_mc.c
330
if (!(region->flags & VFIO_REGION_INFO_FLAG_WRITE))
drivers/vfio/fsl-mc/vfio_fsl_mc.c
333
if (!region->ioaddr) {
drivers/vfio/fsl-mc/vfio_fsl_mc.c
334
region->ioaddr = ioremap(region->addr, region->size);
drivers/vfio/fsl-mc/vfio_fsl_mc.c
335
if (!region->ioaddr)
drivers/vfio/fsl-mc/vfio_fsl_mc.c
345
ret = vfio_fsl_mc_send_command(region->ioaddr, data);
drivers/vfio/fsl-mc/vfio_fsl_mc.c
353
static int vfio_fsl_mc_mmap_mmio(struct vfio_fsl_mc_region region,
drivers/vfio/fsl-mc/vfio_fsl_mc.c
364
if (region.size < PAGE_SIZE || base + size > region.size)
drivers/vfio/fsl-mc/vfio_fsl_mc.c
367
region_cacheable = (region.type & FSL_MC_REGION_CACHEABLE) &&
drivers/vfio/fsl-mc/vfio_fsl_mc.c
368
(region.type & FSL_MC_REGION_SHAREABLE);
drivers/vfio/fsl-mc/vfio_fsl_mc.c
372
vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff;
drivers/vfio/pci/nvgrace-gpu/main.c
101
region = nvgrace_gpu_memregion(index, nvdev);
drivers/vfio/pci/nvgrace-gpu/main.c
102
if (!region)
drivers/vfio/pci/nvgrace-gpu/main.c
105
start_pfn = PHYS_PFN(region->memphys);
drivers/vfio/pci/nvgrace-gpu/main.c
106
num_pages = region->memlength >> PAGE_SHIFT;
drivers/vfio/pci/nvgrace-gpu/main.c
152
struct mem_region *region)
drivers/vfio/pci/nvgrace-gpu/main.c
156
pfn = PHYS_PFN(region->memphys);
drivers/vfio/pci/nvgrace-gpu/main.c
157
nr_pages = region->memlength >> PAGE_SHIFT;
drivers/vfio/pci/nvgrace-gpu/main.c
159
region->pfn_address_space.node.start = pfn;
drivers/vfio/pci/nvgrace-gpu/main.c
160
region->pfn_address_space.node.last = pfn + nr_pages - 1;
drivers/vfio/pci/nvgrace-gpu/main.c
161
region->pfn_address_space.mapping = core_vdev->inode->i_mapping;
drivers/vfio/pci/nvgrace-gpu/main.c
162
region->pfn_address_space.pfn_to_vma_pgoff = nvgrace_gpu_pfn_to_vma_pgoff;
drivers/vfio/pci/nvgrace-gpu/main.c
164
return register_pfn_address_space(&region->pfn_address_space);
drivers/vfio/pci/nvgrace-gpu/main.c
98
struct mem_region *region;
drivers/vfio/pci/pds/dirty.c
100
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
102
vfree(region->host_seq);
drivers/vfio/pci/pds/dirty.c
103
vfree(region->host_ack);
drivers/vfio/pci/pds/dirty.c
104
region->host_seq = NULL;
drivers/vfio/pci/pds/dirty.c
105
region->host_ack = NULL;
drivers/vfio/pci/pds/dirty.c
106
region->bmp_bytes = 0;
drivers/vfio/pci/pds/dirty.c
111
struct pds_vfio_region *region)
drivers/vfio/pci/pds/dirty.c
116
dma_unmap_single(pdsc_dev, region->sgl_addr,
drivers/vfio/pci/pds/dirty.c
117
region->num_sge * sizeof(struct pds_lm_sg_elem),
drivers/vfio/pci/pds/dirty.c
119
kfree(region->sgl);
drivers/vfio/pci/pds/dirty.c
121
region->num_sge = 0;
drivers/vfio/pci/pds/dirty.c
122
region->sgl = NULL;
drivers/vfio/pci/pds/dirty.c
123
region->sgl_addr = 0;
drivers/vfio/pci/pds/dirty.c
134
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
136
if (region->sgl)
drivers/vfio/pci/pds/dirty.c
137
__pds_vfio_dirty_free_sgl(pds_vfio, region);
drivers/vfio/pci/pds/dirty.c
142
struct pds_vfio_region *region,
drivers/vfio/pci/pds/dirty.c
165
region->sgl = sgl;
drivers/vfio/pci/pds/dirty.c
166
region->num_sge = max_sge;
drivers/vfio/pci/pds/dirty.c
167
region->sgl_addr = sgl_addr;
drivers/vfio/pci/pds/dirty.c
195
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
204
err = pds_vfio_dirty_alloc_bitmaps(region,
drivers/vfio/pci/pds/dirty.c
212
err = pds_vfio_dirty_alloc_sgl(pds_vfio, region, page_count);
drivers/vfio/pci/pds/dirty.c
219
region->size = region_size;
drivers/vfio/pci/pds/dirty.c
220
region->start = region_start;
drivers/vfio/pci/pds/dirty.c
221
region->page_size = region_page_size;
drivers/vfio/pci/pds/dirty.c
222
region->dev_bmp_offset_start_byte = dev_bmp_offset_byte;
drivers/vfio/pci/pds/dirty.c
370
struct pds_vfio_region *region,
drivers/vfio/pci/pds/dirty.c
425
struct pds_lm_sg_elem *sg_elem = &region->sgl[i];
drivers/vfio/pci/pds/dirty.c
433
offset += region->dev_bmp_offset_start_byte;
drivers/vfio/pci/pds/dirty.c
434
dma_sync_single_for_device(pdsc_dev, region->sgl_addr, size, dma_dir);
drivers/vfio/pci/pds/dirty.c
435
err = pds_vfio_dirty_seq_ack_cmd(pds_vfio, region->sgl_addr, num_sge,
drivers/vfio/pci/pds/dirty.c
441
num_sge, region->sgl_addr, ERR_PTR(err));
drivers/vfio/pci/pds/dirty.c
442
dma_sync_single_for_cpu(pdsc_dev, region->sgl_addr, size, dma_dir);
drivers/vfio/pci/pds/dirty.c
454
struct pds_vfio_region *region,
drivers/vfio/pci/pds/dirty.c
458
return pds_vfio_dirty_seq_ack(pds_vfio, region, region->host_ack,
drivers/vfio/pci/pds/dirty.c
463
struct pds_vfio_region *region,
drivers/vfio/pci/pds/dirty.c
466
return pds_vfio_dirty_seq_ack(pds_vfio, region, region->host_seq,
drivers/vfio/pci/pds/dirty.c
471
struct pds_vfio_region *region,
drivers/vfio/pci/pds/dirty.c
475
u64 page_size = region->page_size;
drivers/vfio/pci/pds/dirty.c
476
u64 region_start = region->start;
drivers/vfio/pci/pds/dirty.c
482
seq = (__le64 *)((u64)region->host_seq + bmp_offset);
drivers/vfio/pci/pds/dirty.c
483
ack = (__le64 *)((u64)region->host_ack + bmp_offset);
drivers/vfio/pci/pds/dirty.c
512
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
514
if (iova >= region->start &&
drivers/vfio/pci/pds/dirty.c
515
iova < (region->start + region->size))
drivers/vfio/pci/pds/dirty.c
516
return region;
drivers/vfio/pci/pds/dirty.c
527
struct pds_vfio_region *region;
drivers/vfio/pci/pds/dirty.c
540
region = pds_vfio_get_region(pds_vfio, iova);
drivers/vfio/pci/pds/dirty.c
541
if (!region) {
drivers/vfio/pci/pds/dirty.c
547
pages = DIV_ROUND_UP(length, region->page_size);
drivers/vfio/pci/pds/dirty.c
553
pds_vfio->vf_id, iova, length, region->page_size,
drivers/vfio/pci/pds/dirty.c
556
if (!length || ((iova - region->start + length) > region->size)) {
drivers/vfio/pci/pds/dirty.c
563
bmp_bytes = ALIGN(DIV_ROUND_UP(length / region->page_size,
drivers/vfio/pci/pds/dirty.c
572
if (bmp_bytes > region->bmp_bytes) {
drivers/vfio/pci/pds/dirty.c
575
bmp_bytes, region->bmp_bytes);
drivers/vfio/pci/pds/dirty.c
579
bmp_offset = DIV_ROUND_UP((iova - region->start) /
drivers/vfio/pci/pds/dirty.c
580
region->page_size, sizeof(u64));
drivers/vfio/pci/pds/dirty.c
586
err = pds_vfio_dirty_read_seq(pds_vfio, region, bmp_offset, bmp_bytes);
drivers/vfio/pci/pds/dirty.c
590
err = pds_vfio_dirty_process_bitmaps(pds_vfio, region, dirty_bitmap,
drivers/vfio/pci/pds/dirty.c
595
err = pds_vfio_dirty_write_ack(pds_vfio, region, bmp_offset, bmp_bytes);
drivers/vfio/pci/pds/dirty.c
72
static int pds_vfio_dirty_alloc_bitmaps(struct pds_vfio_region *region,
drivers/vfio/pci/pds/dirty.c
87
region->host_seq = host_seq_bmp;
drivers/vfio/pci/pds/dirty.c
88
region->host_ack = host_ack_bmp;
drivers/vfio/pci/pds/dirty.c
89
region->bmp_bytes = bytes;
drivers/vfio/pci/vfio_pci_core.c
1128
info->size = vdev->region[i].size;
drivers/vfio/pci/vfio_pci_core.c
1129
info->flags = vdev->region[i].flags;
drivers/vfio/pci/vfio_pci_core.c
1131
cap_type.type = vdev->region[i].type;
drivers/vfio/pci/vfio_pci_core.c
1132
cap_type.subtype = vdev->region[i].subtype;
drivers/vfio/pci/vfio_pci_core.c
1139
if (vdev->region[i].ops->add_capability) {
drivers/vfio/pci/vfio_pci_core.c
1140
ret = vdev->region[i].ops->add_capability(
drivers/vfio/pci/vfio_pci_core.c
1141
vdev, &vdev->region[i], caps);
drivers/vfio/pci/vfio_pci_core.c
1581
ret = vdev->region[index].ops->rw(vdev, buf,
drivers/vfio/pci/vfio_pci_core.c
1747
struct vfio_pci_region *region = vdev->region + regnum;
drivers/vfio/pci/vfio_pci_core.c
1749
if (region->ops && region->ops->mmap &&
drivers/vfio/pci/vfio_pci_core.c
1750
(region->flags & VFIO_REGION_INFO_FLAG_MMAP))
drivers/vfio/pci/vfio_pci_core.c
1751
return region->ops->mmap(vdev, region, vma);
drivers/vfio/pci/vfio_pci_core.c
2127
kfree(vdev->region);
drivers/vfio/pci/vfio_pci_core.c
641
vdev->region[i].ops->release(vdev, &vdev->region[i]);
drivers/vfio/pci/vfio_pci_core.c
644
kfree(vdev->region);
drivers/vfio/pci/vfio_pci_core.c
645
vdev->region = NULL; /* don't krealloc a freed pointer */
drivers/vfio/pci/vfio_pci_core.c
926
struct vfio_pci_region *region;
drivers/vfio/pci/vfio_pci_core.c
928
region = krealloc(vdev->region,
drivers/vfio/pci/vfio_pci_core.c
929
(vdev->num_regions + 1) * sizeof(*region),
drivers/vfio/pci/vfio_pci_core.c
931
if (!region)
drivers/vfio/pci/vfio_pci_core.c
934
vdev->region = region;
drivers/vfio/pci/vfio_pci_core.c
935
vdev->region[vdev->num_regions].type = type;
drivers/vfio/pci/vfio_pci_core.c
936
vdev->region[vdev->num_regions].subtype = subtype;
drivers/vfio/pci/vfio_pci_core.c
937
vdev->region[vdev->num_regions].ops = ops;
drivers/vfio/pci/vfio_pci_core.c
938
vdev->region[vdev->num_regions].size = size;
drivers/vfio/pci/vfio_pci_core.c
939
vdev->region[vdev->num_regions].flags = flags;
drivers/vfio/pci/vfio_pci_core.c
940
vdev->region[vdev->num_regions].data = data;
drivers/vfio/pci/vfio_pci_igd.c
152
struct vfio_pci_region *region)
drivers/vfio/pci/vfio_pci_igd.c
154
struct igd_opregion_vbt *opregionvbt = region->data;
drivers/vfio/pci/vfio_pci_igd.c
286
struct pci_dev *pdev = vdev->region[i].data;
drivers/vfio/pci/vfio_pci_igd.c
291
if (pos >= vdev->region[i].size || iswrite)
drivers/vfio/pci/vfio_pci_igd.c
294
size = count = min(count, (size_t)(vdev->region[i].size - pos));
drivers/vfio/pci/vfio_pci_igd.c
378
struct vfio_pci_region *region)
drivers/vfio/pci/vfio_pci_igd.c
380
struct pci_dev *pdev = region->data;
drivers/vfio/pci/vfio_pci_igd.c
69
struct igd_opregion_vbt *opregionvbt = vdev->region[i].data;
drivers/vfio/pci/vfio_pci_igd.c
73
if (pos >= vdev->region[i].size || iswrite)
drivers/vfio/pci/vfio_pci_igd.c
76
count = min_t(size_t, count, vdev->region[i].size - pos);
drivers/vfio/platform/vfio_platform_common.c
548
static int vfio_platform_mmap_mmio(struct vfio_platform_region region,
drivers/vfio/platform/vfio_platform_common.c
558
if (region.size < PAGE_SIZE || req_start + req_len > region.size)
drivers/vfio/platform/vfio_platform_common.c
562
vma->vm_pgoff = (region.addr >> PAGE_SHIFT) + pgoff;
drivers/vfio/vfio_iommu_type1.c
1983
struct iommu_resv_region *region;
drivers/vfio/vfio_iommu_type1.c
1986
list_for_each_entry(region, group_resv_regions, list) {
drivers/vfio/vfio_iommu_type1.c
1992
if (region->type == IOMMU_RESV_MSI) {
drivers/vfio/vfio_iommu_type1.c
1997
if (region->type == IOMMU_RESV_SW_MSI) {
drivers/vfio/vfio_iommu_type1.c
1998
*base = region->start;
drivers/vfio/vfio_iommu_type1.c
2018
struct vfio_iova *region;
drivers/vfio/vfio_iommu_type1.c
2020
region = kmalloc_obj(*region);
drivers/vfio/vfio_iommu_type1.c
2021
if (!region)
drivers/vfio/vfio_iommu_type1.c
2024
INIT_LIST_HEAD(&region->list);
drivers/vfio/vfio_iommu_type1.c
2025
region->start = start;
drivers/vfio/vfio_iommu_type1.c
2026
region->end = end;
drivers/vfio/vfio_iommu_type1.c
2028
list_add_tail(&region->list, head);
drivers/vfio/vfio_iommu_type1.c
2113
struct iommu_resv_region *region;
drivers/vfio/vfio_iommu_type1.c
2116
list_for_each_entry(region, resv_regions, list) {
drivers/vfio/vfio_iommu_type1.c
2117
if (region->type == IOMMU_RESV_DIRECT_RELAXABLE)
drivers/vfio/vfio_iommu_type1.c
2120
if (vfio_find_dma(iommu, region->start, region->length))
drivers/vhost/vhost.c
1970
struct vhost_memory_region *region;
drivers/vhost/vhost.c
1998
for (region = newmem->regions;
drivers/vhost/vhost.c
1999
region < newmem->regions + mem.nregions;
drivers/vhost/vhost.c
2000
region++) {
drivers/vhost/vhost.c
2002
region->guest_phys_addr,
drivers/vhost/vhost.c
2003
region->guest_phys_addr +
drivers/vhost/vhost.c
2004
region->memory_size - 1,
drivers/vhost/vhost.c
2005
region->userspace_addr,
drivers/video/fbdev/aty/radeon_accel.c
44
const struct fb_fillrect *region)
drivers/video/fbdev/aty/radeon_accel.c
53
OUTREG(DP_BRUSH_FRGD_CLR, rinfo->pseudo_palette[region->color]);
drivers/video/fbdev/aty/radeon_accel.c
55
OUTREG(DP_BRUSH_FRGD_CLR, region->color);
drivers/video/fbdev/aty/radeon_accel.c
64
OUTREG(DST_Y_X, (region->dy << 16) | region->dx);
drivers/video/fbdev/aty/radeon_accel.c
65
OUTREG(DST_WIDTH_HEIGHT, (region->width << 16) | region->height);
drivers/video/fbdev/aty/radeon_accel.c
68
void radeonfb_fillrect(struct fb_info *info, const struct fb_fillrect *region)
drivers/video/fbdev/aty/radeon_accel.c
77
cfb_fillrect(info, region);
drivers/video/fbdev/aty/radeon_accel.c
86
memcpy(&modded, region, sizeof(struct fb_fillrect));
drivers/video/fbdev/aty/radeonfb.h
499
extern void radeonfb_fillrect(struct fb_info *info, const struct fb_fillrect *region);
drivers/video/fbdev/cirrusfb.c
1732
const struct fb_fillrect *region)
drivers/video/fbdev/cirrusfb.c
1739
cinfo->pseudo_palette[region->color] : region->color;
drivers/video/fbdev/cirrusfb.c
1744
cfb_fillrect(info, region);
drivers/video/fbdev/cirrusfb.c
1751
memcpy(&modded, region, sizeof(struct fb_fillrect));
drivers/video/fbdev/cirrusfb.c
1764
(region->dx * m) / 8, region->dy,
drivers/video/fbdev/cirrusfb.c
1765
(region->width * m) / 8, region->height,
drivers/video/fbdev/core/bitblit.c
238
struct fb_fillrect region;
drivers/video/fbdev/core/bitblit.c
240
region.color = color;
drivers/video/fbdev/core/bitblit.c
241
region.rop = ROP_COPY;
drivers/video/fbdev/core/bitblit.c
244
region.dx = info->var.xoffset + rs;
drivers/video/fbdev/core/bitblit.c
245
region.dy = 0;
drivers/video/fbdev/core/bitblit.c
246
region.width = rw;
drivers/video/fbdev/core/bitblit.c
247
region.height = info->var.yres_virtual;
drivers/video/fbdev/core/bitblit.c
248
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/bitblit.c
252
region.dx = info->var.xoffset;
drivers/video/fbdev/core/bitblit.c
253
region.dy = info->var.yoffset + bs;
drivers/video/fbdev/core/bitblit.c
254
region.width = rs;
drivers/video/fbdev/core/bitblit.c
255
region.height = bh;
drivers/video/fbdev/core/bitblit.c
256
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/bitblit.c
64
struct fb_fillrect region;
drivers/video/fbdev/core/bitblit.c
66
region.color = bg;
drivers/video/fbdev/core/bitblit.c
67
region.dx = sx * vc->vc_font.width;
drivers/video/fbdev/core/bitblit.c
68
region.dy = sy * vc->vc_font.height;
drivers/video/fbdev/core/bitblit.c
69
region.width = width * vc->vc_font.width;
drivers/video/fbdev/core/bitblit.c
70
region.height = height * vc->vc_font.height;
drivers/video/fbdev/core/bitblit.c
71
region.rop = ROP_COPY;
drivers/video/fbdev/core/bitblit.c
73
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_ccw.c
198
struct fb_fillrect region;
drivers/video/fbdev/core/fbcon_ccw.c
200
region.color = color;
drivers/video/fbdev/core/fbcon_ccw.c
201
region.rop = ROP_COPY;
drivers/video/fbdev/core/fbcon_ccw.c
204
region.dx = 0;
drivers/video/fbdev/core/fbcon_ccw.c
205
region.dy = info->var.yoffset;
drivers/video/fbdev/core/fbcon_ccw.c
206
region.height = rw;
drivers/video/fbdev/core/fbcon_ccw.c
207
region.width = info->var.xres_virtual;
drivers/video/fbdev/core/fbcon_ccw.c
208
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_ccw.c
212
region.dx = info->var.xoffset + bs;
drivers/video/fbdev/core/fbcon_ccw.c
213
region.dy = 0;
drivers/video/fbdev/core/fbcon_ccw.c
214
region.height = info->var.yres_virtual;
drivers/video/fbdev/core/fbcon_ccw.c
215
region.width = bh;
drivers/video/fbdev/core/fbcon_ccw.c
216
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_ccw.c
84
struct fb_fillrect region;
drivers/video/fbdev/core/fbcon_ccw.c
87
region.color = bg;
drivers/video/fbdev/core/fbcon_ccw.c
88
region.dx = sy * vc->vc_font.height;
drivers/video/fbdev/core/fbcon_ccw.c
89
region.dy = vyres - ((sx + width) * vc->vc_font.width);
drivers/video/fbdev/core/fbcon_ccw.c
90
region.height = width * vc->vc_font.width;
drivers/video/fbdev/core/fbcon_ccw.c
91
region.width = height * vc->vc_font.height;
drivers/video/fbdev/core/fbcon_ccw.c
92
region.rop = ROP_COPY;
drivers/video/fbdev/core/fbcon_ccw.c
94
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_cw.c
181
struct fb_fillrect region;
drivers/video/fbdev/core/fbcon_cw.c
183
region.color = color;
drivers/video/fbdev/core/fbcon_cw.c
184
region.rop = ROP_COPY;
drivers/video/fbdev/core/fbcon_cw.c
187
region.dx = 0;
drivers/video/fbdev/core/fbcon_cw.c
188
region.dy = info->var.yoffset + rs;
drivers/video/fbdev/core/fbcon_cw.c
189
region.height = rw;
drivers/video/fbdev/core/fbcon_cw.c
190
region.width = info->var.xres_virtual;
drivers/video/fbdev/core/fbcon_cw.c
191
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_cw.c
195
region.dx = info->var.xoffset;
drivers/video/fbdev/core/fbcon_cw.c
196
region.dy = info->var.yoffset;
drivers/video/fbdev/core/fbcon_cw.c
197
region.height = info->var.yres;
drivers/video/fbdev/core/fbcon_cw.c
198
region.width = bh;
drivers/video/fbdev/core/fbcon_cw.c
199
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_cw.c
69
struct fb_fillrect region;
drivers/video/fbdev/core/fbcon_cw.c
72
region.color = bg;
drivers/video/fbdev/core/fbcon_cw.c
73
region.dx = vxres - ((sy + height) * vc->vc_font.height);
drivers/video/fbdev/core/fbcon_cw.c
74
region.dy = sx * vc->vc_font.width;
drivers/video/fbdev/core/fbcon_cw.c
75
region.height = width * vc->vc_font.width;
drivers/video/fbdev/core/fbcon_cw.c
76
region.width = height * vc->vc_font.height;
drivers/video/fbdev/core/fbcon_cw.c
77
region.rop = ROP_COPY;
drivers/video/fbdev/core/fbcon_cw.c
79
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_ud.c
228
struct fb_fillrect region;
drivers/video/fbdev/core/fbcon_ud.c
230
region.color = color;
drivers/video/fbdev/core/fbcon_ud.c
231
region.rop = ROP_COPY;
drivers/video/fbdev/core/fbcon_ud.c
234
region.dy = 0;
drivers/video/fbdev/core/fbcon_ud.c
235
region.dx = info->var.xoffset;
drivers/video/fbdev/core/fbcon_ud.c
236
region.width = rw;
drivers/video/fbdev/core/fbcon_ud.c
237
region.height = info->var.yres_virtual;
drivers/video/fbdev/core/fbcon_ud.c
238
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_ud.c
242
region.dy = info->var.yoffset;
drivers/video/fbdev/core/fbcon_ud.c
243
region.dx = info->var.xoffset;
drivers/video/fbdev/core/fbcon_ud.c
244
region.height = bh;
drivers/video/fbdev/core/fbcon_ud.c
245
region.width = info->var.xres;
drivers/video/fbdev/core/fbcon_ud.c
246
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/core/fbcon_ud.c
70
struct fb_fillrect region;
drivers/video/fbdev/core/fbcon_ud.c
74
region.color = bg;
drivers/video/fbdev/core/fbcon_ud.c
75
region.dy = vyres - ((sy + height) * vc->vc_font.height);
drivers/video/fbdev/core/fbcon_ud.c
76
region.dx = vxres - ((sx + width) * vc->vc_font.width);
drivers/video/fbdev/core/fbcon_ud.c
77
region.width = width * vc->vc_font.width;
drivers/video/fbdev/core/fbcon_ud.c
78
region.height = height * vc->vc_font.height;
drivers/video/fbdev/core/fbcon_ud.c
79
region.rop = ROP_COPY;
drivers/video/fbdev/core/fbcon_ud.c
81
info->fbops->fb_fillrect(info, &region);
drivers/video/fbdev/cyber2000fb.c
1517
cfb->fb.screen_base = cfb->region;
drivers/video/fbdev/cyber2000fb.c
1699
cfb->region = pci_ioremap_bar(dev, 0);
drivers/video/fbdev/cyber2000fb.c
1700
if (!cfb->region) {
drivers/video/fbdev/cyber2000fb.c
1705
cfb->regs = cfb->region + MMIO_OFFSET;
drivers/video/fbdev/cyber2000fb.c
1750
iounmap(cfb->region);
drivers/video/fbdev/cyber2000fb.c
1766
iounmap(cfb->region);
drivers/video/fbdev/cyber2000fb.c
187
cyber2000fb_copyarea(struct fb_info *info, const struct fb_copyarea *region)
drivers/video/fbdev/cyber2000fb.c
194
cfb_copyarea(info, region);
drivers/video/fbdev/cyber2000fb.c
199
cyber2000fb_writew(region->width - 1, CO_REG_PIXWIDTH, cfb);
drivers/video/fbdev/cyber2000fb.c
200
cyber2000fb_writew(region->height - 1, CO_REG_PIXHEIGHT, cfb);
drivers/video/fbdev/cyber2000fb.c
202
src = region->sx + region->sy * cfb->fb.var.xres_virtual;
drivers/video/fbdev/cyber2000fb.c
203
dst = region->dx + region->dy * cfb->fb.var.xres_virtual;
drivers/video/fbdev/cyber2000fb.c
205
if (region->sx < region->dx) {
drivers/video/fbdev/cyber2000fb.c
206
src += region->width - 1;
drivers/video/fbdev/cyber2000fb.c
207
dst += region->width - 1;
drivers/video/fbdev/cyber2000fb.c
211
if (region->sy < region->dy) {
drivers/video/fbdev/cyber2000fb.c
212
src += (region->height - 1) * cfb->fb.var.xres_virtual;
drivers/video/fbdev/cyber2000fb.c
213
dst += (region->height - 1) * cfb->fb.var.xres_virtual;
drivers/video/fbdev/cyber2000fb.c
60
unsigned char __iomem *region;
drivers/video/fbdev/hpfb.c
151
static void hpfb_fillrect(struct fb_info *p, const struct fb_fillrect *region)
drivers/video/fbdev/hpfb.c
155
clr = region->color & 0xff;
drivers/video/fbdev/hpfb.c
162
out_8(fb_regs + WMRR, (region->rop == ROP_COPY ? RR_SET : RR_INVERT));
drivers/video/fbdev/hpfb.c
166
out_8(fb_regs + WMRR, (region->rop == ROP_COPY ? RR_CLEAR : RR_NOOP));
drivers/video/fbdev/hpfb.c
168
topcat_blit(region->dx, region->dy, region->dx, region->dy, region->width, region->height, -1);
drivers/video/fbdev/i810/i810_accel.c
326
void i810fb_copyarea(struct fb_info *info, const struct fb_copyarea *region)
drivers/video/fbdev/i810/i810_accel.c
333
cfb_copyarea(info, region);
drivers/video/fbdev/i810/i810_accel.c
337
dx = region->dx * par->depth;
drivers/video/fbdev/i810/i810_accel.c
338
sx = region->sx * par->depth;
drivers/video/fbdev/i810/i810_accel.c
339
width = region->width * par->depth;
drivers/video/fbdev/i810/i810_accel.c
340
sy = region->sy;
drivers/video/fbdev/i810/i810_accel.c
341
dy = region->dy;
drivers/video/fbdev/i810/i810_accel.c
342
height = region->height;
drivers/video/fbdev/i810/i810_main.h
30
const struct fb_copyarea *region);
drivers/video/fbdev/nvidia/nv_accel.c
298
void nvidiafb_copyarea(struct fb_info *info, const struct fb_copyarea *region)
drivers/video/fbdev/nvidia/nv_accel.c
306
cfb_copyarea(info, region);
drivers/video/fbdev/nvidia/nv_accel.c
311
NVDmaNext(par, (region->sy << 16) | region->sx);
drivers/video/fbdev/nvidia/nv_accel.c
312
NVDmaNext(par, (region->dy << 16) | region->dx);
drivers/video/fbdev/nvidia/nv_accel.c
313
NVDmaNext(par, (region->height << 16) | region->width);
drivers/video/fbdev/nvidia/nv_proto.h
52
const struct fb_copyarea *region);
drivers/video/fbdev/omap/lcdc.c
623
static int alloc_fbmem(struct omapfb_mem_region *region)
drivers/video/fbdev/omap/lcdc.c
633
if (region->size > frame_size)
drivers/video/fbdev/omap/lcdc.c
634
frame_size = region->size;
drivers/video/fbdev/omap/lcdc.c
642
region->size = frame_size;
drivers/video/fbdev/omap/lcdc.c
643
region->paddr = lcdc.vram_phys;
drivers/video/fbdev/omap/lcdc.c
644
region->vaddr = lcdc.vram_virt;
drivers/video/fbdev/omap/lcdc.c
645
region->alloc = 1;
drivers/video/fbdev/omap/lcdc.c
670
return alloc_fbmem(&req_md->region[0]);
drivers/video/fbdev/omap/omapfb.h
55
struct omapfb_mem_region region[OMAPFB_PLANE_NUM];
drivers/video/fbdev/omap/omapfb_main.c
1032
*(u16 *)fbdev->mem_desc.region[0].vaddr = pixval;
drivers/video/fbdev/omap/omapfb_main.c
135
fbdev->mem_desc.region[i].size =
drivers/video/fbdev/omap/omapfb_main.c
150
fbdev->mem_desc.region[0].size = PAGE_ALIGN(def_size);
drivers/video/fbdev/omap/omapfb_main.c
163
fbdev->mem_desc.region[i].paddr,
drivers/video/fbdev/omap/omapfb_main.c
164
fbdev->mem_desc.region[i].vaddr,
drivers/video/fbdev/omap/omapfb_main.c
165
fbdev->mem_desc.region[i].size);
drivers/video/fbdev/omap/omapfb_main.c
1736
vram += fbdev->mem_desc.region[i].size;
drivers/video/fbdev/omap/omapfb_main.c
376
rg = &plane->fbdev->mem_desc.region[plane->idx];
drivers/video/fbdev/omap/omapfb_main.c
523
max_frame_size = fbdev->mem_desc.region[plane->idx].size;
drivers/video/fbdev/omap/omapfb_main.c
780
if (pi->enabled && !fbdev->mem_desc.region[plane->idx].size) {
drivers/video/fbdev/omap/omapfb_main.c
819
struct omapfb_mem_region *rg = &fbdev->mem_desc.region[plane->idx];
drivers/video/fbdev/omap/omapfb_main.c
895
rg = &fbdev->mem_desc.region[plane->idx];
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
154
ofbi->region = old_rg;
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
218
rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
235
if (ofbi2->region != rg)
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
266
rg = omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
28
if (ofbi->id == ofbi->region->id)
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
31
return OMAPFB_MEM_IDX_ENABLED | ofbi->region->id;
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
69
old_rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-ioctl.c
98
ofbi->region = new_rg;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1005
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1009
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1022
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1033
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1057
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1061
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1100
rg = omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1124
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1304
rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1364
rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
152
const struct vrfb *vrfb = &ofbi->region->vrfb;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1555
rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1585
struct omapfb2_mem_region *rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1738
if (ofbi->region->size == 0) {
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
181
return ofbi->region->vrfb.paddr[rot]
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
184
return ofbi->region->paddr;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1890
ofbi->region = &fbdev->regions[i];
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1891
ofbi->region->id = i;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1892
init_rwsem(&ofbi->region->lock);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
191
return ofbi->region->vrfb.paddr[0];
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1926
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1928
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
193
return ofbi->region->paddr;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1940
if (ofbi->region->size == 0)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1963
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
1965
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
199
return ofbi->region->vrfb.vaddr[0];
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
201
return ofbi->region->vaddr;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
446
unsigned long max_frame_size = ofbi->region->size;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
493
struct omapfb2_mem_region *rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
554
fbi->screen_base = ofbi->region->vrfb.vaddr[0];
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
556
fix->smem_start = ofbi->region->vrfb.paddr[0];
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
595
struct omapfb2_mem_region *rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
664
WARN_ON(!atomic_read(&ofbi->region->lock_count));
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
686
if (ofbi->region->size != 0 && check_fb_size(ofbi, var))
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
847
WARN_ON(!atomic_read(&ofbi->region->lock_count));
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
868
if (ofbi->region->size)
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
941
WARN_ON(!atomic_read(&ofbi->region->lock_count));
drivers/video/fbdev/omap2/omapfb/omapfb-main.c
948
if (ofbi->region->size == 0) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
108
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
124
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
262
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
269
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
303
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
307
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
394
omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
398
omapfb_put_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
419
return sysfs_emit(buf, "%lu\n", ofbi->region->size);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
445
rg = ofbi->region;
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
459
if (ofbi2->region != rg)
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
472
if (size != ofbi->region->size) {
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
473
r = omapfb_realloc_fbmem(fbi, size, ofbi->region->type);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
496
return sysfs_emit(buf, "%0x\n", ofbi->region->paddr);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
505
return sysfs_emit(buf, "%p\n", ofbi->region->vaddr);
drivers/video/fbdev/omap2/omapfb/omapfb-sysfs.c
59
rg = omapfb_get_mem_region(ofbi->region);
drivers/video/fbdev/omap2/omapfb/omapfb.h
60
struct omapfb2_mem_region *region;
drivers/video/fbdev/pm2fb.c
1053
const struct fb_fillrect *region)
drivers/video/fbdev/pm2fb.c
1059
((u32 *)info->pseudo_palette)[region->color] : region->color;
drivers/video/fbdev/pm2fb.c
1064
region->rop != ROP_COPY ) {
drivers/video/fbdev/pm2fb.c
1065
cfb_fillrect(info, region);
drivers/video/fbdev/pm2fb.c
1072
memcpy(&modded, region, sizeof(struct fb_fillrect));
drivers/video/fbdev/pm3fb.c
377
const struct fb_fillrect *region)
drivers/video/fbdev/pm3fb.c
384
((u32 *)info->pseudo_palette)[region->color] : region->color;
drivers/video/fbdev/pm3fb.c
389
cfb_fillrect(info, region);
drivers/video/fbdev/pm3fb.c
392
if (region->rop == ROP_COPY )
drivers/video/fbdev/pm3fb.c
401
memcpy(&modded, region, sizeof(struct fb_fillrect));
drivers/video/fbdev/riva/fbdev.c
1439
static void rivafb_copyarea(struct fb_info *info, const struct fb_copyarea *region)
drivers/video/fbdev/riva/fbdev.c
1444
cfb_copyarea(info, region);
drivers/video/fbdev/riva/fbdev.c
1450
(region->sy << 16) | region->sx);
drivers/video/fbdev/riva/fbdev.c
1452
(region->dy << 16) | region->dx);
drivers/video/fbdev/riva/fbdev.c
1455
(region->height << 16) | region->width);
drivers/video/fbdev/savage/savagefb.h
399
const struct fb_copyarea *region);
drivers/video/fbdev/savage/savagefb_accel.c
31
void savagefb_copyarea(struct fb_info *info, const struct fb_copyarea *region)
drivers/video/fbdev/savage/savagefb_accel.c
34
int sx = region->sx, dx = region->dx;
drivers/video/fbdev/savage/savagefb_accel.c
35
int sy = region->sy, dy = region->dy;
drivers/video/fbdev/savage/savagefb_accel.c
38
if (!region->width || !region->height)
drivers/video/fbdev/savage/savagefb_accel.c
47
sx += region->width - 1;
drivers/video/fbdev/savage/savagefb_accel.c
48
dx += region->width - 1;
drivers/video/fbdev/savage/savagefb_accel.c
54
sy += region->height - 1;
drivers/video/fbdev/savage/savagefb_accel.c
55
dy += region->height - 1;
drivers/video/fbdev/savage/savagefb_accel.c
62
BCI_SEND(BCI_W_H(region->width, region->height));
drivers/video/fbdev/skeletonfb.c
492
void xxxfb_fillrect(struct fb_info *p, const struct fb_fillrect *region)
drivers/video/fbdev/vesafb.c
38
struct resource *region;
drivers/video/fbdev/vesafb.c
435
par->region = request_region(0x3c0, 32, "vesafb");
drivers/video/fbdev/vesafb.c
498
if (par->region)
drivers/video/fbdev/vesafb.c
509
if (((struct vesafb_par *)(info->par))->region)
drivers/virt/acrn/mm.c
100
ret = modify_region(vm, region);
drivers/virt/acrn/mm.c
104
kfree(region);
drivers/virt/acrn/mm.c
19
static int modify_region(struct acrn_vm *vm, struct vm_memory_region_op *region)
drivers/virt/acrn/mm.c
30
regions->regions_gpa = virt_to_phys(region);
drivers/virt/acrn/mm.c
55
struct vm_memory_region_op *region;
drivers/virt/acrn/mm.c
58
region = kzalloc_obj(*region);
drivers/virt/acrn/mm.c
59
if (!region)
drivers/virt/acrn/mm.c
62
region->type = ACRN_MEM_REGION_ADD;
drivers/virt/acrn/mm.c
63
region->user_vm_pa = user_gpa;
drivers/virt/acrn/mm.c
64
region->service_vm_pa = service_gpa;
drivers/virt/acrn/mm.c
65
region->size = size;
drivers/virt/acrn/mm.c
66
region->attr = ((mem_type & ACRN_MEM_TYPE_MASK) |
drivers/virt/acrn/mm.c
68
ret = modify_region(vm, region);
drivers/virt/acrn/mm.c
73
kfree(region);
drivers/virt/acrn/mm.c
87
struct vm_memory_region_op *region;
drivers/virt/acrn/mm.c
90
region = kzalloc_obj(*region);
drivers/virt/acrn/mm.c
91
if (!region)
drivers/virt/acrn/mm.c
94
region->type = ACRN_MEM_REGION_DEL;
drivers/virt/acrn/mm.c
95
region->user_vm_pa = user_gpa;
drivers/virt/acrn/mm.c
96
region->service_vm_pa = 0UL;
drivers/virt/acrn/mm.c
97
region->size = size;
drivers/virt/acrn/mm.c
98
region->attr = 0U;
drivers/virtio/virtio_mmio.c
493
struct virtio_shm_region *region, u8 id)
drivers/virtio/virtio_mmio.c
505
region->len = len;
drivers/virtio/virtio_mmio.c
517
region->addr = addr;
drivers/virtio/virtio_pci_modern.c
847
struct virtio_shm_region *region, u8 id)
drivers/virtio/virtio_pci_modern.c
874
region->len = len;
drivers/virtio/virtio_pci_modern.c
875
region->addr = (u64) phys_addr + offset;
fs/ext4/fast_commit.c
1726
struct ext4_fc_alloc_region *region;
fs/ext4/fast_commit.c
1749
region = &state->fc_regions[state->fc_regions_used++];
fs/ext4/fast_commit.c
1750
region->ino = ino;
fs/ext4/fast_commit.c
1751
region->lblk = lblk;
fs/ext4/fast_commit.c
1752
region->pblk = pblk;
fs/ext4/fast_commit.c
1753
region->len = len;
fs/proc/nommu.c
31
static int nommu_region_show(struct seq_file *m, struct vm_region *region)
fs/proc/nommu.c
38
flags = region->vm_flags;
fs/proc/nommu.c
39
file = region->vm_file;
fs/proc/nommu.c
42
struct inode *inode = file_inode(region->vm_file);
fs/proc/nommu.c
50
region->vm_start,
fs/proc/nommu.c
51
region->vm_end,
fs/proc/nommu.c
56
((loff_t)region->vm_pgoff) << PAGE_SHIFT,
fs/proc/task_nommu.c
100
struct vm_region *region;
fs/proc/task_nommu.c
106
region = vma->vm_region;
fs/proc/task_nommu.c
107
if (region) {
fs/proc/task_nommu.c
108
size += kobjsize(region);
fs/proc/task_nommu.c
109
size += region->vm_end - region->vm_start;
fs/proc/task_nommu.c
25
struct vm_region *region;
fs/proc/task_nommu.c
32
region = vma->vm_region;
fs/proc/task_nommu.c
33
if (region) {
fs/proc/task_nommu.c
34
size = kobjsize(region);
fs/proc/task_nommu.c
35
size += region->vm_end - region->vm_start;
fs/proc/task_nommu.c
45
if (region)
fs/proc/task_nommu.c
46
slack = region->vm_end - vma->vm_end;
include/drm/drm_file.h
520
const char *region,
include/drm/drm_file.h
525
const char *region);
include/linux/avf/virtchnl.h
1429
u8 region;
include/linux/cgroup_dmem.h
19
void dmem_cgroup_unregister_region(struct dmem_cgroup_region *region);
include/linux/cgroup_dmem.h
20
int dmem_cgroup_try_charge(struct dmem_cgroup_region *region, u64 size,
include/linux/cgroup_dmem.h
36
static inline void dmem_cgroup_unregister_region(struct dmem_cgroup_region *region)
include/linux/cgroup_dmem.h
39
static inline int dmem_cgroup_try_charge(struct dmem_cgroup_region *region, u64 size,
include/linux/dm-dirty-log.h
101
int (*get_resync_work)(struct dm_dirty_log *log, region_t *region);
include/linux/dm-dirty-log.h
109
region_t region, int in_sync);
include/linux/dm-dirty-log.h
130
int (*is_remote_recovering)(struct dm_dirty_log *log, region_t region);
include/linux/dm-dirty-log.h
58
int (*is_clean)(struct dm_dirty_log *log, region_t region);
include/linux/dm-dirty-log.h
71
int (*in_sync)(struct dm_dirty_log *log, region_t region,
include/linux/dm-dirty-log.h
86
void (*mark_region)(struct dm_dirty_log *log, region_t region);
include/linux/dm-dirty-log.h
87
void (*clear_region)(struct dm_dirty_log *log, region_t region);
include/linux/dm-io.h
83
struct dm_io_region *region, unsigned int long *sync_error_bits,
include/linux/dm-region-hash.h
54
sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region);
include/linux/dm-region-hash.h
67
int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block);
include/linux/dm-region-hash.h
68
void dm_rh_set_state(struct dm_region_hash *rh, region_t region,
include/linux/dm-region-hash.h
79
void dm_rh_dec(struct dm_region_hash *rh, region_t region);
include/linux/fb.h
281
void (*fb_copyarea) (struct fb_info *info, const struct fb_copyarea *region);
include/linux/firewire.h
385
const struct fw_address_region *region);
include/linux/fpga/fpga-region.h
28
int (*get_bridges)(struct fpga_region *region);
include/linux/fpga/fpga-region.h
52
int (*get_bridges)(struct fpga_region *region);
include/linux/fpga/fpga-region.h
61
int fpga_region_program_fpga(struct fpga_region *region);
include/linux/fpga/fpga-region.h
74
void fpga_region_unregister(struct fpga_region *region);
include/linux/iommu.h
309
void (*free)(struct device *dev, struct iommu_resv_region *region);
include/linux/ioport.h
364
#define rename_region(region, newname) do { (region)->name = (newname); } while (0)
include/linux/ipack.h
97
struct ipack_region region[IPACK_SPACE_COUNT];
include/linux/memblock.h
569
#define for_each_mem_region(region) \
include/linux/memblock.h
570
for (region = memblock.memory.regions; \
include/linux/memblock.h
571
region < (memblock.memory.regions + memblock.memory.cnt); \
include/linux/memblock.h
572
region++)
include/linux/memblock.h
578
#define for_each_reserved_mem_region(region) \
include/linux/memblock.h
579
for (region = memblock.reserved.regions; \
include/linux/memblock.h
580
region < (memblock.reserved.regions + memblock.reserved.cnt); \
include/linux/memblock.h
581
region++)
include/linux/pci.h
1215
void pcibios_resource_to_bus(struct pci_bus *bus, struct pci_bus_region *region,
include/linux/pci.h
1218
struct pci_bus_region *region);
include/linux/pci.h
1662
struct pci_bus_region region;
include/linux/pci.h
1664
pcibios_resource_to_bus(pdev->bus, &region, &pdev->resource[bar]);
include/linux/pci.h
1665
return region.start;
include/linux/pci.h
898
static inline pci_bus_addr_t pci_bus_region_size(const struct pci_bus_region *region)
include/linux/pci.h
900
return region->end - region->start + 1;
include/linux/platform_data/cros_ec_commands.h
1684
uint32_t region;
include/linux/platform_data/cros_ec_commands.h
5575
uint8_t region; /* enum ec_flash_region */
include/linux/pruss_driver.h
111
struct pruss_mem_region *region);
include/linux/pruss_driver.h
113
struct pruss_mem_region *region);
include/linux/pruss_driver.h
133
struct pruss_mem_region *region)
include/linux/pruss_driver.h
139
struct pruss_mem_region *region)
include/linux/vfio_pci_core.h
112
struct vfio_pci_region *region;
include/linux/vfio_pci_core.h
42
struct vfio_pci_region *region);
include/linux/vfio_pci_core.h
44
struct vfio_pci_region *region,
include/linux/vfio_pci_core.h
47
struct vfio_pci_region *region,
include/linux/virtio_config.h
137
struct virtio_shm_region *region, u8 id);
include/net/devlink.h
1922
void devl_region_destroy(struct devlink_region *region);
include/net/devlink.h
1923
void devlink_region_destroy(struct devlink_region *region);
include/net/devlink.h
1926
int devlink_region_snapshot_create(struct devlink_region *region,
include/net/dsa.h
1286
void dsa_devlink_region_destroy(struct devlink_region *region);
include/uapi/drm/i915_drm.h
3484
struct drm_i915_gem_memory_class_instance region;
include/uapi/drm/i915_drm.h
857
int region;
include/uapi/drm/i915_drm.h
864
int region;
include/uapi/drm/i915_drm.h
869
int region;
include/uapi/drm/i915_drm.h
878
int region;
include/uapi/drm/radeon_drm.h
735
int region;
include/uapi/drm/radeon_drm.h
742
int region;
include/uapi/drm/radeon_drm.h
747
int region;
include/uapi/linux/ethtool.h
1636
__u32 region;
include/uapi/linux/omap3isp.h
255
struct omap3isp_hist_region region[OMAP3ISP_HIST_MAX_REGIONS];
include/video/sticore.h
69
u32 region; /* complete region value */
io_uring/kbuf.c
448
io_free_region(ctx->user, &bl->region);
io_uring/kbuf.c
661
ret = io_create_region(ctx, &bl->region, &rd, mmap_offset);
io_uring/kbuf.c
664
br = io_region_get_ptr(&bl->region);
io_uring/kbuf.c
693
io_free_region(ctx->user, &bl->region);
io_uring/kbuf.c
756
return &bl->region;
io_uring/kbuf.h
36
struct io_mapped_region region;
io_uring/memmap.c
274
struct io_mapped_region *region;
io_uring/memmap.c
276
region = io_mmap_get_region(ctx, pgoff);
io_uring/memmap.c
277
if (!region)
io_uring/memmap.c
279
return io_region_validate_mmap(ctx, region);
io_uring/memmap.c
301
struct io_mapped_region *region;
io_uring/memmap.c
317
region = io_mmap_get_region(ctx, vma->vm_pgoff);
io_uring/memmap.c
318
return io_region_mmap(ctx, region, vma, page_limit);
io_uring/register.c
678
struct io_mapped_region region = {};
io_uring/register.c
702
ret = io_create_region(ctx, &region, &rd, IORING_MAP_OFF_PARAM_REGION);
io_uring/register.c
706
io_free_region(ctx->user, &region);
io_uring/register.c
711
ctx->cq_wait_arg = io_region_get_ptr(&region);
io_uring/register.c
715
io_region_publish(ctx, &region, &ctx->param_region);
io_uring/zcrx.c
395
ret = io_create_region(ctx, &ifq->region, rd, mmap_offset);
io_uring/zcrx.c
399
ptr = io_region_get_ptr(&ifq->region);
io_uring/zcrx.c
408
io_free_region(ifq->user, &ifq->region);
io_uring/zcrx.c
643
return ifq ? &ifq->region : NULL;
io_uring/zcrx.h
66
struct io_mapped_region region;
kernel/cgroup/dmem.c
137
kref_put(&pool->region->ref, dmemcg_free_region);
kernel/cgroup/dmem.c
234
find_cg_pool_locked(struct dmemcg_state *dmemcs, struct dmem_cgroup_region *region)
kernel/cgroup/dmem.c
239
if (pool->region == region)
kernel/cgroup/dmem.c
271
if (pool->region == limit_pool->region) {
kernel/cgroup/dmem.c
355
alloc_pool_single(struct dmemcg_state *dmemcs, struct dmem_cgroup_region *region,
kernel/cgroup/dmem.c
370
pool->region = region;
kernel/cgroup/dmem.c
374
ppool = find_cg_pool_locked(parent, region);
kernel/cgroup/dmem.c
380
kref_get(&region->ref);
kernel/cgroup/dmem.c
387
list_add_tail(&pool->region_node, &region->pools);
kernel/cgroup/dmem.c
397
get_cg_pool_locked(struct dmemcg_state *dmemcs, struct dmem_cgroup_region *region,
kernel/cgroup/dmem.c
408
pool = find_cg_pool_locked(p, region);
kernel/cgroup/dmem.c
410
pool = alloc_pool_single(p, region, allocpool);
kernel/cgroup/dmem.c
422
retpool = pool = find_cg_pool_locked(dmemcs, region);
kernel/cgroup/dmem.c
428
ppool = find_cg_pool_locked(pp, region);
kernel/cgroup/dmem.c
446
struct dmem_cgroup_region *region = container_of(rcu, typeof(*region), rcu);
kernel/cgroup/dmem.c
449
list_for_each_entry_safe(pool, next, &region->pools, region_node)
kernel/cgroup/dmem.c
451
kfree(region->name);
kernel/cgroup/dmem.c
452
kfree(region);
kernel/cgroup/dmem.c
468
void dmem_cgroup_unregister_region(struct dmem_cgroup_region *region)
kernel/cgroup/dmem.c
472
if (!region)
kernel/cgroup/dmem.c
478
list_del_rcu(&region->region_node);
kernel/cgroup/dmem.c
480
list_for_each_entry_safe(pool, next, &region->pools, region_node) {
kernel/cgroup/dmem.c
491
region->unregistered = true;
kernel/cgroup/dmem.c
494
kref_put(&region->ref, dmemcg_free_region);
kernel/cgroup/dmem.c
545
struct dmem_cgroup_region *region;
kernel/cgroup/dmem.c
547
list_for_each_entry_rcu(region, &dmem_cgroup_regions, region_node, spin_is_locked(&dmemcg_lock))
kernel/cgroup/dmem.c
548
if (!strcmp(name, region->name) &&
kernel/cgroup/dmem.c
549
kref_get_unless_zero(&region->ref))
kernel/cgroup/dmem.c
550
return region;
kernel/cgroup/dmem.c
572
get_cg_pool_unlocked(struct dmemcg_state *cg, struct dmem_cgroup_region *region)
kernel/cgroup/dmem.c
578
pool = find_cg_pool_locked(cg, region);
kernel/cgroup/dmem.c
587
if (!region->unregistered)
kernel/cgroup/dmem.c
588
pool = get_cg_pool_locked(cg, region, &allocpool);
kernel/cgroup/dmem.c
63
struct dmem_cgroup_region *region;
kernel/cgroup/dmem.c
651
int dmem_cgroup_try_charge(struct dmem_cgroup_region *region, u64 size,
kernel/cgroup/dmem.c
670
pool = get_cg_pool_unlocked(cg, region);
kernel/cgroup/dmem.c
699
struct dmem_cgroup_region *region;
kernel/cgroup/dmem.c
702
list_for_each_entry_rcu(region, &dmem_cgroup_regions, region_node) {
kernel/cgroup/dmem.c
703
seq_puts(sf, region->name);
kernel/cgroup/dmem.c
704
seq_printf(sf, " %llu\n", region->size);
kernel/cgroup/dmem.c
710
static int dmemcg_parse_limit(char *options, struct dmem_cgroup_region *region,
kernel/cgroup/dmem.c
737
struct dmem_cgroup_region *region;
kernel/cgroup/dmem.c
759
region = dmemcg_get_region_by_name(region_name);
kernel/cgroup/dmem.c
762
if (!region)
kernel/cgroup/dmem.c
765
err = dmemcg_parse_limit(options, region, &new_limit);
kernel/cgroup/dmem.c
769
pool = get_cg_pool_unlocked(dmemcs, region);
kernel/cgroup/dmem.c
780
kref_put(&region->ref, dmemcg_free_region);
kernel/cgroup/dmem.c
791
struct dmem_cgroup_region *region;
kernel/cgroup/dmem.c
794
list_for_each_entry_rcu(region, &dmem_cgroup_regions, region_node) {
kernel/cgroup/dmem.c
795
struct dmem_cgroup_pool_state *pool = find_cg_pool_locked(dmemcs, region);
kernel/cgroup/dmem.c
798
seq_puts(sf, region->name);
kernel/power/snapshot.c
1006
region = list_entry(nosave_regions.prev,
kernel/power/snapshot.c
1008
if (region->end_pfn == start_pfn) {
kernel/power/snapshot.c
1009
region->end_pfn = end_pfn;
kernel/power/snapshot.c
1014
region = memblock_alloc_or_panic(sizeof(struct nosave_region),
kernel/power/snapshot.c
1016
region->start_pfn = start_pfn;
kernel/power/snapshot.c
1017
region->end_pfn = end_pfn;
kernel/power/snapshot.c
1018
list_add_tail(&region->list, &nosave_regions);
kernel/power/snapshot.c
1084
struct nosave_region *region;
kernel/power/snapshot.c
1089
list_for_each_entry(region, &nosave_regions, list) {
kernel/power/snapshot.c
1093
(unsigned long long) region->start_pfn << PAGE_SHIFT,
kernel/power/snapshot.c
1094
((unsigned long long) region->end_pfn << PAGE_SHIFT)
kernel/power/snapshot.c
1097
for_each_valid_pfn(pfn, region->start_pfn, region->end_pfn) {
kernel/power/snapshot.c
999
struct nosave_region *region;
lib/bitmap-str.c
220
static void bitmap_set_region(const struct region *r, unsigned long *bitmap)
lib/bitmap-str.c
228
static int bitmap_check_region(const struct region *r)
lib/bitmap-str.c
295
static const char *bitmap_parse_region(const char *str, struct region *r)
lib/bitmap-str.c
376
struct region r;
mm/damon/core.c
119
struct damon_region *region;
mm/damon/core.c
121
region = kmem_cache_alloc(damon_region_cache, GFP_KERNEL);
mm/damon/core.c
122
if (!region)
mm/damon/core.c
125
region->ar.start = start;
mm/damon/core.c
126
region->ar.end = end;
mm/damon/core.c
127
region->nr_accesses = 0;
mm/damon/core.c
128
region->nr_accesses_bp = 0;
mm/damon/core.c
129
INIT_LIST_HEAD(&region->list);
mm/damon/core.c
131
region->age = 0;
mm/damon/core.c
132
region->last_nr_accesses = 0;
mm/damon/core.c
134
return region;
mm/damon/stat.c
112
struct damon_region **sorted_regions, *region;
mm/damon/stat.c
122
region = sorted_regions[i];
mm/damon/stat.c
123
accounted_bytes += region->ar.end - region->ar.start;
mm/damon/stat.c
126
damon_stat_idletime(region) *
mm/damon/sysfs-schemes.c
27
struct damon_region *region)
mm/damon/sysfs-schemes.c
2830
struct damon_sysfs_scheme_region *region;
mm/damon/sysfs-schemes.c
2848
region = damon_sysfs_scheme_region_alloc(r);
mm/damon/sysfs-schemes.c
2849
if (!region)
mm/damon/sysfs-schemes.c
2851
region->sz_filter_passed = sz_filter_passed;
mm/damon/sysfs-schemes.c
2852
list_add_tail(&region->list, &sysfs_regions->regions_list);
mm/damon/sysfs-schemes.c
2854
if (kobject_init_and_add(&region->kobj,
mm/damon/sysfs-schemes.c
2858
kobject_put(&region->kobj);
mm/damon/sysfs-schemes.c
34
sysfs_region->ar = region->ar;
mm/damon/sysfs-schemes.c
35
sysfs_region->nr_accesses = region->nr_accesses_bp / 10000;
mm/damon/sysfs-schemes.c
36
sysfs_region->age = region->age;
mm/damon/sysfs-schemes.c
44
struct damon_sysfs_scheme_region *region = container_of(kobj,
mm/damon/sysfs-schemes.c
47
return sysfs_emit(buf, "%lu\n", region->ar.start);
mm/damon/sysfs-schemes.c
53
struct damon_sysfs_scheme_region *region = container_of(kobj,
mm/damon/sysfs-schemes.c
56
return sysfs_emit(buf, "%lu\n", region->ar.end);
mm/damon/sysfs-schemes.c
62
struct damon_sysfs_scheme_region *region = container_of(kobj,
mm/damon/sysfs-schemes.c
65
return sysfs_emit(buf, "%u\n", region->nr_accesses);
mm/damon/sysfs-schemes.c
71
struct damon_sysfs_scheme_region *region = container_of(kobj,
mm/damon/sysfs-schemes.c
74
return sysfs_emit(buf, "%u\n", region->age);
mm/damon/sysfs-schemes.c
80
struct damon_sysfs_scheme_region *region = container_of(kobj,
mm/damon/sysfs-schemes.c
83
return sysfs_emit(buf, "%lu\n", region->sz_filter_passed);
mm/damon/sysfs-schemes.c
88
struct damon_sysfs_scheme_region *region = container_of(kobj,
mm/damon/sysfs-schemes.c
91
list_del(&region->list);
mm/damon/sysfs-schemes.c
92
kfree(region);
mm/damon/sysfs.c
120
struct damon_sysfs_region **regions_arr, *region;
mm/damon/sysfs.c
134
region = damon_sysfs_region_alloc();
mm/damon/sysfs.c
135
if (!region) {
mm/damon/sysfs.c
140
err = kobject_init_and_add(&region->kobj,
mm/damon/sysfs.c
144
kobject_put(&region->kobj);
mm/damon/sysfs.c
149
regions_arr[i] = region;
mm/damon/sysfs.c
31
struct damon_sysfs_region *region = container_of(kobj,
mm/damon/sysfs.c
34
return sysfs_emit(buf, "%lu\n", region->ar.start);
mm/damon/sysfs.c
40
struct damon_sysfs_region *region = container_of(kobj,
mm/damon/sysfs.c
42
int err = kstrtoul(buf, 0, &region->ar.start);
mm/damon/sysfs.c
50
struct damon_sysfs_region *region = container_of(kobj,
mm/damon/sysfs.c
53
return sysfs_emit(buf, "%lu\n", region->ar.end);
mm/damon/sysfs.c
59
struct damon_sysfs_region *region = container_of(kobj,
mm/damon/sysfs.c
61
int err = kstrtoul(buf, 0, &region->ar.end);
mm/memblock.c
2245
struct memblock_region *region;
mm/memblock.c
2256
for_each_mem_region(region) {
mm/memblock.c
2257
nid = memblock_get_region_node(region);
mm/memblock.c
2258
start = region->base;
mm/memblock.c
2259
end = start + region->size;
mm/memblock.c
2261
if (memblock_is_nomap(region))
mm/memblock.c
2264
memblock_set_node(start, region->size, &memblock.reserved, nid);
mm/memblock.c
2278
for_each_reserved_mem_region(region) {
mm/memblock.c
2279
if (!memblock_is_reserved_noinit(region)) {
mm/memblock.c
2280
nid = memblock_get_region_node(region);
mm/memblock.c
2281
start = region->base;
mm/memblock.c
2282
end = start + region->size;
mm/nommu.c
1023
struct vm_region *region;
mm/nommu.c
1048
region = kmem_cache_zalloc(vm_region_jar, GFP_KERNEL);
mm/nommu.c
1049
if (!region)
mm/nommu.c
1056
region->vm_usage = 1;
mm/nommu.c
1057
region->vm_flags = vm_flags;
mm/nommu.c
1058
region->vm_pgoff = pgoff;
mm/nommu.c
1064
region->vm_file = get_file(file);
mm/nommu.c
1136
fput(region->vm_file);
mm/nommu.c
1137
kmem_cache_free(vm_region_jar, region);
mm/nommu.c
1138
region = pregion;
mm/nommu.c
1164
vma->vm_start = region->vm_start = addr;
mm/nommu.c
1165
vma->vm_end = region->vm_end = addr + len;
mm/nommu.c
1170
vma->vm_region = region;
mm/nommu.c
1178
ret = do_mmap_private(vma, region, len, capabilities);
mm/nommu.c
1181
add_nommu_region(region);
mm/nommu.c
1187
memset((void *)region->vm_start, 0,
mm/nommu.c
1188
region->vm_end - region->vm_start);
mm/nommu.c
1208
if (vma->vm_flags & VM_EXEC && !region->vm_icache_flushed) {
mm/nommu.c
1209
flush_icache_user_range(region->vm_start, region->vm_end);
mm/nommu.c
1210
region->vm_icache_flushed = true;
mm/nommu.c
1221
if (region->vm_file)
mm/nommu.c
1222
fput(region->vm_file);
mm/nommu.c
1223
kmem_cache_free(vm_region_jar, region);
mm/nommu.c
1236
kmem_cache_free(vm_region_jar, region);
mm/nommu.c
1310
struct vm_region *region;
mm/nommu.c
1323
region = kmem_cache_alloc(vm_region_jar, GFP_KERNEL);
mm/nommu.c
1324
if (!region)
mm/nommu.c
1332
*region = *vma->vm_region;
mm/nommu.c
1333
new->vm_region = region;
mm/nommu.c
1338
region->vm_top = region->vm_end = new->vm_end = addr;
mm/nommu.c
1340
region->vm_start = new->vm_start = addr;
mm/nommu.c
1341
region->vm_pgoff = new->vm_pgoff += npages;
mm/nommu.c
1376
kmem_cache_free(vm_region_jar, region);
mm/nommu.c
1388
struct vm_region *region;
mm/nommu.c
1403
region = vma->vm_region;
mm/nommu.c
1404
BUG_ON(region->vm_usage != 1);
mm/nommu.c
1407
delete_nommu_region(region);
mm/nommu.c
1408
if (from > region->vm_start) {
mm/nommu.c
1409
to = region->vm_top;
mm/nommu.c
1410
region->vm_top = region->vm_end = from;
mm/nommu.c
1412
region->vm_start = to;
mm/nommu.c
1414
add_nommu_region(region);
mm/nommu.c
1808
struct vm_region *region;
mm/nommu.c
1839
region = vma->vm_region;
mm/nommu.c
1840
r_size = region->vm_top - region->vm_start;
mm/nommu.c
1841
r_top = (region->vm_pgoff << PAGE_SHIFT) + r_size;
mm/nommu.c
1844
region->vm_top -= r_top - newsize;
mm/nommu.c
1845
if (region->vm_end > region->vm_top)
mm/nommu.c
1846
region->vm_end = region->vm_top;
mm/nommu.c
440
struct vm_region *region, *last;
mm/nommu.c
452
region = rb_entry(p, struct vm_region, vm_rb);
mm/nommu.c
455
BUG_ON(region->vm_end <= region->vm_start);
mm/nommu.c
456
BUG_ON(region->vm_top < region->vm_end);
mm/nommu.c
457
BUG_ON(region->vm_start < last->vm_top);
mm/nommu.c
471
static void add_nommu_region(struct vm_region *region)
mm/nommu.c
483
if (region->vm_start < pregion->vm_start)
mm/nommu.c
485
else if (region->vm_start > pregion->vm_start)
mm/nommu.c
487
else if (pregion == region)
mm/nommu.c
493
rb_link_node(&region->vm_rb, parent, p);
mm/nommu.c
494
rb_insert_color(&region->vm_rb, &nommu_region_tree);
mm/nommu.c
502
static void delete_nommu_region(struct vm_region *region)
mm/nommu.c
507
rb_erase(&region->vm_rb, &nommu_region_tree);
mm/nommu.c
530
static void __put_nommu_region(struct vm_region *region)
mm/nommu.c
535
if (--region->vm_usage == 0) {
mm/nommu.c
536
if (region->vm_top > region->vm_start)
mm/nommu.c
537
delete_nommu_region(region);
mm/nommu.c
540
if (region->vm_file)
mm/nommu.c
541
fput(region->vm_file);
mm/nommu.c
545
if (region->vm_flags & VM_MAPPED_COPY)
mm/nommu.c
546
free_page_series(region->vm_start, region->vm_top);
mm/nommu.c
547
kmem_cache_free(vm_region_jar, region);
mm/nommu.c
556
static void put_nommu_region(struct vm_region *region)
mm/nommu.c
559
__put_nommu_region(region);
mm/nommu.c
915
struct vm_region *region,
mm/nommu.c
966
region->vm_flags = vma->vm_flags;
mm/nommu.c
967
region->vm_start = (unsigned long) base;
mm/nommu.c
968
region->vm_end = region->vm_start + len;
mm/nommu.c
969
region->vm_top = region->vm_start + (total << PAGE_SHIFT);
mm/nommu.c
971
vma->vm_start = region->vm_start;
mm/nommu.c
972
vma->vm_end = region->vm_start + len;
mm/nommu.c
996
free_page_series(region->vm_start, region->vm_top);
mm/nommu.c
997
region->vm_start = vma->vm_start = 0;
mm/nommu.c
998
region->vm_end = vma->vm_end = 0;
mm/nommu.c
999
region->vm_top = 0;
net/devlink/region.c
1048
struct devlink_region *region;
net/devlink/region.c
1058
region = kzalloc_obj(*region);
net/devlink/region.c
1059
if (!region)
net/devlink/region.c
1062
region->devlink = devlink;
net/devlink/region.c
1063
region->max_snapshots = region_max_snapshots;
net/devlink/region.c
1064
region->ops = ops;
net/devlink/region.c
1065
region->size = region_size;
net/devlink/region.c
1066
INIT_LIST_HEAD(&region->snapshot_list);
net/devlink/region.c
1067
mutex_init(&region->snapshot_lock);
net/devlink/region.c
1068
list_add_tail(&region->list, &devlink->region_list);
net/devlink/region.c
1069
devlink_nl_region_notify(region, NULL, DEVLINK_CMD_REGION_NEW);
net/devlink/region.c
107
list_for_each_entry(snapshot, &region->snapshot_list, list) {
net/devlink/region.c
1071
return region;
net/devlink/region.c
1090
struct devlink_region *region;
net/devlink/region.c
1093
region = devl_region_create(devlink, ops, region_max_snapshots,
net/devlink/region.c
1096
return region;
net/devlink/region.c
1116
struct devlink_region *region;
net/devlink/region.c
1131
region = kzalloc_obj(*region);
net/devlink/region.c
1132
if (!region) {
net/devlink/region.c
1137
region->devlink = devlink;
net/devlink/region.c
1138
region->port = port;
net/devlink/region.c
1139
region->max_snapshots = region_max_snapshots;
net/devlink/region.c
1140
region->port_ops = ops;
net/devlink/region.c
1141
region->size = region_size;
net/devlink/region.c
1142
INIT_LIST_HEAD(&region->snapshot_list);
net/devlink/region.c
1143
mutex_init(&region->snapshot_lock);
net/devlink/region.c
1144
list_add_tail(&region->list, &port->region_list);
net/devlink/region.c
1145
devlink_nl_region_notify(region, NULL, DEVLINK_CMD_REGION_NEW);
net/devlink/region.c
1148
return region;
net/devlink/region.c
1161
void devl_region_destroy(struct devlink_region *region)
net/devlink/region.c
1163
struct devlink *devlink = region->devlink;
net/devlink/region.c
1169
mutex_lock(&region->snapshot_lock);
net/devlink/region.c
1170
list_for_each_entry_safe(snapshot, ts, &region->snapshot_list, list)
net/devlink/region.c
1171
devlink_region_snapshot_del(region, snapshot);
net/devlink/region.c
1172
mutex_unlock(&region->snapshot_lock);
net/devlink/region.c
1174
list_del(&region->list);
net/devlink/region.c
1175
mutex_destroy(&region->snapshot_lock);
net/devlink/region.c
1177
devlink_nl_region_notify(region, NULL, DEVLINK_CMD_REGION_DEL);
net/devlink/region.c
1178
kfree(region);
net/devlink/region.c
1189
void devlink_region_destroy(struct devlink_region *region)
net/devlink/region.c
1191
struct devlink *devlink = region->devlink;
net/devlink/region.c
1194
devl_region_destroy(region);
net/devlink/region.c
124
struct devlink_region *region)
net/devlink/region.c
1248
int devlink_region_snapshot_create(struct devlink_region *region,
net/devlink/region.c
1253
mutex_lock(&region->snapshot_lock);
net/devlink/region.c
1254
err = __devlink_region_snapshot_create(region, data, snapshot_id);
net/devlink/region.c
1255
mutex_unlock(&region->snapshot_lock);
net/devlink/region.c
137
if (region->port) {
net/devlink/region.c
139
region->port->index);
net/devlink/region.c
144
err = nla_put_string(msg, DEVLINK_ATTR_REGION_NAME, region->ops->name);
net/devlink/region.c
148
err = devlink_nl_put_u64(msg, DEVLINK_ATTR_REGION_SIZE, region->size);
net/devlink/region.c
153
region->max_snapshots);
net/devlink/region.c
157
err = devlink_nl_region_snapshots_id_put(msg, devlink, region);
net/devlink/region.c
170
devlink_nl_region_notify_build(struct devlink_region *region,
net/devlink/region.c
174
struct devlink *devlink = region->devlink;
net/devlink/region.c
193
if (region->port) {
net/devlink/region.c
195
region->port->index);
net/devlink/region.c
201
region->ops->name);
net/devlink/region.c
212
region->size);
net/devlink/region.c
227
static void devlink_nl_region_notify(struct devlink_region *region,
net/devlink/region.c
231
struct devlink *devlink = region->devlink;
net/devlink/region.c
239
msg = devlink_nl_region_notify_build(region, snapshot, cmd, 0, 0);
net/devlink/region.c
248
struct devlink_region *region;
net/devlink/region.c
250
list_for_each_entry(region, &devlink->region_list, list)
net/devlink/region.c
251
devlink_nl_region_notify(region, NULL, DEVLINK_CMD_REGION_NEW);
net/devlink/region.c
256
struct devlink_region *region;
net/devlink/region.c
258
list_for_each_entry_reverse(region, &devlink->region_list, list)
net/devlink/region.c
259
devlink_nl_region_notify(region, NULL, DEVLINK_CMD_REGION_DEL);
net/devlink/region.c
29
struct devlink_region *region;
net/devlink/region.c
37
struct devlink_region *region;
net/devlink/region.c
39
list_for_each_entry(region, &devlink->region_list, list)
net/devlink/region.c
40
if (!strcmp(region->ops->name, region_name))
net/devlink/region.c
41
return region;
net/devlink/region.c
415
__devlink_region_snapshot_create(struct devlink_region *region,
net/devlink/region.c
418
struct devlink *devlink = region->devlink;
net/devlink/region.c
422
lockdep_assert_held(&region->snapshot_lock);
net/devlink/region.c
425
if (region->cur_snapshots == region->max_snapshots)
net/devlink/region.c
428
if (devlink_region_snapshot_get_by_id(region, snapshot_id))
net/devlink/region.c
440
snapshot->region = region;
net/devlink/region.c
443
list_add_tail(&snapshot->list, &region->snapshot_list);
net/devlink/region.c
445
region->cur_snapshots++;
net/devlink/region.c
447
devlink_nl_region_notify(region, snapshot, DEVLINK_CMD_REGION_NEW);
net/devlink/region.c
455
static void devlink_region_snapshot_del(struct devlink_region *region,
net/devlink/region.c
458
struct devlink *devlink = region->devlink;
net/devlink/region.c
460
lockdep_assert_held(&region->snapshot_lock);
net/devlink/region.c
462
devlink_nl_region_notify(region, snapshot, DEVLINK_CMD_REGION_DEL);
net/devlink/region.c
463
region->cur_snapshots--;
net/devlink/region.c
465
region->ops->destructor(snapshot->data);
net/devlink/region.c
474
struct devlink_region *region;
net/devlink/region.c
493
region = devlink_port_region_get_by_name(port, region_name);
net/devlink/region.c
495
region = devlink_region_get_by_name(devlink, region_name);
net/devlink/region.c
497
if (!region)
net/devlink/region.c
50
struct devlink_region *region;
net/devlink/region.c
506
region);
net/devlink/region.c
52
list_for_each_entry(region, &port->region_list, list)
net/devlink/region.c
520
struct devlink_region *region;
net/devlink/region.c
523
list_for_each_entry(region, &port->region_list, list) {
net/devlink/region.c
53
if (!strcmp(region->port_ops->name, region_name))
net/devlink/region.c
532
flags, region);
net/devlink/region.c
54
return region;
net/devlink/region.c
548
struct devlink_region *region;
net/devlink/region.c
554
list_for_each_entry(region, &devlink->region_list, list) {
net/devlink/region.c
563
region);
net/devlink/region.c
594
struct devlink_region *region;
net/devlink/region.c
60
devlink_region_snapshot_get_by_id(struct devlink_region *region, u32 id)
net/devlink/region.c
615
region = devlink_port_region_get_by_name(port, region_name);
net/devlink/region.c
617
region = devlink_region_get_by_name(devlink, region_name);
net/devlink/region.c
619
if (!region)
net/devlink/region.c
622
mutex_lock(&region->snapshot_lock);
net/devlink/region.c
623
snapshot = devlink_region_snapshot_get_by_id(region, snapshot_id);
net/devlink/region.c
625
mutex_unlock(&region->snapshot_lock);
net/devlink/region.c
629
devlink_region_snapshot_del(region, snapshot);
net/devlink/region.c
630
mutex_unlock(&region->snapshot_lock);
net/devlink/region.c
64
list_for_each_entry(snapshot, &region->snapshot_list, list)
net/devlink/region.c
640
struct devlink_region *region;
net/devlink/region.c
663
region = devlink_port_region_get_by_name(port, region_name);
net/devlink/region.c
665
region = devlink_region_get_by_name(devlink, region_name);
net/devlink/region.c
667
if (!region) {
net/devlink/region.c
672
if (!region->ops->snapshot) {
net/devlink/region.c
677
mutex_lock(&region->snapshot_lock);
net/devlink/region.c
679
if (region->cur_snapshots == region->max_snapshots) {
net/devlink/region.c
689
if (devlink_region_snapshot_get_by_id(region, snapshot_id)) {
net/devlink/region.c
707
err = region->port_ops->snapshot(port, region->port_ops,
net/devlink/region.c
710
err = region->ops->snapshot(devlink, region->ops,
net/devlink/region.c
715
err = __devlink_region_snapshot_create(region, data, snapshot_id);
net/devlink/region.c
722
snapshot = devlink_region_snapshot_get_by_id(region,
net/devlink/region.c
729
msg = devlink_nl_region_notify_build(region, snapshot,
net/devlink/region.c
742
mutex_unlock(&region->snapshot_lock);
net/devlink/region.c
746
region->ops->destructor(data);
net/devlink/region.c
749
mutex_unlock(&region->snapshot_lock);
net/devlink/region.c
753
devlink_region_snapshot_del(region, snapshot);
net/devlink/region.c
755
mutex_unlock(&region->snapshot_lock);
net/devlink/region.c
847
struct devlink_region *region = cb_priv;
net/devlink/region.c
849
return region->port_ops->read(region->port, region->port_ops, extack,
net/devlink/region.c
857
struct devlink_region *region = cb_priv;
net/devlink/region.c
859
return region->ops->read(region->devlink, region->ops, extack,
net/devlink/region.c
873
struct devlink_region *region;
net/devlink/region.c
908
region = devlink_port_region_get_by_name(port, region_name);
net/devlink/region.c
910
region = devlink_region_get_by_name(devlink, region_name);
net/devlink/region.c
912
if (!region) {
net/devlink/region.c
926
if (!region->ops->read) {
net/devlink/region.c
936
region_cb_priv = region;
net/devlink/region.c
948
snapshot = devlink_region_snapshot_get_by_id(region, snapshot_id);
net/devlink/region.c
96
struct devlink_region *region)
net/devlink/region.c
968
if (end_offset > region->size)
net/devlink/region.c
969
end_offset = region->size;
net/devlink/region.c
989
if (region->port) {
net/devlink/region.c
991
region->port->index);
net/dsa/devlink.c
296
void dsa_devlink_region_destroy(struct devlink_region *region)
net/dsa/devlink.c
298
devlink_region_destroy(region);
net/netfilter/ipset/ip_set_hash_gen.h
574
r = gc->region++;
net/netfilter/ipset/ip_set_hash_gen.h
576
r = gc->region = 0;
net/netfilter/ipset/ip_set_hash_gen.h
78
u32 region; /* Last gc run position */
net/wireless/util.c
2444
enum nl80211_dfs_regions region = 0;
net/wireless/util.c
2472
region = regdom->dfs_region;
net/wireless/util.c
2523
!(c->radar_detect_regions & BIT(region)))
samples/damon/mtier.c
123
region = damon_new_region(addr.start, addr.end);
samples/damon/mtier.c
124
if (!region)
samples/damon/mtier.c
126
damon_add_region(region, target);
samples/damon/mtier.c
78
struct damon_region *region;
sound/pci/cs46xx/cs46xx.h
1653
} region;
sound/pci/cs46xx/cs46xx_lib.c
2757
struct snd_cs46xx_region *region = entry->private_data;
sound/pci/cs46xx/cs46xx_lib.c
2759
if (copy_to_user_fromio(buf, region->remap_addr + pos, count))
sound/pci/cs46xx/cs46xx_lib.c
2774
struct snd_cs46xx_region *region = &chip->region.idx[idx];
sound/pci/cs46xx/cs46xx_lib.c
2775
if (! snd_card_proc_new(card, region->name, &entry)) {
sound/pci/cs46xx/cs46xx_lib.c
2779
entry->size = region->size;
sound/pci/cs46xx/cs46xx_lib.c
315
dst = chip->region.idx[bank+1].remap_addr + offset;
sound/pci/cs46xx/cs46xx_lib.c
3785
struct snd_cs46xx_region *region;
sound/pci/cs46xx/cs46xx_lib.c
3815
region = &chip->region.name.ba0;
sound/pci/cs46xx/cs46xx_lib.c
3816
strscpy(region->name, "CS46xx_BA0");
sound/pci/cs46xx/cs46xx_lib.c
3817
region->base = chip->ba0_addr;
sound/pci/cs46xx/cs46xx_lib.c
3818
region->size = CS46XX_BA0_SIZE;
sound/pci/cs46xx/cs46xx_lib.c
3820
region = &chip->region.name.data0;
sound/pci/cs46xx/cs46xx_lib.c
3821
strscpy(region->name, "CS46xx_BA1_data0");
sound/pci/cs46xx/cs46xx_lib.c
3822
region->base = chip->ba1_addr + BA1_SP_DMEM0;
sound/pci/cs46xx/cs46xx_lib.c
3823
region->size = CS46XX_BA1_DATA0_SIZE;
sound/pci/cs46xx/cs46xx_lib.c
3825
region = &chip->region.name.data1;
sound/pci/cs46xx/cs46xx_lib.c
3826
strscpy(region->name, "CS46xx_BA1_data1");
sound/pci/cs46xx/cs46xx_lib.c
3827
region->base = chip->ba1_addr + BA1_SP_DMEM1;
sound/pci/cs46xx/cs46xx_lib.c
3828
region->size = CS46XX_BA1_DATA1_SIZE;
sound/pci/cs46xx/cs46xx_lib.c
3830
region = &chip->region.name.pmem;
sound/pci/cs46xx/cs46xx_lib.c
3831
strscpy(region->name, "CS46xx_BA1_pmem");
sound/pci/cs46xx/cs46xx_lib.c
3832
region->base = chip->ba1_addr + BA1_SP_PMEM;
sound/pci/cs46xx/cs46xx_lib.c
3833
region->size = CS46XX_BA1_PRG_SIZE;
sound/pci/cs46xx/cs46xx_lib.c
3835
region = &chip->region.name.reg;
sound/pci/cs46xx/cs46xx_lib.c
3836
strscpy(region->name, "CS46xx_BA1_reg");
sound/pci/cs46xx/cs46xx_lib.c
3837
region->base = chip->ba1_addr + BA1_SP_REG;
sound/pci/cs46xx/cs46xx_lib.c
3838
region->size = CS46XX_BA1_REG_SIZE;
sound/pci/cs46xx/cs46xx_lib.c
3882
region = &chip->region.idx[idx];
sound/pci/cs46xx/cs46xx_lib.c
3883
region->remap_addr = devm_ioremap(&pci->dev, region->base,
sound/pci/cs46xx/cs46xx_lib.c
3884
region->size);
sound/pci/cs46xx/cs46xx_lib.c
3885
if (region->remap_addr == NULL) {
sound/pci/cs46xx/cs46xx_lib.c
3887
"%s ioremap problem\n", region->name);
sound/pci/cs46xx/cs46xx_lib.c
478
dst = chip->region.idx[bank+1].remap_addr + offset;
sound/pci/cs46xx/cs46xx_lib.h
55
writel(val, chip->region.idx[bank+1].remap_addr + offset);
sound/pci/cs46xx/cs46xx_lib.h
62
return readl(chip->region.idx[bank+1].remap_addr + offset);
sound/pci/cs46xx/cs46xx_lib.h
67
writel(val, chip->region.name.ba0.remap_addr + offset);
sound/pci/cs46xx/cs46xx_lib.h
72
return readl(chip->region.name.ba0.remap_addr + offset);
sound/pci/cs46xx/dsp_spos.c
546
void __iomem *dst = chip->region.idx[1].remap_addr + DSP_PARAMETER_BYTE_OFFSET;
sound/pci/cs46xx/dsp_spos.c
605
void __iomem *dst = chip->region.idx[1].remap_addr + DSP_PARAMETER_BYTE_OFFSET;
sound/pci/cs46xx/dsp_spos.c
633
void __iomem *dst = chip->region.idx[2].remap_addr;
sound/pci/cs46xx/dsp_spos.c
865
void __iomem *spdst = chip->region.idx[1].remap_addr +
sound/pci/cs46xx/dsp_spos.c
879
void __iomem *spdst = chip->region.idx[1].remap_addr +
sound/pci/cs46xx/dsp_spos_scb_lib.c
150
void __iomem *dst = chip->region.idx[2].remap_addr + sample_buffer_addr;
sound/pci/cs46xx/dsp_spos_scb_lib.c
64
void __iomem *dst = chip->region.idx[1].remap_addr + DSP_PARAMETER_BYTE_OFFSET;
sound/pci/maestro3.c
932
static u16 snd_m3_assp_read(struct snd_m3 *chip, u16 region, u16 index)
sound/pci/maestro3.c
934
snd_m3_outw(chip, region & MEMTYPE_MASK, DSP_PORT_MEMORY_TYPE);
sound/pci/maestro3.c
939
static void snd_m3_assp_write(struct snd_m3 *chip, u16 region, u16 index, u16 data)
sound/pci/maestro3.c
941
snd_m3_outw(chip, region & MEMTYPE_MASK, DSP_PORT_MEMORY_TYPE);
sound/soc/codecs/wm_adsp.c
1429
struct wm_adsp_buffer_region *region;
sound/soc/codecs/wm_adsp.c
1438
region = &buf->regions[i];
sound/soc/codecs/wm_adsp.c
1440
region->offset = offset;
sound/soc/codecs/wm_adsp.c
1441
region->mem_type = caps->region_defs[i].mem_type;
sound/soc/codecs/wm_adsp.c
1444
&region->base_addr);
sound/soc/codecs/wm_adsp.c
1453
region->cumulative_size = offset;
sound/soc/codecs/wm_adsp.c
1457
i, region->mem_type, region->base_addr,
sound/soc/codecs/wm_adsp.c
1458
region->offset, region->cumulative_size);
sound/soc/renesas/siu_dai.c
719
struct resource *res, *region;
sound/soc/renesas/siu_dai.c
745
region = devm_request_mem_region(&pdev->dev, res->start,
sound/soc/renesas/siu_dai.c
747
if (!region) {
tools/arch/x86/include/uapi/asm/kvm.h
400
__u32 region[1024];
tools/include/uapi/drm/i915_drm.h
3484
struct drm_i915_gem_memory_class_instance region;
tools/include/uapi/drm/i915_drm.h
857
int region;
tools/include/uapi/drm/i915_drm.h
864
int region;
tools/include/uapi/drm/i915_drm.h
869
int region;
tools/include/uapi/drm/i915_drm.h
878
int region;
tools/testing/memblock/tests/alloc_api.c
169
struct region r1;
tools/testing/memblock/tests/alloc_api.c
218
struct region r1, r2;
tools/testing/memblock/tests/alloc_api.c
267
struct region r1, r2;
tools/testing/memblock/tests/alloc_api.c
552
struct region r1;
tools/testing/memblock/tests/alloc_api.c
642
struct region r1;
tools/testing/memblock/tests/alloc_api.c
692
struct region r1, r2;
tools/testing/memblock/tests/alloc_api.c
75
struct region r1;
tools/testing/memblock/tests/alloc_exact_nid_api.c
354
struct region r1;
tools/testing/memblock/tests/alloc_exact_nid_api.c
666
struct region r1;
tools/testing/memblock/tests/alloc_exact_nid_api.c
845
struct region r1, r2;
tools/testing/memblock/tests/alloc_exact_nid_api.c
85
struct region r1;
tools/testing/memblock/tests/alloc_exact_nid_api.c
916
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
1406
struct region r1;
tools/testing/memblock/tests/alloc_nid_api.c
1466
struct region r1;
tools/testing/memblock/tests/alloc_nid_api.c
1951
struct region r1;
tools/testing/memblock/tests/alloc_nid_api.c
2013
struct region r1;
tools/testing/memblock/tests/alloc_nid_api.c
2396
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
2466
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
419
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
483
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
546
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
613
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
892
struct region r1, r2;
tools/testing/memblock/tests/alloc_nid_api.c
962
struct region r1, r2;
tools/testing/memblock/tests/basic_api.c
1004
struct region r = {
tools/testing/memblock/tests/basic_api.c
1127
struct region r = {
tools/testing/memblock/tests/basic_api.c
124
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
128
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1286
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1290
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1335
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1339
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1387
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1391
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1441
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1445
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1496
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1500
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1542
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1586
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1591
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1638
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1642
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1646
struct region r3 = {
tools/testing/memblock/tests/basic_api.c
1719
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1723
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1768
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
177
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1772
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
181
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1820
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1824
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1869
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1873
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1924
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
1928
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
1970
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
2013
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
2018
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
2065
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
2069
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
2073
struct region r3 = {
tools/testing/memblock/tests/basic_api.c
2181
struct region r = {
tools/testing/memblock/tests/basic_api.c
2227
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
2231
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
2282
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
2286
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
229
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
233
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
2343
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
2347
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
2392
struct region r = {
tools/testing/memblock/tests/basic_api.c
278
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
282
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
310
struct region r = {
tools/testing/memblock/tests/basic_api.c
349
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
353
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
357
struct region r3 = {
tools/testing/memblock/tests/basic_api.c
403
struct region r = {
tools/testing/memblock/tests/basic_api.c
436
struct region r = {
tools/testing/memblock/tests/basic_api.c
47
struct region r = {
tools/testing/memblock/tests/basic_api.c
552
struct region r = {
tools/testing/memblock/tests/basic_api.c
588
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
592
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
642
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
646
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
695
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
699
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
746
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
750
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
779
struct region r = {
tools/testing/memblock/tests/basic_api.c
80
struct region r = {
tools/testing/memblock/tests/basic_api.c
818
struct region r1 = {
tools/testing/memblock/tests/basic_api.c
822
struct region r2 = {
tools/testing/memblock/tests/basic_api.c
826
struct region r3 = {
tools/testing/memblock/tests/basic_api.c
872
struct region r = {
tools/testing/memblock/tests/basic_api.c
905
struct region r = {
tools/testing/nvdimm/dax-dev.c
28
if (dev_dax->region->align > PAGE_SIZE)
tools/testing/nvdimm/test/ndtest.c
390
struct ndtest_region *region = nd_region_provider_data(nd_region);
tools/testing/nvdimm/test/ndtest.c
392
return sprintf(buf, "%d\n", region->range_index);
tools/testing/nvdimm/test/ndtest.c
412
struct ndtest_region *region)
tools/testing/nvdimm/test/ndtest.c
418
int i, ndimm = region->mapping[0].dimm;
tools/testing/nvdimm/test/ndtest.c
426
if (!ndtest_alloc_resource(p, region->size, &res.start))
tools/testing/nvdimm/test/ndtest.c
429
res.end = res.start + region->size - 1;
tools/testing/nvdimm/test/ndtest.c
432
ndr_desc->provider_data = region;
tools/testing/nvdimm/test/ndtest.c
449
for (i = 0; i < region->num_mappings; i++) {
tools/testing/nvdimm/test/ndtest.c
450
ndimm = region->mapping[i].dimm;
tools/testing/nvdimm/test/ndtest.c
451
mappings[i].start = region->mapping[i].start;
tools/testing/nvdimm/test/ndtest.c
452
mappings[i].size = region->mapping[i].size;
tools/testing/nvdimm/test/ndtest.c
453
mappings[i].position = region->mapping[i].position;
tools/testing/nvdimm/test/ndtest.c
457
ndr_desc->num_mappings = region->num_mappings;
tools/testing/nvdimm/test/ndtest.c
458
region->region = nvdimm_pmem_region_create(p->bus, ndr_desc);
tools/testing/nvdimm/test/ndtest.c
460
if (!region->region) {
tools/testing/nvdimm/test/ndtest.h
62
struct nd_region *region;
tools/testing/nvdimm/test/nfit.c
634
struct nd_region *region;
tools/testing/nvdimm/test/nfit.c
655
ctx->region = nd_region;
tools/testing/nvdimm/test/nfit.c
671
.region = NULL,
tools/testing/nvdimm/test/nfit.c
682
nd_region = ctx.region;
tools/testing/selftests/kvm/arm64/page_fault_test.c
352
static void setup_uffd_args(struct userspace_mem_region *region,
tools/testing/selftests/kvm/arm64/page_fault_test.c
355
args->hva = (void *)region->region.userspace_addr;
tools/testing/selftests/kvm/arm64/page_fault_test.c
356
args->paging_size = region->region.memory_size;
tools/testing/selftests/kvm/arm64/page_fault_test.c
407
struct userspace_mem_region *region)
tools/testing/selftests/kvm/arm64/page_fault_test.c
409
void *hva = (void *)region->region.userspace_addr;
tools/testing/selftests/kvm/arm64/page_fault_test.c
410
uint64_t paging_size = region->region.memory_size;
tools/testing/selftests/kvm/arm64/page_fault_test.c
411
int ret, fd = region->fd;
tools/testing/selftests/kvm/arm64/page_fault_test.c
427
struct userspace_mem_region *region;
tools/testing/selftests/kvm/arm64/page_fault_test.c
430
region = vm_get_mem_region(vm, MEM_REGION_TEST_DATA);
tools/testing/selftests/kvm/arm64/page_fault_test.c
431
hva = (void *)region->region.userspace_addr;
tools/testing/selftests/kvm/arm64/page_fault_test.c
433
TEST_ASSERT_EQ(run->mmio.phys_addr, region->region.guest_phys_addr);
tools/testing/selftests/kvm/arm64/page_fault_test.c
451
struct userspace_mem_region *region,
tools/testing/selftests/kvm/arm64/page_fault_test.c
456
uint64_t size = region->region.memory_size;
tools/testing/selftests/kvm/arm64/page_fault_test.c
460
kvm_vm_get_dirty_log(vm, region->region.slot, bmap);
tools/testing/selftests/kvm/arm64/page_fault_test.c
476
pte_pg = (pte_gpa - pt_region->region.guest_phys_addr) / getpagesize();
tools/testing/selftests/kvm/arm64/page_fault_test.c
529
struct userspace_mem_region *region;
tools/testing/selftests/kvm/arm64/page_fault_test.c
532
region = vm_get_mem_region(vm, MEM_REGION_TEST_DATA);
tools/testing/selftests/kvm/arm64/page_fault_test.c
533
hva = (void *)region->region.userspace_addr;
tools/testing/selftests/kvm/arm64/page_fault_test.c
554
struct userspace_mem_region *region;
tools/testing/selftests/kvm/arm64/page_fault_test.c
557
region = vm_get_mem_region(vm, MEM_REGION_TEST_DATA);
tools/testing/selftests/kvm/arm64/page_fault_test.c
559
virt_pg_map(vm, TEST_GVA, region->region.guest_phys_addr);
tools/testing/selftests/kvm/arm64/page_fault_test.c
616
struct userspace_mem_region *region = vm_get_mem_region(vm, MEM_REGION_TEST_DATA);
tools/testing/selftests/kvm/arm64/page_fault_test.c
618
ucall_init(vm, region->region.guest_phys_addr + region->region.memory_size);
tools/testing/selftests/kvm/include/kvm_util.h
37
struct kvm_userspace_memory_region2 region;
tools/testing/selftests/kvm/include/x86/sev.h
113
struct userspace_mem_region *region)
tools/testing/selftests/kvm/include/x86/sev.h
116
.addr = region->region.userspace_addr,
tools/testing/selftests/kvm/include/x86/sev.h
117
.size = region->region.memory_size,
tools/testing/selftests/kvm/lib/kvm_util.c
1011
region = (struct userspace_mem_region *) userspace_mem_region_find(
tools/testing/selftests/kvm/lib/kvm_util.c
1013
if (region != NULL)
tools/testing/selftests/kvm/lib/kvm_util.c
1019
(uint64_t) region->region.guest_phys_addr,
tools/testing/selftests/kvm/lib/kvm_util.c
1020
(uint64_t) region->region.memory_size);
tools/testing/selftests/kvm/lib/kvm_util.c
1023
hash_for_each_possible(vm->regions.slot_hash, region, slot_node,
tools/testing/selftests/kvm/lib/kvm_util.c
1025
if (region->region.slot != slot)
tools/testing/selftests/kvm/lib/kvm_util.c
1032
slot, gpa, npages, region->region.slot,
tools/testing/selftests/kvm/lib/kvm_util.c
1033
(uint64_t) region->region.guest_phys_addr,
tools/testing/selftests/kvm/lib/kvm_util.c
1034
(uint64_t) region->region.memory_size);
tools/testing/selftests/kvm/lib/kvm_util.c
1038
region = calloc(1, sizeof(*region));
tools/testing/selftests/kvm/lib/kvm_util.c
1039
TEST_ASSERT(region != NULL, "Insufficient Memory");
tools/testing/selftests/kvm/lib/kvm_util.c
1040
region->mmap_size = mem_size;
tools/testing/selftests/kvm/lib/kvm_util.c
1062
region->mmap_size += alignment;
tools/testing/selftests/kvm/lib/kvm_util.c
1064
region->fd = -1;
tools/testing/selftests/kvm/lib/kvm_util.c
1066
region->fd = kvm_memfd_alloc(region->mmap_size,
tools/testing/selftests/kvm/lib/kvm_util.c
1069
region->mmap_start = kvm_mmap(region->mmap_size, PROT_READ | PROT_WRITE,
tools/testing/selftests/kvm/lib/kvm_util.c
1071
region->fd);
tools/testing/selftests/kvm/lib/kvm_util.c
1074
region->mmap_start == align_ptr_up(region->mmap_start, backing_src_pagesz),
tools/testing/selftests/kvm/lib/kvm_util.c
1076
region->mmap_start, backing_src_pagesz);
tools/testing/selftests/kvm/lib/kvm_util.c
1079
region->host_mem = align_ptr_up(region->mmap_start, alignment);
tools/testing/selftests/kvm/lib/kvm_util.c
1084
ret = madvise(region->host_mem, mem_size,
tools/testing/selftests/kvm/lib/kvm_util.c
1087
region->host_mem, mem_size,
tools/testing/selftests/kvm/lib/kvm_util.c
1091
region->backing_src_type = src_type;
tools/testing/selftests/kvm/lib/kvm_util.c
1109
region->region.guest_memfd = guest_memfd;
tools/testing/selftests/kvm/lib/kvm_util.c
1110
region->region.guest_memfd_offset = guest_memfd_offset;
tools/testing/selftests/kvm/lib/kvm_util.c
1112
region->region.guest_memfd = -1;
tools/testing/selftests/kvm/lib/kvm_util.c
1115
region->unused_phy_pages = sparsebit_alloc();
tools/testing/selftests/kvm/lib/kvm_util.c
1117
region->protected_phy_pages = sparsebit_alloc();
tools/testing/selftests/kvm/lib/kvm_util.c
1118
sparsebit_set_num(region->unused_phy_pages, gpa >> vm->page_shift, npages);
tools/testing/selftests/kvm/lib/kvm_util.c
1119
region->region.slot = slot;
tools/testing/selftests/kvm/lib/kvm_util.c
1120
region->region.flags = flags;
tools/testing/selftests/kvm/lib/kvm_util.c
1121
region->region.guest_phys_addr = gpa;
tools/testing/selftests/kvm/lib/kvm_util.c
1122
region->region.memory_size = npages * vm->page_size;
tools/testing/selftests/kvm/lib/kvm_util.c
1123
region->region.userspace_addr = (uintptr_t) region->host_mem;
tools/testing/selftests/kvm/lib/kvm_util.c
1124
ret = __vm_ioctl(vm, KVM_SET_USER_MEMORY_REGION2, &region->region);
tools/testing/selftests/kvm/lib/kvm_util.c
1129
ret, errno, slot, flags, gpa, region->region.memory_size,
tools/testing/selftests/kvm/lib/kvm_util.c
1130
region->region.guest_memfd);
tools/testing/selftests/kvm/lib/kvm_util.c
1133
vm_userspace_mem_region_gpa_insert(&vm->regions.gpa_tree, region);
tools/testing/selftests/kvm/lib/kvm_util.c
1134
vm_userspace_mem_region_hva_insert(&vm->regions.hva_tree, region);
tools/testing/selftests/kvm/lib/kvm_util.c
1135
hash_add(vm->regions.slot_hash, &region->slot_node, slot);
tools/testing/selftests/kvm/lib/kvm_util.c
1138
if (region->fd >= 0) {
tools/testing/selftests/kvm/lib/kvm_util.c
1139
region->mmap_alias = kvm_mmap(region->mmap_size,
tools/testing/selftests/kvm/lib/kvm_util.c
1142
region->fd);
tools/testing/selftests/kvm/lib/kvm_util.c
1145
region->host_alias = align_ptr_up(region->mmap_alias, alignment);
tools/testing/selftests/kvm/lib/kvm_util.c
1175
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
1177
hash_for_each_possible(vm->regions.slot_hash, region, slot_node,
tools/testing/selftests/kvm/lib/kvm_util.c
1179
if (region->region.slot == memslot)
tools/testing/selftests/kvm/lib/kvm_util.c
1180
return region;
tools/testing/selftests/kvm/lib/kvm_util.c
1207
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
1209
region = memslot2region(vm, slot);
tools/testing/selftests/kvm/lib/kvm_util.c
1211
region->region.flags = flags;
tools/testing/selftests/kvm/lib/kvm_util.c
1213
ret = __vm_ioctl(vm, KVM_SET_USER_MEMORY_REGION2, &region->region);
tools/testing/selftests/kvm/lib/kvm_util.c
1222
struct userspace_mem_region *region = memslot2region(vm, slot);
tools/testing/selftests/kvm/lib/kvm_util.c
1223
struct kvm_userspace_memory_region2 tmp = region->region;
tools/testing/selftests/kvm/lib/kvm_util.c
1227
vm_ioctl(vm, KVM_SET_USER_MEMORY_REGION2, &region->region);
tools/testing/selftests/kvm/lib/kvm_util.c
1246
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
1249
region = memslot2region(vm, slot);
tools/testing/selftests/kvm/lib/kvm_util.c
1251
region->region.guest_phys_addr = new_gpa;
tools/testing/selftests/kvm/lib/kvm_util.c
1253
ret = __vm_ioctl(vm, KVM_SET_USER_MEMORY_REGION2, &region->region);
tools/testing/selftests/kvm/lib/kvm_util.c
1275
struct userspace_mem_region *region = memslot2region(vm, slot);
tools/testing/selftests/kvm/lib/kvm_util.c
1277
region->region.memory_size = 0;
tools/testing/selftests/kvm/lib/kvm_util.c
1278
vm_ioctl(vm, KVM_SET_USER_MEMORY_REGION2, &region->region);
tools/testing/selftests/kvm/lib/kvm_util.c
1280
__vm_mem_region_delete(vm, region);
tools/testing/selftests/kvm/lib/kvm_util.c
1287
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
1296
region = userspace_mem_region_find(vm, gpa, gpa);
tools/testing/selftests/kvm/lib/kvm_util.c
1297
TEST_ASSERT(region && region->region.flags & KVM_MEM_GUEST_MEMFD,
tools/testing/selftests/kvm/lib/kvm_util.c
1300
offset = gpa - region->region.guest_phys_addr;
tools/testing/selftests/kvm/lib/kvm_util.c
1301
fd_offset = region->region.guest_memfd_offset + offset;
tools/testing/selftests/kvm/lib/kvm_util.c
1302
len = min_t(uint64_t, end - gpa, region->region.memory_size - offset);
tools/testing/selftests/kvm/lib/kvm_util.c
1304
ret = fallocate(region->region.guest_memfd, mode, fd_offset, len);
tools/testing/selftests/kvm/lib/kvm_util.c
1307
region->region.guest_memfd, mode, fd_offset);
tools/testing/selftests/kvm/lib/kvm_util.c
1623
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
1627
region = userspace_mem_region_find(vm, gpa, gpa);
tools/testing/selftests/kvm/lib/kvm_util.c
1628
if (!region) {
tools/testing/selftests/kvm/lib/kvm_util.c
1633
return (void *)((uintptr_t)region->host_mem
tools/testing/selftests/kvm/lib/kvm_util.c
1634
+ (gpa - region->region.guest_phys_addr));
tools/testing/selftests/kvm/lib/kvm_util.c
1659
struct userspace_mem_region *region =
tools/testing/selftests/kvm/lib/kvm_util.c
1662
if (hva >= region->host_mem) {
tools/testing/selftests/kvm/lib/kvm_util.c
1663
if (hva <= (region->host_mem
tools/testing/selftests/kvm/lib/kvm_util.c
1664
+ region->region.memory_size - 1))
tools/testing/selftests/kvm/lib/kvm_util.c
1666
region->region.guest_phys_addr
tools/testing/selftests/kvm/lib/kvm_util.c
1667
+ (hva - (uintptr_t)region->host_mem));
tools/testing/selftests/kvm/lib/kvm_util.c
1699
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
1702
region = userspace_mem_region_find(vm, gpa, gpa);
tools/testing/selftests/kvm/lib/kvm_util.c
1703
if (!region)
tools/testing/selftests/kvm/lib/kvm_util.c
1706
if (!region->host_alias)
tools/testing/selftests/kvm/lib/kvm_util.c
1709
offset = gpa - region->region.guest_phys_addr;
tools/testing/selftests/kvm/lib/kvm_util.c
1710
return (void *) ((uintptr_t) region->host_alias + offset);
tools/testing/selftests/kvm/lib/kvm_util.c
1969
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
1976
hash_for_each(vm->regions.slot_hash, ctr, region, slot_node) {
tools/testing/selftests/kvm/lib/kvm_util.c
1979
(uint64_t) region->region.guest_phys_addr,
tools/testing/selftests/kvm/lib/kvm_util.c
1980
(uint64_t) region->region.memory_size,
tools/testing/selftests/kvm/lib/kvm_util.c
1981
region->host_mem);
tools/testing/selftests/kvm/lib/kvm_util.c
1983
sparsebit_dump(stream, region->unused_phy_pages, 0);
tools/testing/selftests/kvm/lib/kvm_util.c
1984
if (region->protected_phy_pages) {
tools/testing/selftests/kvm/lib/kvm_util.c
1986
sparsebit_dump(stream, region->protected_phy_pages, 0);
tools/testing/selftests/kvm/lib/kvm_util.c
2105
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
2115
region = memslot2region(vm, memslot);
tools/testing/selftests/kvm/lib/kvm_util.c
2116
TEST_ASSERT(!protected || region->protected_phy_pages,
tools/testing/selftests/kvm/lib/kvm_util.c
2122
if (!sparsebit_is_set(region->unused_phy_pages, pg)) {
tools/testing/selftests/kvm/lib/kvm_util.c
2123
base = pg = sparsebit_next_set(region->unused_phy_pages, pg);
tools/testing/selftests/kvm/lib/kvm_util.c
2139
sparsebit_clear(region->unused_phy_pages, pg);
tools/testing/selftests/kvm/lib/kvm_util.c
2141
sparsebit_set(region->protected_phy_pages, pg);
tools/testing/selftests/kvm/lib/kvm_util.c
2370
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
2375
region = userspace_mem_region_find(vm, paddr, paddr);
tools/testing/selftests/kvm/lib/kvm_util.c
2376
TEST_ASSERT(region, "No vm physical memory at 0x%lx", paddr);
tools/testing/selftests/kvm/lib/kvm_util.c
2379
return sparsebit_is_set(region->protected_phy_pages, pg);
tools/testing/selftests/kvm/lib/kvm_util.c
516
ucall_init(vm, slot0->region.guest_phys_addr + slot0->region.memory_size);
tools/testing/selftests/kvm/lib/kvm_util.c
596
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
602
hash_for_each(vmp->regions.slot_hash, ctr, region, slot_node) {
tools/testing/selftests/kvm/lib/kvm_util.c
603
int ret = ioctl(vmp->fd, KVM_SET_USER_MEMORY_REGION2, &region->region);
tools/testing/selftests/kvm/lib/kvm_util.c
609
ret, errno, region->region.slot,
tools/testing/selftests/kvm/lib/kvm_util.c
610
region->region.flags,
tools/testing/selftests/kvm/lib/kvm_util.c
611
region->region.guest_phys_addr,
tools/testing/selftests/kvm/lib/kvm_util.c
612
region->region.memory_size);
tools/testing/selftests/kvm/lib/kvm_util.c
723
struct userspace_mem_region *region =
tools/testing/selftests/kvm/lib/kvm_util.c
725
uint64_t existing_start = region->region.guest_phys_addr;
tools/testing/selftests/kvm/lib/kvm_util.c
726
uint64_t existing_end = region->region.guest_phys_addr
tools/testing/selftests/kvm/lib/kvm_util.c
727
+ region->region.memory_size - 1;
tools/testing/selftests/kvm/lib/kvm_util.c
729
return region;
tools/testing/selftests/kvm/lib/kvm_util.c
806
struct userspace_mem_region *region)
tools/testing/selftests/kvm/lib/kvm_util.c
808
rb_erase(&region->gpa_node, &vm->regions.gpa_tree);
tools/testing/selftests/kvm/lib/kvm_util.c
809
rb_erase(&region->hva_node, &vm->regions.hva_tree);
tools/testing/selftests/kvm/lib/kvm_util.c
810
hash_del(&region->slot_node);
tools/testing/selftests/kvm/lib/kvm_util.c
812
sparsebit_free(&region->unused_phy_pages);
tools/testing/selftests/kvm/lib/kvm_util.c
813
sparsebit_free(&region->protected_phy_pages);
tools/testing/selftests/kvm/lib/kvm_util.c
814
kvm_munmap(region->mmap_start, region->mmap_size);
tools/testing/selftests/kvm/lib/kvm_util.c
815
if (region->fd >= 0) {
tools/testing/selftests/kvm/lib/kvm_util.c
817
kvm_munmap(region->mmap_alias, region->mmap_size);
tools/testing/selftests/kvm/lib/kvm_util.c
818
close(region->fd);
tools/testing/selftests/kvm/lib/kvm_util.c
820
if (region->region.guest_memfd >= 0)
tools/testing/selftests/kvm/lib/kvm_util.c
821
close(region->region.guest_memfd);
tools/testing/selftests/kvm/lib/kvm_util.c
823
free(region);
tools/testing/selftests/kvm/lib/kvm_util.c
833
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/kvm_util.c
839
hash_for_each_safe(vmp->regions.slot_hash, ctr, node, region, slot_node)
tools/testing/selftests/kvm/lib/kvm_util.c
840
__vm_mem_region_delete(vmp, region);
tools/testing/selftests/kvm/lib/kvm_util.c
870
struct userspace_mem_region *region)
tools/testing/selftests/kvm/lib/kvm_util.c
879
if (region->region.guest_phys_addr <
tools/testing/selftests/kvm/lib/kvm_util.c
880
cregion->region.guest_phys_addr)
tools/testing/selftests/kvm/lib/kvm_util.c
883
TEST_ASSERT(region->region.guest_phys_addr !=
tools/testing/selftests/kvm/lib/kvm_util.c
884
cregion->region.guest_phys_addr,
tools/testing/selftests/kvm/lib/kvm_util.c
891
rb_link_node(&region->gpa_node, parent, cur);
tools/testing/selftests/kvm/lib/kvm_util.c
892
rb_insert_color(&region->gpa_node, gpa_tree);
tools/testing/selftests/kvm/lib/kvm_util.c
896
struct userspace_mem_region *region)
tools/testing/selftests/kvm/lib/kvm_util.c
905
if (region->host_mem < cregion->host_mem)
tools/testing/selftests/kvm/lib/kvm_util.c
908
TEST_ASSERT(region->host_mem !=
tools/testing/selftests/kvm/lib/kvm_util.c
916
rb_link_node(&region->hva_node, parent, cur);
tools/testing/selftests/kvm/lib/kvm_util.c
917
rb_insert_color(&region->hva_node, hva_tree);
tools/testing/selftests/kvm/lib/kvm_util.c
924
struct kvm_userspace_memory_region region = {
tools/testing/selftests/kvm/lib/kvm_util.c
932
return ioctl(vm->fd, KVM_SET_USER_MEMORY_REGION, &region);
tools/testing/selftests/kvm/lib/kvm_util.c
952
struct kvm_userspace_memory_region2 region = {
tools/testing/selftests/kvm/lib/kvm_util.c
964
return ioctl(vm->fd, KVM_SET_USER_MEMORY_REGION2, &region);
tools/testing/selftests/kvm/lib/kvm_util.c
985
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/x86/processor.c
528
struct userspace_mem_region *region = memslot2region(vm, memslot);
tools/testing/selftests/kvm/lib/x86/processor.c
534
i = (region->region.guest_phys_addr >> vm->page_shift) - 1;
tools/testing/selftests/kvm/lib/x86/processor.c
535
last = i + (region->region.memory_size >> vm->page_shift);
tools/testing/selftests/kvm/lib/x86/processor.c
537
i = sparsebit_next_clear(region->unused_phy_pages, i);
tools/testing/selftests/kvm/lib/x86/sev.c
145
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/x86/sev.c
148
hash_for_each(vm->regions.slot_hash, ctr, region, slot_node)
tools/testing/selftests/kvm/lib/x86/sev.c
149
encrypt_region(vm, region, KVM_SEV_SNP_PAGE_TYPE_NORMAL, true);
tools/testing/selftests/kvm/lib/x86/sev.c
17
static void encrypt_region(struct kvm_vm *vm, struct userspace_mem_region *region,
tools/testing/selftests/kvm/lib/x86/sev.c
20
const struct sparsebit *protected_phy_pages = region->protected_phy_pages;
tools/testing/selftests/kvm/lib/x86/sev.c
21
const vm_paddr_t gpa_base = region->region.guest_phys_addr;
tools/testing/selftests/kvm/lib/x86/sev.c
29
sev_register_encrypted_memory(vm, region);
tools/testing/selftests/kvm/lib/x86/sev.c
87
struct userspace_mem_region *region;
tools/testing/selftests/kvm/lib/x86/sev.c
97
hash_for_each(vm->regions.slot_hash, ctr, region, slot_node)
tools/testing/selftests/kvm/lib/x86/sev.c
98
encrypt_region(vm, region, KVM_SEV_PAGE_TYPE_INVALID, false);
tools/testing/selftests/kvm/s390/cmma_test.c
146
ucall_init(vm, slot0->region.guest_phys_addr + slot0->region.memory_size);
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
34
struct page_region region;
tools/testing/selftests/kvm/s390/shared_zeropage_test.c
38
.vec = (uintptr_t)&region,
tools/testing/selftests/kvm/s390/ucontrol_test.c
442
struct kvm_userspace_memory_region region = {
tools/testing/selftests/kvm/s390/ucontrol_test.c
455
ASSERT_EQ(-1, ioctl(self->vm_fd, KVM_SET_USER_MEMORY_REGION, &region));
tools/testing/selftests/kvm/x86/state_test.c
299
xstate_bv = (void *)&((uint8_t *)state->xsave->region)[512];
tools/testing/selftests/mm/pagemap_ioctl.c
42
#define LEN(region) ((region.end - region.start)/page_size)
tools/testing/selftests/mm/rmap.c
199
if (data->region != MAP_FAILED)
tools/testing/selftests/mm/rmap.c
200
munmap(data->region, data->mapsize);
tools/testing/selftests/mm/rmap.c
201
data->region = MAP_FAILED;
tools/testing/selftests/mm/rmap.c
227
FORCE_READ(*data->region);
tools/testing/selftests/mm/rmap.c
230
int try_to_move_page(char *region)
tools/testing/selftests/mm/rmap.c
237
ret = move_pages(0, 1, (void **)&region, NULL, &status, MPOL_MF_MOVE_ALL);
tools/testing/selftests/mm/rmap.c
255
ret = move_pages(0, 1, (void **)&region, &node, &status, MPOL_MF_MOVE_ALL);
tools/testing/selftests/mm/rmap.c
280
ret = try_to_move_page(data->region);
tools/testing/selftests/mm/rmap.c
287
*data->expected_pfn = pagemap_get_pfn(pagemap_fd, data->region);
tools/testing/selftests/mm/rmap.c
297
if (data->region == MAP_FAILED)
tools/testing/selftests/mm/rmap.c
304
pfn = pagemap_get_pfn(pagemap_fd, data->region);
tools/testing/selftests/mm/rmap.c
316
data->region = mmap(0, data->mapsize, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/rmap.c
318
ASSERT_NE(data->region, MAP_FAILED);
tools/testing/selftests/mm/rmap.c
319
memset(data->region, 0xcf, data->mapsize);
tools/testing/selftests/mm/rmap.c
340
data->region = mmap(0, data->mapsize, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/rmap.c
342
ASSERT_NE(data->region, MAP_FAILED);
tools/testing/selftests/mm/rmap.c
343
memset(data->region, 0xcf, data->mapsize);
tools/testing/selftests/mm/rmap.c
365
data->region = mmap(0, data->mapsize, PROT_READ | PROT_WRITE,
tools/testing/selftests/mm/rmap.c
367
ASSERT_NE(data->region, MAP_FAILED);
tools/testing/selftests/mm/rmap.c
368
memset(data->region, 0xcf, data->mapsize);
tools/testing/selftests/mm/rmap.c
381
data->region = mmap(NULL, data->mapsize, PROT_READ|PROT_WRITE,
tools/testing/selftests/mm/rmap.c
383
if (data->region == MAP_FAILED)
tools/testing/selftests/mm/rmap.c
386
memset(data->region, 0xcf, data->mapsize);
tools/testing/selftests/mm/rmap.c
394
if (data->region == MAP_FAILED)
tools/testing/selftests/mm/rmap.c
400
ret = try_to_move_page(data->region);
tools/testing/selftests/mm/rmap.c
405
*data->expected_pfn = pagemap_get_pfn(pagemap_fd, data->region);
tools/testing/selftests/mm/rmap.c
60
char *region;
tools/testing/selftests/perf_events/mmap.c
115
rb = mmap(region, RB_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED | MAP_FIXED, fd, 0);
tools/testing/selftests/perf_events/mmap.c
134
aux = mmap(region + REGION_AUX_OFFS, AUX_SIZE, PROT_READ | PROT_WRITE,
tools/testing/selftests/perf_events/mmap.c
161
rb = mmap(region, RB_SIZE, PROT_READ | PROT_WRITE, MAP_SHARED | MAP_FIXED, self->fd, 0);
tools/testing/selftests/perf_events/mmap.c
174
aux = mmap(region + REGION_AUX_OFFS, AUX_SIZE, PROT_READ | PROT_WRITE,
tools/testing/selftests/perf_events/mmap.c
182
ASSERT_EQ(munmap(self->region, REGION_SIZE), 0);
tools/testing/selftests/perf_events/mmap.c
38
void *region;
tools/testing/selftests/perf_events/mmap.c
87
void *aux, *region;
tools/testing/selftests/perf_events/mmap.c
96
region = mmap(NULL, REGION_SIZE, PROT_NONE, MAP_ANON | MAP_PRIVATE, -1, 0);
tools/testing/selftests/perf_events/mmap.c
97
ASSERT_NE(region, MAP_FAILED);
tools/testing/selftests/perf_events/mmap.c
98
self->region = region;
tools/testing/selftests/powerpc/mm/pkey_siginfo.c
105
base = ((struct region *) p)->base;
tools/testing/selftests/powerpc/mm/pkey_siginfo.c
106
size = ((struct region *) p)->size;
tools/testing/selftests/powerpc/mm/pkey_siginfo.c
147
base = ((struct region *) p)->base;
tools/testing/selftests/powerpc/mm/pkey_siginfo.c
148
size = ((struct region *) p)->size;
tools/testing/selftests/powerpc/mm/pkey_siginfo.c
149
rights = ((struct region *) p)->rights;
tools/testing/selftests/powerpc/mm/pkey_siginfo.c
246
struct region r;
tools/testing/selftests/powerpc/syscalls/rtas_filter.c
157
struct region rmo_region;
tools/testing/selftests/powerpc/syscalls/rtas_filter.c
51
struct region *next;
tools/testing/selftests/powerpc/syscalls/rtas_filter.c
85
static int read_kregion_bounds(struct region *kregion)
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
211
VFIO_ASSERT_GE(device->driver.region.size, sizeof(*dsa));
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
49
return device->driver.region.vaddr;
tools/testing/selftests/vfio/lib/drivers/ioat/ioat.c
107
VFIO_ASSERT_GE(device->driver.region.size, sizeof(*ioat));
tools/testing/selftests/vfio/lib/drivers/ioat/ioat.c
28
return device->driver.region.vaddr;
tools/testing/selftests/vfio/lib/include/libvfio/iommu.h
38
int __iommu_map(struct iommu *iommu, struct dma_region *region);
tools/testing/selftests/vfio/lib/include/libvfio/iommu.h
40
static inline void iommu_map(struct iommu *iommu, struct dma_region *region)
tools/testing/selftests/vfio/lib/include/libvfio/iommu.h
42
VFIO_ASSERT_EQ(__iommu_map(iommu, region), 0);
tools/testing/selftests/vfio/lib/include/libvfio/iommu.h
45
int __iommu_unmap(struct iommu *iommu, struct dma_region *region, u64 *unmapped);
tools/testing/selftests/vfio/lib/include/libvfio/iommu.h
47
static inline void iommu_unmap(struct iommu *iommu, struct dma_region *region)
tools/testing/selftests/vfio/lib/include/libvfio/iommu.h
49
VFIO_ASSERT_EQ(__iommu_unmap(iommu, region, NULL), 0);
tools/testing/selftests/vfio/lib/include/libvfio/vfio_pci_driver.h
74
struct dma_region region;
tools/testing/selftests/vfio/lib/iommu.c
100
static int vfio_iommu_map(struct iommu *iommu, struct dma_region *region)
tools/testing/selftests/vfio/lib/iommu.c
105
.vaddr = (u64)region->vaddr,
tools/testing/selftests/vfio/lib/iommu.c
106
.iova = region->iova,
tools/testing/selftests/vfio/lib/iommu.c
107
.size = region->size,
tools/testing/selftests/vfio/lib/iommu.c
116
static int iommufd_map(struct iommu *iommu, struct dma_region *region)
tools/testing/selftests/vfio/lib/iommu.c
123
.user_va = (u64)region->vaddr,
tools/testing/selftests/vfio/lib/iommu.c
124
.iova = region->iova,
tools/testing/selftests/vfio/lib/iommu.c
125
.length = region->size,
tools/testing/selftests/vfio/lib/iommu.c
135
int __iommu_map(struct iommu *iommu, struct dma_region *region)
tools/testing/selftests/vfio/lib/iommu.c
140
ret = iommufd_map(iommu, region);
tools/testing/selftests/vfio/lib/iommu.c
142
ret = vfio_iommu_map(iommu, region);
tools/testing/selftests/vfio/lib/iommu.c
147
list_add(&region->link, &iommu->dma_regions);
tools/testing/selftests/vfio/lib/iommu.c
170
static int vfio_iommu_unmap(struct iommu *iommu, struct dma_region *region,
tools/testing/selftests/vfio/lib/iommu.c
173
return __vfio_iommu_unmap(iommu->container_fd, region->iova,
tools/testing/selftests/vfio/lib/iommu.c
174
region->size, 0, unmapped);
tools/testing/selftests/vfio/lib/iommu.c
195
static int iommufd_unmap(struct iommu *iommu, struct dma_region *region,
tools/testing/selftests/vfio/lib/iommu.c
198
return __iommufd_unmap(iommu->iommufd, region->iova, region->size,
tools/testing/selftests/vfio/lib/iommu.c
202
int __iommu_unmap(struct iommu *iommu, struct dma_region *region, u64 *unmapped)
tools/testing/selftests/vfio/lib/iommu.c
207
ret = iommufd_unmap(iommu, region, unmapped);
tools/testing/selftests/vfio/lib/iommu.c
209
ret = vfio_iommu_unmap(iommu, region, unmapped);
tools/testing/selftests/vfio/lib/iommu.c
214
list_del_init(&region->link);
tools/testing/selftests/vfio/lib/iommu.c
71
struct dma_region *region;
tools/testing/selftests/vfio/lib/iommu.c
73
list_for_each_entry(region, &iommu->dma_regions, link) {
tools/testing/selftests/vfio/lib/iommu.c
74
if (vaddr < region->vaddr)
tools/testing/selftests/vfio/lib/iommu.c
77
if (vaddr >= region->vaddr + region->size)
tools/testing/selftests/vfio/lib/iommu.c
81
*iova = region->iova + (vaddr - region->vaddr);
tools/testing/selftests/vfio/lib/vfio_pci_driver.c
52
VFIO_ASSERT_NOT_NULL(driver->region.vaddr);
tools/testing/selftests/vfio/vfio_dma_mapping_mmio_test.c
100
iommu_unmap(iommu, &region);
tools/testing/selftests/vfio/vfio_dma_mapping_mmio_test.c
102
VFIO_ASSERT_NE(__iommu_map(iommu, &region), 0);
tools/testing/selftests/vfio/vfio_dma_mapping_mmio_test.c
103
VFIO_ASSERT_NE(__iommu_unmap(iommu, &region, NULL), 0);
tools/testing/selftests/vfio/vfio_dma_mapping_mmio_test.c
87
struct dma_region region = {
tools/testing/selftests/vfio/vfio_dma_mapping_mmio_test.c
99
iommu_map(iommu, &region);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
138
struct dma_region region;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
144
region.vaddr = mmap(NULL, size, PROT_READ | PROT_WRITE, flags, -1, 0);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
147
if (flags & MAP_HUGETLB && region.vaddr == MAP_FAILED)
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
150
ASSERT_NE(region.vaddr, MAP_FAILED);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
152
region.iova = iova_allocator_alloc(self->iova_allocator, size);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
153
region.size = size;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
155
iommu_map(self->iommu, &region);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
156
printf("Mapped HVA %p (size 0x%lx) at IOVA 0x%lx\n", region.vaddr, size, region.iova);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
158
ASSERT_EQ(region.iova, to_iova(self->device, region.vaddr));
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
160
rc = iommu_mapping_get(device_bdf, region.iova, &mapping);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
168
printf("Found IOMMU mappings for IOVA 0x%lx:\n", region.iova);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
193
rc = __iommu_unmap(self->iommu, &region, &unmapped);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
195
ASSERT_EQ(unmapped, region.size);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
196
printf("Unmapped IOVA 0x%lx\n", region.iova);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
197
ASSERT_NE(0, __to_iova(self->device, region.vaddr, NULL));
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
198
ASSERT_NE(0, iommu_mapping_get(device_bdf, region.iova, &mapping));
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
200
ASSERT_TRUE(!munmap(region.vaddr, size));
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
206
struct dma_region region;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
225
struct dma_region *region = &self->region;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
239
region->vaddr = mmap(NULL, self->mmap_size, PROT_READ | PROT_WRITE,
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
241
ASSERT_NE(region->vaddr, MAP_FAILED);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
249
region->iova = last_iova & ~(region_size - 1);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
250
region->size = region_size;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
257
ASSERT_EQ(munmap(self->region.vaddr, self->mmap_size), 0);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
262
struct dma_region *region = &self->region;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
266
iommu_map(self->iommu, region);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
267
ASSERT_EQ(region->iova, to_iova(self->device, region->vaddr));
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
269
rc = __iommu_unmap(self->iommu, region, &unmapped);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
271
ASSERT_EQ(unmapped, region->size);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
276
struct dma_region *region = &self->region;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
280
iommu_map(self->iommu, region);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
281
ASSERT_EQ(region->iova, to_iova(self->device, region->vaddr));
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
285
ASSERT_EQ(unmapped, region->size);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
290
struct dma_region *region = &self->region;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
293
region->iova = ~(iova_t)0 & ~(region->size - 1);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
294
region->size = self->mmap_size;
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
296
rc = __iommu_map(self->iommu, region);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
299
rc = __iommu_unmap(self->iommu, region, NULL);
tools/testing/selftests/vfio/vfio_pci_driver_test.c
113
region_teardown(self->iommu, &driver->region);
tools/testing/selftests/vfio/vfio_pci_driver_test.c
23
struct dma_region *region, u64 size)
tools/testing/selftests/vfio/vfio_pci_driver_test.c
32
region->vaddr = vaddr;
tools/testing/selftests/vfio/vfio_pci_driver_test.c
33
region->iova = iova_allocator_alloc(iova_allocator, size);
tools/testing/selftests/vfio/vfio_pci_driver_test.c
34
region->size = size;
tools/testing/selftests/vfio/vfio_pci_driver_test.c
36
iommu_map(iommu, region);
tools/testing/selftests/vfio/vfio_pci_driver_test.c
39
static void region_teardown(struct iommu *iommu, struct dma_region *region)
tools/testing/selftests/vfio/vfio_pci_driver_test.c
41
iommu_unmap(iommu, region);
tools/testing/selftests/vfio/vfio_pci_driver_test.c
42
VFIO_ASSERT_EQ(munmap(region->vaddr, region->size), 0);
tools/testing/selftests/vfio/vfio_pci_driver_test.c
83
region_setup(self->iommu, self->iova_allocator, &driver->region, SZ_2M);