arch/sparc/kernel/leon_pci_grpci1.c
435
REGSTORE(regs->iomap, REGLOAD(regs->iomap) & 0x0000ffff);
arch/sparc/kernel/leon_pci_grpci1.c
44
unsigned int iomap; /* 0x14 IO Map */
block/fops.c
440
unsigned int flags, struct iomap *iomap, struct iomap *srcmap)
block/fops.c
448
iomap->bdev = bdev;
block/fops.c
449
iomap->offset = ALIGN_DOWN(offset, bdev_logical_block_size(bdev));
block/fops.c
450
iomap->type = IOMAP_MAPPED;
block/fops.c
451
iomap->addr = iomap->offset;
block/fops.c
452
iomap->length = isize - iomap->offset;
block/fops.c
453
iomap->flags |= IOMAP_F_BUFFER_HEAD; /* noop for !CONFIG_BUFFER_HEAD */
block/fops.c
555
if (offset < wpc->iomap.offset ||
block/fops.c
556
offset >= wpc->iomap.offset + wpc->iomap.length) {
block/fops.c
560
IOMAP_WRITE, &wpc->iomap, NULL);
drivers/ata/libata-sff.c
2154
void __iomem * const *iomap;
drivers/ata/libata-sff.c
2179
host->iomap = iomap = pcim_iomap_table(pdev);
drivers/ata/libata-sff.c
2181
ap->ioaddr.cmd_addr = iomap[base];
drivers/ata/libata-sff.c
2184
((unsigned long)iomap[base + 1] | ATA_PCI_CTL_OFS);
drivers/ata/libata-sff.c
3096
host->iomap = pcim_iomap_table(pdev);
drivers/ata/libata-sff.c
3100
void __iomem *bmdma = host->iomap[4] + 8 * i;
drivers/ata/pata_atp867x.c
418
host->iomap = pcim_iomap_table(pdev);
drivers/ata/pata_atp867x.c
424
host->iomap[i]);
drivers/ata/pata_atp867x.c
69
#define ATP867X_IOBASE(ap) ((ap)->host->iomap[0])
drivers/ata/pata_cs5520.c
121
void __iomem *iomap[5];
drivers/ata/pata_cs5520.c
160
iomap[0] = devm_ioport_map(&pdev->dev, cmd_port[0], 8);
drivers/ata/pata_cs5520.c
161
iomap[1] = devm_ioport_map(&pdev->dev, ctl_port[0], 1);
drivers/ata/pata_cs5520.c
162
iomap[2] = devm_ioport_map(&pdev->dev, cmd_port[1], 8);
drivers/ata/pata_cs5520.c
163
iomap[3] = devm_ioport_map(&pdev->dev, ctl_port[1], 1);
drivers/ata/pata_cs5520.c
164
iomap[4] = pcim_iomap(pdev, 2, 0);
drivers/ata/pata_cs5520.c
166
if (!iomap[0] || !iomap[1] || !iomap[2] || !iomap[3] || !iomap[4])
drivers/ata/pata_cs5520.c
170
ioaddr->cmd_addr = iomap[0];
drivers/ata/pata_cs5520.c
171
ioaddr->ctl_addr = iomap[1];
drivers/ata/pata_cs5520.c
172
ioaddr->altstatus_addr = iomap[1];
drivers/ata/pata_cs5520.c
173
ioaddr->bmdma_addr = iomap[4];
drivers/ata/pata_cs5520.c
181
ioaddr->cmd_addr = iomap[2];
drivers/ata/pata_cs5520.c
182
ioaddr->ctl_addr = iomap[3];
drivers/ata/pata_cs5520.c
183
ioaddr->altstatus_addr = iomap[3];
drivers/ata/pata_cs5520.c
184
ioaddr->bmdma_addr = iomap[4] + 8;
drivers/ata/pata_hpt3x3.c
223
host->iomap = pcim_iomap_table(pdev);
drivers/ata/pata_hpt3x3.c
228
base = host->iomap[4]; /* Bus mastering base */
drivers/ata/pata_macio.c
1089
priv->host->iomap = &priv->tfregs;
drivers/ata/pata_ninja32.c
125
host->iomap = pcim_iomap_table(dev);
drivers/ata/pata_ninja32.c
133
base = host->iomap[0];
drivers/ata/pata_ninja32.c
162
ninja32_program(host->iomap[0]);
drivers/ata/pata_pdc2027x.c
176
return ap->host->iomap[PDC_MMIO_BAR] + ap->port_no * 0x100 + offset;
drivers/ata/pata_pdc2027x.c
461
void __iomem *mmio_base = host->iomap[PDC_MMIO_BAR];
drivers/ata/pata_pdc2027x.c
502
void __iomem *mmio_base = host->iomap[PDC_MMIO_BAR];
drivers/ata/pata_pdc2027x.c
585
void __iomem *mmio_base = host->iomap[PDC_MMIO_BAR];
drivers/ata/pata_pdc2027x.c
708
host->iomap = pcim_iomap_table(pdev);
drivers/ata/pata_pdc2027x.c
714
mmio_base = host->iomap[PDC_MMIO_BAR];
drivers/ata/pata_sil680.c
371
host->iomap = pcim_iomap_table(pdev);
drivers/ata/pata_sil680.c
380
mmio_base = host->iomap[SIL680_MMIO_BAR];
drivers/ata/pdc_adma.c
41
ADMA_REGS((ap)->host->iomap[ADMA_MMIO_BAR], ap->port_no)
drivers/ata/pdc_adma.c
540
writeb(7, host->iomap[ADMA_MMIO_BAR] + ADMA_MODE_LOCK);
drivers/ata/pdc_adma.c
574
host->iomap = pcim_iomap_table(pdev);
drivers/ata/pdc_adma.c
575
mmio_base = host->iomap[ADMA_MMIO_BAR];
drivers/ata/sata_inic162x.c
820
void __iomem * const *iomap;
drivers/ata/sata_inic162x.c
851
host->iomap = iomap = pcim_iomap_table(pdev);
drivers/ata/sata_inic162x.c
852
hpriv->mmio_base = iomap[mmio_bar];
drivers/ata/sata_mv.c
4086
host->iomap = NULL;
drivers/ata/sata_mv.c
4400
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_mv.c
4401
hpriv->base = host->iomap[MV_PRIMARY_BAR];
drivers/ata/sata_nv.c
1028
ap->host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_CK804);
drivers/ata/sata_nv.c
1068
ap->host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_CK804);
drivers/ata/sata_nv.c
1124
mmio = ap->host->iomap[NV_MMIO_BAR] + NV_ADMA_PORT +
drivers/ata/sata_nv.c
1127
pp->gen_block = ap->host->iomap[NV_MMIO_BAR] + NV_ADMA_GEN;
drivers/ata/sata_nv.c
1253
void __iomem *mmio = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1497
irq_stat = readb(host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_CK804);
drivers/ata/sata_nv.c
1580
void __iomem *mmio_base = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1591
void __iomem *mmio_base = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1604
void __iomem *mmio_base = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1617
void __iomem *mmio_base = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1812
void __iomem *mmio = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1831
void __iomem *mmio = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1851
void __iomem *mmio = host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
1924
void __iomem *mmio = ap->host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
2288
irq_stat = readl(host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_MCP55);
drivers/ata/sata_nv.c
2361
base = host->iomap[NV_MMIO_BAR];
drivers/ata/sata_nv.c
899
u8 irq_stat = readb(host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_CK804)
drivers/ata/sata_nv.c
907
u8 irq_stat = readb(host->iomap[NV_MMIO_BAR] + NV_INT_STATUS_CK804)
drivers/ata/sata_promise.c
1095
void __iomem *host_mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_promise.c
1196
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_promise.c
709
void __iomem *host_mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_promise.c
747
void __iomem *host_mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_promise.c
774
void __iomem *host_mmio = ap->host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_promise.c
782
void __iomem *host_mmio = ap->host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_promise.c
923
if (!host || !host->iomap[PDC_MMIO_BAR])
drivers/ata/sata_promise.c
926
host_mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_promise.c
990
void __iomem *host_mmio = ap->host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_qstor.c
163
return host->iomap[QS_MMIO_BAR];
drivers/ata/sata_qstor.c
490
void __iomem *mmio_base = host->iomap[QS_MMIO_BAR];
drivers/ata/sata_qstor.c
566
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_qstor.c
568
rc = qs_set_dma_masks(pdev, host->iomap[QS_MMIO_BAR]);
drivers/ata/sata_qstor.c
575
void __iomem *chan = host->iomap[QS_MMIO_BAR] + offset;
drivers/ata/sata_sil.c
254
void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil.c
280
void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil.c
348
void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil.c
509
void __iomem *mmio_base = host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil.c
538
void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil.c
566
void __iomem *mmio_base = ap->host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil.c
653
void __iomem *mmio_base = host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil.c
759
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_sil.c
765
mmio_base = host->iomap[SIL_MMIO_BAR];
drivers/ata/sata_sil24.c
1140
void __iomem *host_base = host->iomap[SIL24_HOST_BAR];
drivers/ata/sata_sil24.c
1219
void __iomem *host_base = host->iomap[SIL24_HOST_BAR];
drivers/ata/sata_sil24.c
1263
void __iomem * const *iomap;
drivers/ata/sata_sil24.c
1284
iomap = pcim_iomap_table(pdev);
drivers/ata/sata_sil24.c
1288
tmp = readl(iomap[SIL24_HOST_BAR] + HOST_CTRL);
drivers/ata/sata_sil24.c
1301
host->iomap = iomap;
drivers/ata/sata_sil24.c
1331
void __iomem *host_base = host->iomap[SIL24_HOST_BAR];
drivers/ata/sata_sil24.c
468
return ap->host->iomap[SIL24_PORT_BAR] + sil24_port_offset(ap);
drivers/ata/sata_sis.c
286
mmio = host->iomap[SIS_SCR_PCI_BAR];
drivers/ata/sata_svw.c
460
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_svw.c
461
mmio_base = host->iomap[bar_pos];
drivers/ata/sata_sx4.c
1038
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
1097
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
1165
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
1223
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
1364
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
1422
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_sx4.c
1426
void __iomem *base = host->iomap[PDC_MMIO_BAR] + PDC_CHIP0_OFS;
drivers/ata/sata_sx4.c
419
void __iomem *mmio = ap->host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
420
void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR];
drivers/ata/sata_sx4.c
478
void __iomem *mmio = ap->host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
479
void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR];
drivers/ata/sata_sx4.c
529
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
582
void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR];
drivers/ata/sata_sx4.c
597
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
749
if (!host || !host->iomap[PDC_MMIO_BAR])
drivers/ata/sata_sx4.c
752
mmio_base = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
943
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
944
void __iomem *dimm_mmio = host->iomap[PDC_DIMM_BAR];
drivers/ata/sata_sx4.c
993
void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
drivers/ata/sata_sx4.c
994
void __iomem *dimm_mmio = host->iomap[PDC_DIMM_BAR];
drivers/ata/sata_uli.c
135
void __iomem * const *iomap;
drivers/ata/sata_uli.c
166
iomap = host->iomap;
drivers/ata/sata_uli.c
177
ioaddr->cmd_addr = iomap[0] + 8;
drivers/ata/sata_uli.c
180
((unsigned long)iomap[1] | ATA_PCI_CTL_OFS) + 4;
drivers/ata/sata_uli.c
181
ioaddr->bmdma_addr = iomap[4] + 16;
drivers/ata/sata_uli.c
192
ioaddr->cmd_addr = iomap[2] + 8;
drivers/ata/sata_uli.c
195
((unsigned long)iomap[3] | ATA_PCI_CTL_OFS) + 4;
drivers/ata/sata_uli.c
196
ioaddr->bmdma_addr = iomap[4] + 24;
drivers/ata/sata_via.c
439
void __iomem * const * iomap = ap->host->iomap;
drivers/ata/sata_via.c
440
void __iomem *reg_addr = iomap[ap->port_no];
drivers/ata/sata_via.c
441
void __iomem *bmdma_addr = iomap[4] + (ap->port_no * 8);
drivers/ata/sata_via.c
449
ioaddr->scr_addr = vt6421_scr_addr(iomap[5], ap->port_no);
drivers/ata/sata_via.c
479
host->ports[0]->ioaddr.scr_addr = svia_scr_addr(host->iomap[5], 0);
drivers/ata/sata_via.c
480
host->ports[1]->ioaddr.scr_addr = svia_scr_addr(host->iomap[5], 1);
drivers/ata/sata_via.c
504
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_vsc.c
108
mask_addr = ap->host->iomap[VSC_MMIO_BAR] +
drivers/ata/sata_vsc.c
119
mask_addr = ap->host->iomap[VSC_MMIO_BAR] +
drivers/ata/sata_vsc.c
131
mask_addr = ap->host->iomap[VSC_MMIO_BAR] +
drivers/ata/sata_vsc.c
255
status = readl(host->iomap[VSC_MMIO_BAR] + VSC_SATA_INT_STAT_OFFSET);
drivers/ata/sata_vsc.c
357
host->iomap = pcim_iomap_table(pdev);
drivers/ata/sata_vsc.c
359
mmio_base = host->iomap[VSC_MMIO_BAR];
drivers/base/property.c
1040
return fwnode_call_ptr_op(fwnode, iomap, index);
drivers/dma/ioat/init.c
1347
void __iomem * const *iomap;
drivers/dma/ioat/init.c
1361
iomap = pcim_iomap_table(pdev);
drivers/dma/ioat/init.c
1362
if (!iomap)
drivers/dma/ioat/init.c
1365
version = readb(iomap[IOAT_MMIO_BAR] + IOAT_VER_OFFSET);
drivers/dma/ioat/init.c
1373
device = alloc_ioatdma(pdev, iomap[IOAT_MMIO_BAR]);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1175
io_mapping_map_atomic_wc(&ggtt->iomap, offset);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1328
vaddr = (void __force *)io_mapping_map_atomic_wc(&ggtt->iomap,
drivers/gpu/drm/i915/gem/i915_gem_lmem.c
25
return io_mapping_map_wc(&obj->mm.region->iomap, offset, size);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
258
resource_size_t iomap;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
273
iomap = -1;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
275
iomap = obj->mm.region->iomap.base;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
276
iomap -= obj->mm.region->region.start;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
283
obj->mm.pages->sgl, obj_offset, iomap);
drivers/gpu/drm/i915/gem/i915_gem_mman.c
463
ret = remap_io_mapping(area, start, pfn, end - start, &ggtt->iomap);
drivers/gpu/drm/i915/gem/i915_gem_object.c
524
src_map = io_mapping_map_wc(&obj->mm.region->iomap,
drivers/gpu/drm/i915/gem/i915_gem_pages.c
335
resource_size_t iomap = obj->mm.region->iomap.base -
drivers/gpu/drm/i915/gem/i915_gem_pages.c
354
pfns[i++] = (iomap + addr) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gem/i915_gem_shrinker.c
412
if (!vma->iomap || i915_vma_is_active(vma))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
621
s = io_mapping_map_wc(&ggtt->iomap,
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
868
!io_mapping_init_wc(&mem->iomap, mem->io.start, resource_size(&mem->io)))
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
881
io_mapping_fini(&mem->iomap);
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
697
base = obj->mm.region->iomap.base - obj->mm.region->region.start;
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
708
resource_size_t iomap = obj->mm.region->iomap.base -
drivers/gpu/drm/i915/gem/i915_gem_ttm.c
728
ptr = ioremap_wc(iomap + daddr + offset, bytes);
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
326
ttm_kmap_iter_iomap_init(&arg->_dst_iter.io, &dst_reg->iomap,
drivers/gpu/drm/i915/gem/i915_gem_ttm_move.c
331
ttm_kmap_iter_iomap_init(&arg->_src_iter.io, &src_reg->iomap,
drivers/gpu/drm/i915/gt/intel_ggtt.c
1115
if (ggtt->iomap.size)
drivers/gpu/drm/i915/gt/intel_ggtt.c
1116
io_mapping_fini(&ggtt->iomap);
drivers/gpu/drm/i915/gt/intel_ggtt.c
71
if (!io_mapping_init_wc(&ggtt->iomap,
drivers/gpu/drm/i915/gt/intel_gtt.h
372
struct io_mapping iomap; /* Mapping to our CPU mappable region */
drivers/gpu/drm/i915/gt/intel_region_lmem.c
122
io_mapping_fini(&mem->iomap);
drivers/gpu/drm/i915/gt/intel_region_lmem.c
132
if (!io_mapping_init_wc(&mem->iomap,
drivers/gpu/drm/i915/gt/intel_region_lmem.c
144
io_mapping_fini(&mem->iomap);
drivers/gpu/drm/i915/gt/selftest_reset.c
137
s = io_mapping_map_wc(&ggtt->iomap,
drivers/gpu/drm/i915/gt/selftest_reset.c
94
s = io_mapping_map_wc(&ggtt->iomap,
drivers/gpu/drm/i915/gt/uc/intel_uc_fw.c
1323
vaddr = io_mapping_map_atomic_wc(&mr->iomap,
drivers/gpu/drm/i915/gvt/kvmgt.c
784
aperture_va = io_mapping_map_wc(&vgpu->gvt->gt->ggtt->iomap,
drivers/gpu/drm/i915/i915_gem.c
430
if (gtt_user_read(&ggtt->iomap, page_base, page_offset,
drivers/gpu/drm/i915/i915_gem.c
617
if (ggtt_write(&ggtt->iomap, page_base, page_offset,
drivers/gpu/drm/i915/i915_gpu_error.c
1195
s = io_mapping_map_wc(&ggtt->iomap, slot, PAGE_SIZE);
drivers/gpu/drm/i915/i915_gpu_error.c
1220
s = io_mapping_map_wc(&mem->iomap, offset, PAGE_SIZE);
drivers/gpu/drm/i915/i915_mm.c
106
r.prot = __pgprot((pgprot_val(iomap->prot) & _PAGE_CACHE_MASK) |
drivers/gpu/drm/i915/i915_mm.c
96
struct io_mapping *iomap)
drivers/gpu/drm/i915/i915_mm.h
19
struct io_mapping *iomap);
drivers/gpu/drm/i915/i915_mm.h
24
struct io_mapping *iomap)
drivers/gpu/drm/i915/i915_vma.c
1919
if (vma->iomap == NULL)
drivers/gpu/drm/i915/i915_vma.c
1922
if (page_unmask_bits(vma->iomap))
drivers/gpu/drm/i915/i915_vma.c
1925
io_mapping_unmap(vma->iomap);
drivers/gpu/drm/i915/i915_vma.c
1926
vma->iomap = NULL;
drivers/gpu/drm/i915/i915_vma.c
595
ptr = READ_ONCE(vma->iomap);
drivers/gpu/drm/i915/i915_vma.c
607
ptr = io_mapping_map_wc(&i915_vm_to_ggtt(vma->vm)->iomap,
drivers/gpu/drm/i915/i915_vma.c
625
if (unlikely(cmpxchg(&vma->iomap, NULL, ptr))) {
drivers/gpu/drm/i915/i915_vma.c
630
ptr = vma->iomap;
drivers/gpu/drm/i915/i915_vma.c
659
GEM_BUG_ON(vma->iomap == NULL);
drivers/gpu/drm/i915/i915_vma.h
340
return READ_ONCE(vma->iomap);
drivers/gpu/drm/i915/i915_vma_types.h
144
void __iomem *iomap;
drivers/gpu/drm/i915/intel_memory_region.h
64
struct io_mapping iomap;
drivers/gpu/drm/i915/selftests/i915_gem.c
67
s = io_mapping_map_atomic_wc(&ggtt->iomap, slot);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1405
vaddr = io_mapping_map_atomic_wc(&ggtt->iomap, offset);
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1417
vaddr = io_mapping_map_atomic_wc(&ggtt->iomap, offset);
drivers/gpu/drm/ttm/ttm_resource.c
745
addr = io_mapping_map_local_wc(iter_io->iomap, iter_io->cache.offs +
drivers/gpu/drm/ttm/ttm_resource.c
776
struct io_mapping *iomap,
drivers/gpu/drm/ttm/ttm_resource.c
781
iter_io->iomap = iomap;
drivers/net/ethernet/microchip/lan966x/lan966x_main.c
101
lan966x->regs[iomap->id] = begin[iomap->range] + iomap->offset;
drivers/net/ethernet/microchip/lan966x/lan966x_main.c
98
const struct lan966x_main_io_resource *iomap =
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x.c
351
.iomap = lan969x_main_iomap,
drivers/net/ethernet/microchip/sparx5/sparx5_main.c
1112
.iomap = sparx5_main_iomap,
drivers/net/ethernet/microchip/sparx5/sparx5_main.c
270
const struct sparx5_main_io_resource *iomap = sparx5->data->iomap;
drivers/net/ethernet/microchip/sparx5/sparx5_main.c
280
const struct sparx5_main_io_resource *io = &iomap[jdx];
drivers/net/ethernet/microchip/sparx5/sparx5_main.c
302
begin[idx] = iomem[idx] - iomap[range_id[idx]].offset;
drivers/net/ethernet/microchip/sparx5/sparx5_main.c
305
const struct sparx5_main_io_resource *io = &iomap[jdx];
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
363
const struct sparx5_main_io_resource *iomap;
drivers/net/ethernet/ti/davinci_cpdma.c
250
pool->iomap = devm_ioremap(ctlr->dev, pool->phys,
drivers/net/ethernet/ti/davinci_cpdma.c
256
pool->iomap = (void __iomem __force *)pool->cpumap;
drivers/net/ethernet/ti/davinci_cpdma.c
260
if (!pool->iomap)
drivers/net/ethernet/ti/davinci_cpdma.c
263
ret = gen_pool_add_virt(pool->gen_pool, (unsigned long)pool->iomap,
drivers/net/ethernet/ti/davinci_cpdma.c
284
return pool->hw_addr + (__force long)desc - (__force long)pool->iomap;
drivers/net/ethernet/ti/davinci_cpdma.c
290
return dma ? pool->iomap + dma - pool->hw_addr : NULL;
drivers/net/ethernet/ti/davinci_cpdma.c
82
void __iomem *iomap; /* ioremap map */
drivers/of/property.c
1661
.iomap = of_fwnode_iomap,
drivers/pcmcia/pcmcia_resource.c
483
pccard_io_map iomap;
drivers/pcmcia/pcmcia_resource.c
585
iomap.speed = io_speed;
drivers/pcmcia/pcmcia_resource.c
588
iomap.map = i;
drivers/pcmcia/pcmcia_resource.c
589
iomap.flags = MAP_ACTIVE;
drivers/pcmcia/pcmcia_resource.c
592
iomap.flags |= MAP_16BIT; break;
drivers/pcmcia/pcmcia_resource.c
594
iomap.flags |= MAP_AUTOSZ; break;
drivers/pcmcia/pcmcia_resource.c
598
iomap.start = s->io[i].res->start;
drivers/pcmcia/pcmcia_resource.c
599
iomap.stop = s->io[i].res->end;
drivers/pcmcia/pcmcia_resource.c
600
s->ops->set_io_map(s, &iomap);
drivers/phy/microchip/sparx5_serdes.c
2596
.iomap = sparx5_serdes_iomap,
drivers/phy/microchip/sparx5_serdes.c
2611
.iomap = lan969x_serdes_iomap,
drivers/phy/microchip/sparx5_serdes.c
2703
const struct sparx5_serdes_io_resource *iomap =
drivers/phy/microchip/sparx5_serdes.c
2704
&priv->data->iomap[idx];
drivers/phy/microchip/sparx5_serdes.c
2706
priv->regs[iomap->id] = iomem + iomap->offset;
drivers/phy/microchip/sparx5_serdes.h
68
const struct sparx5_serdes_io_resource *iomap;
drivers/scsi/isci/init.c
269
void __iomem * const *iomap;
drivers/scsi/isci/init.c
286
iomap = pcim_iomap_table(pdev);
drivers/scsi/isci/init.c
287
if (!iomap)
drivers/staging/media/ipu3/ipu3.c
622
void __iomem *const *iomap;
drivers/staging/media/ipu3/ipu3.c
651
iomap = pcim_iomap_table(pci_dev);
drivers/staging/media/ipu3/ipu3.c
652
if (!iomap) {
drivers/staging/media/ipu3/ipu3.c
657
imgu->base = iomap[IMGU_PCI_BAR];
fs/btrfs/direct-io.c
356
loff_t length, unsigned int flags, struct iomap *iomap,
fs/btrfs/direct-io.c
357
struct iomap *srcmap)
fs/btrfs/direct-io.c
359
struct iomap_iter *iter = container_of(iomap, struct iomap_iter, iomap);
fs/btrfs/direct-io.c
559
iomap->addr = IOMAP_NULL_ADDR;
fs/btrfs/direct-io.c
560
iomap->type = IOMAP_HOLE;
fs/btrfs/direct-io.c
562
iomap->addr = btrfs_extent_map_block_start(em) + (start - em->start);
fs/btrfs/direct-io.c
563
iomap->type = IOMAP_MAPPED;
fs/btrfs/direct-io.c
565
iomap->offset = start;
fs/btrfs/direct-io.c
566
iomap->bdev = fs_info->fs_devices->latest_dev->bdev;
fs/btrfs/direct-io.c
567
iomap->length = len;
fs/btrfs/direct-io.c
609
ssize_t written, unsigned int flags, struct iomap *iomap)
fs/btrfs/direct-io.c
611
struct iomap_iter *iter = container_of(iomap, struct iomap_iter, iomap);
fs/btrfs/direct-io.c
617
if (!write && (iomap->type == IOMAP_HOLE)) {
fs/buffer.c
2041
const struct iomap *iomap)
fs/buffer.c
2045
bh->b_bdev = iomap->bdev;
fs/buffer.c
2053
if (offset >= iomap->offset + iomap->length)
fs/buffer.c
2056
switch (iomap->type) {
fs/buffer.c
2085
if ((iomap->flags & IOMAP_F_NEW) ||
fs/buffer.c
2097
bh->b_blocknr = (iomap->addr + offset - iomap->offset) >>
fs/buffer.c
2108
get_block_t *get_block, const struct iomap *iomap)
fs/buffer.c
2144
err = iomap_to_bh(inode, block, bh, iomap);
fs/dax.c
1000
return PHYS_PFN(iomap->addr + (pos & PAGE_MASK) - iomap->offset);
fs/dax.c
1005
pgoff_t pgoff = dax_iomap_pgoff(&iter->iomap, iter->pos);
fs/dax.c
1011
rc = dax_direct_access(iter->iomap.dax_dev, pgoff, 1, DAX_ACCESS,
fs/dax.c
1032
(iter->iomap.flags & IOMAP_F_DIRTY);
fs/dax.c
1050
bool shared = iter->iomap.flags & IOMAP_F_SHARED;
fs/dax.c
1240
static int dax_iomap_direct_access(const struct iomap *iomap, loff_t pos,
fs/dax.c
1243
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
1248
length = dax_direct_access(iomap->dax_dev, pgoff, PHYS_PFN(size),
fs/dax.c
1292
const struct iomap *srcmap, void *daddr)
fs/dax.c
1407
struct iomap *iomap = &iter->iomap;
fs/dax.c
1408
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/dax.c
1439
ret = dax_iomap_direct_access(iomap, copy_pos, copy_len, &daddr, NULL);
fs/dax.c
1480
const struct iomap *iomap = &iter->iomap;
fs/dax.c
1481
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/dax.c
1483
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
1487
ret = dax_direct_access(iomap->dax_dev, pgoff, 1, DAX_ACCESS, &kaddr,
fs/dax.c
1493
if (iomap->flags & IOMAP_F_SHARED)
fs/dax.c
1497
dax_flush(iomap->dax_dev, kaddr + offset, size);
fs/dax.c
1503
const struct iomap *iomap = &iter->iomap;
fs/dax.c
1504
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/dax.c
1516
if (iomap->flags & IOMAP_F_SHARED)
fs/dax.c
1524
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
1531
ret = dax_zero_page_range(iomap->dax_dev, pgoff, 1);
fs/dax.c
1581
const struct iomap *iomap = &iomi->iomap;
fs/dax.c
1582
const struct iomap *srcmap = iomap_iter_srcmap(iomi);
fs/dax.c
1585
struct dax_device *dax_dev = iomap->dax_dev;
fs/dax.c
1588
bool cow = write && iomap->flags & IOMAP_F_SHARED;
fs/dax.c
1598
if (iomap->type == IOMAP_HOLE || iomap->type == IOMAP_UNWRITTEN) {
fs/dax.c
1608
if (WARN_ON_ONCE(iomap->type != IOMAP_MAPPED &&
fs/dax.c
1609
!(iomap->flags & IOMAP_F_SHARED)))
fs/dax.c
1617
if (iomap->flags & IOMAP_F_NEW || cow) {
fs/dax.c
1637
pgoff_t pgoff = dax_iomap_pgoff(iomap, pos);
fs/dax.c
1772
switch (iter->iomap.type) {
fs/dax.c
1809
const struct iomap *iomap = &iter->iomap;
fs/dax.c
1810
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/dax.c
1825
(iomap->type == IOMAP_UNWRITTEN || iomap->type == IOMAP_HOLE)) {
fs/dax.c
1831
if (iomap->type != IOMAP_MAPPED && !(iomap->flags & IOMAP_F_SHARED)) {
fs/dax.c
1836
err = dax_iomap_direct_access(iomap, pos, size, &kaddr, &pfn);
fs/dax.c
1842
if (write && iomap->flags & IOMAP_F_SHARED) {
fs/dax.c
1916
(iter.iomap.flags & IOMAP_F_NEW)) {
fs/dax.c
2159
const struct iomap *smap = &it_src->iomap;
fs/dax.c
2160
const struct iomap *dmap = &it_dest->iomap;
fs/dax.c
998
static pgoff_t dax_iomap_pgoff(const struct iomap *iomap, loff_t pos)
fs/erofs/data.c
277
unsigned int flags, struct iomap *iomap, struct iomap *srcmap)
fs/erofs/data.c
279
struct iomap_iter *iter = container_of(iomap, struct iomap_iter, iomap);
fs/erofs/data.c
293
iomap->offset = map.m_la;
fs/erofs/data.c
294
iomap->length = map.m_llen;
fs/erofs/data.c
295
iomap->flags = 0;
fs/erofs/data.c
296
iomap->addr = IOMAP_NULL_ADDR;
fs/erofs/data.c
298
iomap->type = IOMAP_HOLE;
fs/erofs/data.c
312
iomap->dax_dev = mdev.m_dif->dax_dev;
fs/erofs/data.c
314
iomap->bdev = mdev.m_bdev;
fs/erofs/data.c
315
iomap->addr = mdev.m_dif->fsoff + mdev.m_pa;
fs/erofs/data.c
317
iomap->addr += mdev.m_dif->dax_part_off;
fs/erofs/data.c
321
iomap->type = IOMAP_INLINE;
fs/erofs/data.c
331
iomap->inline_data = ptr;
fs/erofs/data.c
336
iomap->type = IOMAP_MAPPED;
fs/erofs/data.c
342
ssize_t written, unsigned int flags, struct iomap *iomap)
fs/erofs/data.c
344
struct iomap_iter *iter = container_of(iomap, struct iomap_iter, iomap);
fs/erofs/data.c
353
DBG_BUGON(iomap->type != IOMAP_INLINE);
fs/erofs/zmap.c
771
struct iomap *iomap, struct iomap *srcmap)
fs/erofs/zmap.c
781
iomap->bdev = inode->i_sb->s_bdev;
fs/erofs/zmap.c
782
iomap->offset = map.m_la;
fs/erofs/zmap.c
783
iomap->length = map.m_llen;
fs/erofs/zmap.c
785
iomap->type = IOMAP_MAPPED;
fs/erofs/zmap.c
786
iomap->addr = map.m_flags & __EROFS_MAP_FRAGMENT ?
fs/erofs/zmap.c
789
iomap->type = IOMAP_HOLE;
fs/erofs/zmap.c
790
iomap->addr = IOMAP_NULL_ADDR;
fs/erofs/zmap.c
800
if (iomap->offset >= inode->i_size)
fs/erofs/zmap.c
801
iomap->length = length + offset - map.m_la;
fs/erofs/zmap.c
803
iomap->flags = 0;
fs/ext2/inode.c
807
unsigned flags, struct iomap *iomap, struct iomap *srcmap)
fs/ext2/inode.c
833
iomap->flags |= IOMAP_F_DIRTY;
fs/ext2/inode.c
840
iomap->flags = 0;
fs/ext2/inode.c
841
iomap->offset = (u64)first_block << blkbits;
fs/ext2/inode.c
843
iomap->dax_dev = sbi->s_daxdev;
fs/ext2/inode.c
845
iomap->bdev = inode->i_sb->s_bdev;
fs/ext2/inode.c
854
iomap->type = IOMAP_HOLE;
fs/ext2/inode.c
855
iomap->addr = IOMAP_NULL_ADDR;
fs/ext2/inode.c
856
iomap->length = 1 << blkbits;
fs/ext2/inode.c
858
iomap->type = IOMAP_MAPPED;
fs/ext2/inode.c
859
iomap->addr = (u64)bno << blkbits;
fs/ext2/inode.c
861
iomap->addr += sbi->s_dax_part_off;
fs/ext2/inode.c
862
iomap->length = (u64)ret << blkbits;
fs/ext2/inode.c
863
iomap->flags |= IOMAP_F_MERGED;
fs/ext2/inode.c
867
iomap->flags |= IOMAP_F_NEW;
fs/ext2/inode.c
873
ssize_t written, unsigned flags, struct iomap *iomap)
fs/ext2/inode.c
882
if (iomap->type == IOMAP_MAPPED &&
fs/ext4/ext4.h
3695
struct iomap;
fs/ext4/ext4.h
3696
extern int ext4_inline_data_iomap(struct inode *inode, struct iomap *iomap);
fs/ext4/extents.c
5111
static int ext4_iomap_xattr_fiemap(struct inode *inode, struct iomap *iomap)
fs/ext4/extents.c
5144
iomap->addr = physical;
fs/ext4/extents.c
5145
iomap->offset = 0;
fs/ext4/extents.c
5146
iomap->length = length;
fs/ext4/extents.c
5147
iomap->type = iomap_type;
fs/ext4/extents.c
5148
iomap->flags = 0;
fs/ext4/extents.c
5155
struct iomap *iomap, struct iomap *srcmap)
fs/ext4/extents.c
5159
error = ext4_iomap_xattr_fiemap(inode, iomap);
fs/ext4/extents.c
5160
if (error == 0 && (offset >= iomap->length))
fs/ext4/inline.c
1812
int ext4_inline_data_iomap(struct inode *inode, struct iomap *iomap)
fs/ext4/inline.c
1832
iomap->addr = addr;
fs/ext4/inline.c
1833
iomap->offset = 0;
fs/ext4/inline.c
1834
iomap->length = min_t(loff_t, ext4_get_inline_size(inode),
fs/ext4/inline.c
1836
iomap->type = IOMAP_INLINE;
fs/ext4/inline.c
1837
iomap->flags = 0;
fs/ext4/inode.c
3455
static void ext4_set_iomap(struct inode *inode, struct iomap *iomap,
fs/ext4/inode.c
3466
iomap->flags = 0;
fs/ext4/inode.c
3469
iomap->flags |= IOMAP_F_DIRTY;
fs/ext4/inode.c
3472
iomap->flags |= IOMAP_F_NEW;
fs/ext4/inode.c
3476
iomap->flags |= IOMAP_F_ATOMIC_BIO;
fs/ext4/inode.c
3479
iomap->dax_dev = EXT4_SB(inode->i_sb)->s_daxdev;
fs/ext4/inode.c
3481
iomap->bdev = inode->i_sb->s_bdev;
fs/ext4/inode.c
3482
iomap->offset = EXT4_LBLK_TO_B(inode, map->m_lblk);
fs/ext4/inode.c
3483
iomap->length = EXT4_LBLK_TO_B(inode, map->m_len);
fs/ext4/inode.c
3487
iomap->flags |= IOMAP_F_MERGED;
fs/ext4/inode.c
3499
iomap->type = IOMAP_UNWRITTEN;
fs/ext4/inode.c
3500
iomap->addr = (u64) map->m_pblk << blkbits;
fs/ext4/inode.c
3502
iomap->addr += EXT4_SB(inode->i_sb)->s_dax_part_off;
fs/ext4/inode.c
3504
iomap->type = IOMAP_MAPPED;
fs/ext4/inode.c
3505
iomap->addr = (u64) map->m_pblk << blkbits;
fs/ext4/inode.c
3507
iomap->addr += EXT4_SB(inode->i_sb)->s_dax_part_off;
fs/ext4/inode.c
3509
iomap->type = IOMAP_DELALLOC;
fs/ext4/inode.c
3510
iomap->addr = IOMAP_NULL_ADDR;
fs/ext4/inode.c
3512
iomap->type = IOMAP_HOLE;
fs/ext4/inode.c
3513
iomap->addr = IOMAP_NULL_ADDR;
fs/ext4/inode.c
3759
unsigned flags, struct iomap *iomap, struct iomap *srcmap)
fs/ext4/inode.c
3832
ext4_set_iomap(inode, iomap, &map, offset, length, flags);
fs/ext4/inode.c
3843
struct iomap *iomap, struct iomap *srcmap)
fs/ext4/inode.c
3853
ret = ext4_inline_data_iomap(inode, iomap);
fs/ext4/inode.c
3855
if (ret == 0 && offset >= iomap->length)
fs/ext4/inode.c
3887
ext4_set_iomap(inode, iomap, &map, offset, length, flags);
fs/f2fs/data.c
4502
unsigned int flags, struct iomap *iomap,
fs/f2fs/data.c
4503
struct iomap *srcmap)
fs/f2fs/data.c
4514
if (flags & IOMAP_WRITE && iomap->private) {
fs/f2fs/data.c
4515
map.m_last_pblk = (unsigned long)iomap->private;
fs/f2fs/data.c
4516
iomap->private = NULL;
fs/f2fs/data.c
4531
iomap->offset = F2FS_BLK_TO_BYTES(map.m_lblk);
fs/f2fs/data.c
4551
iomap->length = F2FS_BLK_TO_BYTES(map.m_len);
fs/f2fs/data.c
4552
iomap->type = IOMAP_MAPPED;
fs/f2fs/data.c
4553
iomap->flags |= IOMAP_F_MERGED;
fs/f2fs/data.c
4554
iomap->bdev = map.m_bdev;
fs/f2fs/data.c
4555
iomap->addr = F2FS_BLK_TO_BYTES(map.m_pblk);
fs/f2fs/data.c
4558
iomap->private = (void *)map.m_last_pblk;
fs/f2fs/data.c
4564
iomap->length = F2FS_BLK_TO_BYTES(next_pgofs) -
fs/f2fs/data.c
4565
iomap->offset;
fs/f2fs/data.c
4566
iomap->type = IOMAP_HOLE;
fs/f2fs/data.c
4568
iomap->length = F2FS_BLK_TO_BYTES(map.m_len);
fs/f2fs/data.c
4569
iomap->type = IOMAP_UNWRITTEN;
fs/f2fs/data.c
4573
iomap->addr = IOMAP_NULL_ADDR;
fs/f2fs/data.c
4577
iomap->flags |= IOMAP_F_NEW;
fs/f2fs/data.c
4580
iomap->flags |= IOMAP_F_DIRTY;
fs/fuse/dax.c
388
static void fuse_fill_iomap_hole(struct iomap *iomap, loff_t length)
fs/fuse/dax.c
390
iomap->addr = IOMAP_NULL_ADDR;
fs/fuse/dax.c
391
iomap->length = length;
fs/fuse/dax.c
392
iomap->type = IOMAP_HOLE;
fs/fuse/dax.c
396
struct iomap *iomap, struct fuse_dax_mapping *dmap,
fs/fuse/dax.c
410
iomap->addr = dmap->window_offset + offset;
fs/fuse/dax.c
411
iomap->length = len;
fs/fuse/dax.c
413
iomap->length = ALIGN(len, PAGE_SIZE);
fs/fuse/dax.c
414
iomap->type = IOMAP_MAPPED;
fs/fuse/dax.c
423
WARN_ON_ONCE(iomap->private);
fs/fuse/dax.c
424
iomap->private = dmap;
fs/fuse/dax.c
427
fuse_fill_iomap_hole(iomap, length);
fs/fuse/dax.c
433
struct iomap *iomap)
fs/fuse/dax.c
480
fuse_fill_iomap(inode, pos, length, iomap, dmap, flags);
fs/fuse/dax.c
494
fuse_fill_iomap(inode, pos, length, iomap, alloc_dmap, flags);
fs/fuse/dax.c
501
struct iomap *iomap)
fs/fuse/dax.c
552
fuse_fill_iomap(inode, pos, length, iomap, dmap, flags);
fs/fuse/dax.c
562
unsigned int flags, struct iomap *iomap,
fs/fuse/dax.c
563
struct iomap *srcmap)
fs/fuse/dax.c
576
iomap->offset = pos;
fs/fuse/dax.c
577
iomap->flags = 0;
fs/fuse/dax.c
578
iomap->bdev = NULL;
fs/fuse/dax.c
579
iomap->dax_dev = fc->dax->dev;
fs/fuse/dax.c
607
flags, iomap);
fs/fuse/dax.c
609
fuse_fill_iomap(inode, pos, length, iomap, dmap, flags);
fs/fuse/dax.c
621
iomap);
fs/fuse/dax.c
629
fuse_fill_iomap_hole(iomap, length);
fs/fuse/dax.c
631
__func__, pos, length, iomap->length);
fs/fuse/dax.c
637
struct iomap *iomap)
fs/fuse/dax.c
639
struct fuse_dax_mapping *dmap = iomap->private;
fs/fuse/file.c
2282
.iomap.type = IOMAP_MAPPED,
fs/fuse/file.c
2304
.iomap.type = IOMAP_MAPPED,
fs/fuse/file.c
850
unsigned int flags, struct iomap *iomap,
fs/fuse/file.c
851
struct iomap *srcmap)
fs/fuse/file.c
853
iomap->type = IOMAP_MAPPED;
fs/fuse/file.c
854
iomap->length = length;
fs/fuse/file.c
855
iomap->offset = offset;
fs/gfs2/bmap.c
1016
struct iomap *iomap,
fs/gfs2/bmap.c
1027
if (unstuff || iomap->type == IOMAP_HOLE) {
fs/gfs2/bmap.c
1033
gfs2_write_calc_reserv(ip, iomap->length, &data_blocks,
fs/gfs2/bmap.c
1054
iomap->length >> inode->i_blkbits);
fs/gfs2/bmap.c
1063
ret = __gfs2_iomap_get(inode, iomap->offset,
fs/gfs2/bmap.c
1064
iomap->length, flags, iomap, mp);
fs/gfs2/bmap.c
1069
if (iomap->type == IOMAP_HOLE) {
fs/gfs2/bmap.c
1070
ret = __gfs2_iomap_alloc(inode, iomap, mp);
fs/gfs2/bmap.c
1074
punch_hole(ip, iomap->offset, iomap->length);
fs/gfs2/bmap.c
1098
unsigned flags, struct iomap *iomap,
fs/gfs2/bmap.c
1099
struct iomap *srcmap)
fs/gfs2/bmap.c
1106
iomap->flags |= IOMAP_F_BUFFER_HEAD;
fs/gfs2/bmap.c
1109
ret = __gfs2_iomap_get(inode, pos, length, flags, iomap, &mp);
fs/gfs2/bmap.c
1120
if (iomap->type != IOMAP_MAPPED)
fs/gfs2/bmap.c
1126
if (iomap->type == IOMAP_HOLE)
fs/gfs2/bmap.c
1133
ret = gfs2_iomap_begin_write(inode, pos, length, flags, iomap, &mp);
fs/gfs2/bmap.c
1138
if (iomap->type == IOMAP_INLINE) {
fs/gfs2/bmap.c
1139
iomap->private = metapath_dibh(&mp);
fs/gfs2/bmap.c
1140
get_bh(iomap->private);
fs/gfs2/bmap.c
1145
trace_gfs2_iomap_end(ip, iomap, ret);
fs/gfs2/bmap.c
1150
ssize_t written, unsigned flags, struct iomap *iomap)
fs/gfs2/bmap.c
1155
if (iomap->private)
fs/gfs2/bmap.c
1156
brelse(iomap->private);
fs/gfs2/bmap.c
1164
if (iomap->type == IOMAP_HOLE)
fs/gfs2/bmap.c
1182
if (length != written && (iomap->flags & IOMAP_F_NEW)) {
fs/gfs2/bmap.c
1185
loff_t hend = iomap->offset + iomap->length;
fs/gfs2/bmap.c
1196
if (iomap->flags & IOMAP_F_SIZE_CHANGED)
fs/gfs2/bmap.c
1234
struct iomap iomap = { };
fs/gfs2/bmap.c
1243
ret = gfs2_iomap_get(inode, pos, length, &iomap);
fs/gfs2/bmap.c
1245
ret = gfs2_iomap_alloc(inode, pos, length, &iomap);
fs/gfs2/bmap.c
1249
if (iomap.length > bh_map->b_size) {
fs/gfs2/bmap.c
1250
iomap.length = bh_map->b_size;
fs/gfs2/bmap.c
1251
iomap.flags &= ~IOMAP_F_GFS2_BOUNDARY;
fs/gfs2/bmap.c
1253
if (iomap.addr != IOMAP_NULL_ADDR)
fs/gfs2/bmap.c
1254
map_bh(bh_map, inode->i_sb, iomap.addr >> inode->i_blkbits);
fs/gfs2/bmap.c
1255
bh_map->b_size = iomap.length;
fs/gfs2/bmap.c
1256
if (iomap.flags & IOMAP_F_GFS2_BOUNDARY)
fs/gfs2/bmap.c
1258
if (iomap.flags & IOMAP_F_NEW)
fs/gfs2/bmap.c
1270
struct iomap iomap = { };
fs/gfs2/bmap.c
1275
&iomap);
fs/gfs2/bmap.c
1278
if (iomap.type != IOMAP_MAPPED)
fs/gfs2/bmap.c
1280
*dblock = iomap.addr >> blkbits;
fs/gfs2/bmap.c
1281
len = iomap.length >> blkbits;
fs/gfs2/bmap.c
1291
struct iomap iomap = { };
fs/gfs2/bmap.c
1296
&iomap);
fs/gfs2/bmap.c
1299
if (iomap.type != IOMAP_MAPPED)
fs/gfs2/bmap.c
1301
*dblock = iomap.addr >> blkbits;
fs/gfs2/bmap.c
1302
len = iomap.length >> blkbits;
fs/gfs2/bmap.c
1305
*new = iomap.flags & IOMAP_F_NEW;
fs/gfs2/bmap.c
1426
struct iomap *iomap)
fs/gfs2/bmap.c
1431
ret = __gfs2_iomap_get(inode, pos, length, 0, iomap, &mp);
fs/gfs2/bmap.c
1437
struct iomap *iomap)
fs/gfs2/bmap.c
1442
ret = __gfs2_iomap_get(inode, pos, length, IOMAP_WRITE, iomap, &mp);
fs/gfs2/bmap.c
1443
if (!ret && iomap->type == IOMAP_HOLE)
fs/gfs2/bmap.c
1444
ret = __gfs2_iomap_alloc(inode, iomap, &mp);
fs/gfs2/bmap.c
2492
if (offset < wpc->iomap.offset ||
fs/gfs2/bmap.c
2493
offset >= wpc->iomap.offset + wpc->iomap.length) {
fs/gfs2/bmap.c
2496
memset(&wpc->iomap, 0, sizeof(wpc->iomap));
fs/gfs2/bmap.c
2497
ret = gfs2_iomap_get(wpc->inode, offset, INT_MAX, &wpc->iomap);
fs/gfs2/bmap.c
585
struct metapath *mp, struct iomap *iomap)
fs/gfs2/bmap.c
600
iomap->length = hole_size << inode->i_blkbits;
fs/gfs2/bmap.c
661
static int __gfs2_iomap_alloc(struct inode *inode, struct iomap *iomap,
fs/gfs2/bmap.c
669
size_t dblks = iomap->length >> inode->i_blkbits;
fs/gfs2/bmap.c
765
iomap->addr = bn << inode->i_blkbits;
fs/gfs2/bmap.c
766
iomap->flags |= IOMAP_F_MERGED | IOMAP_F_NEW;
fs/gfs2/bmap.c
771
} while (iomap->addr == IOMAP_NULL_ADDR);
fs/gfs2/bmap.c
773
iomap->type = IOMAP_MAPPED;
fs/gfs2/bmap.c
774
iomap->length = (u64)dblks << inode->i_blkbits;
fs/gfs2/bmap.c
841
unsigned flags, struct iomap *iomap,
fs/gfs2/bmap.c
872
iomap->length = max_size;
fs/gfs2/bmap.c
879
iomap->offset = pos;
fs/gfs2/bmap.c
880
iomap->length = length;
fs/gfs2/bmap.c
884
iomap->length = size;
fs/gfs2/bmap.c
886
iomap->addr = (ip->i_no_addr << inode->i_blkbits) +
fs/gfs2/bmap.c
888
iomap->type = IOMAP_INLINE;
fs/gfs2/bmap.c
889
iomap->inline_data = dibh->b_data + sizeof(struct gfs2_dinode);
fs/gfs2/bmap.c
895
iomap->offset = lblock << inode->i_blkbits;
fs/gfs2/bmap.c
898
iomap->length = len << inode->i_blkbits;
fs/gfs2/bmap.c
921
iomap->addr = be64_to_cpu(*ptr) << inode->i_blkbits;
fs/gfs2/bmap.c
922
iomap->length = len << inode->i_blkbits;
fs/gfs2/bmap.c
923
iomap->type = IOMAP_MAPPED;
fs/gfs2/bmap.c
924
iomap->flags |= IOMAP_F_MERGED;
fs/gfs2/bmap.c
926
iomap->flags |= IOMAP_F_GFS2_BOUNDARY;
fs/gfs2/bmap.c
929
iomap->bdev = inode->i_sb->s_bdev;
fs/gfs2/bmap.c
939
ret = gfs2_hole_size(inode, lblock, len, mp, iomap);
fs/gfs2/bmap.c
941
iomap->length = size - iomap->offset;
fs/gfs2/bmap.c
950
if (alloc_size < iomap->length)
fs/gfs2/bmap.c
951
iomap->length = alloc_size;
fs/gfs2/bmap.c
954
ret = gfs2_hole_size(inode, lblock, len, mp, iomap);
fs/gfs2/bmap.c
957
iomap->addr = IOMAP_NULL_ADDR;
fs/gfs2/bmap.c
958
iomap->type = IOMAP_HOLE;
fs/gfs2/bmap.h
54
struct iomap *iomap);
fs/gfs2/bmap.h
56
struct iomap *iomap);
fs/gfs2/file.c
1198
struct iomap iomap = { };
fs/gfs2/file.c
1200
error = gfs2_iomap_alloc(inode, offset, end - offset, &iomap);
fs/gfs2/file.c
1203
offset = iomap.offset + iomap.length;
fs/gfs2/file.c
1204
if (!(iomap.flags & IOMAP_F_NEW))
fs/gfs2/file.c
1206
error = sb_issue_zeroout(sb, iomap.addr >> inode->i_blkbits,
fs/gfs2/file.c
1207
iomap.length >> inode->i_blkbits,
fs/gfs2/file.c
394
struct iomap iomap = { };
fs/gfs2/file.c
396
if (gfs2_iomap_alloc(folio->mapping->host, pos, length, &iomap))
fs/gfs2/file.c
399
if (length < iomap.length)
fs/gfs2/file.c
400
iomap.length = length;
fs/gfs2/file.c
401
length -= iomap.length;
fs/gfs2/file.c
402
pos += iomap.length;
fs/gfs2/quota.c
400
struct iomap iomap = { };
fs/gfs2/quota.c
416
i_blocksize(inode), &iomap);
fs/gfs2/quota.c
420
if (iomap.type != IOMAP_MAPPED)
fs/gfs2/quota.c
423
error = gfs2_meta_read(ip->i_gl, iomap.addr >> inode->i_blkbits,
fs/gfs2/trace_gfs2.h
513
TP_PROTO(const struct gfs2_inode *ip, struct iomap *iomap, int ret),
fs/gfs2/trace_gfs2.h
515
TP_ARGS(ip, iomap, ret),
fs/gfs2/trace_gfs2.h
531
__entry->offset = iomap->offset;
fs/gfs2/trace_gfs2.h
532
__entry->length = iomap->length;
fs/gfs2/trace_gfs2.h
533
__entry->pblock = iomap->addr == IOMAP_NULL_ADDR ? 0 :
fs/gfs2/trace_gfs2.h
534
(iomap->addr >> ip->i_inode.i_blkbits);
fs/gfs2/trace_gfs2.h
535
__entry->flags = iomap->flags;
fs/gfs2/trace_gfs2.h
536
__entry->type = iomap->type;
fs/hpfs/file.c
123
unsigned flags, struct iomap *iomap, struct iomap *srcmap)
fs/hpfs/file.c
133
iomap->bdev = inode->i_sb->s_bdev;
fs/hpfs/file.c
134
iomap->offset = offset;
fs/hpfs/file.c
145
iomap->type = IOMAP_MAPPED;
fs/hpfs/file.c
146
iomap->flags = IOMAP_F_MERGED;
fs/hpfs/file.c
147
iomap->addr = (u64)s << blkbits;
fs/hpfs/file.c
148
iomap->length = (u64)n_secs << blkbits;
fs/hpfs/file.c
150
iomap->type = IOMAP_HOLE;
fs/hpfs/file.c
151
iomap->addr = IOMAP_NULL_ADDR;
fs/hpfs/file.c
152
iomap->length = 1 << blkbits;
fs/internal.h
10
struct iomap;
fs/internal.h
37
get_block_t *get_block, const struct iomap *iomap);
fs/iomap/bio.c
101
bio = bio_alloc(iomap->bdev, bio_max_segs(nr_vecs), REQ_OP_READ,
fs/iomap/bio.c
109
bio = bio_alloc(iomap->bdev, 1, REQ_OP_READ, orig_gfp);
fs/iomap/bio.c
129
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/iomap/bio.c
82
const struct iomap *iomap = &iter->iomap;
fs/iomap/bio.c
89
sector = iomap_sector(iomap, pos);
fs/iomap/buffered-io.c
1057
const struct iomap *iomap = &iter->iomap;
fs/iomap/buffered-io.c
1061
BUG_ON(!iomap_inline_data_valid(iomap));
fs/iomap/buffered-io.c
1063
if (WARN_ON_ONCE(!iomap->inline_data))
fs/iomap/buffered-io.c
1068
memcpy(iomap_inline_data(iomap, pos), addr, copied);
fs/iomap/buffered-io.c
1082
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/iomap/buffered-io.c
1150
if (iter->iomap.flags & IOMAP_F_STALE)
fs/iomap/buffered-io.c
1172
iter->iomap.flags |= IOMAP_F_SIZE_CHANGED;
fs/iomap/buffered-io.c
1237
struct iomap *iomap, iomap_punch_t punch)
fs/iomap/buffered-io.c
1260
1 << blkbits, iomap);
fs/iomap/buffered-io.c
1267
struct iomap *iomap, iomap_punch_t punch)
fs/iomap/buffered-io.c
1275
iomap);
fs/iomap/buffered-io.c
1280
iomap, punch);
fs/iomap/buffered-io.c
1308
struct iomap *iomap, iomap_punch_t punch)
fs/iomap/buffered-io.c
1323
start_byte, end_byte, iomap, punch);
fs/iomap/buffered-io.c
1392
loff_t end_byte, unsigned flags, struct iomap *iomap,
fs/iomap/buffered-io.c
1446
data_end, iomap, punch);
fs/iomap/buffered-io.c
1454
iomap);
fs/iomap/buffered-io.c
1461
struct iomap *iomap = &iter->iomap;
fs/iomap/buffered-io.c
1478
if (iomap->flags & IOMAP_F_STALE)
fs/iomap/buffered-io.c
1531
i->iomap.flags |= IOMAP_F_STALE;
fs/iomap/buffered-io.c
1551
if (iter->iomap.flags & IOMAP_F_STALE)
fs/iomap/buffered-io.c
1648
const struct iomap *srcmap = iomap_iter_srcmap(&iter);
fs/iomap/buffered-io.c
1650
if (WARN_ON_ONCE((iter.iomap.flags & IOMAP_F_FOLIO_BATCH) &&
fs/iomap/buffered-io.c
1654
if (!(iter.iomap.flags & IOMAP_F_FOLIO_BATCH) &&
fs/iomap/buffered-io.c
1697
if (iter->iomap.flags & IOMAP_F_BUFFER_HEAD) {
fs/iomap/buffered-io.c
1699
&iter->iomap);
fs/iomap/buffered-io.c
1793
if (wpc->iomap.type != IOMAP_HOLE)
fs/iomap/buffered-io.c
354
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/iomap/buffered-io.c
373
const struct iomap *iomap = iomap_iter_srcmap(iter);
fs/iomap/buffered-io.c
374
size_t size = i_size_read(iter->inode) - iomap->offset;
fs/iomap/buffered-io.c
375
size_t offset = offset_in_folio(folio, iomap->offset);
fs/iomap/buffered-io.c
377
if (WARN_ON_ONCE(!iomap->inline_data))
fs/iomap/buffered-io.c
383
if (WARN_ON_ONCE(size > iomap->length)) {
fs/iomap/buffered-io.c
385
iomap->offset, size, -EIO, GFP_NOFS);
fs/iomap/buffered-io.c
391
folio_fill_tail(folio, offset, iomap->inline_data, size);
fs/iomap/buffered-io.c
512
const struct iomap *iomap = &iter->iomap;
fs/iomap/buffered-io.c
522
if (iomap->type == IOMAP_INLINE) {
fs/iomap/buffered-io.c
871
if (iter->iomap.flags & IOMAP_F_FOLIO_BATCH) {
fs/iomap/buffered-io.c
884
iter->iomap.flags |= IOMAP_F_STALE;
fs/iomap/buffered-io.c
947
const struct iomap *srcmap = iomap_iter_srcmap(iter);
fs/iomap/buffered-io.c
969
WARN_ON_ONCE(!(iter->iomap.flags & IOMAP_F_FOLIO_BATCH));
fs/iomap/buffered-io.c
985
&iter->iomap);
fs/iomap/buffered-io.c
987
iter->iomap.flags |= IOMAP_F_STALE;
fs/iomap/direct-io.c
316
bio->bi_iter.bi_sector = iomap_sector(&iter->iomap, pos);
fs/iomap/direct-io.c
347
bio->bi_iter.bi_sector = iomap_sector(&iter->iomap, pos);
fs/iomap/direct-io.c
392
const struct iomap *iomap = &iter->iomap;
fs/iomap/direct-io.c
411
alignment = bdev_logical_block_size(iomap->bdev);
fs/iomap/direct-io.c
419
switch (iomap->type) {
fs/iomap/direct-io.c
436
if (iomap->flags & IOMAP_F_ATOMIC_BIO) {
fs/iomap/direct-io.c
447
if (iomap->flags & IOMAP_F_SHARED) {
fs/iomap/direct-io.c
456
if (iomap->flags & IOMAP_F_NEW) {
fs/iomap/direct-io.c
473
!(iomap->flags & IOMAP_F_DIRTY) &&
fs/iomap/direct-io.c
474
(!bdev_write_cache(iomap->bdev) ||
fs/iomap/direct-io.c
475
bdev_fua(iomap->bdev)))
fs/iomap/direct-io.c
57
return bio_alloc_bioset(iter->iomap.bdev, nr_vecs, opf,
fs/iomap/direct-io.c
585
const struct iomap *iomap = &iomi->iomap;
fs/iomap/direct-io.c
587
void *inline_data = iomap_inline_data(iomap, iomi->pos);
fs/iomap/direct-io.c
59
return bio_alloc(iter->iomap.bdev, nr_vecs, opf, GFP_KERNEL);
fs/iomap/direct-io.c
595
if (WARN_ON_ONCE(!iomap_inline_data_valid(iomap)))
fs/iomap/direct-io.c
602
memset(iomap_inline_data(iomap, size), 0, pos - size);
fs/iomap/direct-io.c
620
switch (iter->iomap.type) {
fs/iomap/direct-io.c
78
WARN_ON_ONCE(iter->iomap.flags & IOMAP_F_ANON_WRITE);
fs/iomap/fiemap.c
10
const struct iomap *iomap, u32 flags)
fs/iomap/fiemap.c
111
if (iter.iomap.type == IOMAP_MAPPED)
fs/iomap/fiemap.c
112
bno = iomap_sector(&iter.iomap, iter.pos) >> blkshift;
fs/iomap/fiemap.c
12
switch (iomap->type) {
fs/iomap/fiemap.c
29
if (iomap->flags & IOMAP_F_MERGED)
fs/iomap/fiemap.c
31
if (iomap->flags & IOMAP_F_SHARED)
fs/iomap/fiemap.c
34
return fiemap_fill_next_extent(fi, iomap->offset,
fs/iomap/fiemap.c
35
iomap->addr != IOMAP_NULL_ADDR ? iomap->addr : 0,
fs/iomap/fiemap.c
36
iomap->length, flags);
fs/iomap/fiemap.c
40
struct fiemap_extent_info *fi, struct iomap *prev)
fs/iomap/fiemap.c
44
if (iter->iomap.type == IOMAP_HOLE)
fs/iomap/fiemap.c
48
*prev = iter->iomap;
fs/iomap/fiemap.c
67
struct iomap prev = {
fs/iomap/ioend.c
138
if (WARN_ON_ONCE(wpc->iomap.flags & IOMAP_F_ANON_WRITE))
fs/iomap/ioend.c
157
bio = bio_alloc_bioset(wpc->iomap.bdev, BIO_MAX_VECS,
fs/iomap/ioend.c
160
bio->bi_iter.bi_sector = iomap_sector(&wpc->iomap, pos);
fs/iomap/ioend.c
179
if (!(wpc->iomap.flags & IOMAP_F_ANON_WRITE) &&
fs/iomap/ioend.c
180
iomap_sector(&wpc->iomap, pos) != bio_end_sector(&ioend->io_bio))
fs/iomap/ioend.c
210
wpc->iomap.offset + wpc->iomap.length - pos);
fs/iomap/ioend.c
213
trace_iomap_add_to_ioend(wpc->inode, pos, dirty_len, &wpc->iomap);
fs/iomap/ioend.c
217
switch (wpc->iomap.type) {
fs/iomap/ioend.c
230
if (wpc->iomap.flags & IOMAP_F_SHARED)
fs/iomap/ioend.c
234
if (pos == wpc->iomap.offset && (wpc->iomap.flags & IOMAP_F_BOUNDARY))
fs/iomap/iter.c
11
if (iter->iomap.flags & IOMAP_F_FOLIO_BATCH) {
fs/iomap/iter.c
111
&iter->iomap, &iter->srcmap);
fs/iomap/iter.c
14
iter->iomap.flags &= ~IOMAP_F_FOLIO_BATCH;
fs/iomap/iter.c
18
memset(&iter->iomap, 0, sizeof(iter->iomap));
fs/iomap/iter.c
34
WARN_ON_ONCE(iter->iomap.offset > iter->pos);
fs/iomap/iter.c
35
WARN_ON_ONCE(iter->iomap.length == 0);
fs/iomap/iter.c
36
WARN_ON_ONCE(iter->iomap.offset + iter->iomap.length <= iter->pos);
fs/iomap/iter.c
37
WARN_ON_ONCE(iter->iomap.flags & IOMAP_F_STALE);
fs/iomap/iter.c
41
trace_iomap_iter_dstmap(iter->inode, &iter->iomap);
fs/iomap/iter.c
63
bool stale = iter->iomap.flags & IOMAP_F_STALE;
fs/iomap/iter.c
70
if (!iter->iomap.length)
fs/iomap/iter.c
84
advanced, iter->flags, &iter->iomap);
fs/iomap/seek.c
14
switch (iter->iomap.type) {
fs/iomap/seek.c
60
switch (iter->iomap.type) {
fs/iomap/swapfile.c
110
if (iomap->flags & IOMAP_F_DIRTY)
fs/iomap/swapfile.c
112
if (iomap->flags & IOMAP_F_SHARED)
fs/iomap/swapfile.c
116
if (iomap->bdev != isi->sis->bdev)
fs/iomap/swapfile.c
119
if (isi->iomap.length == 0) {
fs/iomap/swapfile.c
12
struct iomap iomap; /* accumulated iomap */
fs/iomap/swapfile.c
121
memcpy(&isi->iomap, iomap, sizeof(isi->iomap));
fs/iomap/swapfile.c
122
} else if (isi->iomap.addr + isi->iomap.length == iomap->addr) {
fs/iomap/swapfile.c
124
isi->iomap.length += iomap->length;
fs/iomap/swapfile.c
130
memcpy(&isi->iomap, iomap, sizeof(isi->iomap));
fs/iomap/swapfile.c
167
iter.status = iomap_swapfile_iter(&iter, &iter.iomap, &isi);
fs/iomap/swapfile.c
171
if (isi.iomap.length) {
fs/iomap/swapfile.c
29
struct iomap *iomap = &isi->iomap;
fs/iomap/swapfile.c
45
first_ppage = ALIGN(iomap->addr, PAGE_SIZE) >> PAGE_SHIFT;
fs/iomap/swapfile.c
46
next_ppage = ALIGN_DOWN(iomap->addr + iomap->length, PAGE_SIZE) >>
fs/iomap/swapfile.c
61
if (iomap->offset == 0)
fs/iomap/swapfile.c
95
struct iomap *iomap, struct iomap_swapfile_info *isi)
fs/iomap/swapfile.c
97
switch (iomap->type) {
fs/iomap/trace.h
131
TP_PROTO(struct inode *inode, struct iomap *iomap),
fs/iomap/trace.h
132
TP_ARGS(inode, iomap),
fs/iomap/trace.h
146
__entry->addr = iomap->addr;
fs/iomap/trace.h
147
__entry->offset = iomap->offset;
fs/iomap/trace.h
148
__entry->length = iomap->length;
fs/iomap/trace.h
149
__entry->type = iomap->type;
fs/iomap/trace.h
150
__entry->flags = iomap->flags;
fs/iomap/trace.h
151
__entry->bdev = iomap->bdev ? iomap->bdev->bd_dev : 0;
fs/iomap/trace.h
169
TP_PROTO(struct inode *inode, struct iomap *iomap), \
fs/iomap/trace.h
170
TP_ARGS(inode, iomap))
fs/iomap/trace.h
176
struct iomap *iomap),
fs/iomap/trace.h
177
TP_ARGS(inode, pos, dirty_len, iomap),
fs/iomap/trace.h
18
#define TRACE_SYSTEM iomap
fs/iomap/trace.h
195
__entry->addr = iomap->addr;
fs/iomap/trace.h
196
__entry->offset = iomap->offset;
fs/iomap/trace.h
197
__entry->length = iomap->length;
fs/iomap/trace.h
198
__entry->type = iomap->type;
fs/iomap/trace.h
199
__entry->flags = iomap->flags;
fs/iomap/trace.h
200
__entry->bdev = iomap->bdev ? iomap->bdev->bd_dev : 0;
fs/nfsd/blocklayout.c
180
struct iomap *iomaps, int nr_iomaps)
fs/nfsd/blocklayout.c
240
struct iomap *iomaps;
fs/nfsd/blocklayout.c
31
struct iomap iomap;
fs/nfsd/blocklayout.c
35
error = sb->s_export_op->map_blocks(inode, offset, length, &iomap,
fs/nfsd/blocklayout.c
382
struct iomap *iomaps;
fs/nfsd/blocklayout.c
43
switch (iomap.type) {
fs/nfsd/blocklayout.c
49
bex->soff = iomap.addr;
fs/nfsd/blocklayout.c
62
bex->soff = iomap.addr;
fs/nfsd/blocklayout.c
74
WARN(1, "pnfsd: filesystem returned %d extent\n", iomap.type);
fs/nfsd/blocklayout.c
82
bex->foff = iomap.offset;
fs/nfsd/blocklayout.c
83
bex->len = iomap.length;
fs/nfsd/blocklayoutxdr.c
152
nfsd4_block_decode_layoutupdate(struct xdr_stream *xdr, struct iomap **iomapp,
fs/nfsd/blocklayoutxdr.c
155
struct iomap *iomaps;
fs/nfsd/blocklayoutxdr.c
246
nfsd4_scsi_decode_layoutupdate(struct xdr_stream *xdr, struct iomap **iomapp,
fs/nfsd/blocklayoutxdr.c
249
struct iomap *iomaps;
fs/nfsd/blocklayoutxdr.h
72
struct iomap **iomapp, int *nr_iomapsp, u32 block_size);
fs/nfsd/blocklayoutxdr.h
74
struct iomap **iomapp, int *nr_iomapsp, u32 block_size);
fs/nfsd/blocklayoutxdr.h
8
struct iomap;
fs/ntfs3/inode.c
1006
struct iomap *iomap = &wpc->iomap;
fs/ntfs3/inode.c
1008
if (iomap->offset + iomap->length <= offset || offset < iomap->offset) {
fs/ntfs3/inode.c
1021
IOMAP_WRITE, iomap, WB_NO_DA);
fs/ntfs3/inode.c
616
const struct iomap *iomap = &iter->iomap;
fs/ntfs3/inode.c
623
sector = iomap_sector(iomap, pos);
fs/ntfs3/inode.c
635
bio = bio_alloc(iomap->bdev, bio_max_segs(nr_vecs), REQ_OP_READ,
fs/ntfs3/inode.c
643
bio = bio_alloc(iomap->bdev, 1, REQ_OP_READ, orig_gfp);
fs/ntfs3/inode.c
760
#define WB_NO_DA (struct iomap *)1
fs/ntfs3/inode.c
774
unsigned int flags, struct iomap *iomap,
fs/ntfs3/inode.c
775
struct iomap *srcmap)
fs/ntfs3/inode.c
839
iomap->flags = new_clst ? IOMAP_F_NEW : 0;
fs/ntfs3/inode.c
851
iomap->private = iomap->inline_data = res;
fs/ntfs3/inode.c
852
iomap->type = IOMAP_INLINE;
fs/ntfs3/inode.c
853
iomap->offset = 0;
fs/ntfs3/inode.c
854
iomap->length = clen; /* resident size in bytes. */
fs/ntfs3/inode.c
863
iomap->bdev = inode->i_sb->s_bdev;
fs/ntfs3/inode.c
864
iomap->offset = offset;
fs/ntfs3/inode.c
865
iomap->length = ((loff_t)clen << cluster_bits) - off;
fs/ntfs3/inode.c
873
iomap->type = IOMAP_DELALLOC;
fs/ntfs3/inode.c
874
iomap->addr = IOMAP_NULL_ADDR;
fs/ntfs3/inode.c
878
iomap->addr = ((loff_t)lcn << cluster_bits) + off;
fs/ntfs3/inode.c
879
if (length && iomap->length > length)
fs/ntfs3/inode.c
880
iomap->length = length;
fs/ntfs3/inode.c
882
endbyte = offset + iomap->length;
fs/ntfs3/inode.c
885
iomap->addr = IOMAP_NULL_ADDR;
fs/ntfs3/inode.c
886
iomap->type = IOMAP_HOLE;
fs/ntfs3/inode.c
892
iomap->type = IOMAP_MAPPED;
fs/ntfs3/inode.c
894
iomap->type = IOMAP_MAPPED;
fs/ntfs3/inode.c
896
iomap->length = ni->i_valid - offset;
fs/ntfs3/inode.c
898
iomap->type = IOMAP_MAPPED;
fs/ntfs3/inode.c
900
iomap->type = IOMAP_UNWRITTEN;
fs/ntfs3/inode.c
905
(iomap->type == IOMAP_MAPPED || iomap->type == IOMAP_DELALLOC)) {
fs/ntfs3/inode.c
914
if (iomap->length > tail)
fs/ntfs3/inode.c
915
iomap->length = tail;
fs/ntfs3/inode.c
923
struct iomap *iomap)
fs/ntfs3/inode.c
930
if (iomap->type == IOMAP_INLINE) {
fs/ntfs3/inode.c
950
iomap_inline_data(iomap, pos), written);
fs/ntfs3/inode.c
960
(iomap->type == IOMAP_MAPPED || iomap->type == IOMAP_DELALLOC)) {
fs/ntfs3/inode.c
967
if (iomap->type == IOMAP_INLINE) {
fs/ntfs3/inode.c
968
kfree(iomap->private);
fs/ntfs3/inode.c
969
iomap->private = NULL;
fs/xfs/libxfs/xfs_bmap.c
4359
struct iomap *iomap,
fs/xfs/libxfs/xfs_bmap.c
4407
xfs_bmbt_to_iomap(ip, iomap, &bma.got, 0, flags,
fs/xfs/libxfs/xfs_bmap.c
4455
xfs_bmbt_to_iomap(ip, iomap, &bma.got, 0, flags,
fs/xfs/libxfs/xfs_bmap.c
4491
struct iomap *iomap,
fs/xfs/libxfs/xfs_bmap.c
4504
iomap, seq);
fs/xfs/libxfs/xfs_bmap.c
4507
} while (iomap->offset + iomap->length <= offset);
fs/xfs/libxfs/xfs_bmap.h
227
xfs_off_t offset, struct iomap *iomap, unsigned int *seq);
fs/xfs/xfs_aops.c
291
if (offset < wpc->iomap.offset ||
fs/xfs/xfs_aops.c
292
offset >= wpc->iomap.offset + wpc->iomap.length)
fs/xfs/xfs_aops.c
299
if (wpc->iomap.flags & IOMAP_F_SHARED)
fs/xfs/xfs_aops.c
310
trace_xfs_wb_data_iomap_invalid(ip, &wpc->iomap,
fs/xfs/xfs_aops.c
316
trace_xfs_wb_cow_iomap_invalid(ip, &wpc->iomap,
fs/xfs/xfs_aops.c
434
xfs_bmbt_to_iomap(ip, &wpc->iomap, &imap, 0, 0, XFS_WPC(wpc)->data_seq);
fs/xfs/xfs_aops.c
450
&wpc->iomap, seq);
fs/xfs/xfs_aops.c
473
if (cow_offset < wpc->iomap.offset + wpc->iomap.length)
fs/xfs/xfs_aops.c
474
wpc->iomap.length = cow_offset - wpc->iomap.offset;
fs/xfs/xfs_aops.c
477
ASSERT(wpc->iomap.offset <= offset);
fs/xfs/xfs_aops.c
478
ASSERT(wpc->iomap.offset + wpc->iomap.length > offset);
fs/xfs/xfs_aops.c
610
xfs_bmbt_to_iomap(ip, &wpc->iomap, &imap, 0, 0, 0);
fs/xfs/xfs_aops.c
622
wpc->iomap.type = IOMAP_MAPPED;
fs/xfs/xfs_aops.c
623
wpc->iomap.flags = IOMAP_F_DIRTY;
fs/xfs/xfs_aops.c
624
wpc->iomap.bdev = mp->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_aops.c
625
wpc->iomap.offset = offset;
fs/xfs/xfs_aops.c
626
wpc->iomap.length = XFS_FSB_TO_B(mp, count_fsb);
fs/xfs/xfs_aops.c
627
wpc->iomap.flags = IOMAP_F_ANON_WRITE;
fs/xfs/xfs_aops.c
629
trace_xfs_zoned_map_blocks(ip, offset, wpc->iomap.length);
fs/xfs/xfs_iomap.c
1015
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags,
fs/xfs/xfs_iomap.c
102
struct iomap *iomap,
fs/xfs/xfs_iomap.c
1029
return xfs_bmbt_to_iomap(ip, iomap, &cmap, flags, IOMAP_F_SHARED, seq);
fs/xfs/xfs_iomap.c
1053
struct iomap *iomap,
fs/xfs/xfs_iomap.c
1054
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
1080
iomap->type = IOMAP_MAPPED;
fs/xfs/xfs_iomap.c
1081
iomap->flags = IOMAP_F_DIRTY;
fs/xfs/xfs_iomap.c
1082
iomap->bdev = ip->i_mount->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_iomap.c
1083
iomap->offset = offset;
fs/xfs/xfs_iomap.c
1084
iomap->length = length;
fs/xfs/xfs_iomap.c
1085
iomap->flags = IOMAP_F_ANON_WRITE;
fs/xfs/xfs_iomap.c
1123
struct iomap *iomap,
fs/xfs/xfs_iomap.c
1124
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
117
iomap->addr = IOMAP_NULL_ADDR;
fs/xfs/xfs_iomap.c
118
iomap->type = IOMAP_HOLE;
fs/xfs/xfs_iomap.c
121
iomap->addr = IOMAP_NULL_ADDR;
fs/xfs/xfs_iomap.c
122
iomap->type = IOMAP_DELALLOC;
fs/xfs/xfs_iomap.c
1254
return xfs_bmbt_to_iomap(ip, iomap, &cmap, flags, IOMAP_F_SHARED, seq);
fs/xfs/xfs_iomap.c
1258
error = xfs_bmapi_convert_delalloc(ip, XFS_COW_FORK, offset, iomap,
fs/xfs/xfs_iomap.c
126
iomap->addr = BBTOB(daddr);
fs/xfs/xfs_iomap.c
128
iomap->addr += target->bt_dax_part_off;
fs/xfs/xfs_iomap.c
1285
struct iomap *iomap)
fs/xfs/xfs_iomap.c
131
iomap->type = IOMAP_UNWRITTEN;
fs/xfs/xfs_iomap.c
133
iomap->type = IOMAP_MAPPED;
fs/xfs/xfs_iomap.c
142
iomap->flags |= IOMAP_F_BOUNDARY;
fs/xfs/xfs_iomap.c
144
iomap->offset = XFS_FSB_TO_B(mp, imap->br_startoff);
fs/xfs/xfs_iomap.c
145
iomap->length = XFS_FSB_TO_B(mp, imap->br_blockcount);
fs/xfs/xfs_iomap.c
147
iomap->dax_dev = target->bt_daxdev;
fs/xfs/xfs_iomap.c
149
iomap->bdev = target->bt_bdev;
fs/xfs/xfs_iomap.c
150
iomap->flags = iomap_flags;
fs/xfs/xfs_iomap.c
1588
struct iomap *iomap,
fs/xfs/xfs_iomap.c
1589
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
1592
container_of(iomap, struct iomap_iter, iomap);
fs/xfs/xfs_iomap.c
161
iomap->flags |= IOMAP_F_DIRTY;
fs/xfs/xfs_iomap.c
165
iomap->validity_cookie = sequence_cookie;
fs/xfs/xfs_iomap.c
1658
xfs_hole_to_iomap(ip, iomap, offset_fsb,
fs/xfs/xfs_iomap.c
172
struct iomap *iomap,
fs/xfs/xfs_iomap.c
1745
error = xfs_bmbt_to_iomap(ip, iomap, &got, flags, iomap_flags,
fs/xfs/xfs_iomap.c
1758
struct iomap *iomap,
fs/xfs/xfs_iomap.c
1759
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
1761
struct iomap_iter *iter = container_of(iomap, struct iomap_iter,
fs/xfs/xfs_iomap.c
1762
iomap);
fs/xfs/xfs_iomap.c
178
iomap->addr = IOMAP_NULL_ADDR;
fs/xfs/xfs_iomap.c
1782
count, flags, iomap, srcmap);
fs/xfs/xfs_iomap.c
1787
flags, iomap, srcmap);
fs/xfs/xfs_iomap.c
179
iomap->type = IOMAP_HOLE;
fs/xfs/xfs_iomap.c
180
iomap->offset = XFS_FSB_TO_B(ip->i_mount, offset_fsb);
fs/xfs/xfs_iomap.c
181
iomap->length = XFS_FSB_TO_B(ip->i_mount, end_fsb - offset_fsb);
fs/xfs/xfs_iomap.c
182
iomap->bdev = target->bt_bdev;
fs/xfs/xfs_iomap.c
1823
xfs_hole_to_iomap(ip, iomap, offset_fsb, imap.br_startoff);
fs/xfs/xfs_iomap.c
183
iomap->dax_dev = target->bt_daxdev;
fs/xfs/xfs_iomap.c
1996
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags, iomap_flags, seq);
fs/xfs/xfs_iomap.c
2002
iomap, NULL);
fs/xfs/xfs_iomap.c
2023
return xfs_bmbt_to_iomap(ip, iomap, &cmap, flags, iomap_flags, seq);
fs/xfs/xfs_iomap.c
2035
struct iomap *iomap)
fs/xfs/xfs_iomap.c
2038
container_of(iomap, struct iomap_iter, iomap);
fs/xfs/xfs_iomap.c
2041
(iomap->flags & IOMAP_F_SHARED) ?
fs/xfs/xfs_iomap.c
2053
struct iomap *iomap)
fs/xfs/xfs_iomap.c
2058
if (iomap->type != IOMAP_DELALLOC || !(iomap->flags & IOMAP_F_NEW))
fs/xfs/xfs_iomap.c
2079
iomap, xfs_buffered_write_delalloc_punch);
fs/xfs/xfs_iomap.c
2083
iomap, xfs_buffered_write_delalloc_punch);
fs/xfs/xfs_iomap.c
2101
struct iomap *iomap,
fs/xfs/xfs_iomap.c
2102
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
2132
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags,
fs/xfs/xfs_iomap.c
2146
struct iomap *iomap,
fs/xfs/xfs_iomap.c
2147
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
2194
error = xfs_bmbt_to_iomap(ip, iomap, &cmap, flags,
fs/xfs/xfs_iomap.c
2201
iomap->type = IOMAP_UNWRITTEN;
fs/xfs/xfs_iomap.c
2218
error = xfs_bmbt_to_iomap(ip, iomap, &imap, flags, 0, seq);
fs/xfs/xfs_iomap.c
2234
struct iomap *iomap,
fs/xfs/xfs_iomap.c
2235
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
2268
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags, IOMAP_F_XATTR, seq);
fs/xfs/xfs_iomap.c
78
const struct iomap *iomap)
fs/xfs/xfs_iomap.c
82
if (iomap->type == IOMAP_HOLE)
fs/xfs/xfs_iomap.c
85
if (iomap->validity_cookie !=
fs/xfs/xfs_iomap.c
850
struct iomap *iomap,
fs/xfs/xfs_iomap.c
851
struct iomap *srcmap)
fs/xfs/xfs_iomap.c
86
xfs_iomap_inode_sequence(ip, iomap->flags)) {
fs/xfs/xfs_iomap.c
87
trace_xfs_iomap_invalid(ip, iomap);
fs/xfs/xfs_iomap.c
984
return xfs_bmbt_to_iomap(ip, iomap, &imap, flags, iomap_flags, seq);
fs/xfs/xfs_iomap.h
23
int xfs_bmbt_to_iomap(struct xfs_inode *ip, struct iomap *iomap,
fs/xfs/xfs_pnfs.c
112
struct iomap *iomap,
fs/xfs/xfs_pnfs.c
211
error = xfs_bmbt_to_iomap(ip, iomap, &imap, 0, 0, seq);
fs/xfs/xfs_pnfs.c
258
struct iomap *maps,
fs/xfs/xfs_pnfs.h
8
struct iomap *iomap, bool write, u32 *device_generation);
fs/xfs/xfs_pnfs.h
9
int xfs_fs_commit_blocks(struct inode *inode, struct iomap *maps, int nr_maps,
fs/xfs/xfs_trace.h
4079
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap, unsigned int wpcseq, int whichfork),
fs/xfs/xfs_trace.h
4080
TP_ARGS(ip, iomap, wpcseq, whichfork),
fs/xfs/xfs_trace.h
4095
__entry->addr = iomap->addr;
fs/xfs/xfs_trace.h
4096
__entry->pos = iomap->offset;
fs/xfs/xfs_trace.h
4097
__entry->len = iomap->length;
fs/xfs/xfs_trace.h
4098
__entry->type = iomap->type;
fs/xfs/xfs_trace.h
4099
__entry->flags = iomap->flags;
fs/xfs/xfs_trace.h
4116
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap, unsigned int wpcseq, int whichfork), \
fs/xfs/xfs_trace.h
4117
TP_ARGS(ip, iomap, wpcseq, whichfork))
fs/xfs/xfs_trace.h
4122
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap),
fs/xfs/xfs_trace.h
4123
TP_ARGS(ip, iomap),
fs/xfs/xfs_trace.h
4138
__entry->addr = iomap->addr;
fs/xfs/xfs_trace.h
4139
__entry->pos = iomap->offset;
fs/xfs/xfs_trace.h
4140
__entry->len = iomap->length;
fs/xfs/xfs_trace.h
4141
__entry->validity_cookie = iomap->validity_cookie;
fs/xfs/xfs_trace.h
4142
__entry->type = iomap->type;
fs/xfs/xfs_trace.h
4143
__entry->flags = iomap->flags;
fs/xfs/xfs_trace.h
4144
__entry->inodeseq = xfs_iomap_inode_sequence(ip, iomap->flags);
fs/xfs/xfs_trace.h
4159
TP_PROTO(struct xfs_inode *ip, const struct iomap *iomap), \
fs/xfs/xfs_trace.h
4160
TP_ARGS(ip, iomap))
fs/zonefs/file.c
100
iomap->length = isize - iomap->offset;
fs/zonefs/file.c
104
trace_zonefs_iomap_begin(inode, iomap);
fs/zonefs/file.c
139
if (offset < wpc->iomap.offset ||
fs/zonefs/file.c
140
offset >= wpc->iomap.offset + wpc->iomap.length) {
fs/zonefs/file.c
145
&wpc->iomap, NULL);
fs/zonefs/file.c
29
struct iomap *iomap, struct iomap *srcmap)
fs/zonefs/file.c
41
iomap->bdev = inode->i_sb->s_bdev;
fs/zonefs/file.c
42
iomap->offset = ALIGN_DOWN(offset, sb->s_blocksize);
fs/zonefs/file.c
44
if (iomap->offset >= isize) {
fs/zonefs/file.c
45
iomap->type = IOMAP_HOLE;
fs/zonefs/file.c
46
iomap->addr = IOMAP_NULL_ADDR;
fs/zonefs/file.c
47
iomap->length = length;
fs/zonefs/file.c
49
iomap->type = IOMAP_MAPPED;
fs/zonefs/file.c
50
iomap->addr = (z->z_sector << SECTOR_SHIFT) + iomap->offset;
fs/zonefs/file.c
51
iomap->length = isize - iomap->offset;
fs/zonefs/file.c
55
trace_zonefs_iomap_begin(inode, iomap);
fs/zonefs/file.c
66
struct iomap *iomap, struct iomap *srcmap)
fs/zonefs/file.c
91
iomap->bdev = inode->i_sb->s_bdev;
fs/zonefs/file.c
92
iomap->offset = ALIGN_DOWN(offset, sb->s_blocksize);
fs/zonefs/file.c
93
iomap->addr = (z->z_sector << SECTOR_SHIFT) + iomap->offset;
fs/zonefs/file.c
95
if (iomap->offset >= isize) {
fs/zonefs/file.c
96
iomap->type = IOMAP_UNWRITTEN;
fs/zonefs/file.c
97
iomap->length = z->z_capacity - iomap->offset;
fs/zonefs/file.c
99
iomap->type = IOMAP_MAPPED;
fs/zonefs/trace.h
76
TP_PROTO(struct inode *inode, struct iomap *iomap),
fs/zonefs/trace.h
77
TP_ARGS(inode, iomap),
fs/zonefs/trace.h
88
__entry->addr = iomap->addr;
fs/zonefs/trace.h
89
__entry->offset = iomap->offset;
fs/zonefs/trace.h
90
__entry->length = iomap->length;
include/drm/ttm/ttm_resource.h
368
struct io_mapping *iomap;
include/drm/ttm/ttm_resource.h
506
struct io_mapping *iomap,
include/linux/dax.h
15
struct iomap;
include/linux/exportfs.h
11
struct iomap;
include/linux/exportfs.h
295
u64 len, struct iomap *iomap,
include/linux/exportfs.h
297
int (*commit_blocks)(struct inode *inode, struct iomap *iomaps,
include/linux/fwnode.h
182
void __iomem *(*iomap)(struct fwnode_handle *fwnode, int index);
include/linux/io-mapping.h
131
io_mapping_init_wc(struct io_mapping *iomap,
include/linux/io-mapping.h
135
iomap->iomem = ioremap_wc(base, size);
include/linux/io-mapping.h
136
if (!iomap->iomem)
include/linux/io-mapping.h
139
iomap->base = base;
include/linux/io-mapping.h
140
iomap->size = size;
include/linux/io-mapping.h
141
iomap->prot = pgprot_writecombine(PAGE_KERNEL);
include/linux/io-mapping.h
143
return iomap;
include/linux/io-mapping.h
207
struct io_mapping *iomap;
include/linux/io-mapping.h
209
iomap = kmalloc_obj(*iomap);
include/linux/io-mapping.h
210
if (!iomap)
include/linux/io-mapping.h
213
if (!io_mapping_init_wc(iomap, base, size)) {
include/linux/io-mapping.h
214
kfree(iomap);
include/linux/io-mapping.h
218
return iomap;
include/linux/io-mapping.h
222
io_mapping_free(struct io_mapping *iomap)
include/linux/io-mapping.h
224
io_mapping_fini(iomap);
include/linux/io-mapping.h
225
kfree(iomap);
include/linux/io-mapping.h
42
io_mapping_init_wc(struct io_mapping *iomap,
include/linux/io-mapping.h
51
iomap->base = base;
include/linux/io-mapping.h
52
iomap->size = size;
include/linux/io-mapping.h
53
iomap->prot = prot;
include/linux/io-mapping.h
54
return iomap;
include/linux/iomap.h
123
static inline sector_t iomap_sector(const struct iomap *iomap, loff_t pos)
include/linux/iomap.h
125
if (iomap->flags & IOMAP_F_ANON_WRITE)
include/linux/iomap.h
127
return (iomap->addr + pos - iomap->offset) >> SECTOR_SHIFT;
include/linux/iomap.h
133
static inline void *iomap_inline_data(const struct iomap *iomap, loff_t pos)
include/linux/iomap.h
135
return iomap->inline_data + pos - iomap->offset;
include/linux/iomap.h
143
static inline bool iomap_inline_data_valid(const struct iomap *iomap)
include/linux/iomap.h
145
return iomap->length <= PAGE_SIZE - offset_in_page(iomap->inline_data);
include/linux/iomap.h
174
bool (*iomap_valid)(struct inode *inode, const struct iomap *iomap);
include/linux/iomap.h
213
unsigned flags, struct iomap *iomap,
include/linux/iomap.h
214
struct iomap *srcmap);
include/linux/iomap.h
223
ssize_t written, unsigned flags, struct iomap *iomap);
include/linux/iomap.h
248
struct iomap iomap;
include/linux/iomap.h
249
struct iomap srcmap;
include/linux/iomap.h
269
u64 end = iter->iomap.offset + iter->iomap.length;
include/linux/iomap.h
304
static inline const struct iomap *iomap_iter_srcmap(const struct iomap_iter *i)
include/linux/iomap.h
308
return &i->iomap;
include/linux/iomap.h
340
return (iter->iomap.flags & IOMAP_F_SHARED) &&
include/linux/iomap.h
370
struct iomap *iomap);
include/linux/iomap.h
372
loff_t end_byte, unsigned flags, struct iomap *iomap,
include/linux/iomap.h
463
struct iomap iomap;
include/linux/libata.h
621
void __iomem * const *iomap;