Symbol: block
arch/arm/kernel/setup.c
444
int block;
arch/arm/kernel/setup.c
452
block = cpuid_feature_extract(CPUID_EXT_ISAR0, 24);
arch/arm/kernel/setup.c
453
if (block >= 2)
arch/arm/kernel/setup.c
455
if (block >= 1)
arch/arm/kernel/setup.c
459
block = cpuid_feature_extract(CPUID_EXT_MMFR0, 0);
arch/arm/kernel/setup.c
460
if (block >= 5)
arch/arm/kernel/setup.c
466
block = cpuid_feature_extract_field(isar5, 4);
arch/arm/kernel/setup.c
467
if (block >= 2)
arch/arm/kernel/setup.c
469
if (block >= 1)
arch/arm/kernel/setup.c
472
block = cpuid_feature_extract_field(isar5, 8);
arch/arm/kernel/setup.c
473
if (block >= 1)
arch/arm/kernel/setup.c
476
block = cpuid_feature_extract_field(isar5, 12);
arch/arm/kernel/setup.c
477
if (block >= 1)
arch/arm/kernel/setup.c
480
block = cpuid_feature_extract_field(isar5, 16);
arch/arm/kernel/setup.c
481
if (block >= 1)
arch/arm/kernel/setup.c
486
block = cpuid_feature_extract_field(isar6, 12);
arch/arm/kernel/setup.c
487
if (block >= 1)
arch/arm/kernel/setup.c
492
block = cpuid_feature_extract_field(pfr2, 4);
arch/arm/kernel/setup.c
493
if (block >= 1)
arch/mips/alchemy/common/usb.c
265
static inline int au1300_usb_control(int block, int enable)
arch/mips/alchemy/common/usb.c
271
switch (block) {
arch/mips/alchemy/common/usb.c
360
static inline int au1200_usb_control(int block, int enable)
arch/mips/alchemy/common/usb.c
365
switch (block) {
arch/mips/alchemy/common/usb.c
458
static inline int au1000_usb_control(int block, int enable, unsigned long rb,
arch/mips/alchemy/common/usb.c
463
switch (block) {
arch/mips/alchemy/common/usb.c
478
int alchemy_usb_control(int block, int enable)
arch/mips/alchemy/common/usb.c
488
ret = au1000_usb_control(block, enable,
arch/mips/alchemy/common/usb.c
492
ret = au1000_usb_control(block, enable,
arch/mips/alchemy/common/usb.c
496
ret = au1200_usb_control(block, enable);
arch/mips/alchemy/common/usb.c
499
ret = au1300_usb_control(block, enable);
arch/mips/alchemy/common/vss.c
23
static inline void __enable_block(int block)
arch/mips/alchemy/common/vss.c
25
void __iomem *base = (void __iomem *)VSS_ADDR(block);
arch/mips/alchemy/common/vss.c
54
static inline void __disable_block(int block)
arch/mips/alchemy/common/vss.c
56
void __iomem *base = (void __iomem *)VSS_ADDR(block);
arch/mips/alchemy/common/vss.c
70
void au1300_vss_block_control(int block, int enable)
arch/mips/alchemy/common/vss.c
80
__enable_block(block);
arch/mips/alchemy/common/vss.c
82
__disable_block(block);
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
229
cvmx_write_csr(CVMX_GMXX_RXX_INT_EN(index, block), gmx_rx_int_en.u64);
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
236
void __cvmx_interrupt_pcsx_intx_en_reg_enable(int index, int block)
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
239
cvmx_write_csr(CVMX_PCSX_INTX_REG(index, block),
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
240
cvmx_read_csr(CVMX_PCSX_INTX_REG(index, block)));
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
272
cvmx_write_csr(CVMX_PCSX_INTX_EN_REG(index, block), pcs_int_en_reg.u64);
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
53
void __cvmx_interrupt_gmxx_rxx_int_en_enable(int index, int block)
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
56
cvmx_write_csr(CVMX_GMXX_RXX_INT_REG(index, block),
arch/mips/cavium-octeon/executive/cvmx-interrupt-decodes.c
57
cvmx_read_csr(CVMX_GMXX_RXX_INT_REG(index, block)));
arch/mips/cavium-octeon/executive/cvmx-interrupt-rsl.c
42
void __cvmx_interrupt_gmxx_rxx_int_en_enable(int index, int block);
arch/mips/cavium-octeon/executive/cvmx-interrupt-rsl.c
50
void __cvmx_interrupt_asxx_enable(int block)
arch/mips/cavium-octeon/executive/cvmx-interrupt-rsl.c
65
csr.u64 = cvmx_read_csr(CVMX_ASXX_INT_EN(block));
arch/mips/cavium-octeon/executive/cvmx-interrupt-rsl.c
69
cvmx_write_csr(CVMX_ASXX_INT_EN(block), csr.u64);
arch/mips/cavium-octeon/octeon-irq.c
2612
int block = intsn >> 12;
arch/mips/cavium-octeon/octeon-irq.c
2615
domain = ciu3_info->domain[block];
arch/mips/cavium-octeon/octeon-irq.c
2616
if (ciu3_info->intsn2hw[block])
arch/mips/cavium-octeon/octeon-irq.c
2617
hw = ciu3_info->intsn2hw[block](domain, intsn);
arch/mips/cavium-octeon/octeon-irq.c
2996
struct irq_domain *octeon_irq_get_block_domain(int node, uint8_t block)
arch/mips/cavium-octeon/octeon-irq.c
3001
return ciu3_info->domain[block];
arch/mips/include/asm/mach-au1x00/au1000.h
797
int alchemy_usb_control(int block, int enable);
arch/mips/include/asm/mach-au1x00/au1000.h
892
extern void au1300_vss_block_control(int block, int enable);
arch/mips/include/asm/mach-cavium-octeon/kernel-entry-init.h
26
# a3 = address of boot descriptor block
arch/mips/include/asm/mips-cm.h
484
unsigned int vp, unsigned int block);
arch/mips/include/asm/mips-cm.h
497
unsigned int vp, unsigned int block) { }
arch/mips/include/asm/mips-cm.h
511
static inline void mips_cm_lock_other_cpu(unsigned int cpu, unsigned int block)
arch/mips/include/asm/mips-cm.h
515
mips_cm_lock_other(cpu_cluster(d), cpu_core(d), cpu_vpe_id(d), block);
arch/mips/include/asm/octeon/cvmx-asxx-defs.h
58
void __cvmx_interrupt_asxx_enable(int block);
arch/mips/include/asm/octeon/cvmx-pcsx-defs.h
337
void __cvmx_interrupt_pcsx_intx_en_reg_enable(int index, int block);
arch/mips/include/asm/octeon/octeon.h
362
struct irq_domain *octeon_irq_get_block_domain(int node, uint8_t block);
arch/mips/kernel/mips-cm.c
318
unsigned int vp, unsigned int block)
arch/mips/kernel/mips-cm.c
335
val |= FIELD_PREP(CM_GCR_Cx_OTHER_BLOCK, block);
arch/mips/kernel/mips-cm.c
338
WARN_ON(block != CM_GCR_Cx_OTHER_BLOCK_LOCAL);
arch/mips/kernel/mips-cm.c
354
WARN_ON(block != CM_GCR_Cx_OTHER_BLOCK_LOCAL);
arch/powerpc/include/asm/kvm_book3s.h
688
int block = (id / KVM_MAX_VCPUS) * (MAX_SMT_THREADS / stride);
arch/powerpc/include/asm/kvm_book3s.h
691
if (WARN_ONCE(block >= MAX_SMT_THREADS, "VCPU ID too large to pack"))
arch/powerpc/include/asm/kvm_book3s.h
693
packed_id = (id % KVM_MAX_VCPUS) + block_offsets[block];
arch/powerpc/include/asm/rheap.h
30
rh_block_t *block;
arch/powerpc/include/asm/rheap.h
57
rh_block_t * block);
arch/powerpc/lib/rheap.c
268
info->block = NULL;
arch/powerpc/lib/rheap.c
288
kfree(info->block);
arch/powerpc/lib/rheap.c
301
rh_block_t * block)
arch/powerpc/lib/rheap.c
313
info->block = block;
arch/powerpc/lib/rheap.c
323
for (i = 0, blk = block; i < max_blocks; i++, blk++)
arch/powerpc/lib/rheap.c
47
rh_block_t *block, *blk;
arch/powerpc/lib/rheap.c
57
block = kmalloc_objs(rh_block_t, max_blocks, GFP_ATOMIC);
arch/powerpc/lib/rheap.c
58
if (block == NULL)
arch/powerpc/lib/rheap.c
64
memcpy(block, info->block,
arch/powerpc/lib/rheap.c
67
delta = (char *)block - (char *)info->block;
arch/powerpc/lib/rheap.c
70
blks = (unsigned long)info->block;
arch/powerpc/lib/rheap.c
71
blke = (unsigned long)(info->block + info->max_blocks);
arch/powerpc/lib/rheap.c
73
for (i = 0, blk = block; i < info->max_blocks; i++, blk++)
arch/powerpc/lib/rheap.c
82
kfree(info->block);
arch/powerpc/lib/rheap.c
85
info->block = block;
arch/powerpc/lib/rheap.c
91
blk = block + info->max_blocks - new_blocks;
arch/powerpc/mm/kasan/8xx.c
12
kasan_init_shadow_8M(unsigned long k_start, unsigned long k_end, void *block)
arch/powerpc/mm/kasan/8xx.c
17
for (k_cur = k_start; k_cur != k_end; k_cur = k_next, pmd++, block += SZ_4M) {
arch/powerpc/mm/kasan/8xx.c
30
pte_t pte = pte_mkhuge(pfn_pte(PHYS_PFN(__pa(block + i * PAGE_SIZE)), PAGE_KERNEL));
arch/powerpc/mm/kasan/8xx.c
46
void *block;
arch/powerpc/mm/kasan/8xx.c
48
block = memblock_alloc(k_end - k_start, SZ_8M);
arch/powerpc/mm/kasan/8xx.c
49
if (!block)
arch/powerpc/mm/kasan/8xx.c
53
kasan_init_shadow_8M(k_start, ALIGN_DOWN(k_end, SZ_8M), block);
arch/powerpc/mm/kasan/8xx.c
67
void *va = block + k_cur - k_start;
arch/powerpc/mm/kasan/init_32.c
61
void *block;
arch/powerpc/mm/kasan/init_32.c
68
block = memblock_alloc(k_end - k_start, PAGE_SIZE);
arch/powerpc/mm/kasan/init_32.c
69
if (!block)
arch/powerpc/mm/kasan/init_32.c
74
void *va = block + k_cur - k_start;
arch/s390/crypto/aes_s390.c
326
u8 block[16];
arch/s390/crypto/aes_s390.c
352
memset(pcc_param.block, 0, sizeof(pcc_param.block));
arch/s390/crypto/paes_s390.c
1191
u8 block[16];
arch/s390/kernel/sthyi.c
232
struct diag204_x_phys_block *block,
arch/s390/kernel/sthyi.c
237
for (i = 0; i < block->hdr.cpus; i++) {
arch/s390/kernel/sthyi.c
238
switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) {
arch/s390/kernel/sthyi.c
240
if (block->cpus[i].weight == DED_WEIGHT)
arch/s390/kernel/sthyi.c
246
if (block->cpus[i].weight == DED_WEIGHT)
arch/s390/kernel/sthyi.c
260
struct diag204_x_part_block *block)
arch/s390/kernel/sthyi.c
265
for (i = 0; i < block->hdr.rcpus; i++) {
arch/s390/kernel/sthyi.c
266
if (!(block->cpus[i].cflag & DIAG204_CPU_ONLINE))
arch/s390/kernel/sthyi.c
269
switch (cpu_id(block->cpus[i].ctidx, diag224_buf)) {
arch/s390/kernel/sthyi.c
272
if (block->cpus[i].cur_weight < DED_WEIGHT)
arch/s390/kernel/sthyi.c
273
weight_cp |= block->cpus[i].cur_weight;
arch/s390/kernel/sthyi.c
277
if (block->cpus[i].cur_weight < DED_WEIGHT)
arch/s390/kernel/sthyi.c
278
weight_ifl |= block->cpus[i].cur_weight;
arch/s390/kernel/sthyi.c
287
capped |= block->cpus[i].cflag & DIAG204_CPU_CAPPED;
arch/s390/kernel/sthyi.c
288
cpu_inf->lpar_cap |= block->cpus[i].cpu_type_cap;
arch/s390/kernel/sthyi.c
289
cpu_inf->lpar_grp_cap |= block->cpus[i].group_cpu_type_cap;
arch/s390/kernel/sthyi.c
291
if (block->cpus[i].weight == DED_WEIGHT)
arch/s390/kernel/sthyi.c
303
return (struct diag204_x_part_block *)&block->cpus[i];
arch/s390/mm/pfault.c
176
goto block;
arch/s390/mm/pfault.c
195
block:
arch/sparc/include/asm/viking.h
165
static inline void viking_get_dcache_ptag(int set, int block,
arch/sparc/include/asm/viking.h
168
unsigned long ptag = ((set & 0x7f) << 5) | ((block & 0x3) << 26) |
arch/sparc/mm/init_64.c
2608
void *block = vmemmap_alloc_block(PMD_SIZE, node);
arch/sparc/mm/init_64.c
2610
if (!block)
arch/sparc/mm/init_64.c
2613
pmd_val(*pmd) = pte_base | __pa(block);
arch/um/drivers/random.c
34
static int rng_dev_read(struct hwrng *rng, void *buf, size_t max, bool block)
arch/um/drivers/random.c
40
if (block && ret == -EAGAIN) {
arch/x86/kernel/cpu/mce/amd.c
1048
if (b->block < ARRAY_SIZE(smca_umc_block_names))
arch/x86/kernel/cpu/mce/amd.c
1049
return smca_umc_block_names[b->block];
arch/x86/kernel/cpu/mce/amd.c
1052
if (b && b->block) {
arch/x86/kernel/cpu/mce/amd.c
1053
snprintf(buf_mcatype, MAX_MCATYPE_NAME_LEN, "th_block_%u", b->block);
arch/x86/kernel/cpu/mce/amd.c
1072
unsigned int bank, unsigned int block,
arch/x86/kernel/cpu/mce/amd.c
1079
if ((bank >= this_cpu_read(mce_num_banks)) || (block >= NR_BLOCKS))
arch/x86/kernel/cpu/mce/amd.c
1086
if (block)
arch/x86/kernel/cpu/mce/amd.c
1100
b->block = block;
arch/x86/kernel/cpu/mce/amd.c
1123
address = get_block_address(address, low, high, bank, ++block, cpu);
arch/x86/kernel/cpu/mce/amd.c
1127
err = allocate_threshold_blocks(cpu, tb, bank, block, address);
arch/x86/kernel/cpu/mce/amd.c
233
unsigned int block;
arch/x86/kernel/cpu/mce/amd.c
415
b->bank, b->block, b->address, hi, lo);
arch/x86/kernel/cpu/mce/amd.c
422
b->cpu, apic, b->bank, b->block, b->address, hi, lo);
arch/x86/kernel/cpu/mce/amd.c
482
struct threshold_block *block, *tmp;
arch/x86/kernel/cpu/mce/amd.c
490
list_for_each_entry_safe(block, tmp, &thr_banks[bank]->miscj, miscj) {
arch/x86/kernel/cpu/mce/amd.c
491
tr.b = block;
arch/x86/kernel/cpu/mce/amd.c
531
unsigned int bank, unsigned int block,
arch/x86/kernel/cpu/mce/amd.c
536
if ((bank >= per_cpu(mce_num_banks, cpu)) || (block >= NR_BLOCKS))
arch/x86/kernel/cpu/mce/amd.c
540
if (!block)
arch/x86/kernel/cpu/mce/amd.c
546
return MSR_AMD64_SMCA_MCx_MISCy(bank, block - 1);
arch/x86/kernel/cpu/mce/amd.c
550
switch (block) {
arch/x86/kernel/cpu/mce/amd.c
565
static int prepare_threshold_block(unsigned int bank, unsigned int block, u32 addr,
arch/x86/kernel/cpu/mce/amd.c
572
if (!block)
arch/x86/kernel/cpu/mce/amd.c
578
b.block = block;
arch/x86/kernel/cpu/mce/amd.c
718
unsigned int bank, block, cpu = smp_processor_id();
arch/x86/kernel/cpu/mce/amd.c
738
for (block = 0; block < NR_BLOCKS; ++block) {
arch/x86/kernel/cpu/mce/amd.c
739
address = get_block_address(address, low, high, bank, block, cpu);
arch/x86/kernel/cpu/mce/amd.c
753
offset = prepare_threshold_block(bank, block, address, offset, high);
arch/x86/kernel/cpu/mtrr/generic.c
868
int block = -1, range;
arch/x86/kernel/cpu/mtrr/generic.c
872
while (fixed_range_blocks[++block].ranges) {
arch/x86/kernel/cpu/mtrr/generic.c
873
for (range = 0; range < fixed_range_blocks[block].ranges; range++)
arch/x86/kernel/cpu/mtrr/generic.c
874
set_fixed_range(fixed_range_blocks[block].base_msr + range,
arch/x86/kernel/step.c
204
static void enable_step(struct task_struct *child, bool block)
arch/x86/kernel/step.c
213
if (enable_single_step(child) && block)
block/disk-events.c
106
if (WARN_ON_ONCE(ev->block <= 0))
block/disk-events.c
109
if (--ev->block)
block/disk-events.c
16
int block; /* event blocking depth */
block/disk-events.c
160
if (!ev->block)
block/disk-events.c
203
if (!ev->block && intv)
block/disk-events.c
449
ev->block = 1;
block/disk-events.c
487
WARN_ON_ONCE(disk->ev && disk->ev->block != 1);
block/disk-events.c
89
cancel = !ev->block++;
block/partitions/acorn.c
309
unsigned long block)
block/partitions/acorn.c
312
unsigned char *data = read_part_sector(state, block, &sect);
block/partitions/ldm.c
655
static u64 ldm_get_vnum (const u8 *block)
block/partitions/ldm.c
660
BUG_ON (!block);
block/partitions/ldm.c
662
length = *block++;
block/partitions/ldm.c
666
tmp = (tmp << 8) | *block++;
block/partitions/ldm.c
690
static int ldm_get_vstr (const u8 *block, u8 *buffer, int buflen)
block/partitions/ldm.c
694
BUG_ON (!block || !buffer);
block/partitions/ldm.c
696
length = block[0];
block/partitions/ldm.c
701
memcpy (buffer, block + 1, length);
crypto/cast6_generic.c
153
static inline void Q(u32 *block, const u8 *Kr, const u32 *Km)
crypto/cast6_generic.c
156
block[2] ^= F1(block[3], Kr[0], Km[0]);
crypto/cast6_generic.c
157
block[1] ^= F2(block[2], Kr[1], Km[1]);
crypto/cast6_generic.c
158
block[0] ^= F3(block[1], Kr[2], Km[2]);
crypto/cast6_generic.c
159
block[3] ^= F1(block[0], Kr[3], Km[3]);
crypto/cast6_generic.c
163
static inline void QBAR(u32 *block, const u8 *Kr, const u32 *Km)
crypto/cast6_generic.c
166
block[3] ^= F1(block[0], Kr[3], Km[3]);
crypto/cast6_generic.c
167
block[0] ^= F3(block[1], Kr[2], Km[2]);
crypto/cast6_generic.c
168
block[1] ^= F2(block[2], Kr[1], Km[1]);
crypto/cast6_generic.c
169
block[2] ^= F1(block[3], Kr[0], Km[0]);
crypto/cast6_generic.c
175
u32 block[4];
crypto/cast6_generic.c
179
block[0] = get_unaligned_be32(inbuf);
crypto/cast6_generic.c
180
block[1] = get_unaligned_be32(inbuf + 4);
crypto/cast6_generic.c
181
block[2] = get_unaligned_be32(inbuf + 8);
crypto/cast6_generic.c
182
block[3] = get_unaligned_be32(inbuf + 12);
crypto/cast6_generic.c
184
Km = c->Km[0]; Kr = c->Kr[0]; Q(block, Kr, Km);
crypto/cast6_generic.c
185
Km = c->Km[1]; Kr = c->Kr[1]; Q(block, Kr, Km);
crypto/cast6_generic.c
186
Km = c->Km[2]; Kr = c->Kr[2]; Q(block, Kr, Km);
crypto/cast6_generic.c
187
Km = c->Km[3]; Kr = c->Kr[3]; Q(block, Kr, Km);
crypto/cast6_generic.c
188
Km = c->Km[4]; Kr = c->Kr[4]; Q(block, Kr, Km);
crypto/cast6_generic.c
189
Km = c->Km[5]; Kr = c->Kr[5]; Q(block, Kr, Km);
crypto/cast6_generic.c
190
Km = c->Km[6]; Kr = c->Kr[6]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
191
Km = c->Km[7]; Kr = c->Kr[7]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
192
Km = c->Km[8]; Kr = c->Kr[8]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
193
Km = c->Km[9]; Kr = c->Kr[9]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
194
Km = c->Km[10]; Kr = c->Kr[10]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
195
Km = c->Km[11]; Kr = c->Kr[11]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
197
put_unaligned_be32(block[0], outbuf);
crypto/cast6_generic.c
198
put_unaligned_be32(block[1], outbuf + 4);
crypto/cast6_generic.c
199
put_unaligned_be32(block[2], outbuf + 8);
crypto/cast6_generic.c
200
put_unaligned_be32(block[3], outbuf + 12);
crypto/cast6_generic.c
212
u32 block[4];
crypto/cast6_generic.c
216
block[0] = get_unaligned_be32(inbuf);
crypto/cast6_generic.c
217
block[1] = get_unaligned_be32(inbuf + 4);
crypto/cast6_generic.c
218
block[2] = get_unaligned_be32(inbuf + 8);
crypto/cast6_generic.c
219
block[3] = get_unaligned_be32(inbuf + 12);
crypto/cast6_generic.c
221
Km = c->Km[11]; Kr = c->Kr[11]; Q(block, Kr, Km);
crypto/cast6_generic.c
222
Km = c->Km[10]; Kr = c->Kr[10]; Q(block, Kr, Km);
crypto/cast6_generic.c
223
Km = c->Km[9]; Kr = c->Kr[9]; Q(block, Kr, Km);
crypto/cast6_generic.c
224
Km = c->Km[8]; Kr = c->Kr[8]; Q(block, Kr, Km);
crypto/cast6_generic.c
225
Km = c->Km[7]; Kr = c->Kr[7]; Q(block, Kr, Km);
crypto/cast6_generic.c
226
Km = c->Km[6]; Kr = c->Kr[6]; Q(block, Kr, Km);
crypto/cast6_generic.c
227
Km = c->Km[5]; Kr = c->Kr[5]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
228
Km = c->Km[4]; Kr = c->Kr[4]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
229
Km = c->Km[3]; Kr = c->Kr[3]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
230
Km = c->Km[2]; Kr = c->Kr[2]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
231
Km = c->Km[1]; Kr = c->Kr[1]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
232
Km = c->Km[0]; Kr = c->Kr[0]; QBAR(block, Kr, Km);
crypto/cast6_generic.c
234
put_unaligned_be32(block[0], outbuf);
crypto/cast6_generic.c
235
put_unaligned_be32(block[1], outbuf + 4);
crypto/cast6_generic.c
236
put_unaligned_be32(block[2], outbuf + 8);
crypto/cast6_generic.c
237
put_unaligned_be32(block[3], outbuf + 12);
crypto/ccm.c
66
static int set_msg_len(u8 *block, unsigned int msglen, int csize)
crypto/ccm.c
70
memset(block, 0, csize);
crypto/ccm.c
71
block += csize;
crypto/ccm.c
79
memcpy(block - csize, (u8 *)&data + 4 - csize, csize);
crypto/md4.c
135
le32_to_cpu_array(ctx->block, ARRAY_SIZE(ctx->block));
crypto/md4.c
136
md4_transform(ctx->hash, ctx->block);
crypto/md4.c
155
const u32 avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f);
crypto/md4.c
160
memcpy((char *)mctx->block + (sizeof(mctx->block) - avail),
crypto/md4.c
165
memcpy((char *)mctx->block + (sizeof(mctx->block) - avail),
crypto/md4.c
172
while (len >= sizeof(mctx->block)) {
crypto/md4.c
173
memcpy(mctx->block, data, sizeof(mctx->block));
crypto/md4.c
175
data += sizeof(mctx->block);
crypto/md4.c
176
len -= sizeof(mctx->block);
crypto/md4.c
179
memcpy(mctx->block, data, len);
crypto/md4.c
188
char *p = (char *)mctx->block + offset;
crypto/md4.c
195
p = (char *)mctx->block;
crypto/md4.c
200
mctx->block[14] = mctx->byte_count << 3;
crypto/md4.c
201
mctx->block[15] = mctx->byte_count >> 29;
crypto/md4.c
202
le32_to_cpu_array(mctx->block, (sizeof(mctx->block) -
crypto/md4.c
204
md4_transform(mctx->hash, mctx->block);
crypto/md4.c
38
u32 block[MD4_BLOCK_WORDS];
crypto/md5.c
23
static_assert(offsetof(struct md5_ctx, buf) == offsetof(struct md5_state, block));
crypto/rmd160.c
297
} block = {};
crypto/rmd160.c
304
memcpy(&block, src, len);
crypto/rmd160.c
305
block.u8[len] = 0x80;
crypto/rmd160.c
306
block.l64[bit_offset] = cpu_to_le64(rctx->byte_count << 3);
crypto/rmd160.c
308
rmd160_transform(rctx->state, block.l32);
crypto/rmd160.c
311
block.l32 + RMD160_BLOCK_SIZE / 4);
crypto/rmd160.c
312
memzero_explicit(&block, sizeof(block));
crypto/wp512.c
783
u64 block[8]; /* mu(buffer) */
crypto/wp512.c
788
block[i] = get_unaligned_be64(buffer + i * 8);
crypto/wp512.c
790
state[0] = block[0] ^ (K[0] = wctx->hash[0]);
crypto/wp512.c
791
state[1] = block[1] ^ (K[1] = wctx->hash[1]);
crypto/wp512.c
792
state[2] = block[2] ^ (K[2] = wctx->hash[2]);
crypto/wp512.c
793
state[3] = block[3] ^ (K[3] = wctx->hash[3]);
crypto/wp512.c
794
state[4] = block[4] ^ (K[4] = wctx->hash[4]);
crypto/wp512.c
795
state[5] = block[5] ^ (K[5] = wctx->hash[5]);
crypto/wp512.c
796
state[6] = block[6] ^ (K[6] = wctx->hash[6]);
crypto/wp512.c
797
state[7] = block[7] ^ (K[7] = wctx->hash[7]);
crypto/wp512.c
975
wctx->hash[0] ^= state[0] ^ block[0];
crypto/wp512.c
976
wctx->hash[1] ^= state[1] ^ block[1];
crypto/wp512.c
977
wctx->hash[2] ^= state[2] ^ block[2];
crypto/wp512.c
978
wctx->hash[3] ^= state[3] ^ block[3];
crypto/wp512.c
979
wctx->hash[4] ^= state[4] ^ block[4];
crypto/wp512.c
980
wctx->hash[5] ^= state[5] ^ block[5];
crypto/wp512.c
981
wctx->hash[6] ^= state[6] ^ block[6];
crypto/wp512.c
982
wctx->hash[7] ^= state[7] ^ block[7];
drivers/accel/habanalabs/common/habanalabs.h
248
void (*fn)(struct hl_device *hdev, int block, int inst, u32 offset,
drivers/acpi/acpica/dbdisply.c
732
u32 block = 0;
drivers/acpi/acpica/dbdisply.c
743
block = 0;
drivers/acpi/acpica/dbdisply.c
767
block, gpe_block, gpe_block->node, buffer,
drivers/acpi/acpica/dbdisply.c
909
block++;
drivers/acpi/arm64/gtdt.c
206
static int __init gtdt_parse_timer_block(struct acpi_gtdt_timer_block *block,
drivers/acpi/arm64/gtdt.c
213
if (!block->timer_count) {
drivers/acpi/arm64/gtdt.c
218
if (block->timer_count > ARCH_TIMER_MEM_MAX_FRAMES) {
drivers/acpi/arm64/gtdt.c
220
block->timer_count);
drivers/acpi/arm64/gtdt.c
224
timer_mem->cntctlbase = (phys_addr_t)block->block_address;
drivers/acpi/arm64/gtdt.c
232
gtdt_frame = (void *)block + block->timer_offset;
drivers/acpi/arm64/gtdt.c
233
if (gtdt_frame + block->timer_count != (void *)block + block->header.length)
drivers/acpi/arm64/gtdt.c
239
for (i = 0; i < block->timer_count; i++, gtdt_frame++) {
drivers/ata/libata-core.c
668
u64 block = 0;
drivers/ata/libata-core.c
672
block |= (u64)tf->hob_lbah << 40;
drivers/ata/libata-core.c
673
block |= (u64)tf->hob_lbam << 32;
drivers/ata/libata-core.c
674
block |= (u64)tf->hob_lbal << 24;
drivers/ata/libata-core.c
676
block |= (tf->device & 0xf) << 24;
drivers/ata/libata-core.c
678
block |= tf->lbah << 16;
drivers/ata/libata-core.c
679
block |= tf->lbam << 8;
drivers/ata/libata-core.c
680
block |= tf->lbal;
drivers/ata/libata-core.c
694
block = (cyl * dev->heads + head) * dev->sectors + sect - 1;
drivers/ata/libata-core.c
697
return block;
drivers/ata/libata-core.c
740
int ata_build_rw_tf(struct ata_queued_cmd *qc, u64 block, u32 n_block,
drivers/ata/libata-core.c
751
if (!lba_48_ok(block, n_block))
drivers/ata/libata-core.c
766
tf->hob_lbah = (block >> 40) & 0xff;
drivers/ata/libata-core.c
767
tf->hob_lbam = (block >> 32) & 0xff;
drivers/ata/libata-core.c
768
tf->hob_lbal = (block >> 24) & 0xff;
drivers/ata/libata-core.c
769
tf->lbah = (block >> 16) & 0xff;
drivers/ata/libata-core.c
770
tf->lbam = (block >> 8) & 0xff;
drivers/ata/libata-core.c
771
tf->lbal = block & 0xff;
drivers/ata/libata-core.c
793
lba_28_ok(block, n_block)) {
drivers/ata/libata-core.c
795
tf->device |= (block >> 24) & 0xf;
drivers/ata/libata-core.c
796
} else if (lba_48_ok(block, n_block)) {
drivers/ata/libata-core.c
805
tf->hob_lbah = (block >> 40) & 0xff;
drivers/ata/libata-core.c
806
tf->hob_lbam = (block >> 32) & 0xff;
drivers/ata/libata-core.c
807
tf->hob_lbal = (block >> 24) & 0xff;
drivers/ata/libata-core.c
818
tf->lbah = (block >> 16) & 0xff;
drivers/ata/libata-core.c
819
tf->lbam = (block >> 8) & 0xff;
drivers/ata/libata-core.c
820
tf->lbal = block & 0xff;
drivers/ata/libata-core.c
828
if (!lba_28_ok(block, n_block))
drivers/ata/libata-core.c
835
track = (u32)block / dev->sectors;
drivers/ata/libata-core.c
838
sect = (u32)block % dev->sectors + 1;
drivers/ata/libata-scsi.c
1406
u64 block;
drivers/ata/libata-scsi.c
1419
scsi_10_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
1426
scsi_16_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
1435
if (block >= dev_sectors)
drivers/ata/libata-scsi.c
1437
if ((block + n_block) > dev_sectors)
drivers/ata/libata-scsi.c
1443
if (lba_28_ok(block, n_block)) {
drivers/ata/libata-scsi.c
1446
tf->device |= (block >> 24) & 0xf;
drivers/ata/libata-scsi.c
1447
} else if (lba_48_ok(block, n_block)) {
drivers/ata/libata-scsi.c
1457
tf->hob_lbah = (block >> 40) & 0xff;
drivers/ata/libata-scsi.c
1458
tf->hob_lbam = (block >> 32) & 0xff;
drivers/ata/libata-scsi.c
1459
tf->hob_lbal = (block >> 24) & 0xff;
drivers/ata/libata-scsi.c
1466
tf->lbah = (block >> 16) & 0xff;
drivers/ata/libata-scsi.c
1467
tf->lbam = (block >> 8) & 0xff;
drivers/ata/libata-scsi.c
1468
tf->lbal = block & 0xff;
drivers/ata/libata-scsi.c
1475
if (!lba_28_ok(block, n_block))
drivers/ata/libata-scsi.c
1479
track = (u32)block / dev->sectors;
drivers/ata/libata-scsi.c
1482
sect = (u32)block % dev->sectors + 1;
drivers/ata/libata-scsi.c
1556
u64 block;
drivers/ata/libata-scsi.c
1577
scsi_10_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
1589
scsi_6_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
1605
scsi_16_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
1631
rc = ata_build_rw_tf(qc, block, n_block, tf_flags, dld, class);
drivers/ata/libata-scsi.c
3485
u64 block;
drivers/ata/libata-scsi.c
3509
scsi_16_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
3536
size = ata_format_dsm_trim_descr(scmd, trmax, block, n_block);
drivers/ata/libata-scsi.c
3757
u64 block;
drivers/ata/libata-scsi.c
3766
scsi_16_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
3807
tf->lbah = (block >> 16) & 0xff;
drivers/ata/libata-scsi.c
3808
tf->lbam = (block >> 8) & 0xff;
drivers/ata/libata-scsi.c
3809
tf->lbal = block & 0xff;
drivers/ata/libata-scsi.c
3810
tf->hob_lbah = (block >> 40) & 0xff;
drivers/ata/libata-scsi.c
3811
tf->hob_lbam = (block >> 32) & 0xff;
drivers/ata/libata-scsi.c
3812
tf->hob_lbal = (block >> 24) & 0xff;
drivers/ata/libata-scsi.c
3840
u64 block;
drivers/ata/libata-scsi.c
3856
scsi_16_lba_len(cdb, &block, &n_block);
drivers/ata/libata-scsi.c
3869
block = 0;
drivers/ata/libata-scsi.c
3870
} else if (block >= dev->n_sectors) {
drivers/ata/libata-scsi.c
3891
tf->lbah = (block >> 16) & 0xff;
drivers/ata/libata-scsi.c
3892
tf->lbam = (block >> 8) & 0xff;
drivers/ata/libata-scsi.c
3893
tf->lbal = block & 0xff;
drivers/ata/libata-scsi.c
3894
tf->hob_lbah = (block >> 40) & 0xff;
drivers/ata/libata-scsi.c
3895
tf->hob_lbam = (block >> 32) & 0xff;
drivers/ata/libata-scsi.c
3896
tf->hob_lbal = (block >> 24) & 0xff;
drivers/ata/libata.h
66
extern int ata_build_rw_tf(struct ata_queued_cmd *qc, u64 block, u32 n_block,
drivers/base/power/wakeup.c
962
bool pm_get_wakeup_count(unsigned int *count, bool block)
drivers/base/power/wakeup.c
966
if (block) {
drivers/base/regmap/internal.h
265
int regcache_sync_block(struct regmap *map, void *block,
drivers/base/regmap/regcache-rbtree.c
217
kfree(rbtree_node->block);
drivers/base/regmap/regcache-rbtree.c
23
void *block;
drivers/base/regmap/regcache-rbtree.c
280
blk = krealloc_array(rbnode->block, blklen, map->cache_word_size, map->alloc_flags);
drivers/base/regmap/regcache-rbtree.c
284
rbnode->block = blk;
drivers/base/regmap/regcache-rbtree.c
348
rbnode->block = kmalloc_array(rbnode->blklen, map->cache_word_size,
drivers/base/regmap/regcache-rbtree.c
350
if (!rbnode->block)
drivers/base/regmap/regcache-rbtree.c
362
kfree(rbnode->block);
drivers/base/regmap/regcache-rbtree.c
498
ret = regcache_sync_block(map, rbnode->block,
drivers/base/regmap/regcache-rbtree.c
51
return regcache_get_val(map, rbnode->block, idx);
drivers/base/regmap/regcache-rbtree.c
59
regcache_set_val(map, rbnode->block, idx, val);
drivers/base/regmap/regcache.c
781
static int regcache_sync_block_single(struct regmap *map, void *block,
drivers/base/regmap/regcache.c
796
val = regcache_get_val(map, block, i);
drivers/base/regmap/regcache.c
833
static int regcache_sync_block_raw(struct regmap *map, void *block,
drivers/base/regmap/regcache.c
856
val = regcache_get_val(map, block, i);
drivers/base/regmap/regcache.c
866
data = regcache_get_val_addr(map, block, i);
drivers/base/regmap/regcache.c
875
int regcache_sync_block(struct regmap *map, void *block,
drivers/base/regmap/regcache.c
881
return regcache_sync_block_raw(map, block, cache_present,
drivers/base/regmap/regcache.c
884
return regcache_sync_block_single(map, block, cache_present,
drivers/block/amiflop.c
1463
unsigned int cnt, block, track, sector;
drivers/block/amiflop.c
1472
block = blk_rq_pos(rq) + cnt;
drivers/block/amiflop.c
1473
track = block / (floppy->dtype->sects * floppy->type->sect_mult);
drivers/block/amiflop.c
1474
sector = block % (floppy->dtype->sects * floppy->type->sect_mult);
drivers/block/ataflop.c
1487
int block = ReqBlock + ReqCnt;
drivers/block/ataflop.c
1489
ReqTrack = block / UDT->spt;
drivers/block/ataflop.c
1490
ReqSector = block - ReqTrack * UDT->spt + 1;
drivers/block/floppy.c
2280
int block;
drivers/block/floppy.c
2295
block = current_count_sectors + blk_rq_pos(req);
drivers/block/floppy.c
2296
INFBOUND(drive_state[current_drive].maxblock, block);
drivers/block/floppy.c
2297
if (block > _floppy->sect)
drivers/cdrom/gdrom.c
572
int block, block_cnt;
drivers/cdrom/gdrom.c
583
block = blk_rq_pos(req)/GD_TO_BLK + GD_SESSION_OFFSET;
drivers/cdrom/gdrom.c
590
read_command->cmd[2] = (block >> 16) & 0xFF;
drivers/cdrom/gdrom.c
591
read_command->cmd[3] = (block >> 8) & 0xFF;
drivers/cdrom/gdrom.c
592
read_command->cmd[4] = block & 0xFF;
drivers/char/hw_random/n2-drv.c
114
int block = 0, busy = 0;
drivers/char/hw_random/n2-drv.c
130
if (++block >= N2RNG_BLOCK_LIMIT)
drivers/char/hw_random/n2-drv.c
180
int block = 0, hcheck = 0;
drivers/char/hw_random/n2-drv.c
188
if (++block >= N2RNG_BLOCK_LIMIT)
drivers/char/hw_random/n2-drv.c
227
int block = 0;
drivers/char/hw_random/n2-drv.c
237
if (++block >= N2RNG_BLOCK_LIMIT)
drivers/char/hw_random/n2-drv.c
256
int block = 0, busy = 0;
drivers/char/hw_random/n2-drv.c
265
if (++block >= N2RNG_BLOCK_LIMIT)
drivers/char/random.c
1421
u8 block[BLAKE2S_BLOCK_SIZE];
drivers/char/random.c
1429
copied = copy_from_iter(block, sizeof(block), iter);
drivers/char/random.c
1431
mix_pool_bytes(block, copied);
drivers/char/random.c
1432
if (!iov_iter_count(iter) || copied != sizeof(block))
drivers/char/random.c
1435
BUILD_BUG_ON(PAGE_SIZE % sizeof(block) != 0);
drivers/char/random.c
1443
memzero_explicit(block, sizeof(block));
drivers/char/random.c
437
u8 block[CHACHA_BLOCK_SIZE];
drivers/char/random.c
461
chacha20_block(&chacha_state, block);
drivers/char/random.c
465
copied = copy_to_iter(block, sizeof(block), iter);
drivers/char/random.c
467
if (!iov_iter_count(iter) || copied != sizeof(block))
drivers/char/random.c
470
BUILD_BUG_ON(PAGE_SIZE % sizeof(block) != 0);
drivers/char/random.c
478
memzero_explicit(block, sizeof(block));
drivers/char/random.c
670
} block;
drivers/char/random.c
673
for (i = 0; i < ARRAY_SIZE(block.rdseed);) {
drivers/char/random.c
674
longs = arch_get_random_seed_longs(&block.rdseed[i], ARRAY_SIZE(block.rdseed) - i);
drivers/char/random.c
679
longs = arch_get_random_longs(&block.rdseed[i], ARRAY_SIZE(block.rdseed) - i);
drivers/char/random.c
684
block.rdseed[i++] = random_get_entropy();
drivers/char/random.c
693
block.counter = 0;
drivers/char/random.c
694
blake2s(seed, sizeof(seed), (const u8 *)&block, sizeof(block), next_key, sizeof(next_key));
drivers/char/random.c
703
++block.counter;
drivers/char/random.c
704
blake2s(seed, sizeof(seed), (const u8 *)&block, sizeof(block), buf, i);
drivers/char/random.c
710
memzero_explicit(&block, sizeof(block));
drivers/clk/renesas/r9a09g077-cpg.c
23
#define RZT2H_REG_CONF(block, offset) (((block) << RZT2H_REG_BLOCK_SHIFT) | \
drivers/clk/renesas/renesas-cpg-mssr.c
93
#define RZT2H_MSTPCR(block, offset) (((block) << RZT2H_MSTPCR_BLOCK_SHIFT) | \
drivers/comedi/drivers/jr3_pci.c
353
struct jr3_block __iomem *block = dev->mmio;
drivers/comedi/drivers/jr3_pci.c
386
lo = &block[subdev].program_lo[addr];
drivers/comedi/drivers/jr3_pci.c
387
hi = &block[subdev].program_hi[addr];
drivers/comedi/drivers/jr3_pci.c
610
struct jr3_block __iomem *block = dev->mmio;
drivers/comedi/drivers/jr3_pci.c
619
spriv->sensor = &block[s->index].sensor;
drivers/comedi/drivers/jr3_pci.c
645
struct jr3_block __iomem *block = dev->mmio;
drivers/comedi/drivers/jr3_pci.c
646
struct jr3_sensor __iomem *sensor0 = &block[0].sensor;
drivers/comedi/drivers/jr3_pci.c
663
struct jr3_block __iomem *block;
drivers/comedi/drivers/jr3_pci.c
685
if (pci_resource_len(pcidev, 0) < board->n_subdevs * sizeof(*block))
drivers/comedi/drivers/jr3_pci.c
692
block = dev->mmio;
drivers/comedi/drivers/jr3_pci.c
717
writel(0, &block[i].reset);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
70
memcpy(octx->block, op->buf, op->len);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
96
memcpy(op->buf, ictx->block, op->len);
drivers/crypto/ccree/cc_aead.c
1391
static int set_msg_len(u8 *block, unsigned int msglen, unsigned int csize)
drivers/crypto/ccree/cc_aead.c
1395
memset(block, 0, csize);
drivers/crypto/ccree/cc_aead.c
1396
block += csize;
drivers/crypto/ccree/cc_aead.c
1404
memcpy(block - csize, (u8 *)&data + 4 - csize, csize);
drivers/crypto/chelsio/chcr_algo.c
2788
static int set_msg_len(u8 *block, unsigned int msglen, int csize)
drivers/crypto/chelsio/chcr_algo.c
2792
memset(block, 0, csize);
drivers/crypto/chelsio/chcr_algo.c
2793
block += csize;
drivers/crypto/chelsio/chcr_algo.c
2801
memcpy(block - csize, (u8 *)&data + 4 - csize, csize);
drivers/crypto/hisilicon/sgl.c
100
block[i].size = block_size;
drivers/crypto/hisilicon/sgl.c
104
block[i].sgl = dma_alloc_coherent(dev, remain_sgl * sgl_size,
drivers/crypto/hisilicon/sgl.c
105
&block[i].sgl_dma,
drivers/crypto/hisilicon/sgl.c
107
if (!block[i].sgl) {
drivers/crypto/hisilicon/sgl.c
112
block[i].size = remain_sgl * sgl_size;
drivers/crypto/hisilicon/sgl.c
125
dma_free_coherent(dev, block_size, block[j].sgl,
drivers/crypto/hisilicon/sgl.c
126
block[j].sgl_dma);
drivers/crypto/hisilicon/sgl.c
142
struct mem_block *block;
drivers/crypto/hisilicon/sgl.c
148
block = pool->mem_block;
drivers/crypto/hisilicon/sgl.c
151
dma_free_coherent(dev, block[i].size, block[i].sgl,
drivers/crypto/hisilicon/sgl.c
152
block[i].sgl_dma);
drivers/crypto/hisilicon/sgl.c
161
struct mem_block *block;
drivers/crypto/hisilicon/sgl.c
164
block = pool->mem_block;
drivers/crypto/hisilicon/sgl.c
168
*hw_sgl_dma = block[block_index].sgl_dma + pool->sgl_size * offset;
drivers/crypto/hisilicon/sgl.c
169
return (void *)block[block_index].sgl + pool->sgl_size * offset;
drivers/crypto/hisilicon/sgl.c
62
struct mem_block *block;
drivers/crypto/hisilicon/sgl.c
89
block = pool->mem_block;
drivers/crypto/hisilicon/sgl.c
92
block[i].sgl = dma_alloc_coherent(dev, block_size,
drivers/crypto/hisilicon/sgl.c
93
&block[i].sgl_dma,
drivers/crypto/hisilicon/sgl.c
95
if (!block[i].sgl) {
drivers/crypto/inside-secure/eip93/eip93-hash.c
27
struct mkt_hash_block *block, *tmp;
drivers/crypto/inside-secure/eip93/eip93-hash.c
29
list_for_each_entry_safe(block, tmp, &rctx->blocks, list) {
drivers/crypto/inside-secure/eip93/eip93-hash.c
30
dma_unmap_single(eip93->dev, block->data_dma,
drivers/crypto/inside-secure/eip93/eip93-hash.c
32
kfree(block);
drivers/crypto/inside-secure/eip93/eip93-hash.c
321
struct mkt_hash_block *block;
drivers/crypto/inside-secure/eip93/eip93-hash.c
335
block = kzalloc_obj(*block, GFP_ATOMIC);
drivers/crypto/inside-secure/eip93/eip93-hash.c
336
if (!block) {
drivers/crypto/inside-secure/eip93/eip93-hash.c
342
block->data + offset,
drivers/crypto/inside-secure/eip93/eip93-hash.c
350
memcpy(block->data, rctx->data, offset);
drivers/crypto/inside-secure/eip93/eip93-hash.c
355
list_add(&block->list, &rctx->blocks);
drivers/crypto/inside-secure/eip93/eip93-hash.c
370
list_for_each_entry_reverse(block, &rctx->blocks, list) {
drivers/crypto/inside-secure/eip93/eip93-hash.c
372
list_is_first(&block->list, &rctx->blocks);
drivers/crypto/inside-secure/eip93/eip93-hash.c
374
ret = eip93_send_hash_req(async, block->data,
drivers/crypto/inside-secure/eip93/eip93-hash.c
375
&block->data_dma,
drivers/crypto/marvell/cesa/hash.c
912
&out_state->byte_count, out_state->block);
drivers/crypto/marvell/cesa/hash.c
920
in_state->block);
drivers/crypto/marvell/octeontx/otx_cptpf_ucode.c
49
char block[TAR_BLOCK_LEN];
drivers/crypto/nx/nx-aes-ccm.c
104
static int set_msg_len(u8 *block, unsigned int msglen, int csize)
drivers/crypto/nx/nx-aes-ccm.c
108
memset(block, 0, csize);
drivers/crypto/nx/nx-aes-ccm.c
109
block += csize;
drivers/crypto/nx/nx-aes-ccm.c
117
memcpy(block - csize, (u8 *)&data + 4 - csize, csize);
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
258
u32 block, conf_reg = 0;
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
260
block = crypto_tfm_alg_blocksize(tfm);
drivers/crypto/rockchip/rk3288_crypto_skcipher.c
262
if (block == DES_BLOCK_SIZE) {
drivers/crypto/stm32/stm32-cryp.c
1923
u32 block[AES_BLOCK_32];
drivers/crypto/stm32/stm32-cryp.c
1925
readsl(cryp->regs + cryp->caps->dout, block, cryp->hw_blocksize / sizeof(u32));
drivers/crypto/stm32/stm32-cryp.c
1926
memcpy_to_scatterwalk(&cryp->out_walk, block, min(cryp->hw_blocksize,
drivers/crypto/stm32/stm32-cryp.c
1933
u32 block[AES_BLOCK_32] = {0};
drivers/crypto/stm32/stm32-cryp.c
1935
memcpy_from_scatterwalk(block, &cryp->in_walk, min(cryp->hw_blocksize,
drivers/crypto/stm32/stm32-cryp.c
1937
writesl(cryp->regs + cryp->caps->din, block, cryp->hw_blocksize / sizeof(u32));
drivers/crypto/stm32/stm32-cryp.c
1944
u32 cfg, block[AES_BLOCK_32] = {0};
drivers/crypto/stm32/stm32-cryp.c
1981
readsl(cryp->regs + cryp->caps->dout, block, cryp->hw_blocksize / sizeof(u32));
drivers/crypto/stm32/stm32-cryp.c
1983
memcpy_to_scatterwalk(&cryp->out_walk, block, min(cryp->hw_blocksize,
drivers/crypto/stm32/stm32-cryp.c
1998
writesl(cryp->regs + cryp->caps->din, block, AES_BLOCK_32);
drivers/crypto/stm32/stm32-cryp.c
2033
u32 block[AES_BLOCK_32] = {0};
drivers/crypto/stm32/stm32-cryp.c
2078
readsl(cryp->regs + cryp->caps->dout, block, cryp->hw_blocksize / sizeof(u32));
drivers/crypto/stm32/stm32-cryp.c
2080
memcpy_to_scatterwalk(&cryp->out_walk, block, min(cryp->hw_blocksize,
drivers/crypto/stm32/stm32-cryp.c
2099
for (i = 0; i < ARRAY_SIZE(block); i++) {
drivers/crypto/stm32/stm32-cryp.c
2100
block[i] ^= cstmp1[i];
drivers/crypto/stm32/stm32-cryp.c
2101
block[i] ^= cstmp2[i];
drivers/crypto/stm32/stm32-cryp.c
2102
stm32_cryp_write(cryp, cryp->caps->din, block[i]);
drivers/crypto/stm32/stm32-cryp.c
2157
u32 block[AES_BLOCK_32] = {0};
drivers/crypto/stm32/stm32-cryp.c
2162
memcpy_from_scatterwalk(block, &cryp->in_walk, written);
drivers/crypto/stm32/stm32-cryp.c
2164
writesl(cryp->regs + cryp->caps->din, block, AES_BLOCK_32);
drivers/crypto/stm32/stm32-cryp.c
649
u32 block[AES_BLOCK_32] = {0};
drivers/crypto/stm32/stm32-cryp.c
650
u8 *b8 = (u8 *)block;
drivers/crypto/stm32/stm32-cryp.c
670
memcpy_from_scatterwalk((char *)block + len, &cryp->in_walk, written);
drivers/crypto/stm32/stm32-cryp.c
672
writesl(cryp->regs + cryp->caps->din, block, AES_BLOCK_32);
drivers/crypto/tegra/tegra-se-aes.c
947
static int tegra_ccm_set_msg_len(u8 *block, unsigned int msglen, int csize)
drivers/crypto/tegra/tegra-se-aes.c
951
memset(block, 0, csize);
drivers/crypto/tegra/tegra-se-aes.c
952
block += csize;
drivers/crypto/tegra/tegra-se-aes.c
960
memcpy(block - csize, (u8 *)&data + 4 - csize, csize);
drivers/dma/dw/dw.c
49
u32 block;
drivers/dma/dw/dw.c
52
block = dwc->block_size;
drivers/dma/dw/dw.c
55
block = bytes >> width;
drivers/dma/dw/dw.c
59
return block;
drivers/dma/dw/dw.c
62
static size_t dw_dma_block2bytes(struct dw_dma_chan *dwc, u32 block, u32 width)
drivers/dma/dw/dw.c
64
return DWC_CTLH_BLOCK_TS(block) << width;
drivers/dma/dw/idma32.c
184
u32 block;
drivers/dma/dw/idma32.c
187
block = dwc->block_size;
drivers/dma/dw/idma32.c
190
block = bytes;
drivers/dma/dw/idma32.c
194
return block;
drivers/dma/dw/idma32.c
197
static size_t idma32_block2bytes(struct dw_dma_chan *dwc, u32 block, u32 width)
drivers/dma/dw/idma32.c
199
return IDMA32C_CTLH_BLOCK_TS(block);
drivers/dma/dw/regs.h
332
size_t (*block2bytes)(struct dw_dma_chan *dwc, u32 block, u32 width);
drivers/dma/fsl-qdma.c
1015
reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQMR(fsl_queue->id));
drivers/dma/fsl-qdma.c
1017
qdma_writel(fsl_chan->qdma, reg, block + FSL_QDMA_BCQMR(fsl_queue->id));
drivers/dma/fsl-qdma.c
590
void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
drivers/dma/fsl-qdma.c
597
block = fsl_qdma->block_base +
drivers/dma/fsl-qdma.c
600
qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQMR(i));
drivers/dma/fsl-qdma.c
612
block = fsl_qdma->block_base +
drivers/dma/fsl-qdma.c
616
qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BSQMR);
drivers/dma/fsl-qdma.c
623
block + FSL_QDMA_BCQIDR(0));
drivers/dma/fsl-qdma.c
631
__iomem void *block,
drivers/dma/fsl-qdma.c
647
reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQSR);
drivers/dma/fsl-qdma.c
683
reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
drivers/dma/fsl-qdma.c
690
qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
drivers/dma/fsl-qdma.c
698
reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
drivers/dma/fsl-qdma.c
704
qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
drivers/dma/fsl-qdma.c
772
void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
drivers/dma/fsl-qdma.c
781
block = fsl_qdma->block_base +
drivers/dma/fsl-qdma.c
784
intr = qdma_readl(fsl_qdma, block + FSL_QDMA_BCQIDR(0));
drivers/dma/fsl-qdma.c
787
intr = fsl_qdma_queue_transfer_complete(fsl_qdma, block, id);
drivers/dma/fsl-qdma.c
793
qdma_writel(fsl_qdma, 0, block + FSL_QDMA_BCQIER(0));
drivers/dma/fsl-qdma.c
799
block + FSL_QDMA_BCQIDR(0));
drivers/dma/fsl-qdma.c
877
void __iomem *block, *ctrl = fsl_qdma->ctrl_base;
drivers/dma/fsl-qdma.c
893
block = fsl_qdma->block_base +
drivers/dma/fsl-qdma.c
896
block + FSL_QDMA_BCQIDR(0));
drivers/dma/fsl-qdma.c
900
block = fsl_qdma->block_base +
drivers/dma/fsl-qdma.c
913
block + FSL_QDMA_BCQDPA_SADDR(i));
drivers/dma/fsl-qdma.c
915
block + FSL_QDMA_BCQEPA_SADDR(i));
drivers/dma/fsl-qdma.c
921
qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BCQMR(i));
drivers/dma/fsl-qdma.c
931
block + FSL_QDMA_SQCCMR);
drivers/dma/fsl-qdma.c
941
block + FSL_QDMA_SQEPAR);
drivers/dma/fsl-qdma.c
943
block + FSL_QDMA_SQDPAR);
drivers/dma/fsl-qdma.c
946
block + FSL_QDMA_BCQIER(0));
drivers/dma/fsl-qdma.c
949
block + FSL_QDMA_BSQICR);
drivers/dma/fsl-qdma.c
952
block + FSL_QDMA_CQIER);
drivers/dma/fsl-qdma.c
959
qdma_writel(fsl_qdma, reg, block + FSL_QDMA_BSQMR);
drivers/dma/fsl-qdma.c
960
reg = qdma_readl(fsl_qdma, block + FSL_QDMA_BSQMR);
drivers/dma/fsl-qdma.c
997
void __iomem *block = fsl_queue->block_base;
drivers/dma/fsl-qdma.c
999
reg = qdma_readl(fsl_chan->qdma, block + FSL_QDMA_BCQSR(fsl_queue->id));
drivers/dma/ppc4xx/adma.c
129
static void print_cb(struct ppc440spe_adma_chan *chan, void *block)
drivers/dma/ppc4xx/adma.c
138
cdb = block;
drivers/dma/ppc4xx/adma.c
153
cb = block;
drivers/dma/xilinx/xdma.c
146
static inline void *xdma_blk_last_desc(struct xdma_desc_block *block)
drivers/dma/xilinx/xdma.c
148
return block->virt_addr + (XDMA_DESC_ADJACENT - 1) * XDMA_DESC_SIZE;
drivers/dma/xilinx/xdma.c
157
struct xdma_desc_block *block;
drivers/dma/xilinx/xdma.c
164
block = &sw_desc->desc_blocks[i - 1];
drivers/dma/xilinx/xdma.c
165
desc = xdma_blk_last_desc(block);
drivers/dma/xilinx/xdma.c
172
desc->next_desc = cpu_to_le64(block[1].dma_addr);
drivers/dma/xilinx/xdma.c
178
block = &sw_desc->desc_blocks[sw_desc->dblk_num - 2];
drivers/dma/xilinx/xdma.c
179
desc = xdma_blk_last_desc(block);
drivers/dma/xilinx/xdma.c
184
block = &sw_desc->desc_blocks[sw_desc->dblk_num - 1];
drivers/dma/xilinx/xdma.c
185
desc = block->virt_addr + last_blk_desc * XDMA_DESC_SIZE;
drivers/dma/xilinx/xdma.c
195
struct xdma_desc_block *block;
drivers/dma/xilinx/xdma.c
199
block = sw_desc->desc_blocks;
drivers/dma/xilinx/xdma.c
201
desc = block->virt_addr + i * XDMA_DESC_SIZE;
drivers/dma/xilinx/xdma.c
202
desc->next_desc = cpu_to_le64(block->dma_addr + ((i + 1) * XDMA_DESC_SIZE));
drivers/dma/xilinx/xdma.c
204
desc = block->virt_addr + i * XDMA_DESC_SIZE;
drivers/dma/xilinx/xdma.c
205
desc->next_desc = cpu_to_le64(block->dma_addr);
drivers/dma/xilinx/xdma.c
329
struct xdma_desc_block *block;
drivers/dma/xilinx/xdma.c
355
block = &desc->desc_blocks[completed_blocks];
drivers/dma/xilinx/xdma.c
356
val = lower_32_bits(block->dma_addr);
drivers/dma/xilinx/xdma.c
361
val = upper_32_bits(block->dma_addr);
drivers/edac/edac_device.c
117
for (block = 0; block < nr_blocks; block++) {
drivers/edac/edac_device.c
118
blk = &blk_p[block];
drivers/edac/edac_device.c
121
"%s%d", blk_name, block + off_val);
drivers/edac/edac_device.c
124
instance, inst, block, blk, blk->name);
drivers/edac/edac_device.c
483
struct edac_device_block *block = NULL;
drivers/edac/edac_device.c
508
block = instance->blocks + block_nr;
drivers/edac/edac_device.c
509
block->counters.ce_count += count;
drivers/edac/edac_device.c
520
block ? block->name : "N/A", count, msg);
drivers/edac/edac_device.c
529
struct edac_device_block *block = NULL;
drivers/edac/edac_device.c
554
block = instance->blocks + block_nr;
drivers/edac/edac_device.c
555
block->counters.ue_count += count;
drivers/edac/edac_device.c
566
block ? block->name : "N/A", count, msg);
drivers/edac/edac_device.c
571
block ? block->name : "N/A", count, msg);
drivers/edac/edac_device.c
64
unsigned instance, block;
drivers/edac/edac_device_sysfs.c
418
struct edac_device_block *block = to_block(kobj);
drivers/edac/edac_device_sysfs.c
420
return sprintf(data, "%u\n", block->counters.ue_count);
drivers/edac/edac_device_sysfs.c
426
struct edac_device_block *block = to_block(kobj);
drivers/edac/edac_device_sysfs.c
428
return sprintf(data, "%u\n", block->counters.ce_count);
drivers/edac/edac_device_sysfs.c
434
struct edac_device_block *block;
drivers/edac/edac_device_sysfs.c
439
block = to_block(kobj);
drivers/edac/edac_device_sysfs.c
444
kobject_put(&block->instance->ctl->kobj);
drivers/edac/edac_device_sysfs.c
496
struct edac_device_block *block)
drivers/edac/edac_device_sysfs.c
504
instance->name, instance, block->name, block);
drivers/edac/edac_device_sysfs.c
506
&block->kobj, &block->kobj.parent);
drivers/edac/edac_device_sysfs.c
509
memset(&block->kobj, 0, sizeof(struct kobject));
drivers/edac/edac_device_sysfs.c
521
err = kobject_init_and_add(&block->kobj, &ktype_block_ctrl,
drivers/edac/edac_device_sysfs.c
523
"%s", block->name);
drivers/edac/edac_device_sysfs.c
525
edac_dbg(1, "Failed to register instance '%s'\n", block->name);
drivers/edac/edac_device_sysfs.c
534
sysfs_attrib = block->block_attributes;
drivers/edac/edac_device_sysfs.c
535
if (sysfs_attrib && block->nr_attribs) {
drivers/edac/edac_device_sysfs.c
536
for (i = 0; i < block->nr_attribs; i++, sysfs_attrib++) {
drivers/edac/edac_device_sysfs.c
540
sysfs_attrib, &block->kobj);
drivers/edac/edac_device_sysfs.c
543
err = sysfs_create_file(&block->kobj,
drivers/edac/edac_device_sysfs.c
549
kobject_uevent(&block->kobj, KOBJ_ADD);
drivers/edac/edac_device_sysfs.c
555
kobject_put(&block->kobj);
drivers/edac/edac_device_sysfs.c
565
struct edac_device_block *block)
drivers/edac/edac_device_sysfs.c
573
sysfs_attrib = block->block_attributes;
drivers/edac/edac_device_sysfs.c
574
if (sysfs_attrib && block->nr_attribs) {
drivers/edac/edac_device_sysfs.c
575
for (i = 0; i < block->nr_attribs; i++, sysfs_attrib++) {
drivers/edac/edac_device_sysfs.c
578
sysfs_remove_file(&block->kobj,
drivers/edac/edac_device_sysfs.c
586
kobject_put(&block->kobj);
drivers/firewire/core-card.c
44
int fw_compute_block_crc(__be32 *block)
drivers/firewire/core-card.c
49
length = (be32_to_cpu(block[0]) >> 16) & 0xff;
drivers/firewire/core-card.c
50
crc = crc_itu_t(0, (u8 *)&block[1], length * 4);
drivers/firewire/core-card.c
51
*block |= cpu_to_be32(crc);
drivers/firewire/core-device.c
100
block += 3;
drivers/firewire/core-device.c
103
c = block[i / 4] >> (24 - 8 * (i % 4));
drivers/firewire/core-device.c
84
static int textual_leaf_to_string(const u32 *block, char *buf, size_t size)
drivers/firewire/core-device.c
92
quadlets = min(block[0] >> 16, 256U);
drivers/firewire/core-device.c
96
if (block[1] != 0 || block[2] != 0)
drivers/firewire/core.h
129
int fw_compute_block_crc(__be32 *block);
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
334
struct wmfw_coeff_item *block;
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
351
block = (struct wmfw_coeff_item *)&bin->data[le32_to_cpu(header->len)];
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
354
KUNIT_ASSERT_EQ(test, le16_to_cpu(block->type), param->block_type);
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
355
KUNIT_ASSERT_EQ(test, le32_to_cpu(block->len), sizeof(payload));
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
357
block->len = cpu_to_le32(0x8000);
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
362
block->len = cpu_to_le32(0xffff);
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
367
block->len = cpu_to_le32(0x7fffffff);
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
372
block->len = cpu_to_le32(0x80000000);
drivers/firmware/cirrus/test/cs_dsp_test_bin_error.c
377
block->len = cpu_to_le32(0xffffffff);
drivers/firmware/efi/libstub/arm32-stub.c
27
int block;
drivers/firmware/efi/libstub/arm32-stub.c
58
block = cpuid_feature_extract(CPUID_EXT_MMFR0, 0);
drivers/firmware/efi/libstub/arm32-stub.c
59
if (block < 5) {
drivers/gpio/gpio-sch311x.c
134
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
136
if (block->config_regs[offset] == 0) /* GPIO is not available */
drivers/gpio/gpio-sch311x.c
139
if (!request_region(block->runtime_reg + block->config_regs[offset],
drivers/gpio/gpio-sch311x.c
142
block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
150
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
152
if (block->config_regs[offset] == 0) /* GPIO is not available */
drivers/gpio/gpio-sch311x.c
155
release_region(block->runtime_reg + block->config_regs[offset], 1);
drivers/gpio/gpio-sch311x.c
160
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
163
spin_lock(&block->lock);
drivers/gpio/gpio-sch311x.c
164
data = inb(block->runtime_reg + block->data_reg);
drivers/gpio/gpio-sch311x.c
165
spin_unlock(&block->lock);
drivers/gpio/gpio-sch311x.c
170
static void __sch311x_gpio_set(struct sch311x_gpio_block *block,
drivers/gpio/gpio-sch311x.c
173
u8 data = inb(block->runtime_reg + block->data_reg);
drivers/gpio/gpio-sch311x.c
178
outb(data, block->runtime_reg + block->data_reg);
drivers/gpio/gpio-sch311x.c
184
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
186
spin_lock(&block->lock);
drivers/gpio/gpio-sch311x.c
187
__sch311x_gpio_set(block, offset, value);
drivers/gpio/gpio-sch311x.c
188
spin_unlock(&block->lock);
drivers/gpio/gpio-sch311x.c
195
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
198
spin_lock(&block->lock);
drivers/gpio/gpio-sch311x.c
199
data = inb(block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
201
outb(data, block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
202
spin_unlock(&block->lock);
drivers/gpio/gpio-sch311x.c
210
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
213
spin_lock(&block->lock);
drivers/gpio/gpio-sch311x.c
215
data = inb(block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
217
outb(data, block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
218
__sch311x_gpio_set(block, offset, value);
drivers/gpio/gpio-sch311x.c
220
spin_unlock(&block->lock);
drivers/gpio/gpio-sch311x.c
226
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
229
spin_lock(&block->lock);
drivers/gpio/gpio-sch311x.c
230
data = inb(block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
231
spin_unlock(&block->lock);
drivers/gpio/gpio-sch311x.c
242
struct sch311x_gpio_block *block = gpiochip_get_data(chip);
drivers/gpio/gpio-sch311x.c
248
spin_lock(&block->lock);
drivers/gpio/gpio-sch311x.c
249
data = inb(block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
251
outb(data, block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
252
spin_unlock(&block->lock);
drivers/gpio/gpio-sch311x.c
255
spin_lock(&block->lock);
drivers/gpio/gpio-sch311x.c
256
data = inb(block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
258
outb(data, block->runtime_reg + block->config_regs[offset]);
drivers/gpio/gpio-sch311x.c
259
spin_unlock(&block->lock);
drivers/gpio/gpio-sch311x.c
271
struct sch311x_gpio_block *block;
drivers/gpio/gpio-sch311x.c
287
block = &priv->blocks[i];
drivers/gpio/gpio-sch311x.c
289
spin_lock_init(&block->lock);
drivers/gpio/gpio-sch311x.c
291
block->chip.label = DRV_NAME;
drivers/gpio/gpio-sch311x.c
292
block->chip.owner = THIS_MODULE;
drivers/gpio/gpio-sch311x.c
293
block->chip.request = sch311x_gpio_request;
drivers/gpio/gpio-sch311x.c
294
block->chip.free = sch311x_gpio_free;
drivers/gpio/gpio-sch311x.c
295
block->chip.direction_input = sch311x_gpio_direction_in;
drivers/gpio/gpio-sch311x.c
296
block->chip.direction_output = sch311x_gpio_direction_out;
drivers/gpio/gpio-sch311x.c
297
block->chip.get_direction = sch311x_gpio_get_direction;
drivers/gpio/gpio-sch311x.c
298
block->chip.set_config = sch311x_gpio_set_config;
drivers/gpio/gpio-sch311x.c
299
block->chip.get = sch311x_gpio_get;
drivers/gpio/gpio-sch311x.c
300
block->chip.set = sch311x_gpio_set;
drivers/gpio/gpio-sch311x.c
301
block->chip.ngpio = 8;
drivers/gpio/gpio-sch311x.c
302
block->chip.parent = &pdev->dev;
drivers/gpio/gpio-sch311x.c
303
block->chip.base = sch311x_gpio_blocks[i].base;
drivers/gpio/gpio-sch311x.c
304
block->config_regs = sch311x_gpio_blocks[i].config_regs;
drivers/gpio/gpio-sch311x.c
305
block->data_reg = sch311x_gpio_blocks[i].data_reg;
drivers/gpio/gpio-sch311x.c
306
block->runtime_reg = pdata->runtime_reg;
drivers/gpio/gpio-sch311x.c
308
err = devm_gpiochip_add_data(&pdev->dev, &block->chip, block);
drivers/gpio/gpio-ts5500.c
266
const struct ts5500_dio *block = priv->pinout;
drivers/gpio/gpio-ts5500.c
267
const struct ts5500_dio line = block[offset];
drivers/gpio/gpio-ts5500.c
318
enum ts5500_blocks block = platform_get_device_id(pdev)->driver_data;
drivers/gpio/gpio-ts5500.c
347
switch (block) {
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1395
#define RREG32_AUDIO_ENDPT(block, reg) adev->audio_endpt_rreg(adev, (block), (reg))
drivers/gpu/drm/amd/amdgpu/amdgpu.h
1396
#define WREG32_AUDIO_ENDPT(block, reg, v) adev->audio_endpt_wreg(adev, (block), (reg), (v))
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
750
enum amdgpu_ras_block block, uint16_t pasid,
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
753
amdgpu_umc_pasid_poison_handler(adev, block, pasid, pasid_fn, data, reset);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
757
enum amdgpu_ras_block block, uint32_t reset)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
759
amdgpu_umc_pasid_poison_handler(adev, block, 0, NULL, NULL, reset);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
347
enum amdgpu_ras_block block, uint32_t reset);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
350
enum amdgpu_ras_block block, uint16_t pasid,
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1526
uint32_t block, uint32_t reg)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1530
reg, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1547
uint32_t block,
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1552
reg, block, v);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
191
enum amd_ip_block_type block)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
193
return (adev->init_lvl->hwini_ip_block_mask & (1U << block)) != 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3873
struct amdgpu_ip_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3875
block = &adev->ip_blocks[i];
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3876
block->status.hw = false;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3880
if (block->version->type != ip_order[j] ||
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3881
!block->status.valid)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3884
r = block->version->funcs->hw_init(&adev->ip_blocks[i]);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3887
block->version->funcs->name);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3890
block->status.hw = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3899
struct amdgpu_ip_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3915
block = amdgpu_device_ip_get_ip_block(adev, ip_order[i]);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3917
if (!block)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3920
if (block->status.valid && !block->status.hw) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3921
if (block->version->type == AMD_IP_BLOCK_TYPE_SMC) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3922
r = amdgpu_ip_block_resume(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3924
r = block->version->funcs->hw_init(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3929
block->version->funcs->name);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
3932
block->status.hw = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
933
if (amdgpu_ras_is_supported(adev, ras_block->block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
982
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__GFX;
drivers/gpu/drm/amd/amdgpu/amdgpu_hdp.c
43
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__HDP;
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.c
28
enum amd_hw_ip_block_type block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.c
33
switch (block) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.c
38
dev_inst = adev->ip_map.dev_inst[block][inst];
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.c
51
enum amd_hw_ip_block_type block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.c
59
dev_inst = amdgpu_logical_to_dev_inst(adev, block, log_inst);
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.h
91
enum amd_hw_ip_block_type block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ip.h
94
enum amd_hw_ip_block_type block,
drivers/gpu/drm/amd/amdgpu/amdgpu_jpeg.c
296
if (amdgpu_ras_is_supported(adev, ras_block->block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_jpeg.c
330
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__JPEG;
drivers/gpu/drm/amd/amdgpu/amdgpu_mca.c
101
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__MCA;
drivers/gpu/drm/amd/amdgpu/amdgpu_mca.c
125
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__MCA;
drivers/gpu/drm/amd/amdgpu/amdgpu_mca.c
149
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__MCA;
drivers/gpu/drm/amd/amdgpu/amdgpu_mmhub.c
40
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__MMHUB;
drivers/gpu/drm/amd/amdgpu/amdgpu_nbio.c
41
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__PCIE_BIF;
drivers/gpu/drm/amd/amdgpu/amdgpu_nbio.c
73
if (amdgpu_ras_is_supported(adev, ras_block->block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1012
.block = i,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
103
if (ras_block->block >= AMDGPU_RAS_BLOCK_COUNT ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1035
.block = AMDGPU_RAS_BLOCK__MCA,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
104
ras_block->block >= ARRAY_SIZE(ras_block_string))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1058
enum amdgpu_ras_block block)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1063
if (block_obj->ras_comm.block == block)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
107
if (ras_block->block == AMDGPU_RAS_BLOCK__MCA)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1070
enum amdgpu_ras_block block, uint32_t sub_block_index)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1075
if (block >= AMDGPU_RAS_BLOCK__LAST)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1086
if (obj->ras_block_match(obj, block, sub_block_index) == 0)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1089
if (amdgpu_ras_block_match_default(obj, block) == 0)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
110
return ras_block_string[ras_block->block];
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1373
head.block = blk;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1444
enum amdgpu_ras_block blk = info ? info->head.block : AMDGPU_RAS_BLOCK_COUNT;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1457
if (info->head.block == AMDGPU_RAS_BLOCK__UMC) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1460
block_obj = amdgpu_ras_get_ras_block(adev, info->head.block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1470
if ((info->head.block == AMDGPU_RAS_BLOCK__SDMA) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1471
(info->head.block == AMDGPU_RAS_BLOCK__GFX) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1472
(info->head.block == AMDGPU_RAS_BLOCK__MMHUB)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1585
req.block_id = info->head.block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1608
enum amdgpu_ras_block block)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1610
struct amdgpu_ras_block_object *block_obj = amdgpu_ras_get_ras_block(adev, block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1616
ras_block_str(block));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1620
if (!amdgpu_ras_is_supported(adev, block) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1640
enum amdgpu_ras_block block)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1642
struct amdgpu_ras_block_object *block_obj = amdgpu_ras_get_ras_block(adev, block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1644
if (amdgpu_ras_reset_error_count(adev, block) == -EOPNOTSUPP)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1647
if ((block == AMDGPU_RAS_BLOCK__GFX) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1648
(block == AMDGPU_RAS_BLOCK__MMHUB)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1666
inject_req.block_id = info->head.block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1683
.block_id = amdgpu_ras_block_to_ta(info->head.block),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1691
info->head.block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1712
info->head.block != AMDGPU_RAS_BLOCK__GFX) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1719
if (info->head.block == AMDGPU_RAS_BLOCK__GFX)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
1768
if (amdgpu_ras_reset_error_status(adev, query_info->head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2083
if (amdgpu_sriov_vf(adev) && !amdgpu_virt_ras_telemetry_block_en(adev, head->block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2267
if (amdgpu_ras_is_supported(adev, obj->head.block) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2410
amdgpu_ras_get_ras_block(adev, obj->head.block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2423
amdgpu_ras_set_err_poison(adev, block_obj->ras_comm.block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2439
amdgpu_umc_poison_handler(adev, obj->head.block, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2535
if (obj->head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2540
if (obj->head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2569
ih_info.block = info->head.block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2696
if (info.head.block == AMDGPU_RAS_BLOCK__PCIE_BIF)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2705
if ((info.head.block == AMDGPU_RAS_BLOCK__UMC) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2718
if (amdgpu_ras_reset_error_status(adev, info.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2733
if ((info->head.block != AMDGPU_RAS_BLOCK__GFX) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2734
(info->head.block != AMDGPU_RAS_BLOCK__MMHUB))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2738
info->head.block,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
289
if (amdgpu_ras_reset_error_status(obj->adev, info.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
3507
enum amdgpu_ras_block block, uint16_t pasid,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
3515
poison_msg.block = block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
3621
info.head.block = AMDGPU_RAS_BLOCK__UMC;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
393
data->head.block = block_id;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4435
if (amdgpu_ras_reset_error_status(adev, ras_block->block) != 0)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4462
if (!amdgpu_ras_is_supported(adev, ras_block->block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
448
switch (data->head.block) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4580
if (!amdgpu_ras_is_supported(adev, obj->head.block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4639
if (!amdgpu_ras_is_supported(adev, obj->ras_comm.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4689
if (amdgpu_ras_is_supported(adev, obj->ras_comm.block) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4760
enum amdgpu_ras_block block)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4766
set_bit(block, &ras->ras_err_state);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4769
bool amdgpu_ras_is_err_state(struct amdgpu_device *adev, int block)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4775
if (block == AMDGPU_RAS_BLOCK__ANY)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
4778
return test_bit(block, &ras->ras_err_state) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5049
unsigned int block)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5054
if (block >= AMDGPU_RAS_BLOCK_COUNT)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5057
ret = ras && (adev->ras_enabled & (1 << block));
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5066
(block == AMDGPU_RAS_BLOCK__GFX ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5067
block == AMDGPU_RAS_BLOCK__SDMA ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5068
block == AMDGPU_RAS_BLOCK__VCN ||
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5069
block == AMDGPU_RAS_BLOCK__JPEG) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5070
(amdgpu_ras_mask & (1 << block)) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5072
amdgpu_ras_get_ras_block(adev, block, 0))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5668
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5684
list_for_each_entry(block, &vres->blocks, link) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5691
region->start = amdgpu_vram_mgr_block_start(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
5692
region->size = amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
582
if (!amdgpu_ras_is_supported(adev, data.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
594
if (data.head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
717
if (amdgpu_ras_reset_error_status(obj->adev, info.head.block))
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
721
if (info.head.block == AMDGPU_RAS_BLOCK__UMC)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
755
if (head->block >= AMDGPU_RAS_BLOCK_COUNT)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
758
if (head->block == AMDGPU_RAS_BLOCK__MCA) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
764
obj = &con->objs[head->block];
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
793
if (head->block >= AMDGPU_RAS_BLOCK_COUNT)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
796
if (head->block == AMDGPU_RAS_BLOCK__MCA) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
802
obj = &con->objs[head->block];
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
822
return adev->ras_hw_enabled & BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
830
return con->features & BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
861
con->features |= BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
864
con->features &= ~BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
886
if (head->block != AMDGPU_RAS_BLOCK__GFX &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
891
if (head->block == AMDGPU_RAS_BLOCK__GFX &&
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
900
.block_id = amdgpu_ras_block_to_ta(head->block),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
905
.block_id = amdgpu_ras_block_to_ta(head->block),
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
966
if (head->block == AMDGPU_RAS_BLOCK__GFX)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
967
con->features |= BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
972
if (adev->ras_enabled && head->block == AMDGPU_RAS_BLOCK__GFX)
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
973
con->features &= ~BIT(head->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
1018
enum amdgpu_ras_block block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
1020
bool amdgpu_ras_is_err_state(struct amdgpu_device *adev, int block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
1032
enum amdgpu_ras_block block, uint16_t pasid,
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
414
enum amdgpu_ras_block block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
470
enum amdgpu_ras_block block;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
763
enum amdgpu_ras_block block, uint32_t sub_block_index);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
815
amdgpu_ras_block_to_ta(enum amdgpu_ras_block block) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
816
switch (block) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
858
WARN_ONCE(1, "RAS ERROR: unexpected block id %d\n", block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
912
enum amdgpu_ras_block block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
914
enum amdgpu_ras_block block);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h
957
int amdgpu_ras_is_supported(struct amdgpu_device *adev, unsigned int block);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
128
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
146
block = cur->node;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
148
next = block->link.next;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
149
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
151
cur->node = block;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
152
cur->start = amdgpu_vram_mgr_block_start(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
153
cur->size = min(amdgpu_vram_mgr_block_size(block), cur->remaining);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
178
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
182
block = cur->node;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
184
if (!amdgpu_vram_mgr_is_cleared(block))
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
58
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
73
block = list_first_entry_or_null(head,
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
76
if (!block)
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
79
while (start >= amdgpu_vram_mgr_block_size(block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
80
start -= amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
82
next = block->link.next;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
84
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
87
cur->start = amdgpu_vram_mgr_block_start(block) + start;
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
88
cur->size = min(amdgpu_vram_mgr_block_size(block) - start, size);
drivers/gpu/drm/amd/amdgpu/amdgpu_res_cursor.h
90
cur->node = block;
drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c
102
if (amdgpu_ras_is_supported(adev, ras_block->block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c
331
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__SDMA;
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
239
enum amdgpu_ras_block block, uint16_t pasid,
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
260
.block = AMDGPU_RAS_BLOCK__UMC,
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
280
ih_info.block = block;
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
291
block, pasid, pasid_fn, data, reset);
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
300
adev->virt.ops->ras_poison_handler(adev, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
310
enum amdgpu_ras_block block, uint32_t reset)
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
313
block, 0, NULL, NULL, reset);
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
341
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__UMC;
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c
365
if (amdgpu_ras_is_supported(adev, ras_block->block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.h
157
enum amdgpu_ras_block block, uint32_t reset);
drivers/gpu/drm/amd/amdgpu/amdgpu_umc.h
159
enum amdgpu_ras_block block, uint16_t pasid,
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c
1279
adev->virt.ops->ras_poison_handler(adev, ras_if->block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c
1296
if (amdgpu_ras_is_supported(adev, ras_block->block)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c
1330
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__VCN;
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1675
amdgpu_ras_block_to_sriov(struct amdgpu_device *adev, enum amdgpu_ras_block block) {
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1676
switch (block) {
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1717
block);
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1772
int amdgpu_virt_req_ras_err_count(struct amdgpu_device *adev, enum amdgpu_ras_block block,
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1777
sriov_block = amdgpu_ras_block_to_sriov(adev, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1789
err_data->ue_count = adev->virt.count_cache.block[sriov_block].ue_count;
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1790
err_data->ce_count = adev->virt.count_cache.block[sriov_block].ce_count;
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1791
err_data->de_count = adev->virt.count_cache.block[sriov_block].de_count;
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1906
enum amdgpu_ras_block block)
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.c
1910
sriov_block = amdgpu_ras_block_to_sriov(adev, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h
111
enum amdgpu_ras_block block);
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h
489
int amdgpu_virt_req_ras_err_count(struct amdgpu_device *adev, enum amdgpu_ras_block block,
drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h
494
enum amdgpu_ras_block block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
257
struct drm_buddy_block *block)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
259
u64 start = amdgpu_vram_mgr_block_start(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
260
u64 end = start + amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
282
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
291
list_for_each_entry(block, &vres->blocks, link)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
292
usage += amdgpu_vram_mgr_vis_size(adev, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
304
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
313
block = amdgpu_vram_mgr_first_block(&rsv->allocated);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
314
if (!block)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
320
vis_usage = amdgpu_vram_mgr_vis_size(adev, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
406
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
412
list_for_each_entry(block, &vres->blocks, link) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
413
start = amdgpu_vram_mgr_block_start(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
414
size = amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
454
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
588
list_for_each_entry(block, &vres->blocks, link) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
591
start = amdgpu_vram_mgr_block_start(block) +
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
592
amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
601
vis_usage += amdgpu_vram_mgr_vis_size(adev, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
63
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
641
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
649
list_for_each_entry(block, &vres->blocks, link)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
650
vis_usage += amdgpu_vram_mgr_vis_size(adev, block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
66
block = amdgpu_vram_mgr_first_block(head);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
67
if (!block)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
70
while (head != block->link.next) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
71
start = amdgpu_vram_mgr_block_start(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
72
size = amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
74
block = list_entry(block->link.next, struct drm_buddy_block, link);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
75
if (start + size != amdgpu_vram_mgr_block_start(block))
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
818
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
821
list_for_each_entry(block, &mgr->blocks, link) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
823
amdgpu_vram_mgr_block_start(block) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
825
(amdgpu_vram_mgr_block_size(block) >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
84
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
851
struct drm_buddy_block *block;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
854
list_for_each_entry(block, &mgr->blocks, link) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
856
amdgpu_vram_mgr_block_start(block) >> PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
858
(amdgpu_vram_mgr_block_size(block) >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
87
list_for_each_entry(block, head, link)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
88
size += amdgpu_vram_mgr_block_size(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.h
60
static inline u64 amdgpu_vram_mgr_block_start(struct drm_buddy_block *block)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.h
62
return drm_buddy_block_offset(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.h
65
static inline u64 amdgpu_vram_mgr_block_size(struct drm_buddy_block *block)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.h
67
return (u64)PAGE_SIZE << drm_buddy_block_order(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.h
70
static inline bool amdgpu_vram_mgr_is_cleared(struct drm_buddy_block *block)
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.h
72
return drm_buddy_block_is_clear(block);
drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c
1617
ras->ras_block.ras_comm.block = AMDGPU_RAS_BLOCK__XGMI_WAFL;
drivers/gpu/drm/amd/amdgpu/amdgv_sriovmsg.h
458
} block[RAS_TELEMETRY_GPU_BLOCK_COUNT];
drivers/gpu/drm/amd/amdgpu/gfx_v11_0_3.c
72
adev->virt.ops->ras_poison_handler(adev, ras_if->block);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6784
block_info.block_id = amdgpu_ras_block_to_ta(info->head.block);
drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c
775
char block[5] = { mc_client >> 24, (mc_client >> 16) & 0xff,
drivers/gpu/drm/amd/amdgpu/gmc_v7_0.c
786
"write" : "read", block, mc_client, mc_id);
drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c
1009
char block[5] = { mc_client >> 24, (mc_client >> 16) & 0xff,
drivers/gpu/drm/amd/amdgpu/gmc_v8_0.c
1020
"write" : "read", block, mc_client, mc_id);
drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c
1461
if (amdgpu_ras_is_supported(adev, ras_block->block) &&
drivers/gpu/drm/amd/amdgpu/jpeg_v5_0_1.c
1085
if (amdgpu_ras_is_supported(adev, ras_block->block) &&
drivers/gpu/drm/amd/amdgpu/mca_v3_0.c
41
enum amdgpu_ras_block block, uint32_t sub_block_index)
drivers/gpu/drm/amd/amdgpu/mca_v3_0.c
46
if ((block_obj->ras_comm.block == block) &&
drivers/gpu/drm/amd/amdgpu/mxgpu_ai.c
459
enum amdgpu_ras_block block)
drivers/gpu/drm/amd/amdgpu/mxgpu_nv.c
545
enum amdgpu_ras_block block)
drivers/gpu/drm/amd/amdgpu/mxgpu_nv.c
552
IDH_RAS_POISON, block, 0, 0);
drivers/gpu/drm/amd/amdgpu/nbio_v7_4.c
662
.block = AMDGPU_RAS_BLOCK__PCIE_BIF,
drivers/gpu/drm/amd/amdgpu/nbio_v7_9.c
686
.block = AMDGPU_RAS_BLOCK__PCIE_BIF,
drivers/gpu/drm/amd/amdgpu/soc15.c
1335
amdgpu_ras_is_supported(adev, adev->nbio.ras_if->block)) {
drivers/gpu/drm/amd/amdgpu/vcn_v4_0_3.c
2204
if (amdgpu_ras_is_supported(adev, ras_block->block) &&
drivers/gpu/drm/amd/amdgpu/vcn_v5_0_1.c
1793
if (amdgpu_ras_is_supported(adev, ras_block->block) &&
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v11.c
219
enum amdgpu_ras_block block = 0;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v11.c
240
block = AMDGPU_RAS_BLOCK__GFX;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v11.c
246
block = AMDGPU_RAS_BLOCK__GFX;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v11.c
255
amdgpu_amdkfd_ras_poison_consumption_handler(dev->adev, block, reset);
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v12_1.c
185
enum amdgpu_ras_block block = 0;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v12_1.c
206
block = AMDGPU_RAS_BLOCK__GFX;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v12_1.c
212
block = AMDGPU_RAS_BLOCK__GFX;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v12_1.c
223
amdgpu_amdkfd_ras_poison_consumption_handler(node->adev, block, reset);
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v9.c
148
enum amdgpu_ras_block block = 0;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v9.c
170
block = AMDGPU_RAS_BLOCK__GFX;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v9.c
192
block = AMDGPU_RAS_BLOCK__MMHUB;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v9.c
200
block = AMDGPU_RAS_BLOCK__SDMA;
drivers/gpu/drm/amd/amdkfd/kfd_int_process_v9.c
242
block, pasid, NULL, NULL, reset);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c
936
dm_helpers_probe_acpi_edid(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_helpers.c
940
unsigned short start = block * EDID_LENGTH;
drivers/gpu/drm/amd/display/dc/clk_mgr/dce110/dce110_clk_mgr.c
39
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/clk_mgr/dce110/dce110_clk_mgr.c
40
.reg_name = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/clk_mgr/dce112/dce112_clk_mgr.c
41
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/clk_mgr/dce112/dce112_clk_mgr.c
42
.reg_name = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
64
#define CLK_SRI(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
65
.reg_name = mm ## block ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn32/dcn32_clk_mgr.c
115
#define CLK_SR_DCN321(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn32/dcn32_clk_mgr.c
116
.reg_name = mm ## block ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn401/dcn401_clk_mgr.c
57
#define CLK_SR_DCN401(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn401/dcn401_clk_mgr.c
58
.reg_name = mm ## block ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/dccg/dcn20/dcn20_dccg.h
62
#define DCCG_SFII(block, reg_name, field_prefix, field_name, inst, post_fix)\
drivers/gpu/drm/amd/display/dc/dccg/dcn20/dcn20_dccg.h
63
.field_prefix ## _ ## field_name[inst] = block ## inst ## _ ## reg_name ## __ ## field_prefix ## inst ## _ ## field_name ## post_fix
drivers/gpu/drm/amd/display/dc/dccg/dcn314/dcn314_dccg.h
32
#define DCCG_SFII(block, reg_name, field_prefix, field_name, inst, post_fix)\
drivers/gpu/drm/amd/display/dc/dccg/dcn314/dcn314_dccg.h
33
.field_prefix ## _ ## field_name[inst] = block ## inst ## _ ## reg_name ## __ ## field_prefix ## inst ## _ ## field_name ## post_fix
drivers/gpu/drm/amd/display/dc/dccg/dcn32/dcn32_dccg.h
31
#define DCCG_SFII(block, reg_name, field_prefix, field_name, inst, post_fix)\
drivers/gpu/drm/amd/display/dc/dccg/dcn32/dcn32_dccg.h
32
.field_prefix ## _ ## field_name[inst] = block ## inst ## _ ## reg_name ## __ ## field_prefix ## inst ## _ ## field_name ## post_fix
drivers/gpu/drm/amd/display/dc/dccg/dcn35/dcn35_dccg.h
30
#define DCCG_SFII(block, reg_name, field_prefix, field_name, inst, post_fix)\
drivers/gpu/drm/amd/display/dc/dccg/dcn35/dcn35_dccg.h
31
.field_prefix ## _ ## field_name[inst] = block ## inst ## _ ## reg_name ## __ ## field_prefix ## inst ## _ ## field_name ## post_fix
drivers/gpu/drm/amd/display/dc/dccg/dcn401/dcn401_dccg.h
31
#define DCCG_SFII(block, reg_name, field_prefix, field_name, inst, post_fix)\
drivers/gpu/drm/amd/display/dc/dccg/dcn401/dcn401_dccg.h
32
.field_prefix ## _ ## field_name[inst] = block ## inst ## _ ## reg_name ## __ ## field_prefix ## inst ## _ ## field_name ## post_fix
drivers/gpu/drm/amd/display/dc/dce/dce_panel_cntl.h
32
#define DCE_PANEL_CNTL_SR(reg_name, block)\
drivers/gpu/drm/amd/display/dc/dce/dce_panel_cntl.h
33
.reg_name = mm ## block ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/dce/dce_panel_cntl.h
45
#define DCN_PANEL_CNTL_SR(reg_name, block)\
drivers/gpu/drm/amd/display/dc/dce/dce_panel_cntl.h
46
.reg_name = BASE(mm ## block ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/dce/dce_panel_cntl.h
47
mm ## block ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
38
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
39
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
40
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
43
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
44
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
45
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/dm_services.h
165
#define get_reg_field_value_soc15(reg_value, block, reg_num, reg_name, reg_field)\
drivers/gpu/drm/amd/display/dc/dm_services.h
168
block ## reg_num ## _ ## reg_name ## __ ## reg_field ## _MASK,\
drivers/gpu/drm/amd/display/dc/dm_services.h
169
block ## reg_num ## _ ## reg_name ## __ ## reg_field ## __SHIFT)
drivers/gpu/drm/amd/display/dc/dm_services.h
171
#define set_reg_field_value_soc15(reg_value, value, block, reg_num, reg_name, reg_field)\
drivers/gpu/drm/amd/display/dc/dm_services.h
175
block ## reg_num ## _ ## reg_name ## __ ## reg_field ## _MASK,\
drivers/gpu/drm/amd/display/dc/dm_services.h
176
block ## reg_num ## _ ## reg_name ## __ ## reg_field ## __SHIFT)
drivers/gpu/drm/amd/display/dc/gpio/dce110/hw_factory_dce110.c
48
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dce110/hw_factory_dce110.c
49
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_factory_dce120.c
63
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_factory_dce120.c
64
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_factory_dce120.c
65
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_translate_dce120.c
54
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_translate_dce120.c
55
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_translate_dce120.c
56
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_factory_dcn10.c
60
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_factory_dcn10.c
61
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_factory_dcn10.c
62
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_translate_dcn10.c
54
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_translate_dcn10.c
55
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_translate_dcn10.c
56
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn20/hw_factory_dcn20.c
65
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn20/hw_factory_dcn20.c
66
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn20/hw_factory_dcn20.c
67
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn21/hw_factory_dcn21.c
63
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn21/hw_factory_dcn21.c
64
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn21/hw_factory_dcn21.c
65
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn30/hw_factory_dcn30.c
72
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn30/hw_factory_dcn30.c
73
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn30/hw_factory_dcn30.c
74
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn315/hw_factory_dcn315.c
69
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn315/hw_factory_dcn315.c
70
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn315/hw_factory_dcn315.c
71
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn32/hw_factory_dcn32.c
65
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn32/hw_factory_dcn32.c
66
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn32/hw_factory_dcn32.c
67
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/gpio/dcn401/hw_factory_dcn401.c
45
#define REGI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/gpio/dcn401/hw_factory_dcn401.c
46
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/gpio/dcn401/hw_factory_dcn401.c
47
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/inc/hw/clk_mgr_internal.h
84
#define CLK_SRI(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/inc/hw/clk_mgr_internal.h
85
.reg_name = CLK_BASE(mm ## block ## _ ## inst ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/inc/hw/clk_mgr_internal.h
86
mm ## block ## _ ## inst ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
71
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
72
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
73
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
76
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
77
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
79
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
81
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
82
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
84
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
86
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
88
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
168
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
169
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
170
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
173
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
174
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
176
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
178
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
179
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
181
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
183
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
185
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
171
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
172
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
173
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
176
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
177
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
179
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
181
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
182
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
184
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
186
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
188
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
121
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
122
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
123
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
125
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
126
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
128
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
130
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
131
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
133
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
135
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
137
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
178
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
179
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
180
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
186
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
187
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
189
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
191
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
192
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
194
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
196
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
198
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
185
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
186
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
187
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
193
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
194
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
196
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
198
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
199
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
201
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
203
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
205
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
170
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
171
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
172
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
178
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
179
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
180
.enable_mask = block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
182
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
183
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
185
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
186
.ack_mask = block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
187
.ack_value = block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
116
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
117
BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
118
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
121
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
122
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
123
.enable_mask = block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
125
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
126
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
128
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
129
.ack_mask = block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
130
.ack_value = block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
173
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
174
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
175
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
181
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
182
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
184
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
186
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
187
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
189
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
191
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
193
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
175
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
176
(BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
177
reg ## block ## id ## _ ## reg_name)
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
183
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
184
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
186
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
188
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
189
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
191
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
193
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
195
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
180
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
181
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
182
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
188
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
189
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
191
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
193
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
194
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
196
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
198
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
200
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
184
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
185
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
186
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
192
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
193
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
195
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
197
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
198
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
200
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
202
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
204
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
172
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
173
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
174
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
180
#define IRQ_REG_ENTRY(base, block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
181
REG_STRUCT[base + reg_num].enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
183
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
185
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
187
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK, \
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
188
REG_STRUCT[base + reg_num].ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
190
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
192
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
151
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
152
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
153
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
159
#define IRQ_REG_ENTRY(base, block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
160
REG_STRUCT[base + reg_num].enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
162
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
164
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
166
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK, \
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
167
REG_STRUCT[base + reg_num].ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
169
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
171
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
150
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
151
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
152
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
158
#define IRQ_REG_ENTRY(base, block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
159
REG_STRUCT[base + reg_num].enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
161
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
163
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
165
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK, \
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
166
REG_STRUCT[base + reg_num].ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
168
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
170
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
164
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
165
BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
166
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
172
#define IRQ_REG_ENTRY(block, reg_num, reg1, mask1, reg2, mask2)\
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
173
.enable_reg = SRI(reg1, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
175
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
177
block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
178
~block ## reg_num ## _ ## reg1 ## __ ## mask1 ## _MASK \
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
180
.ack_reg = SRI(reg2, block, reg_num),\
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
182
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK,\
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
184
block ## reg_num ## _ ## reg2 ## __ ## mask2 ## _MASK \
drivers/gpu/drm/amd/display/dc/mpc/dcn30/dcn30_mpc.h
38
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/mpc/dcn30/dcn30_mpc.h
39
.RMU##_##reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/mpc/dcn30/dcn30_mpc.h
40
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce100/dce100_resource.c
141
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce100/dce100_resource.c
142
.reg_name = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce100/dce100_resource.c
505
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce100/dce100_resource.c
506
.reg_name[id] = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce110/dce110_resource.c
149
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce110/dce110_resource.c
150
.reg_name = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce110/dce110_resource.c
544
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce110/dce110_resource.c
545
.reg_name[id] = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce112/dce112_resource.c
150
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce112/dce112_resource.c
151
.reg_name = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce112/dce112_resource.c
525
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce112/dce112_resource.c
526
.reg_name[id] = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
141
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
142
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
143
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
784
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
785
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
786
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce60/dce60_resource.c
158
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce60/dce60_resource.c
159
.reg_name = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce60/dce60_resource.c
623
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce60/dce60_resource.c
624
.reg_name[id] = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce80/dce80_resource.c
157
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce80/dce80_resource.c
158
.reg_name = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dce80/dce80_resource.c
629
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dce80/dce80_resource.c
630
.reg_name[id] = mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
117
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
118
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
119
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
122
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
123
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
124
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
126
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
127
.reg_name[id] = BASE(mm ## reg_name ## 0 ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
128
mm ## reg_name ## 0 ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
133
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
134
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
135
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
137
#define SRI2_DWB(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
143
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
146
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
147
.var_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
148
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
150
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
151
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
152
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
154
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
155
.block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
156
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
158
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
159
.reg_name[id] = BASE(mm ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
160
mm ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
256
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
257
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
258
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
260
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
261
.var_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
262
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
264
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
265
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
266
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
268
#define SRI_IX(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
269
.reg_name = ix ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
271
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
272
.block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
273
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
275
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
276
.reg_name[id] = BASE(mm ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
277
mm ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
106
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
107
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
108
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
110
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
111
.var_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
112
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
114
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
115
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
116
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
118
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
119
.block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
120
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
122
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
123
.reg_name[id] = BASE(mm ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
124
mm ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
122
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
123
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
124
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
126
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
130
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
131
.var_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
132
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
134
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
135
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
136
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
138
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
139
.RMU##_##reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
140
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
142
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
143
.reg_name[id] = BASE(mm ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
144
mm ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
146
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
149
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
150
.block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
151
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
153
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
154
.reg_name[id] = BASE(mm ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
155
mm ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
186
#define CLK_SRI(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
187
.reg_name = CLK_BASE(mm ## block ## _ ## inst ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
188
mm ## block ## _ ## inst ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
121
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
122
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
123
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
125
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
129
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
130
.var_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
131
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
133
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
134
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
135
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
142
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
143
.RMU##_##reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
144
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
146
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
147
.reg_name[id] = BASE(mm ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
148
mm ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
150
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
153
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
154
.block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
155
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
157
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
158
.reg_name[id] = BASE(mm ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
159
mm ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
190
#define CLK_SRI(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
191
.reg_name = CLK_BASE(mm ## block ## _ ## inst ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
192
mm ## block ## _ ## inst ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
186
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
187
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
189
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
192
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
193
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
194
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
196
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
197
.block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
198
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
200
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
201
.reg_name[id] = BASE(mm ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
202
mm ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
204
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
205
.reg_name[id] = BASE(mm ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
206
mm ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
208
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
211
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
212
.RMU##_##reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
213
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
182
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
183
.reg_name = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
185
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
188
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
189
.reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
190
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
192
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
193
.block ## _ ## reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
194
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
196
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
197
.reg_name[id] = BASE(mm ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
198
mm ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
200
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
201
.reg_name[id] = BASE(mm ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
202
mm ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
204
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
207
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
208
.RMU##_##reg_name[id] = BASE(mm ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
209
mm ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
133
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
134
.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
135
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
137
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
141
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
142
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
143
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
145
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
146
.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
147
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
149
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
150
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
151
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
153
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
154
.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
155
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
157
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
160
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
161
.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
162
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
164
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
165
.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
166
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
197
#define CLK_SRI(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
198
.reg_name = CLK_BASE(reg ## block ## _ ## inst ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
199
reg ## block ## _ ## inst ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
150
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
151
.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
152
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
154
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
158
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
159
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
160
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
162
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
163
.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
164
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
166
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
167
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
168
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
170
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
171
.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
172
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
174
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
177
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
178
.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
179
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
181
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
182
.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
183
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
214
#define CLK_SRI(reg_name, block, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
215
.reg_name = CLK_BASE(reg ## block ## _ ## inst ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
216
reg ## block ## _ ## inst ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
167
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
168
.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
169
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
171
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
175
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
176
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
177
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
179
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
180
.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
181
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
183
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
184
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
185
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
187
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
188
.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
189
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
191
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
194
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
195
.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
196
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
198
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
199
.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
200
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
153
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
154
.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
155
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
157
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
161
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
162
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
163
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
165
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
166
.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
167
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
169
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
170
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
171
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
173
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
174
.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
175
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
177
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
180
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
181
.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
182
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
184
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
185
.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
186
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
127
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
128
REG_STRUCT.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
129
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
131
#define SRI_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
132
REG_STRUCT[id].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
133
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
138
#define SRI_ARR_I2C(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
139
REG_STRUCT[id-1].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
140
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
142
#define SRI_ARR_ALPHABET(reg_name, block, index, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
143
REG_STRUCT[index].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
144
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
146
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
149
#define SRI2_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
153
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
154
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
155
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
157
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
158
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
159
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
161
#define SRII_ARR_2(reg_name, block, id, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
162
REG_STRUCT[inst].reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
163
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
165
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
166
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
167
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
169
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
170
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
171
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
173
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
176
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
177
REG_STRUCT.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
178
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
180
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
181
REG_STRUCT.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
182
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
127
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
128
REG_STRUCT.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
129
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
131
#define SRI_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
132
REG_STRUCT[id].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
133
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
138
#define SRI_ARR_I2C(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
139
REG_STRUCT[id-1].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
140
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
142
#define SRI_ARR_ALPHABET(reg_name, block, index, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
143
REG_STRUCT[index].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
144
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
146
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
149
#define SRI2_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
153
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
154
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
155
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
157
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
158
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
159
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
161
#define SRII_ARR_2(reg_name, block, id, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
162
REG_STRUCT[inst].reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
163
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
165
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
166
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
167
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
169
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
170
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
171
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
173
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
174
REG_STRUCT.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
175
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
177
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
180
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
181
REG_STRUCT.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
182
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
142
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
143
REG_STRUCT.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
144
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
146
#define SRI_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
147
REG_STRUCT[id].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
148
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
153
#define SRI_ARR_I2C(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
154
REG_STRUCT[id-1].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
155
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
157
#define SRI_ARR_ALPHABET(reg_name, block, index, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
158
REG_STRUCT[index].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
159
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
161
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
165
#define SRI2_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
169
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
170
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
171
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
173
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
174
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
175
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
177
#define SRII_ARR_2(reg_name, block, id, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
178
REG_STRUCT[inst].reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
179
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
181
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
182
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
183
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
185
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
186
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
187
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
189
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
192
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
193
REG_STRUCT.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
194
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
196
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
197
REG_STRUCT.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
198
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
122
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
123
REG_STRUCT.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
124
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
126
#define SRI_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
127
REG_STRUCT[id].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
128
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
133
#define SRI_ARR_I2C(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
134
REG_STRUCT[id-1].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
135
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
137
#define SRI_ARR_ALPHABET(reg_name, block, index, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
138
REG_STRUCT[index].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
139
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
141
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
145
#define SRI2_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
149
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
150
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
151
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
153
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
154
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
155
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
157
#define SRII_ARR_2(reg_name, block, id, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
158
REG_STRUCT[inst].reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
159
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
161
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
162
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
163
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
165
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
166
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
167
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
169
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
172
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
173
REG_STRUCT.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
174
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
176
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
177
REG_STRUCT.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
178
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
127
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
128
REG_STRUCT.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
129
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
131
#define SRI_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
132
REG_STRUCT[id].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
133
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
138
#define SRI_ARR_I2C(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
139
REG_STRUCT[id-1].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
140
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
142
#define SRI_ARR_ALPHABET(reg_name, block, index, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
143
REG_STRUCT[index].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
144
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
146
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
150
#define SRI2_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
154
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
155
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
156
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
158
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
159
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
160
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
162
#define SRII_ARR_2(reg_name, block, id, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
163
REG_STRUCT[inst].reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
164
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
166
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
167
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
168
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
170
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
171
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
172
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
174
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
177
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
178
REG_STRUCT.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
179
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
181
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
182
REG_STRUCT.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
183
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
108
#define SRI(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
109
REG_STRUCT.reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
110
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
112
#define SRI_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
113
REG_STRUCT[id].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
114
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
119
#define SRI_ARR_US(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
120
REG_STRUCT[id].reg_name = BASE(reg ## block ## id ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
121
reg ## block ## id ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
125
#define SRI_ARR_I2C(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
126
REG_STRUCT[id-1].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
127
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
129
#define SRI_ARR_ALPHABET(reg_name, block, index, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
130
REG_STRUCT[index].reg_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
131
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
133
#define SRI2(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
136
#define SRI2_ARR(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
140
#define SRIR(var_name, reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
141
.var_name = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
142
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
144
#define SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
145
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
146
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
148
#define SRII_ARR_2(reg_name, block, id, inst)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
149
REG_STRUCT[inst].reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
150
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
152
#define SRII_MPC_RMU(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
153
.RMU##_##reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
154
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
156
#define SRII_DWB(reg_name, temp_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
157
REG_STRUCT.reg_name[id] = BASE(reg ## block ## id ## _ ## temp_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
158
reg ## block ## id ## _ ## temp_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
160
#define DCCG_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
161
REG_STRUCT.block ## _ ## reg_name[id] = BASE(reg ## block ## id ## _ ## reg_name ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
162
reg ## block ## id ## _ ## reg_name
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
164
#define SF_DWB2(reg_name, block, id, field_name, post_fix) \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
167
#define VUPDATE_SRII(reg_name, block, id)\
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
168
REG_STRUCT.reg_name[id] = BASE(reg ## reg_name ## _ ## block ## id ## _BASE_IDX) + \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
169
reg ## reg_name ## _ ## block ## id
drivers/gpu/drm/amd/include/kgd_pp_interface.h
347
#define PP_CG_MSG_ID(group, block, support, state) \
drivers/gpu/drm/amd/include/kgd_pp_interface.h
348
((group) << PP_GROUP_SHIFT | (block) << PP_BLOCK_SHIFT | \
drivers/gpu/drm/amd/ras/ras_mgr/amdgpu_ras_mgr.c
519
if (ih_info && (ih_info->block == AMDGPU_RAS_BLOCK__UMC)) {
drivers/gpu/drm/amd/ras/ras_mgr/amdgpu_ras_mgr.h
36
uint32_t block;
drivers/gpu/drm/amd/ras/ras_mgr/amdgpu_ras_process.c
107
req.block = ih_info->block;
drivers/gpu/drm/amd/ras/rascore/ras.h
339
enum ras_block_id block, struct ras_ecc_count *ecc_count);
drivers/gpu/drm/amd/ras/rascore/ras_cmd.c
342
static enum ras_ta_block __get_ras_ta_block(enum ras_block_id block)
drivers/gpu/drm/amd/ras/rascore/ras_cmd.c
344
switch (block) {
drivers/gpu/drm/amd/ras/rascore/ras_core.c
480
enum ras_block_id block, struct ras_ecc_count *ecc_count)
drivers/gpu/drm/amd/ras/rascore/ras_core.c
484
if (!ecc_count || (block >= RAS_BLOCK_ID__LAST) || !ras_core)
drivers/gpu/drm/amd/ras/rascore/ras_core.c
487
ret = ras_aca_get_block_ecc_count(ras_core, block, ecc_count);
drivers/gpu/drm/amd/ras/rascore/ras_core.c
489
ras_aca_clear_block_new_ecc_count(ras_core, block);
drivers/gpu/drm/amd/ras/rascore/ras_process.c
142
req.seqno, ras_core_get_ras_block_name(req.block));
drivers/gpu/drm/amd/ras/rascore/ras_process.c
146
req.seqno, ras_core_get_ras_block_name(req.block));
drivers/gpu/drm/amd/ras/rascore/ras_process.h
30
uint32_t block;
drivers/gpu/drm/arm/malidp_hw.c
1158
static void malidp_hw_clear_irq(struct malidp_hw_device *hwdev, u8 block, u32 irq)
drivers/gpu/drm/arm/malidp_hw.c
1160
u32 base = malidp_get_block_base(hwdev, block);
drivers/gpu/drm/arm/malidp_hw.h
291
u8 block)
drivers/gpu/drm/arm/malidp_hw.h
293
switch (block) {
drivers/gpu/drm/arm/malidp_hw.h
304
u8 block, u32 irq)
drivers/gpu/drm/arm/malidp_hw.h
306
u32 base = malidp_get_block_base(hwdev, block);
drivers/gpu/drm/arm/malidp_hw.h
312
u8 block, u32 irq)
drivers/gpu/drm/arm/malidp_hw.h
314
u32 base = malidp_get_block_base(hwdev, block);
drivers/gpu/drm/ast/ast_dp.c
109
offset = (i + block * EDID_LENGTH) / 4;
drivers/gpu/drm/ast/ast_dp.c
84
static int ast_astdp_read_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/ast/ast_dp.c
91
if (block > 0)
drivers/gpu/drm/ast/ast_dp501.c
317
static int ast_dp512_read_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/ast/ast_dp501.c
323
if (block > (512 / EDID_LENGTH))
drivers/gpu/drm/ast/ast_dp501.c
326
offset = AST_DP501_EDID_DATA + block * EDID_LENGTH;
drivers/gpu/drm/bridge/adv7511/adv7511_drv.c
516
static int adv7511_get_edid_block(void *data, u8 *buf, unsigned int block,
drivers/gpu/drm/bridge/adv7511/adv7511_drv.c
528
if (adv7511->current_edid_segment != block / 2) {
drivers/gpu/drm/bridge/adv7511/adv7511_drv.c
539
block);
drivers/gpu/drm/bridge/adv7511/adv7511_drv.c
572
adv7511->current_edid_segment = block / 2;
drivers/gpu/drm/bridge/adv7511/adv7511_drv.c
575
if (block % 2 == 0)
drivers/gpu/drm/bridge/analogix/analogix_dp_core.h
190
enum analog_power_block block,
drivers/gpu/drm/bridge/analogix/analogix_dp_reg.c
250
enum analog_power_block block,
drivers/gpu/drm/bridge/analogix/analogix_dp_reg.c
260
switch (block) {
drivers/gpu/drm/bridge/cadence/cdns-mhdp8546-core.c
415
unsigned int block, size_t length)
drivers/gpu/drm/bridge/cadence/cdns-mhdp8546-core.c
424
msg[0] = block / 2;
drivers/gpu/drm/bridge/cadence/cdns-mhdp8546-core.c
425
msg[1] = block % 2;
drivers/gpu/drm/bridge/cadence/cdns-mhdp8546-core.c
446
if (reg[0] == length && reg[1] == block / 2)
drivers/gpu/drm/bridge/cadence/cdns-mhdp8546-core.c
454
block, ret);
drivers/gpu/drm/bridge/ite-it6263.c
523
static int it6263_read_edid(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/bridge/ite-it6263.c
527
unsigned int start = (block % 2) * EDID_LENGTH;
drivers/gpu/drm/bridge/ite-it6263.c
528
unsigned int segment = block >> 1;
drivers/gpu/drm/bridge/ite-it6505.c
1293
static int it6505_get_edid_block(void *data, u8 *buf, unsigned int block,
drivers/gpu/drm/bridge/ite-it6505.c
1302
DRM_DEV_DEBUG_DRIVER(dev, "block number = %d", block);
drivers/gpu/drm/bridge/ite-it6505.c
1306
block * EDID_LENGTH + offset,
drivers/gpu/drm/bridge/ite-it66121.c
518
unsigned int block, size_t len)
drivers/gpu/drm/bridge/ite-it66121.c
525
offset = (block % 2) * len;
drivers/gpu/drm/bridge/ite-it66121.c
526
block = block / 2;
drivers/gpu/drm/bridge/ite-it66121.c
549
ret = regmap_write(ctx->regmap, IT66121_DDC_SEGMENT_REG, block);
drivers/gpu/drm/bridge/lontium-lt9611.c
616
lt9611_get_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/bridge/lontium-lt9611.c
626
if (block > 1)
drivers/gpu/drm/bridge/lontium-lt9611.c
629
if (block == 0) {
drivers/gpu/drm/bridge/lontium-lt9611.c
637
block %= 2;
drivers/gpu/drm/bridge/lontium-lt9611.c
638
memcpy(buf, lt9611->edid_buf + (block * 128), len);
drivers/gpu/drm/bridge/lontium-lt9611uxc.c
387
static int lt9611uxc_get_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/bridge/lontium-lt9611uxc.c
395
if (block >= EDID_NUM_BLOCKS)
drivers/gpu/drm/bridge/lontium-lt9611uxc.c
402
regmap_write(lt9611uxc->regmap, 0xb00a, block * EDID_BLOCK_SIZE);
drivers/gpu/drm/bridge/megachips-stdpxxxx-ge-b850v3-fw.c
68
static int stdp2690_read_block(void *context, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/bridge/megachips-stdpxxxx-ge-b850v3-fw.c
72
unsigned char start = block * EDID_LENGTH;
drivers/gpu/drm/drm_buddy.c
1006
block = list_first_entry(blocks,
drivers/gpu/drm/drm_buddy.c
1010
block_start = drm_buddy_block_offset(block);
drivers/gpu/drm/drm_buddy.c
1011
block_end = block_start + drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
1013
if (WARN_ON(!drm_buddy_block_is_allocated(block)))
drivers/gpu/drm/drm_buddy.c
1016
if (new_size > drm_buddy_block_size(mm, block))
drivers/gpu/drm/drm_buddy.c
1022
if (new_size == drm_buddy_block_size(mm, block))
drivers/gpu/drm/drm_buddy.c
103
struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
1039
list_del(&block->link);
drivers/gpu/drm/drm_buddy.c
1040
mark_free(mm, block);
drivers/gpu/drm/drm_buddy.c
1041
mm->avail += drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
1042
if (drm_buddy_block_is_clear(block))
drivers/gpu/drm/drm_buddy.c
1043
mm->clear_avail += drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
1046
parent = block->parent;
drivers/gpu/drm/drm_buddy.c
1047
block->parent = NULL;
drivers/gpu/drm/drm_buddy.c
1049
list_add(&block->tmp_link, &dfs);
drivers/gpu/drm/drm_buddy.c
105
unsigned int order = drm_buddy_block_order(block);
drivers/gpu/drm/drm_buddy.c
1052
mark_allocated(mm, block);
drivers/gpu/drm/drm_buddy.c
1053
mm->avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
1054
if (drm_buddy_block_is_clear(block))
drivers/gpu/drm/drm_buddy.c
1055
mm->clear_avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
1056
list_add(&block->link, blocks);
drivers/gpu/drm/drm_buddy.c
1059
block->parent = parent;
drivers/gpu/drm/drm_buddy.c
109
tree = get_block_tree(block);
drivers/gpu/drm/drm_buddy.c
1105
struct drm_buddy_block *block = NULL;
drivers/gpu/drm/drm_buddy.c
112
rb_erase(&block->rb, root);
drivers/gpu/drm/drm_buddy.c
113
RB_CLEAR_NODE(&block->rb);
drivers/gpu/drm/drm_buddy.c
116
static void clear_reset(struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
1169
block = __drm_buddy_alloc_blocks(mm, start,
drivers/gpu/drm/drm_buddy.c
1173
if (!IS_ERR(block))
drivers/gpu/drm/drm_buddy.c
118
block->header &= ~DRM_BUDDY_HEADER_CLEAR;
drivers/gpu/drm/drm_buddy.c
1180
block = __drm_buddy_alloc_blocks(mm, start,
drivers/gpu/drm/drm_buddy.c
1184
if (!IS_ERR(block)) {
drivers/gpu/drm/drm_buddy.c
1205
mark_allocated(mm, block);
drivers/gpu/drm/drm_buddy.c
1206
mm->avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
1207
if (drm_buddy_block_is_clear(block))
drivers/gpu/drm/drm_buddy.c
1208
mm->clear_avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
1209
kmemleak_update_trace(block);
drivers/gpu/drm/drm_buddy.c
121
static void mark_cleared(struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
1210
list_add_tail(&block->link, &allocated);
drivers/gpu/drm/drm_buddy.c
1229
block = list_last_entry(&allocated, typeof(*block), link);
drivers/gpu/drm/drm_buddy.c
123
block->header |= DRM_BUDDY_HEADER_CLEAR;
drivers/gpu/drm/drm_buddy.c
1230
list_move(&block->link, &temp);
drivers/gpu/drm/drm_buddy.c
1232
trim_size = drm_buddy_block_size(mm, block) -
drivers/gpu/drm/drm_buddy.c
1262
struct drm_buddy_block *block,
drivers/gpu/drm/drm_buddy.c
1265
u64 start = drm_buddy_block_offset(block);
drivers/gpu/drm/drm_buddy.c
1266
u64 size = drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
127
struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
1286
struct drm_buddy_block *block, *tmp;
drivers/gpu/drm/drm_buddy.c
129
block->header &= ~DRM_BUDDY_HEADER_STATE;
drivers/gpu/drm/drm_buddy.c
1294
rbtree_postorder_for_each_entry_safe(block, tmp, root, rb) {
drivers/gpu/drm/drm_buddy.c
1295
BUG_ON(!drm_buddy_block_is_free(block));
drivers/gpu/drm/drm_buddy.c
130
block->header |= DRM_BUDDY_ALLOCATED;
drivers/gpu/drm/drm_buddy.c
132
rbtree_remove(mm, block);
drivers/gpu/drm/drm_buddy.c
136
struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
140
block->header &= ~DRM_BUDDY_HEADER_STATE;
drivers/gpu/drm/drm_buddy.c
141
block->header |= DRM_BUDDY_FREE;
drivers/gpu/drm/drm_buddy.c
143
tree = get_block_tree(block);
drivers/gpu/drm/drm_buddy.c
144
rbtree_insert(mm, block, tree);
drivers/gpu/drm/drm_buddy.c
148
struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
150
block->header &= ~DRM_BUDDY_HEADER_STATE;
drivers/gpu/drm/drm_buddy.c
151
block->header |= DRM_BUDDY_SPLIT;
drivers/gpu/drm/drm_buddy.c
153
rbtree_remove(mm, block);
drivers/gpu/drm/drm_buddy.c
167
__get_buddy(struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
171
parent = block->parent;
drivers/gpu/drm/drm_buddy.c
175
if (parent->left == block)
drivers/gpu/drm/drm_buddy.c
182
struct drm_buddy_block *block,
drivers/gpu/drm/drm_buddy.c
188
while ((parent = block->parent)) {
drivers/gpu/drm/drm_buddy.c
191
buddy = __get_buddy(block);
drivers/gpu/drm/drm_buddy.c
201
if (drm_buddy_block_is_clear(block) !=
drivers/gpu/drm/drm_buddy.c
205
if (drm_buddy_block_is_clear(block))
drivers/gpu/drm/drm_buddy.c
213
drm_block_free(mm, block);
drivers/gpu/drm/drm_buddy.c
216
block = parent;
drivers/gpu/drm/drm_buddy.c
219
order = drm_buddy_block_order(block);
drivers/gpu/drm/drm_buddy.c
220
mark_free(mm, block);
drivers/gpu/drm/drm_buddy.c
244
struct drm_buddy_block *block, *buddy;
drivers/gpu/drm/drm_buddy.c
247
block = rbtree_get_free_block(iter);
drivers/gpu/drm/drm_buddy.c
250
if (!block || !block->parent)
drivers/gpu/drm/drm_buddy.c
253
block_start = drm_buddy_block_offset(block);
drivers/gpu/drm/drm_buddy.c
254
block_end = block_start + drm_buddy_block_size(mm, block) - 1;
drivers/gpu/drm/drm_buddy.c
259
buddy = __get_buddy(block);
drivers/gpu/drm/drm_buddy.c
263
WARN_ON(drm_buddy_block_is_clear(block) ==
drivers/gpu/drm/drm_buddy.c
273
rbtree_remove(mm, block);
drivers/gpu/drm/drm_buddy.c
274
if (drm_buddy_block_is_clear(block))
drivers/gpu/drm/drm_buddy.c
275
mm->clear_avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
277
order = __drm_buddy_free(mm, block, true);
drivers/gpu/drm/drm_buddy.c
32
struct drm_buddy_block *block;
drivers/gpu/drm/drm_buddy.c
36
block = kmem_cache_zalloc(slab_blocks, GFP_KERNEL);
drivers/gpu/drm/drm_buddy.c
37
if (!block)
drivers/gpu/drm/drm_buddy.c
40
block->header = offset;
drivers/gpu/drm/drm_buddy.c
41
block->header |= order;
drivers/gpu/drm/drm_buddy.c
42
block->parent = parent;
drivers/gpu/drm/drm_buddy.c
424
struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
426
unsigned int block_order = drm_buddy_block_order(block) - 1;
drivers/gpu/drm/drm_buddy.c
427
u64 offset = drm_buddy_block_offset(block);
drivers/gpu/drm/drm_buddy.c
429
BUG_ON(!drm_buddy_block_is_free(block));
drivers/gpu/drm/drm_buddy.c
430
BUG_ON(!drm_buddy_block_order(block));
drivers/gpu/drm/drm_buddy.c
432
block->left = drm_block_alloc(mm, block, block_order, offset);
drivers/gpu/drm/drm_buddy.c
433
if (!block->left)
drivers/gpu/drm/drm_buddy.c
436
block->right = drm_block_alloc(mm, block, block_order,
drivers/gpu/drm/drm_buddy.c
438
if (!block->right) {
drivers/gpu/drm/drm_buddy.c
439
drm_block_free(mm, block->left);
drivers/gpu/drm/drm_buddy.c
44
RB_CLEAR_NODE(&block->rb);
drivers/gpu/drm/drm_buddy.c
443
mark_split(mm, block);
drivers/gpu/drm/drm_buddy.c
445
if (drm_buddy_block_is_clear(block)) {
drivers/gpu/drm/drm_buddy.c
446
mark_cleared(block->left);
drivers/gpu/drm/drm_buddy.c
447
mark_cleared(block->right);
drivers/gpu/drm/drm_buddy.c
448
clear_reset(block);
drivers/gpu/drm/drm_buddy.c
451
mark_free(mm, block->left);
drivers/gpu/drm/drm_buddy.c
452
mark_free(mm, block->right);
drivers/gpu/drm/drm_buddy.c
46
BUG_ON(block->header & DRM_BUDDY_HEADER_UNUSED);
drivers/gpu/drm/drm_buddy.c
468
drm_get_buddy(struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
47
return block;
drivers/gpu/drm/drm_buddy.c
470
return __get_buddy(block);
drivers/gpu/drm/drm_buddy.c
505
struct drm_buddy_block *block, *tmp;
drivers/gpu/drm/drm_buddy.c
507
rbtree_postorder_for_each_entry_safe(block, tmp, root, rb) {
drivers/gpu/drm/drm_buddy.c
508
rbtree_remove(mm, block);
drivers/gpu/drm/drm_buddy.c
51
struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
510
mark_cleared(block);
drivers/gpu/drm/drm_buddy.c
511
mm->clear_avail += drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
513
clear_reset(block);
drivers/gpu/drm/drm_buddy.c
514
mm->clear_avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
517
rbtree_insert(mm, block, dst_tree);
drivers/gpu/drm/drm_buddy.c
53
kmem_cache_free(slab_blocks, block);
drivers/gpu/drm/drm_buddy.c
530
struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
532
BUG_ON(!drm_buddy_block_is_allocated(block));
drivers/gpu/drm/drm_buddy.c
533
mm->avail += drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
534
if (drm_buddy_block_is_clear(block))
drivers/gpu/drm/drm_buddy.c
535
mm->clear_avail += drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
537
__drm_buddy_free(mm, block, false);
drivers/gpu/drm/drm_buddy.c
546
struct drm_buddy_block *block, *on;
drivers/gpu/drm/drm_buddy.c
550
list_for_each_entry_safe(block, on, objects, link) {
drivers/gpu/drm/drm_buddy.c
552
mark_cleared(block);
drivers/gpu/drm/drm_buddy.c
554
clear_reset(block);
drivers/gpu/drm/drm_buddy.c
555
drm_buddy_free_block(mm, block);
drivers/gpu/drm/drm_buddy.c
57
get_block_tree(struct drm_buddy_block *block)
drivers/gpu/drm/drm_buddy.c
589
static bool block_incompatible(struct drm_buddy_block *block, unsigned int flags)
drivers/gpu/drm/drm_buddy.c
59
return drm_buddy_block_is_clear(block) ?
drivers/gpu/drm/drm_buddy.c
593
return needs_clear != drm_buddy_block_is_clear(block);
drivers/gpu/drm/drm_buddy.c
604
struct drm_buddy_block *block;
drivers/gpu/drm/drm_buddy.c
619
block = list_first_entry_or_null(&dfs,
drivers/gpu/drm/drm_buddy.c
622
if (!block)
drivers/gpu/drm/drm_buddy.c
625
list_del(&block->tmp_link);
drivers/gpu/drm/drm_buddy.c
627
if (drm_buddy_block_order(block) < order)
drivers/gpu/drm/drm_buddy.c
630
block_start = drm_buddy_block_offset(block);
drivers/gpu/drm/drm_buddy.c
631
block_end = block_start + drm_buddy_block_size(mm, block) - 1;
drivers/gpu/drm/drm_buddy.c
636
if (drm_buddy_block_is_allocated(block))
drivers/gpu/drm/drm_buddy.c
648
if (!fallback && block_incompatible(block, flags))
drivers/gpu/drm/drm_buddy.c
652
order == drm_buddy_block_order(block)) {
drivers/gpu/drm/drm_buddy.c
656
if (drm_buddy_block_is_free(block))
drivers/gpu/drm/drm_buddy.c
657
return block;
drivers/gpu/drm/drm_buddy.c
662
if (!drm_buddy_block_is_split(block)) {
drivers/gpu/drm/drm_buddy.c
663
err = split_block(mm, block);
drivers/gpu/drm/drm_buddy.c
668
list_add(&block->right->tmp_link, &dfs);
drivers/gpu/drm/drm_buddy.c
669
list_add(&block->left->tmp_link, &dfs);
drivers/gpu/drm/drm_buddy.c
680
buddy = __get_buddy(block);
drivers/gpu/drm/drm_buddy.c
682
(drm_buddy_block_is_free(block) &&
drivers/gpu/drm/drm_buddy.c
684
__drm_buddy_free(mm, block, false);
drivers/gpu/drm/drm_buddy.c
694
struct drm_buddy_block *block;
drivers/gpu/drm/drm_buddy.c
697
block = __alloc_range_bias(mm, start, end, order,
drivers/gpu/drm/drm_buddy.c
699
if (IS_ERR(block))
drivers/gpu/drm/drm_buddy.c
703
return block;
drivers/gpu/drm/drm_buddy.c
711
struct drm_buddy_block *max_block = NULL, *block = NULL;
drivers/gpu/drm/drm_buddy.c
717
block = rbtree_last_free_block(root);
drivers/gpu/drm/drm_buddy.c
718
if (!block)
drivers/gpu/drm/drm_buddy.c
722
max_block = block;
drivers/gpu/drm/drm_buddy.c
726
if (drm_buddy_block_offset(block) >
drivers/gpu/drm/drm_buddy.c
728
max_block = block;
drivers/gpu/drm/drm_buddy.c
740
struct drm_buddy_block *block = NULL;
drivers/gpu/drm/drm_buddy.c
750
block = get_maxblock(mm, order, tree);
drivers/gpu/drm/drm_buddy.c
751
if (block)
drivers/gpu/drm/drm_buddy.c
753
tmp = drm_buddy_block_order(block);
drivers/gpu/drm/drm_buddy.c
758
block = rbtree_last_free_block(root);
drivers/gpu/drm/drm_buddy.c
759
if (block)
drivers/gpu/drm/drm_buddy.c
764
if (!block) {
drivers/gpu/drm/drm_buddy.c
771
block = rbtree_last_free_block(root);
drivers/gpu/drm/drm_buddy.c
772
if (block)
drivers/gpu/drm/drm_buddy.c
776
if (!block)
drivers/gpu/drm/drm_buddy.c
780
BUG_ON(!drm_buddy_block_is_free(block));
drivers/gpu/drm/drm_buddy.c
783
err = split_block(mm, block);
drivers/gpu/drm/drm_buddy.c
787
block = block->right;
drivers/gpu/drm/drm_buddy.c
790
return block;
drivers/gpu/drm/drm_buddy.c
794
__drm_buddy_free(mm, block, false);
drivers/gpu/drm/drm_buddy.c
80
static bool drm_buddy_block_offset_less(const struct drm_buddy_block *block,
drivers/gpu/drm/drm_buddy.c
804
struct drm_buddy_block *block;
drivers/gpu/drm/drm_buddy.c
817
block = list_first_entry_or_null(dfs,
drivers/gpu/drm/drm_buddy.c
820
if (!block)
drivers/gpu/drm/drm_buddy.c
823
list_del(&block->tmp_link);
drivers/gpu/drm/drm_buddy.c
825
block_start = drm_buddy_block_offset(block);
drivers/gpu/drm/drm_buddy.c
826
block_end = block_start + drm_buddy_block_size(mm, block) - 1;
drivers/gpu/drm/drm_buddy.c
83
return drm_buddy_block_offset(block) < drm_buddy_block_offset(node);
drivers/gpu/drm/drm_buddy.c
831
if (drm_buddy_block_is_allocated(block)) {
drivers/gpu/drm/drm_buddy.c
837
if (drm_buddy_block_is_free(block)) {
drivers/gpu/drm/drm_buddy.c
838
mark_allocated(mm, block);
drivers/gpu/drm/drm_buddy.c
839
total_allocated += drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
840
mm->avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
841
if (drm_buddy_block_is_clear(block))
drivers/gpu/drm/drm_buddy.c
842
mm->clear_avail -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/drm_buddy.c
843
list_add_tail(&block->link, &allocated);
drivers/gpu/drm/drm_buddy.c
851
if (!drm_buddy_block_is_split(block)) {
drivers/gpu/drm/drm_buddy.c
852
err = split_block(mm, block);
drivers/gpu/drm/drm_buddy.c
857
list_add(&block->right->tmp_link, dfs);
drivers/gpu/drm/drm_buddy.c
858
list_add(&block->left->tmp_link, dfs);
drivers/gpu/drm/drm_buddy.c
86
static bool rbtree_block_offset_less(struct rb_node *block,
drivers/gpu/drm/drm_buddy.c
876
buddy = __get_buddy(block);
drivers/gpu/drm/drm_buddy.c
878
(drm_buddy_block_is_free(block) &&
drivers/gpu/drm/drm_buddy.c
880
__drm_buddy_free(mm, block, false);
drivers/gpu/drm/drm_buddy.c
89
return drm_buddy_block_offset_less(rbtree_get_free_block(block),
drivers/gpu/drm/drm_buddy.c
915
struct drm_buddy_block *block;
drivers/gpu/drm/drm_buddy.c
938
block = rbtree_get_free_block(iter);
drivers/gpu/drm/drm_buddy.c
94
struct drm_buddy_block *block,
drivers/gpu/drm/drm_buddy.c
941
rhs_offset = drm_buddy_block_offset(block);
drivers/gpu/drm/drm_buddy.c
952
lhs_offset = drm_buddy_block_offset(block) - lhs_size;
drivers/gpu/drm/drm_buddy.c
97
rb_add(&block->rb,
drivers/gpu/drm/drm_buddy.c
98
&mm->free_trees[tree][drm_buddy_block_order(block)],
drivers/gpu/drm/drm_buddy.c
997
struct drm_buddy_block *block;
drivers/gpu/drm/drm_displayid.c
117
const struct displayid_block *block;
drivers/gpu/drm/drm_displayid.c
122
block = (const struct displayid_block *)&iter->section[iter->idx];
drivers/gpu/drm/drm_displayid.c
124
if (iter->idx + sizeof(*block) <= iter->length &&
drivers/gpu/drm/drm_displayid.c
125
iter->idx + sizeof(*block) + block->num_bytes <= iter->length)
drivers/gpu/drm/drm_displayid.c
126
return block;
drivers/gpu/drm/drm_displayid.c
134
const struct displayid_block *block;
drivers/gpu/drm/drm_displayid.c
141
block = displayid_iter_block(iter);
drivers/gpu/drm/drm_displayid.c
142
if (WARN_ON(!block)) {
drivers/gpu/drm/drm_displayid.c
149
iter->idx += sizeof(*block) + block->num_bytes;
drivers/gpu/drm/drm_displayid.c
151
block = displayid_iter_block(iter);
drivers/gpu/drm/drm_displayid.c
152
if (block)
drivers/gpu/drm/drm_displayid.c
153
return block;
drivers/gpu/drm/drm_displayid.c
180
block = displayid_iter_block(iter);
drivers/gpu/drm/drm_displayid.c
181
if (block)
drivers/gpu/drm/drm_displayid.c
182
return block;
drivers/gpu/drm/drm_edid.c
1749
const void *block = NULL;
drivers/gpu/drm/drm_edid.c
1755
block = drm_edid_block_data(iter->drm_edid, iter->index++);
drivers/gpu/drm/drm_edid.c
1757
return block;
drivers/gpu/drm/drm_edid.c
1806
const u8 *block = _block;
drivers/gpu/drm/drm_edid.c
1811
csum += block[i];
drivers/gpu/drm/drm_edid.c
1820
const struct edid *block = _block;
drivers/gpu/drm/drm_edid.c
1822
return block->checksum;
drivers/gpu/drm/drm_edid.c
1827
const u8 *block = _block;
drivers/gpu/drm/drm_edid.c
1829
return block[0];
drivers/gpu/drm/drm_edid.c
1872
const struct edid *block = _block;
drivers/gpu/drm/drm_edid.c
1874
if (!block)
drivers/gpu/drm/drm_edid.c
1878
int score = drm_edid_header_is_valid(block);
drivers/gpu/drm/drm_edid.c
1881
if (edid_block_is_zero(block))
drivers/gpu/drm/drm_edid.c
1891
if (edid_block_compute_checksum(block) != edid_block_get_checksum(block)) {
drivers/gpu/drm/drm_edid.c
1892
if (edid_block_is_zero(block))
drivers/gpu/drm/drm_edid.c
1899
if (block->version != 1)
drivers/gpu/drm/drm_edid.c
1913
static bool edid_block_valid(const void *block, bool base)
drivers/gpu/drm/drm_edid.c
1915
return edid_block_status_valid(edid_block_check(block, base),
drivers/gpu/drm/drm_edid.c
1916
edid_block_tag(block));
drivers/gpu/drm/drm_edid.c
1920
const struct edid *block,
drivers/gpu/drm/drm_edid.c
1945
if (edid_block_status_valid(status, edid_block_tag(block))) {
drivers/gpu/drm/drm_edid.c
1947
block_num, edid_block_tag(block),
drivers/gpu/drm/drm_edid.c
1948
edid_block_compute_checksum(block));
drivers/gpu/drm/drm_edid.c
1951
block_num, edid_block_tag(block),
drivers/gpu/drm/drm_edid.c
1952
edid_block_compute_checksum(block));
drivers/gpu/drm/drm_edid.c
1957
block->version);
drivers/gpu/drm/drm_edid.c
1966
static void edid_block_dump(const char *level, const void *block, int block_num)
drivers/gpu/drm/drm_edid.c
1971
status = edid_block_check(block, block_num == 0);
drivers/gpu/drm/drm_edid.c
1974
else if (!edid_block_status_valid(status, edid_block_tag(block)))
drivers/gpu/drm/drm_edid.c
1980
block, EDID_LENGTH, false);
drivers/gpu/drm/drm_edid.c
1990
struct edid *block = _block;
drivers/gpu/drm/drm_edid.c
1995
if (WARN_ON(!block))
drivers/gpu/drm/drm_edid.c
1998
status = edid_block_check(block, is_base_block);
drivers/gpu/drm/drm_edid.c
2001
edid_header_fix(block);
drivers/gpu/drm/drm_edid.c
2004
status = edid_block_check(block, is_base_block);
drivers/gpu/drm/drm_edid.c
2021
edid_block_status_print(status, block, block_num);
drivers/gpu/drm/drm_edid.c
2024
valid = edid_block_status_valid(status, edid_block_tag(block));
drivers/gpu/drm/drm_edid.c
2028
edid_block_dump(KERN_NOTICE, block, block_num);
drivers/gpu/drm/drm_edid.c
2050
void *block = (void *)edid_block_data(edid, i);
drivers/gpu/drm/drm_edid.c
2052
if (!drm_edid_block_valid(block, i, true, NULL))
drivers/gpu/drm/drm_edid.c
2080
const void *block = drm_edid_block_data(drm_edid, i);
drivers/gpu/drm/drm_edid.c
2082
if (!edid_block_valid(block, i == 0))
drivers/gpu/drm/drm_edid.c
2143
drm_do_probe_ddc_edid(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/drm_edid.c
2146
unsigned char start = block * EDID_LENGTH;
drivers/gpu/drm/drm_edid.c
2147
unsigned char segment = block >> 1;
drivers/gpu/drm/drm_edid.c
2331
typedef int read_block_fn(void *context, u8 *buf, unsigned int block, size_t len);
drivers/gpu/drm/drm_edid.c
2333
static enum edid_block_status edid_block_read(void *block, unsigned int block_num,
drivers/gpu/drm/drm_edid.c
2342
if (read_block(context, block, block_num, EDID_LENGTH))
drivers/gpu/drm/drm_edid.c
2345
status = edid_block_check(block, is_base_block);
drivers/gpu/drm/drm_edid.c
2347
edid_header_fix(block);
drivers/gpu/drm/drm_edid.c
2350
status = edid_block_check(block, is_base_block);
drivers/gpu/drm/drm_edid.c
2355
if (edid_block_status_valid(status, edid_block_tag(block)))
drivers/gpu/drm/drm_edid.c
2422
void *block = (void *)edid_block_data(edid, i);
drivers/gpu/drm/drm_edid.c
2424
status = edid_block_read(block, i, read_block, context);
drivers/gpu/drm/drm_edid.c
2426
edid_block_status_print(status, block, i);
drivers/gpu/drm/drm_edid.c
2428
if (!edid_block_status_valid(status, edid_block_tag(block))) {
drivers/gpu/drm/drm_edid.c
4207
const struct displayid_block *block;
drivers/gpu/drm/drm_edid.c
4228
displayid_iter_for_each(block, &iter) {
drivers/gpu/drm/drm_edid.c
4229
if (block->tag == DATA_BLOCK_CTA) {
drivers/gpu/drm/drm_edid.c
5112
const struct displayid_block *block;
drivers/gpu/drm/drm_edid.c
5114
displayid_iter_for_each(block, &iter->displayid_iter) {
drivers/gpu/drm/drm_edid.c
5115
if (block->tag != DATA_BLOCK_CTA)
drivers/gpu/drm/drm_edid.c
5122
iter->index = sizeof(*block);
drivers/gpu/drm/drm_edid.c
5123
iter->end = iter->index + block->num_bytes;
drivers/gpu/drm/drm_edid.c
5125
return block;
drivers/gpu/drm/drm_edid.c
6531
const struct displayid_block *block)
drivers/gpu/drm/drm_edid.c
6534
(struct displayid_vesa_vendor_specific_block *)block;
drivers/gpu/drm/drm_edid.c
6537
if (block->num_bytes < 3) {
drivers/gpu/drm/drm_edid.c
6540
connector->base.id, connector->name, block->num_bytes);
drivers/gpu/drm/drm_edid.c
6547
if (sizeof(*vesa) != sizeof(*block) + block->num_bytes) {
drivers/gpu/drm/drm_edid.c
6593
const struct displayid_block *block;
drivers/gpu/drm/drm_edid.c
6597
displayid_iter_for_each(block, &iter) {
drivers/gpu/drm/drm_edid.c
6598
if (block->tag == DATA_BLOCK_2_VENDOR_SPECIFIC)
drivers/gpu/drm/drm_edid.c
6599
drm_parse_vesa_mso_data(connector, block);
drivers/gpu/drm/drm_edid.c
6650
const struct displayid_block *block;
drivers/gpu/drm/drm_edid.c
6654
displayid_iter_for_each(block, &iter) {
drivers/gpu/drm/drm_edid.c
6836
const struct displayid_block *block)
drivers/gpu/drm/drm_edid.c
6838
struct displayid_detailed_timing_block *det = (struct displayid_detailed_timing_block *)block;
drivers/gpu/drm/drm_edid.c
6843
bool type_7 = block->tag == DATA_BLOCK_2_TYPE_7_DETAILED_TIMING;
drivers/gpu/drm/drm_edid.c
6845
if (block->num_bytes % 20)
drivers/gpu/drm/drm_edid.c
6848
num_timings = block->num_bytes / 20;
drivers/gpu/drm/drm_edid.c
6892
const struct displayid_block *block)
drivers/gpu/drm/drm_edid.c
6894
const struct displayid_formula_timing_block *formula_block = (struct displayid_formula_timing_block *)block;
drivers/gpu/drm/drm_edid.c
6898
bool type_10 = block->tag == DATA_BLOCK_2_TYPE_10_FORMULA_TIMING;
drivers/gpu/drm/drm_edid.c
6905
if (block->num_bytes % timing_size)
drivers/gpu/drm/drm_edid.c
6908
num_timings = block->num_bytes / timing_size;
drivers/gpu/drm/drm_edid.c
6925
const struct displayid_block *block;
drivers/gpu/drm/drm_edid.c
6930
displayid_iter_for_each(block, &iter) {
drivers/gpu/drm/drm_edid.c
6931
if (block->tag == DATA_BLOCK_TYPE_1_DETAILED_TIMING ||
drivers/gpu/drm/drm_edid.c
6932
block->tag == DATA_BLOCK_2_TYPE_7_DETAILED_TIMING)
drivers/gpu/drm/drm_edid.c
6933
num_modes += add_displayid_detailed_1_modes(connector, block);
drivers/gpu/drm/drm_edid.c
6934
else if (block->tag == DATA_BLOCK_2_TYPE_9_FORMULA_TIMING ||
drivers/gpu/drm/drm_edid.c
6935
block->tag == DATA_BLOCK_2_TYPE_10_FORMULA_TIMING)
drivers/gpu/drm/drm_edid.c
6936
num_modes += add_displayid_formula_modes(connector, block);
drivers/gpu/drm/drm_edid.c
7498
const struct displayid_block *block)
drivers/gpu/drm/drm_edid.c
7500
const struct displayid_tiled_block *tile = (struct displayid_tiled_block *)block;
drivers/gpu/drm/drm_edid.c
7553
const struct displayid_block *block)
drivers/gpu/drm/drm_edid.c
7556
block->tag == DATA_BLOCK_TILED_DISPLAY) ||
drivers/gpu/drm/drm_edid.c
7558
block->tag == DATA_BLOCK_2_TILED_DISPLAY_TOPOLOGY);
drivers/gpu/drm/drm_edid.c
7564
const struct displayid_block *block;
drivers/gpu/drm/drm_edid.c
7570
displayid_iter_for_each(block, &iter) {
drivers/gpu/drm/drm_edid.c
7571
if (displayid_is_tiled_block(&iter, block))
drivers/gpu/drm/drm_edid.c
7572
drm_parse_tiled_block(connector, block);
drivers/gpu/drm/gud/gud_connector.c
202
static int gud_connector_get_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/gud/gud_connector.c
205
size_t start = block * EDID_LENGTH;
drivers/gpu/drm/gud/gud_pipe.c
107
u8 r, g, b, pix, *block = dst; /* Assign to silence compiler warning */
drivers/gpu/drm/gud/gud_pipe.c
127
block = dst++;
drivers/gpu/drm/gud/gud_pipe.c
128
*block = 0;
drivers/gpu/drm/gud/gud_pipe.c
145
*block |= pix << pixshift;
drivers/gpu/drm/gud/gud_pipe.c
59
u8 pix, *pix8, *block = dst; /* Assign to silence compiler warning */
drivers/gpu/drm/gud/gud_pipe.c
87
block = dst++;
drivers/gpu/drm/gud/gud_pipe.c
88
*block = 0;
drivers/gpu/drm/gud/gud_pipe.c
92
*block |= pix << pixshift;
drivers/gpu/drm/i915/display/intel_bios.c
135
const void *block;
drivers/gpu/drm/i915/display/intel_bios.c
137
block = find_raw_section(bdb, section_id);
drivers/gpu/drm/i915/display/intel_bios.c
138
if (!block)
drivers/gpu/drm/i915/display/intel_bios.c
141
return block - bdb;
drivers/gpu/drm/i915/display/intel_bios.c
371
const void *block;
drivers/gpu/drm/i915/display/intel_bios.c
385
block = find_raw_section(bdb, BDB_LFP_DATA);
drivers/gpu/drm/i915/display/intel_bios.c
386
if (!block)
drivers/gpu/drm/i915/display/intel_bios.c
391
block_size = get_blocksize(block);
drivers/gpu/drm/i915/display/intel_bios.c
442
offset = block - bdb;
drivers/gpu/drm/i915/display/intel_bios.c
463
const void *block;
drivers/gpu/drm/i915/display/intel_bios.c
466
block = find_raw_section(bdb, section_id);
drivers/gpu/drm/i915/display/intel_bios.c
469
if (!block && section_id == BDB_LFP_DATA_PTRS) {
drivers/gpu/drm/i915/display/intel_bios.c
472
block = temp_block + 3;
drivers/gpu/drm/i915/display/intel_bios.c
474
if (!block)
drivers/gpu/drm/i915/display/intel_bios.c
480
block_size = get_blocksize(block);
drivers/gpu/drm/i915/display/intel_bios.c
486
if (section_id == BDB_MIPI_SEQUENCE && *(const u8 *)block >= 3)
drivers/gpu/drm/i915/display/intel_bios.c
496
memcpy(entry->data, block - 3, block_size + 3);
drivers/gpu/drm/i915/display/intel_display_power_well.c
1780
tgl_tc_cold_request(struct intel_display *display, bool block)
drivers/gpu/drm/i915/display/intel_display_power_well.c
1789
if (block)
drivers/gpu/drm/i915/display/intel_display_power_well.c
1800
if (block &&
drivers/gpu/drm/i915/display/intel_display_power_well.c
1814
drm_err(display->drm, "TC cold %sblock failed\n", block ? "" : "un");
drivers/gpu/drm/i915/display/intel_display_power_well.c
1817
block ? "" : "un");
drivers/gpu/drm/i915/display/intel_dmc.c
880
bool block)
drivers/gpu/drm/i915/display/intel_dmc.c
883
PIPEDMC_BLOCK_PKGC_SW_BLOCK_PKGC_ALWAYS, block ?
drivers/gpu/drm/i915/display/intel_dmc.h
27
bool block);
drivers/gpu/drm/i915/display/intel_dp_test.c
200
const struct edid *block = drm_edid_raw(intel_connector->detect_edid);
drivers/gpu/drm/i915/display/intel_dp_test.c
203
block += block->extensions;
drivers/gpu/drm/i915/display/intel_dp_test.c
206
block->checksum) <= 0)
drivers/gpu/drm/i915/display/intel_parent.c
84
display->parent->pc8->block(display->drm);
drivers/gpu/drm/i915/gvt/handlers.c
2833
struct gvt_mmio_block *block = gvt->mmio.mmio_block;
drivers/gpu/drm/i915/gvt/handlers.c
2837
for (i = 0; i < num; i++, block++) {
drivers/gpu/drm/i915/gvt/handlers.c
2838
if (offset >= i915_mmio_reg_offset(block->offset) &&
drivers/gpu/drm/i915/gvt/handlers.c
2839
offset < i915_mmio_reg_offset(block->offset) + block->size)
drivers/gpu/drm/i915/gvt/handlers.c
2840
return block;
drivers/gpu/drm/i915/gvt/handlers.c
2913
struct gvt_mmio_block *block = gvt->mmio.mmio_block;
drivers/gpu/drm/i915/gvt/handlers.c
2916
ret = krealloc(block,
drivers/gpu/drm/i915/gvt/handlers.c
2917
(gvt->mmio.num_mmio_block + 1) * sizeof(*block),
drivers/gpu/drm/i915/gvt/handlers.c
2922
gvt->mmio.mmio_block = block = ret;
drivers/gpu/drm/i915/gvt/handlers.c
2924
block += gvt->mmio.num_mmio_block;
drivers/gpu/drm/i915/gvt/handlers.c
2926
memset(block, 0, sizeof(*block));
drivers/gpu/drm/i915/gvt/handlers.c
2928
block->offset = _MMIO(offset);
drivers/gpu/drm/i915/gvt/handlers.c
2929
block->size = size;
drivers/gpu/drm/i915/gvt/handlers.c
2958
struct gvt_mmio_block *block;
drivers/gpu/drm/i915/gvt/handlers.c
2960
block = find_mmio_block(gvt, VGT_PVINFO_PAGE);
drivers/gpu/drm/i915/gvt/handlers.c
2961
if (!block) {
drivers/gpu/drm/i915/gvt/handlers.c
2967
block->read = pvinfo_mmio_read;
drivers/gpu/drm/i915/gvt/handlers.c
2968
block->write = pvinfo_mmio_write;
drivers/gpu/drm/i915/gvt/handlers.c
3051
struct gvt_mmio_block *block = gvt->mmio.mmio_block;
drivers/gpu/drm/i915/gvt/handlers.c
3061
for (i = 0; i < gvt->mmio.num_mmio_block; i++, block++) {
drivers/gpu/drm/i915/gvt/handlers.c
3063
if (i915_mmio_reg_offset(block->offset) == VGT_PVINFO_PAGE)
drivers/gpu/drm/i915/gvt/handlers.c
3066
for (j = 0; j < block->size; j += 4) {
drivers/gpu/drm/i915/gvt/handlers.c
3067
ret = handler(gvt, i915_mmio_reg_offset(block->offset) + j, data);
drivers/gpu/drm/i915/i915_display_pc8.c
29
.block = i915_display_pc8_block,
drivers/gpu/drm/i915/i915_scatterlist.c
172
struct drm_buddy_block *block;
drivers/gpu/drm/i915/i915_scatterlist.c
202
list_for_each_entry(block, blocks, link) {
drivers/gpu/drm/i915/i915_scatterlist.c
205
block_size = min_t(u64, size, drm_buddy_block_size(mm, block));
drivers/gpu/drm/i915/i915_scatterlist.c
206
offset = drm_buddy_block_offset(block);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
104
struct drm_buddy_block *block;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
106
list_for_each_entry(block, &bman_res->blocks, link) {
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
108
drm_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
112
(drm_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
160
struct drm_buddy_block *block;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
177
list_for_each_entry(block, &bman_res->blocks, link) {
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
179
drm_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
181
(drm_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
198
struct drm_buddy_block *block;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
210
list_for_each_entry(block, &bman_res->blocks, link) {
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
212
drm_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
214
(drm_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
227
struct drm_buddy_block *block;
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
242
list_for_each_entry(block, &bman->reserved, link)
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
243
drm_buddy_block_print(&bman->mm, block, printer);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
450
struct drm_buddy_block *block;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
489
list_for_each_entry(block, blocks, link) {
drivers/gpu/drm/i915/selftests/intel_memory_region.c
490
if (drm_buddy_block_size(mm, block) > size)
drivers/gpu/drm/i915/selftests/intel_memory_region.c
491
size = drm_buddy_block_size(mm, block);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
531
struct drm_buddy_block *block;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
535
list_for_each_entry(block, &bman_res->blocks, link) {
drivers/gpu/drm/i915/selftests/intel_memory_region.c
536
u64 start = drm_buddy_block_offset(block);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
537
u64 end = start + drm_buddy_block_size(mm, block);
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
1112
char block[12] = "unknown";
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
1121
snprintf(block, sizeof(block), "%x", info->fsynr1);
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
1123
return adreno_fault_handler(gpu, iova, flags, info, block, scratch);
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
1824
const char *block = "unknown";
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
1834
block = a6xx_fault_block(gpu, info->fsynr1 & 0xff);
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
1836
return adreno_fault_handler(gpu, iova, flags, info, block, scratch);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
174
static int debugbus_read(struct msm_gpu *gpu, u32 block, u32 offset,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1788
const struct a6xx_shader_block *block = obj->handle;
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1794
print_name(p, " - type: ", block->name);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1798
drm_printf(p, " size: %d\n", block->size);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1803
print_ascii85(p, block->size << 2,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1804
obj->data + (block->size * i));
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
181
A7XX_DBGC_CFG_DBGBUS_SEL_D_PING_BLK_SEL(block);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1811
const struct gen7_shader_block *block = obj->handle;
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1818
print_name(p, " - type: ", a7xx_statetype_names[block->statetype]);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1819
print_name(p, " - pipe: ", a7xx_pipe_names[block->pipeid]);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1820
drm_printf(p, " - location: %d\n", block->location);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1822
for (i = 0; i < block->num_sps; i++) {
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1825
for (j = 0; j < block->num_usptps; j++) {
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1827
drm_printf(p, " size: %d\n", block->size);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1832
print_ascii85(p, block->size << 2, data);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1834
data += block->size;
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
184
A6XX_DBGC_CFG_DBGBUS_SEL_D_PING_BLK_SEL(block);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1932
static void a6xx_show_debugbus_block(const struct a6xx_debugbus_block *block,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1935
if (block) {
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1936
print_name(p, " - debugbus-block: ", block->name);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1942
drm_printf(p, " count: %d\n", block->count << 1);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1944
print_ascii85(p, block->count << 3, data);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
208
static int cx_debugbus_read(struct msm_gpu *gpu, void __iomem *cxdbg, u32 block, u32 offset,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
215
A7XX_CX_DBGC_CFG_DBGBUS_SEL_A_PING_BLK_SEL(block);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
218
A6XX_CX_DBGC_CFG_DBGBUS_SEL_A_PING_BLK_SEL(block);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
316
const struct a6xx_debugbus_block *block,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
322
obj->data = state_kcalloc(a6xx_state, block->count, sizeof(u64));
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
326
obj->handle = block;
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
328
for (ptr = obj->data, i = 0; i < block->count; i++)
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
329
ptr += debugbus_read(gpu, block->id, i, ptr);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
335
const struct a6xx_debugbus_block *block,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
341
obj->data = state_kcalloc(a6xx_state, block->count, sizeof(u64));
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
345
obj->handle = block;
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
347
for (ptr = obj->data, i = 0; i < block->count; i++)
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
348
ptr += cx_debugbus_read(gpu, cxdbg, block->id, i, ptr);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
869
const struct a6xx_shader_block *block,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
875
size_t datasize = block->size * A6XX_NUM_SHADER_BANKS * sizeof(u32);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
883
(block->type << 8) | i);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
886
block->size, out);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
888
out += block->size * sizeof(u32);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
896
obj->handle = block;
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
903
const struct gen7_shader_block *block,
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
910
size_t datasize = block->size * block->num_sps * block->num_usptps * sizeof(u32);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
920
for (i = 0; i < block->num_sps; i++) {
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
921
for (j = 0; j < block->num_usptps; j++) {
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
923
A7XX_SP_READ_SEL_LOCATION(block->location) |
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
924
A7XX_SP_READ_SEL_PIPE(block->pipeid) |
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
925
A7XX_SP_READ_SEL_STATETYPE(block->statetype) |
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
930
block->size, out);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
932
out += block->size * sizeof(u32);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
941
obj->handle = block;
drivers/gpu/drm/msm/adreno/a8xx_gpu.c
869
const char *block = "unknown";
drivers/gpu/drm/msm/adreno/a8xx_gpu.c
879
block = a8xx_fault_block(gpu, info->fsynr1 & 0xff);
drivers/gpu/drm/msm/adreno/a8xx_gpu.c
881
return adreno_fault_handler(gpu, iova, flags, info, block, scratch);
drivers/gpu/drm/msm/adreno/adreno_gpu.c
284
struct adreno_smmu_fault_info *info, const char *block,
drivers/gpu/drm/msm/adreno/adreno_gpu.c
331
type, block,
drivers/gpu/drm/msm/adreno/adreno_gpu.c
349
fault_info.block = block;
drivers/gpu/drm/msm/adreno/adreno_gpu.c
955
drm_printf(p, " - source=%s\n", info->block);
drivers/gpu/drm/msm/adreno/adreno_gpu.h
661
struct adreno_smmu_fault_info *info, const char *block,
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
152
struct msm_disp_state_block *block, *tmp;
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
159
list_for_each_entry_safe(block, tmp, &disp_state->blocks, node) {
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
160
list_del(&block->node);
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
161
kvfree(block->state);
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
162
kfree(block);
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
74
struct msm_disp_state_block *block, *tmp;
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
87
list_for_each_entry_safe(block, tmp, &state->blocks, node) {
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
88
drm_printf(p, "====================%s================\n", block->name);
drivers/gpu/drm/msm/disp/msm_disp_snapshot_util.c
89
msm_disp_state_print_regs(block->state, block->size, block->base_addr, p);
drivers/gpu/drm/msm/msm_gpu.h
104
const char *block;
drivers/gpu/drm/nouveau/include/nvkm/core/intr.h
24
void (*block)(struct nvkm_intr *, int leaf, u32 mask);
drivers/gpu/drm/nouveau/include/nvkm/core/mm.h
34
int nvkm_mm_init(struct nvkm_mm *, u8 heap, u32 offset, u32 length, u32 block);
drivers/gpu/drm/nouveau/nvkm/core/intr.c
123
if (intr->func->block)
drivers/gpu/drm/nouveau/nvkm/core/intr.c
124
intr->func->block(intr, leaf, mask);
drivers/gpu/drm/nouveau/nvkm/core/intr.c
329
for (i = 0; intr->func->block && i < intr->leaves; i++) {
drivers/gpu/drm/nouveau/nvkm/core/intr.c
330
intr->func->block(intr, i, ~0);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
240
nvkm_mm_init(struct nvkm_mm *mm, u8 heap, u32 offset, u32 length, u32 block)
drivers/gpu/drm/nouveau/nvkm/core/mm.c
257
BUG_ON(block != mm->block_size);
drivers/gpu/drm/nouveau/nvkm/core/mm.c
261
mm->block_size = block;
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ga100.c
433
.block = ga100_runl_block,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c
433
.block = gf100_runl_block,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
466
.block = gk104_runl_block,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk110.c
88
.block = gk104_runl_block,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gm107.c
61
.block = gk104_runl_block,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gp100.c
49
.block = gk104_runl_block,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
212
.block = gk104_runl_block,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
303
runl->func->block(runl, ~0);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.h
56
void (*block)(struct nvkm_runl *, u32 engm);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/tu102.c
92
.block = gk104_runl_block,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
340
.block = r535_runl_block,
drivers/gpu/drm/nouveau/nvkm/subdev/mc/gp100.c
78
intr->func->block(intr, i, 0xffffffff);
drivers/gpu/drm/nouveau/nvkm/subdev/mc/gp100.c
86
.block = gp100_mc_intr_block,
drivers/gpu/drm/nouveau/nvkm/subdev/mc/gt215.c
77
.block = gt215_mc_intr_block,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
659
u64 block;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
677
block = (part >> page[i].shift) << page[i].shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
679
block = (size >> page[i].shift) << page[i].shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
681
block = (size >> page[i].shift) << page[i].shift;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
686
int ret = nvkm_vmm_ptes_sparse_get(vmm, &page[i], addr, block);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
693
nvkm_vmm_ptes_sparse_put(vmm, &page[i], addr, block);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
696
size -= block;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
697
addr += block;
drivers/gpu/drm/nouveau/nvkm/subdev/vfn/tu102.c
94
.block = tu102_vfn_intr_block,
drivers/gpu/drm/omapdrm/dss/hdmi4_core.c
104
if (block)
drivers/gpu/drm/omapdrm/dss/hdmi4_core.c
77
int hdmi4_core_ddc_read(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/omapdrm/dss/hdmi4_core.c
91
REG_FLD_MOD(base, HDMI_CORE_DDC_SEGM, block / 2, 7, 0);
drivers/gpu/drm/omapdrm/dss/hdmi4_core.c
97
REG_FLD_MOD(base, HDMI_CORE_DDC_OFFSET, block % 2 ? 0x80 : 0, 7, 0);
drivers/gpu/drm/omapdrm/dss/hdmi4_core.h
253
int hdmi4_core_ddc_read(void *data, u8 *buf, unsigned int block, size_t len);
drivers/gpu/drm/omapdrm/dss/hdmi5_core.c
115
int hdmi5_core_ddc_read(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/omapdrm/dss/hdmi5_core.c
121
u8 seg_ptr = block / 2;
drivers/gpu/drm/omapdrm/dss/hdmi5_core.c
122
u8 edidbase = ((block % 2) * EDID_LENGTH);
drivers/gpu/drm/omapdrm/dss/hdmi5_core.h
285
int hdmi5_core_ddc_read(void *data, u8 *buf, unsigned int block, size_t len);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1067
struct tiler_block *block;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1105
list_for_each_entry(block, &omap_dmm->alloc_head, alloc_node) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1106
if (block->area.tcm == omap_dmm->tcm[lut_idx]) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1107
if (block->fmt != TILFMT_PAGE) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1108
fill_map(map, xdiv, ydiv, &block->area,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1115
&block->area);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1118
ydiv, &block->area.p0) == ' ';
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1120
&block->area.p1) == ' ';
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1122
tcm_for_each_slice(a, block->area, p)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1126
&block->area.p0,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1129
&block->area.p1,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1132
&block->area);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
509
int tiler_pin(struct tiler_block *block, struct page **pages,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
514
ret = fill(&block->area, pages, npages, roll, wait);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
517
tiler_unpin(block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
522
int tiler_unpin(struct tiler_block *block)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
524
return fill(&block->area, NULL, 0, 0, false);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
533
struct tiler_block *block;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
539
block = kzalloc_obj(*block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
540
if (!block)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
555
block->fmt = fmt;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
558
&block->area);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
560
kfree(block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
566
list_add(&block->alloc_node, &omap_dmm->alloc_head);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
569
return block;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
574
struct tiler_block *block = kzalloc_obj(*block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
578
if (!block)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
581
block->fmt = TILFMT_PAGE;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
584
&block->area)) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
585
kfree(block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
590
list_add(&block->alloc_node, &omap_dmm->alloc_head);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
593
return block;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
597
int tiler_release(struct tiler_block *block)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
599
int ret = tcm_free(&block->area);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
602
if (block->area.tcm)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
606
list_del(&block->alloc_node);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
609
kfree(block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
669
dma_addr_t tiler_ssptr(struct tiler_block *block)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
671
BUG_ON(!validfmt(block->fmt));
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
673
return TILVIEW_8BIT + tiler_get_address(block->fmt, 0,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
674
block->area.p0.x * geom[block->fmt].slot_w,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
675
block->area.p0.y * geom[block->fmt].slot_h);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
678
dma_addr_t tiler_tsptr(struct tiler_block *block, u32 orient,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
681
struct tcm_pt *p = &block->area.p0;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
682
BUG_ON(!validfmt(block->fmt));
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
684
return tiler_get_address(block->fmt, orient,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
685
(p->x * geom[block->fmt].slot_w) + x,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
686
(p->y * geom[block->fmt].slot_h) + y);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
730
struct tiler_block *block, *_block;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
741
list_for_each_entry_safe(block, _block, &omap_dmm->alloc_head,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
743
list_del(&block->alloc_node);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
744
kfree(block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.h
82
int tiler_pin(struct tiler_block *block, struct page **pages,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.h
84
int tiler_unpin(struct tiler_block *block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.h
90
int tiler_release(struct tiler_block *block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.h
93
dma_addr_t tiler_ssptr(struct tiler_block *block);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.h
94
dma_addr_t tiler_tsptr(struct tiler_block *block, u32 orient,
drivers/gpu/drm/omapdrm/omap_gem.c
1107
if (omap_obj->block) {
drivers/gpu/drm/omapdrm/omap_gem.c
1112
ret = tiler_pin(omap_obj->block,
drivers/gpu/drm/omapdrm/omap_gem.c
115
struct tiler_block *block; /* the reserved tiler block */
drivers/gpu/drm/omapdrm/omap_gem.c
1150
if (omap_obj->block) {
drivers/gpu/drm/omapdrm/omap_gem.c
1151
struct tcm_area *area = &omap_obj->block->area;
drivers/gpu/drm/omapdrm/omap_gem.c
1492
struct tiler_block *block;
drivers/gpu/drm/omapdrm/omap_gem.c
1495
block = tiler_reserve_2d(fmts[i], w, h, PAGE_SIZE);
drivers/gpu/drm/omapdrm/omap_gem.c
1496
if (IS_ERR(block)) {
drivers/gpu/drm/omapdrm/omap_gem.c
1499
i, j, PTR_ERR(block));
drivers/gpu/drm/omapdrm/omap_gem.c
1502
entry->dma_addr = tiler_ssptr(block);
drivers/gpu/drm/omapdrm/omap_gem.c
1503
entry->block = block;
drivers/gpu/drm/omapdrm/omap_gem.c
457
err = tiler_pin(entry->block, pages, ARRAY_SIZE(pages), 0, true);
drivers/gpu/drm/omapdrm/omap_gem.c
650
if (omap_obj->block) {
drivers/gpu/drm/omapdrm/omap_gem.c
655
ret = tiler_pin(omap_obj->block, omap_obj->pages, npages,
drivers/gpu/drm/omapdrm/omap_gem.c
748
struct tiler_block *block;
drivers/gpu/drm/omapdrm/omap_gem.c
751
BUG_ON(omap_obj->block);
drivers/gpu/drm/omapdrm/omap_gem.c
754
block = tiler_reserve_2d(fmt, omap_obj->width, omap_obj->height,
drivers/gpu/drm/omapdrm/omap_gem.c
757
block = tiler_reserve_1d(obj->size);
drivers/gpu/drm/omapdrm/omap_gem.c
760
if (IS_ERR(block)) {
drivers/gpu/drm/omapdrm/omap_gem.c
761
ret = PTR_ERR(block);
drivers/gpu/drm/omapdrm/omap_gem.c
767
ret = tiler_pin(block, omap_obj->pages, npages, omap_obj->roll, true);
drivers/gpu/drm/omapdrm/omap_gem.c
769
tiler_release(block);
drivers/gpu/drm/omapdrm/omap_gem.c
774
omap_obj->dma_addr = tiler_ssptr(block);
drivers/gpu/drm/omapdrm/omap_gem.c
775
omap_obj->block = block;
drivers/gpu/drm/omapdrm/omap_gem.c
83
struct tiler_block *block;
drivers/gpu/drm/omapdrm/omap_gem.c
859
ret = tiler_unpin(omap_obj->block);
drivers/gpu/drm/omapdrm/omap_gem.c
864
ret = tiler_release(omap_obj->block);
drivers/gpu/drm/omapdrm/omap_gem.c
870
omap_obj->block = NULL;
drivers/gpu/drm/omapdrm/omap_gem.c
904
if ((refcount_read(&omap_obj->pin_cnt) > 0) && omap_obj->block &&
drivers/gpu/drm/omapdrm/omap_gem.c
906
*dma_addr = tiler_tsptr(omap_obj->block, orient, x, y);
drivers/gpu/drm/radeon/cik.c
5657
char block[5] = { mc_client >> 24, (mc_client >> 16) & 0xff,
drivers/gpu/drm/radeon/cik.c
5668
block, mc_client, mc_id);
drivers/gpu/drm/radeon/cik.c
6279
u32 block, bool enable)
drivers/gpu/drm/radeon/cik.c
6282
if (block & RADEON_CG_BLOCK_GFX) {
drivers/gpu/drm/radeon/cik.c
6295
if (block & RADEON_CG_BLOCK_MC) {
drivers/gpu/drm/radeon/cik.c
6302
if (block & RADEON_CG_BLOCK_SDMA) {
drivers/gpu/drm/radeon/cik.c
6307
if (block & RADEON_CG_BLOCK_BIF) {
drivers/gpu/drm/radeon/cik.c
6311
if (block & RADEON_CG_BLOCK_UVD) {
drivers/gpu/drm/radeon/cik.c
6316
if (block & RADEON_CG_BLOCK_HDP) {
drivers/gpu/drm/radeon/cik.c
6321
if (block & RADEON_CG_BLOCK_VCE) {
drivers/gpu/drm/radeon/cik.h
31
void cik_update_cg(struct radeon_device *rdev, u32 block, bool enable);
drivers/gpu/drm/radeon/kv_dpm.c
65
u32 block, bool enable);
drivers/gpu/drm/radeon/ni.c
2511
char *block;
drivers/gpu/drm/radeon/ni.c
2522
block = "CB";
drivers/gpu/drm/radeon/ni.c
2532
block = "CB_FMASK";
drivers/gpu/drm/radeon/ni.c
2542
block = "CB_CMASK";
drivers/gpu/drm/radeon/ni.c
2552
block = "CB_IMMED";
drivers/gpu/drm/radeon/ni.c
2562
block = "DB";
drivers/gpu/drm/radeon/ni.c
2572
block = "DB_HTILE";
drivers/gpu/drm/radeon/ni.c
2582
block = "SX";
drivers/gpu/drm/radeon/ni.c
2592
block = "DB_STEN";
drivers/gpu/drm/radeon/ni.c
2602
block = "TC_TFETCH";
drivers/gpu/drm/radeon/ni.c
2612
block = "TC_VFETCH";
drivers/gpu/drm/radeon/ni.c
2622
block = "VC";
drivers/gpu/drm/radeon/ni.c
2625
block = "CP";
drivers/gpu/drm/radeon/ni.c
2629
block = "SH";
drivers/gpu/drm/radeon/ni.c
2632
block = "VGT";
drivers/gpu/drm/radeon/ni.c
2635
block = "IH";
drivers/gpu/drm/radeon/ni.c
2638
block = "RLC";
drivers/gpu/drm/radeon/ni.c
2641
block = "DMA";
drivers/gpu/drm/radeon/ni.c
2644
block = "HDP";
drivers/gpu/drm/radeon/ni.c
2647
block = "unknown";
drivers/gpu/drm/radeon/ni.c
2654
block, mc_id);
drivers/gpu/drm/radeon/radeon.h
770
void radeon_irq_kms_enable_afmt(struct radeon_device *rdev, int block);
drivers/gpu/drm/radeon/radeon.h
771
void radeon_irq_kms_disable_afmt(struct radeon_device *rdev, int block);
drivers/gpu/drm/radeon/radeon_audio.h
32
#define RREG32_ENDPOINT(block, reg) \
drivers/gpu/drm/radeon/radeon_audio.h
33
radeon_audio_endpoint_rreg(rdev, (block), (reg))
drivers/gpu/drm/radeon/radeon_audio.h
34
#define WREG32_ENDPOINT(block, reg, v) \
drivers/gpu/drm/radeon/radeon_audio.h
35
radeon_audio_endpoint_wreg(rdev, (block), (reg), (v))
drivers/gpu/drm/radeon/radeon_irq_kms.c
490
void radeon_irq_kms_enable_afmt(struct radeon_device *rdev, int block)
drivers/gpu/drm/radeon/radeon_irq_kms.c
498
rdev->irq.afmt[block] = true;
drivers/gpu/drm/radeon/radeon_irq_kms.c
512
void radeon_irq_kms_disable_afmt(struct radeon_device *rdev, int block)
drivers/gpu/drm/radeon/radeon_irq_kms.c
520
rdev->irq.afmt[block] = false;
drivers/gpu/drm/radeon/si.c
4807
char *block;
drivers/gpu/drm/radeon/si.c
4819
block = "CB";
drivers/gpu/drm/radeon/si.c
4829
block = "CB_FMASK";
drivers/gpu/drm/radeon/si.c
4839
block = "CB_CMASK";
drivers/gpu/drm/radeon/si.c
4849
block = "CB_IMMED";
drivers/gpu/drm/radeon/si.c
4859
block = "DB";
drivers/gpu/drm/radeon/si.c
4869
block = "DB_HTILE";
drivers/gpu/drm/radeon/si.c
4879
block = "DB_STEN";
drivers/gpu/drm/radeon/si.c
4893
block = "TC";
drivers/gpu/drm/radeon/si.c
4897
block = "CP";
drivers/gpu/drm/radeon/si.c
4903
block = "SH";
drivers/gpu/drm/radeon/si.c
4907
block = "VGT";
drivers/gpu/drm/radeon/si.c
4910
block = "IH";
drivers/gpu/drm/radeon/si.c
4914
block = "RLC";
drivers/gpu/drm/radeon/si.c
4918
block = "DMA0";
drivers/gpu/drm/radeon/si.c
4921
block = "DMA1";
drivers/gpu/drm/radeon/si.c
4925
block = "HDP";
drivers/gpu/drm/radeon/si.c
4928
block = "unknown";
drivers/gpu/drm/radeon/si.c
4941
block = "CB";
drivers/gpu/drm/radeon/si.c
4951
block = "CB_FMASK";
drivers/gpu/drm/radeon/si.c
4961
block = "CB_CMASK";
drivers/gpu/drm/radeon/si.c
4971
block = "CB_IMMED";
drivers/gpu/drm/radeon/si.c
4981
block = "DB";
drivers/gpu/drm/radeon/si.c
4991
block = "DB_HTILE";
drivers/gpu/drm/radeon/si.c
5001
block = "DB_STEN";
drivers/gpu/drm/radeon/si.c
5011
block = "TC";
drivers/gpu/drm/radeon/si.c
5015
block = "CP";
drivers/gpu/drm/radeon/si.c
5021
block = "SH";
drivers/gpu/drm/radeon/si.c
5024
block = "VGT";
drivers/gpu/drm/radeon/si.c
5027
block = "IH";
drivers/gpu/drm/radeon/si.c
5031
block = "RLC";
drivers/gpu/drm/radeon/si.c
5035
block = "DMA0";
drivers/gpu/drm/radeon/si.c
5038
block = "DMA1";
drivers/gpu/drm/radeon/si.c
5042
block = "HDP";
drivers/gpu/drm/radeon/si.c
5045
block = "unknown";
drivers/gpu/drm/radeon/si.c
5053
block, mc_id);
drivers/gpu/drm/radeon/si.c
5598
u32 block, bool enable)
drivers/gpu/drm/radeon/si.c
5600
if (block & RADEON_CG_BLOCK_GFX) {
drivers/gpu/drm/radeon/si.c
5613
if (block & RADEON_CG_BLOCK_MC) {
drivers/gpu/drm/radeon/si.c
5618
if (block & RADEON_CG_BLOCK_SDMA) {
drivers/gpu/drm/radeon/si.c
5622
if (block & RADEON_CG_BLOCK_BIF) {
drivers/gpu/drm/radeon/si.c
5626
if (block & RADEON_CG_BLOCK_UVD) {
drivers/gpu/drm/radeon/si.c
5632
if (block & RADEON_CG_BLOCK_HDP) {
drivers/gpu/drm/rockchip/cdn-dp-reg.c
437
unsigned int block, size_t length)
drivers/gpu/drm/rockchip/cdn-dp-reg.c
444
msg[0] = block / 2;
drivers/gpu/drm/rockchip/cdn-dp-reg.c
445
msg[1] = block % 2;
drivers/gpu/drm/rockchip/cdn-dp-reg.c
466
if (reg[0] == length && reg[1] == block / 2)
drivers/gpu/drm/rockchip/cdn-dp-reg.c
471
DRM_DEV_ERROR(dp->dev, "get block[%d] edid failed: %d\n", block,
drivers/gpu/drm/rockchip/cdn-dp-reg.h
467
unsigned int block, size_t length);
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
569
static int drm_sysfb_get_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
573
size_t off = block * EDID_LENGTH;
drivers/gpu/drm/tests/drm_buddy_test.c
109
list_for_each_entry_safe(block, tmp, &allocated_blocks, link) {
drivers/gpu/drm/tests/drm_buddy_test.c
111
list_move_tail(&block->link, &free_list);
drivers/gpu/drm/tests/drm_buddy_test.c
116
list_for_each_entry_safe_reverse(block, tmp, &allocated_blocks, link)
drivers/gpu/drm/tests/drm_buddy_test.c
117
list_move(&block->link, &reverse_list);
drivers/gpu/drm/tests/drm_buddy_test.c
133
struct drm_buddy_block *block;
drivers/gpu/drm/tests/drm_buddy_test.c
26
struct drm_buddy_block *block, *tmp;
drivers/gpu/drm/tests/drm_buddy_test.c
358
list_for_each_entry(block, &allocated, link)
drivers/gpu/drm/tests/drm_buddy_test.c
359
KUNIT_EXPECT_EQ(test, drm_buddy_block_is_clear(block), false);
drivers/gpu/drm/tests/drm_buddy_test.c
369
struct drm_buddy_block *block;
drivers/gpu/drm/tests/drm_buddy_test.c
418
list_for_each_entry(block, &clean, link)
drivers/gpu/drm/tests/drm_buddy_test.c
419
KUNIT_EXPECT_EQ(test, drm_buddy_block_is_clear(block), true);
drivers/gpu/drm/tests/drm_buddy_test.c
421
list_for_each_entry(block, &dirty, link)
drivers/gpu/drm/tests/drm_buddy_test.c
422
KUNIT_EXPECT_EQ(test, drm_buddy_block_is_clear(block), false);
drivers/gpu/drm/tests/drm_buddy_test.c
478
list_for_each_entry(block, &allocated, link) {
drivers/gpu/drm/tests/drm_buddy_test.c
480
KUNIT_EXPECT_EQ(test, drm_buddy_block_is_clear(block), false);
drivers/gpu/drm/tests/drm_buddy_test.c
481
total += drm_buddy_block_size(&mm, block);
drivers/gpu/drm/tests/drm_buddy_test.c
521
struct drm_buddy_block *block;
drivers/gpu/drm/tests/drm_buddy_test.c
594
list_for_each_entry(block, &allocated, link)
drivers/gpu/drm/tests/drm_buddy_test.c
595
total += drm_buddy_block_size(&mm, block);
drivers/gpu/drm/tests/drm_buddy_test.c
606
struct drm_buddy_block *block;
drivers/gpu/drm/tests/drm_buddy_test.c
62
list_for_each_entry_safe(block, tmp, &allocated_blocks, link) {
drivers/gpu/drm/tests/drm_buddy_test.c
630
block = list_first_entry_or_null(&blocks, typeof(*block), link);
drivers/gpu/drm/tests/drm_buddy_test.c
631
if (block) {
drivers/gpu/drm/tests/drm_buddy_test.c
632
list_del(&block->link);
drivers/gpu/drm/tests/drm_buddy_test.c
633
drm_buddy_free_block(&mm, block);
drivers/gpu/drm/tests/drm_buddy_test.c
64
list_move_tail(&block->link, &clear_list);
drivers/gpu/drm/tests/drm_buddy_test.c
644
block = list_first_entry_or_null(&tmp, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
645
KUNIT_ASSERT_TRUE_MSG(test, block, "alloc_blocks has no blocks\n");
drivers/gpu/drm/tests/drm_buddy_test.c
647
list_move_tail(&block->link, &blocks);
drivers/gpu/drm/tests/drm_buddy_test.c
656
block = list_first_entry_or_null(&tmp, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
657
KUNIT_ASSERT_TRUE_MSG(test, block, "alloc_blocks has no blocks\n");
drivers/gpu/drm/tests/drm_buddy_test.c
659
list_move_tail(&block->link, &holes);
drivers/gpu/drm/tests/drm_buddy_test.c
66
list_move_tail(&block->link, &dirty_list);
drivers/gpu/drm/tests/drm_buddy_test.c
687
struct drm_buddy_block *block, *bn;
drivers/gpu/drm/tests/drm_buddy_test.c
714
block = list_first_entry_or_null(&tmp, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
715
KUNIT_ASSERT_TRUE_MSG(test, block, "alloc_blocks has no blocks\n");
drivers/gpu/drm/tests/drm_buddy_test.c
717
list_move_tail(&block->link, &blocks);
drivers/gpu/drm/tests/drm_buddy_test.c
726
block = list_first_entry_or_null(&tmp, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
727
KUNIT_ASSERT_TRUE_MSG(test, block, "alloc_blocks has no blocks\n");
drivers/gpu/drm/tests/drm_buddy_test.c
729
list_move_tail(&block->link, &blocks);
drivers/gpu/drm/tests/drm_buddy_test.c
739
block = list_last_entry(&blocks, typeof(*block), link);
drivers/gpu/drm/tests/drm_buddy_test.c
740
list_del(&block->link);
drivers/gpu/drm/tests/drm_buddy_test.c
741
drm_buddy_free_block(&mm, block);
drivers/gpu/drm/tests/drm_buddy_test.c
745
list_for_each_entry_safe(block, bn, &blocks, link) {
drivers/gpu/drm/tests/drm_buddy_test.c
746
list_del(&block->link);
drivers/gpu/drm/tests/drm_buddy_test.c
747
drm_buddy_free_block(&mm, block);
drivers/gpu/drm/tests/drm_buddy_test.c
755
block = list_first_entry_or_null(&tmp, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
756
KUNIT_ASSERT_TRUE_MSG(test, block, "alloc_blocks has no blocks\n");
drivers/gpu/drm/tests/drm_buddy_test.c
758
list_del(&block->link);
drivers/gpu/drm/tests/drm_buddy_test.c
759
drm_buddy_free_block(&mm, block);
drivers/gpu/drm/tests/drm_buddy_test.c
770
block = list_first_entry_or_null(&tmp, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
771
KUNIT_ASSERT_TRUE_MSG(test, block, "alloc_blocks has no blocks\n");
drivers/gpu/drm/tests/drm_buddy_test.c
773
list_del(&block->link);
drivers/gpu/drm/tests/drm_buddy_test.c
774
drm_buddy_free_block(&mm, block);
drivers/gpu/drm/tests/drm_buddy_test.c
782
struct drm_buddy_block *block;
drivers/gpu/drm/tests/drm_buddy_test.c
809
block = list_first_entry_or_null(&tmp, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
810
KUNIT_ASSERT_TRUE_MSG(test, block, "alloc_blocks has no blocks\n");
drivers/gpu/drm/tests/drm_buddy_test.c
812
list_move_tail(&block->link, &blocks);
drivers/gpu/drm/tests/drm_buddy_test.c
828
struct drm_buddy_block *block;
drivers/gpu/drm/tests/drm_buddy_test.c
843
block = list_first_entry_or_null(&allocated, struct drm_buddy_block, link);
drivers/gpu/drm/tests/drm_buddy_test.c
844
KUNIT_EXPECT_TRUE(test, block);
drivers/gpu/drm/tests/drm_buddy_test.c
846
KUNIT_EXPECT_EQ_MSG(test, drm_buddy_block_order(block), mm.max_order,
drivers/gpu/drm/tests/drm_buddy_test.c
848
drm_buddy_block_order(block), mm.max_order);
drivers/gpu/drm/tests/drm_buddy_test.c
850
KUNIT_EXPECT_EQ_MSG(test, drm_buddy_block_size(&mm, block),
drivers/gpu/drm/tests/drm_buddy_test.c
853
drm_buddy_block_size(&mm, block),
drivers/gpu/drm/tiny/bochs.c
196
static int bochs_get_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/tiny/bochs.c
199
size_t i, start = block * EDID_LENGTH;
drivers/gpu/drm/tiny/gm12u320.c
256
int block, dst_offset, len, remain, ret, x1, x2, y1, y2;
drivers/gpu/drm/tiny/gm12u320.c
288
block = dst_offset / DATA_BLOCK_CONTENT_SIZE;
drivers/gpu/drm/tiny/gm12u320.c
300
gm12u320->data_buf[block] + dst_offset,
drivers/gpu/drm/tiny/gm12u320.c
304
block++;
drivers/gpu/drm/tiny/gm12u320.c
307
gm12u320->data_buf[block] + dst_offset,
drivers/gpu/drm/tiny/gm12u320.c
327
int block, block_size, len;
drivers/gpu/drm/tiny/gm12u320.c
332
for (block = 0; block < GM12U320_BLOCK_COUNT; block++) {
drivers/gpu/drm/tiny/gm12u320.c
333
if (block == GM12U320_BLOCK_COUNT - 1)
drivers/gpu/drm/tiny/gm12u320.c
342
gm12u320->cmd_buf[20] = 0xfc - block * 4;
drivers/gpu/drm/tiny/gm12u320.c
344
block | (gm12u320->fb_update.frame << 7);
drivers/gpu/drm/tiny/gm12u320.c
356
gm12u320->data_buf[block], block_size,
drivers/gpu/drm/udl/udl_edid.c
12
static int udl_read_edid_block(void *data, u8 *buf, unsigned int block, size_t len)
drivers/gpu/drm/udl/udl_edid.c
31
int bval = (i + block * EDID_LENGTH) << 8;
drivers/gpu/drm/virtio/virtgpu_vq.c
895
unsigned int block, size_t len)
drivers/gpu/drm/virtio/virtgpu_vq.c
898
size_t start = block * EDID_LENGTH;
drivers/gpu/drm/xe/xe_res_cursor.h
107
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_res_cursor.h
113
block = list_first_entry_or_null(head,
drivers/gpu/drm/xe/xe_res_cursor.h
116
if (!block)
drivers/gpu/drm/xe/xe_res_cursor.h
119
while (start >= drm_buddy_block_size(mm, block)) {
drivers/gpu/drm/xe/xe_res_cursor.h
120
start -= drm_buddy_block_size(mm, block);
drivers/gpu/drm/xe/xe_res_cursor.h
122
next = block->link.next;
drivers/gpu/drm/xe/xe_res_cursor.h
124
block = list_entry(next, struct drm_buddy_block,
drivers/gpu/drm/xe/xe_res_cursor.h
129
cur->start = drm_buddy_block_offset(block) + start;
drivers/gpu/drm/xe/xe_res_cursor.h
130
cur->size = min(drm_buddy_block_size(mm, block) - start,
drivers/gpu/drm/xe/xe_res_cursor.h
133
cur->node = block;
drivers/gpu/drm/xe/xe_res_cursor.h
262
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_res_cursor.h
295
block = cur->node;
drivers/gpu/drm/xe/xe_res_cursor.h
297
next = block->link.next;
drivers/gpu/drm/xe/xe_res_cursor.h
298
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/xe/xe_res_cursor.h
301
while (start >= drm_buddy_block_size(cur->mm, block)) {
drivers/gpu/drm/xe/xe_res_cursor.h
302
start -= drm_buddy_block_size(cur->mm, block);
drivers/gpu/drm/xe/xe_res_cursor.h
304
next = block->link.next;
drivers/gpu/drm/xe/xe_res_cursor.h
305
block = list_entry(next, struct drm_buddy_block, link);
drivers/gpu/drm/xe/xe_res_cursor.h
308
cur->start = drm_buddy_block_offset(block) + start;
drivers/gpu/drm/xe/xe_res_cursor.h
309
cur->size = min(drm_buddy_block_size(cur->mm, block) - start,
drivers/gpu/drm/xe/xe_res_cursor.h
311
cur->node = block;
drivers/gpu/drm/xe/xe_svm.c
1036
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_svm.c
1076
list_for_each_entry(block, blocks, link)
drivers/gpu/drm/xe/xe_svm.c
1077
block->private = vr;
drivers/gpu/drm/xe/xe_svm.c
761
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_svm.c
764
list_for_each_entry(block, blocks, link) {
drivers/gpu/drm/xe/xe_svm.c
765
struct xe_vram_region *vr = block->private;
drivers/gpu/drm/xe/xe_svm.c
768
drm_buddy_block_offset(block));
drivers/gpu/drm/xe/xe_svm.c
771
for (i = 0; i < drm_buddy_block_size(buddy, block) >> PAGE_SHIFT; ++i)
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
135
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
137
list_for_each_entry(block, &vres->blocks, link) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
138
u64 start = drm_buddy_block_offset(block);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
141
u64 end = start + drm_buddy_block_size(mm, block);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
161
struct drm_buddy_block *block = list_first_entry(&vres->blocks,
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
162
typeof(*block),
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
165
vres->base.start = drm_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
227
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
235
list_for_each_entry(block, &vres->blocks, link) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
237
drm_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
239
(drm_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
257
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
265
list_for_each_entry(block, &vres->blocks, link) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
267
drm_buddy_block_offset(block) >> PAGE_SHIFT;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
269
(drm_buddy_block_size(mm, block) >> PAGE_SHIFT);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
28
struct drm_buddy_block *block;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
31
block = xe_ttm_vram_mgr_first_block(head);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
32
if (!block)
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
35
while (head != block->link.next) {
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
36
start = drm_buddy_block_offset(block);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
37
size = drm_buddy_block_size(mm, block);
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
39
block = list_entry(block->link.next, struct drm_buddy_block,
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
41
if (start + size != drm_buddy_block_offset(block))
drivers/hid/hid-cp2112.c
688
read_length = data->block[0];
drivers/hid/hid-cp2112.c
696
if (data->block[0] > I2C_SMBUS_BLOCK_MAX + 1)
drivers/hid/hid-cp2112.c
700
data->block + 1,
drivers/hid/hid-cp2112.c
701
data->block[0]);
drivers/hid/hid-cp2112.c
714
if (data->block[0] > I2C_SMBUS_BLOCK_MAX + 1)
drivers/hid/hid-cp2112.c
718
data->block,
drivers/hid/hid-cp2112.c
719
data->block[0] + 1);
drivers/hid/hid-cp2112.c
730
if (data->block[0] > I2C_SMBUS_BLOCK_MAX + 1)
drivers/hid/hid-cp2112.c
734
command, data->block,
drivers/hid/hid-cp2112.c
735
data->block[0] + 1);
drivers/hid/hid-cp2112.c
811
memcpy(data->block + 1, buf, read_length);
drivers/hid/hid-cp2112.c
819
memcpy(data->block, buf, read_length);
drivers/hid/hid-ft260.c
716
ret = ft260_i2c_read(dev, addr, data->block,
drivers/hid/hid-ft260.c
717
data->block[0] + 1,
drivers/hid/hid-ft260.c
720
ret = ft260_smbus_write(dev, addr, cmd, data->block,
drivers/hid/hid-ft260.c
721
data->block[0] + 1,
drivers/hid/hid-ft260.c
732
ret = ft260_i2c_read(dev, addr, data->block + 1,
drivers/hid/hid-ft260.c
733
data->block[0],
drivers/hid/hid-ft260.c
736
ret = ft260_smbus_write(dev, addr, cmd, data->block + 1,
drivers/hid/hid-ft260.c
737
data->block[0],
drivers/hid/hid-mcp2221.c
540
mcp->rxbuf = data->block;
drivers/hid/hid-mcp2221.c
546
if (!data->block[0]) {
drivers/hid/hid-mcp2221.c
550
ret = mcp_smbus_write(mcp, addr, command, data->block,
drivers/hid/hid-mcp2221.c
551
data->block[0] + 1,
drivers/hid/hid-mcp2221.c
563
mcp->rxbuf = data->block;
drivers/hid/hid-mcp2221.c
569
if (!data->block[0]) {
drivers/hid/hid-mcp2221.c
574
&data->block[1], data->block[0],
drivers/hid/hid-mcp2221.c
590
ret = mcp_smbus_write(mcp, addr, command, data->block,
drivers/hid/hid-mcp2221.c
591
data->block[0] + 1,
drivers/hid/hid-mcp2221.c
599
data->block);
drivers/hwmon/occ/common.c
1058
struct occ_sensor_data_block *block = &poll->block;
drivers/hwmon/occ/common.c
1064
block = (struct occ_sensor_data_block *)((u8 *)block + offset);
drivers/hwmon/occ/common.c
1066
offset = (block->header.num_sensors *
drivers/hwmon/occ/common.c
1067
block->header.sensor_length) + sizeof(block->header);
drivers/hwmon/occ/common.c
1077
old_offset, offset - 1, block->header.eye_catcher,
drivers/hwmon/occ/common.c
1078
block->header.num_sensors);
drivers/hwmon/occ/common.c
1081
if (strncmp(block->header.eye_catcher, "TEMP", 4) == 0)
drivers/hwmon/occ/common.c
1083
else if (strncmp(block->header.eye_catcher, "FREQ", 4) == 0)
drivers/hwmon/occ/common.c
1085
else if (strncmp(block->header.eye_catcher, "POWR", 4) == 0)
drivers/hwmon/occ/common.c
1087
else if (strncmp(block->header.eye_catcher, "CAPS", 4) == 0)
drivers/hwmon/occ/common.c
1089
else if (strncmp(block->header.eye_catcher, "EXTN", 4) == 0)
drivers/hwmon/occ/common.c
1093
block->header.eye_catcher);
drivers/hwmon/occ/common.c
1097
sensor->num_sensors = block->header.num_sensors;
drivers/hwmon/occ/common.c
1098
sensor->version = block->header.sensor_format;
drivers/hwmon/occ/common.c
1099
sensor->data = &block->data;
drivers/hwmon/occ/common.h
61
struct occ_sensor_data_block block;
drivers/hwmon/pmbus/pmbus_core.c
2578
data.block[0] = 2;
drivers/hwmon/pmbus/pmbus_core.c
2579
data.block[1] = attr->reg;
drivers/hwmon/pmbus/pmbus_core.c
2580
data.block[2] = 0x01;
drivers/hwmon/pmbus/pmbus_core.c
2591
if (data.block[0] != 5)
drivers/hwmon/pmbus/pmbus_core.c
2594
m = data.block[1] | (data.block[2] << 8);
drivers/hwmon/pmbus/pmbus_core.c
2595
b = data.block[3] | (data.block[4] << 8);
drivers/hwmon/pmbus/pmbus_core.c
2596
R = data.block[5];
drivers/hwtracing/intel_th/msu-sink.c
56
void *block;
drivers/hwtracing/intel_th/msu-sink.c
71
block = dma_alloc_coherent(priv->dev->parent->parent,
drivers/hwtracing/intel_th/msu-sink.c
74
if (!block)
drivers/hwtracing/intel_th/msu-sink.c
77
sg_set_buf(sg_ptr, block, PAGE_SIZE);
drivers/hwtracing/intel_th/msu.c
100
struct scatterlist *block;
drivers/hwtracing/intel_th/msu.c
1005
void *block;
drivers/hwtracing/intel_th/msu.c
1013
block = dma_alloc_coherent(msc_dev(win->msc)->parent->parent,
drivers/hwtracing/intel_th/msu.c
1016
if (!block)
drivers/hwtracing/intel_th/msu.c
1019
sg_set_buf(sg_ptr, block, PAGE_SIZE);
drivers/hwtracing/intel_th/msu.c
446
return sg_virt(iter->block);
drivers/hwtracing/intel_th/msu.c
495
iter->block = iter->start_block;
drivers/hwtracing/intel_th/msu.c
545
if (iter->wrap_count && iter->block == iter->start_block) {
drivers/hwtracing/intel_th/msu.c
558
if (sg_is_last(iter->block))
drivers/hwtracing/intel_th/msu.c
559
iter->block = msc_win_base_sg(iter->win);
drivers/hwtracing/intel_th/msu.c
561
iter->block = sg_next(iter->block);
drivers/hwtracing/intel_th/msu.c
564
if (!iter->wrap_count && iter->block == iter->start_block)
drivers/hwtracing/intel_th/msu.c
621
if (iter->block == iter->start_block && iter->wrap_count == 2) {
drivers/i2c/busses/i2c-ali1535.c
402
len = data->block[0];
drivers/i2c/busses/i2c-ali1535.c
405
data->block[0] = len;
drivers/i2c/busses/i2c-ali1535.c
409
data->block[0] = len;
drivers/i2c/busses/i2c-ali1535.c
415
outb_p(data->block[i], SMBBLKDAT);
drivers/i2c/busses/i2c-ali1535.c
447
data->block[0] = len;
drivers/i2c/busses/i2c-ali1535.c
450
for (i = 1; i <= data->block[0]; i++) {
drivers/i2c/busses/i2c-ali1535.c
451
data->block[i] = inb_p(SMBBLKDAT);
drivers/i2c/busses/i2c-ali1535.c
453
len, i, data->block[i]);
drivers/i2c/busses/i2c-ali1563.c
195
len = data->block[0];
drivers/i2c/busses/i2c-ali1563.c
201
outb_p(data->block[1], SMB_BLK_DAT);
drivers/i2c/busses/i2c-ali1563.c
209
outb_p(data->block[i + 1], SMB_BLK_DAT);
drivers/i2c/busses/i2c-ali1563.c
224
data->block[i+1] = inb_p(SMB_BLK_DAT);
drivers/i2c/busses/i2c-ali15x3.c
389
len = data->block[0];
drivers/i2c/busses/i2c-ali15x3.c
392
data->block[0] = len;
drivers/i2c/busses/i2c-ali15x3.c
396
data->block[0] = len;
drivers/i2c/busses/i2c-ali15x3.c
402
outb_p(data->block[i], SMBBLKDAT);
drivers/i2c/busses/i2c-ali15x3.c
435
data->block[0] = len;
drivers/i2c/busses/i2c-ali15x3.c
438
for (i = 1; i <= data->block[0]; i++) {
drivers/i2c/busses/i2c-ali15x3.c
439
data->block[i] = inb_p(SMBBLKDAT);
drivers/i2c/busses/i2c-ali15x3.c
441
len, i, data->block[i]);
drivers/i2c/busses/i2c-amd756.c
222
len = data->block[0];
drivers/i2c/busses/i2c-amd756.c
230
outb_p(data->block[i],
drivers/i2c/busses/i2c-amd756.c
262
data->block[0] = inw_p(SMB_HOST_DATA) & 0x3f;
drivers/i2c/busses/i2c-amd756.c
263
if(data->block[0] > 32)
drivers/i2c/busses/i2c-amd756.c
264
data->block[0] = 32;
drivers/i2c/busses/i2c-amd756.c
266
for (i = 1; i <= data->block[0]; i++)
drivers/i2c/busses/i2c-amd756.c
267
data->block[i] = inb_p(SMB_HOST_BLOCK_DATA);
drivers/i2c/busses/i2c-amd8111.c
252
len = min_t(u8, data->block[0],
drivers/i2c/busses/i2c-amd8111.c
260
data->block[i + 1]);
drivers/i2c/busses/i2c-amd8111.c
269
len = min_t(u8, data->block[0],
drivers/i2c/busses/i2c-amd8111.c
281
data->block[i + 1]);
drivers/i2c/busses/i2c-amd8111.c
305
len = min_t(u8, data->block[0],
drivers/i2c/busses/i2c-amd8111.c
315
data->block[i + 1]);
drivers/i2c/busses/i2c-amd8111.c
388
data->block + i + 1);
drivers/i2c/busses/i2c-amd8111.c
392
data->block[0] = len;
drivers/i2c/busses/i2c-highlander.c
297
dev->buf = &data->block[1];
drivers/i2c/busses/i2c-highlander.c
298
dev->buf_len = data->block[0];
drivers/i2c/busses/i2c-i801.c
536
len = data->block[0];
drivers/i2c/busses/i2c-i801.c
539
iowrite8_rep(SMBBLKDAT(priv), data->block + 1, len);
drivers/i2c/busses/i2c-i801.c
554
data->block[0] = len;
drivers/i2c/busses/i2c-i801.c
556
ioread8_rep(SMBBLKDAT(priv), data->block + 1, len);
drivers/i2c/busses/i2c-i801.c
685
len = data->block[0];
drivers/i2c/busses/i2c-i801.c
689
iowrite8(data->block[1], SMBBLKDAT(priv));
drivers/i2c/busses/i2c-i801.c
705
priv->data = &data->block[1];
drivers/i2c/busses/i2c-i801.c
738
data->block[0] = len;
drivers/i2c/busses/i2c-i801.c
742
data->block[i] = ioread8(SMBBLKDAT(priv));
drivers/i2c/busses/i2c-i801.c
748
iowrite8(data->block[i+1], SMBBLKDAT(priv));
drivers/i2c/busses/i2c-i801.c
832
data->block[0] = SMBUS_LEN_SENTINEL;
drivers/i2c/busses/i2c-i801.c
833
else if (data->block[0] < 1 || data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-i801.c
855
if (data->block[0] < 1 || data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-isch.c
216
len = data->block[0];
drivers/i2c/busses/i2c-isch.c
221
sch_io_wr8(priv, SMBBLKDAT + i - 1, data->block[i]);
drivers/i2c/busses/i2c-isch.c
252
data->block[0] = sch_io_rd8(priv, SMBHSTDAT0);
drivers/i2c/busses/i2c-isch.c
253
if (data->block[0] == 0 || data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-isch.c
255
for (i = 1; i <= data->block[0]; i++)
drivers/i2c/busses/i2c-isch.c
256
data->block[i] = sch_io_rd8(priv, SMBBLKDAT + i - 1);
drivers/i2c/busses/i2c-ismt.c
356
memcpy(data->block, dma_buffer, desc->rxbytes);
drivers/i2c/busses/i2c-ismt.c
359
memcpy(&data->block[1], dma_buffer, desc->rxbytes);
drivers/i2c/busses/i2c-ismt.c
360
data->block[0] = desc->rxbytes;
drivers/i2c/busses/i2c-ismt.c
521
if (data->block[0] < 1 || data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-ismt.c
524
dma_size = data->block[0] + 1;
drivers/i2c/busses/i2c-ismt.c
529
memcpy(&dma_buffer[1], &data->block[1], dma_size - 1);
drivers/i2c/busses/i2c-ismt.c
543
if (data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-ismt.c
548
desc->wr_len_cmd = data->block[0] + 1;
drivers/i2c/busses/i2c-ismt.c
553
memcpy(&dma_buffer[1], &data->block[1], data->block[0]);
drivers/i2c/busses/i2c-ismt.c
558
if (data->block[0] < 1)
drivers/i2c/busses/i2c-ismt.c
559
data->block[0] = 1;
drivers/i2c/busses/i2c-ismt.c
561
if (data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-ismt.c
562
data->block[0] = I2C_SMBUS_BLOCK_MAX;
drivers/i2c/busses/i2c-ismt.c
567
dma_size = data->block[0] + 1;
drivers/i2c/busses/i2c-ismt.c
572
memcpy(&dma_buffer[1], &data->block[1], dma_size - 1);
drivers/i2c/busses/i2c-ismt.c
576
dma_size = data->block[0];
drivers/i2c/busses/i2c-microchip-corei2c.c
496
data_len = data->block[0];
drivers/i2c/busses/i2c-microchip-corei2c.c
499
msgs[CORE_I2C_SMBUS_MSG_WR].buf[i + 1] = data->block[i];
drivers/i2c/busses/i2c-microchip-corei2c.c
525
data->block[i] = rx_buf[i];
drivers/i2c/busses/i2c-mlxbf.c
2067
byte_cnt = data->block[0];
drivers/i2c/busses/i2c-mlxbf.c
2068
mlxbf_i2c_smbus_i2c_block_func(&request, &command, data->block,
drivers/i2c/busses/i2c-mlxbf.c
2075
byte_cnt = read ? I2C_SMBUS_BLOCK_MAX : data->block[0];
drivers/i2c/busses/i2c-mlxbf.c
2076
mlxbf_i2c_smbus_block_func(&request, &command, data->block,
drivers/i2c/busses/i2c-mlxbf.c
2090
byte_cnt = data->block[0];
drivers/i2c/busses/i2c-mlxbf.c
2092
data->block, &byte_cnt,
drivers/i2c/busses/i2c-mpc.c
490
i2c->block = 0;
drivers/i2c/busses/i2c-mpc.c
657
time_left = wait_event_timeout(i2c->waitq, !i2c->block, i2c->adap.timeout);
drivers/i2c/busses/i2c-mpc.c
677
i2c->block = 1;
drivers/i2c/busses/i2c-mpc.c
97
u32 block;
drivers/i2c/busses/i2c-mv64xxx.c
133
u32 block;
drivers/i2c/busses/i2c-mv64xxx.c
413
drv_data->block = 0;
drivers/i2c/busses/i2c-mv64xxx.c
432
drv_data->block = 0;
drivers/i2c/busses/i2c-mv64xxx.c
495
drv_data->block = 0;
drivers/i2c/busses/i2c-mv64xxx.c
557
!drv_data->block, drv_data->adapter.timeout);
drivers/i2c/busses/i2c-mv64xxx.c
568
if (abort && drv_data->block) {
drivers/i2c/busses/i2c-mv64xxx.c
573
!drv_data->block, drv_data->adapter.timeout);
drivers/i2c/busses/i2c-mv64xxx.c
575
if ((time_left <= 0) && drv_data->block) {
drivers/i2c/busses/i2c-mv64xxx.c
579
"time_left: %d\n", drv_data->block,
drivers/i2c/busses/i2c-mv64xxx.c
592
while (READ_ONCE(drv_data->block) &&
drivers/i2c/busses/i2c-mv64xxx.c
610
drv_data->block = 1;
drivers/i2c/busses/i2c-mv64xxx.c
684
drv_data->block = 1;
drivers/i2c/busses/i2c-nforce2.c
208
len = data->block[0];
drivers/i2c/busses/i2c-nforce2.c
217
outb_p(data->block[i + 1],
drivers/i2c/busses/i2c-nforce2.c
258
data->block[i + 1] = inb_p(NVIDIA_SMB_DATA + i);
drivers/i2c/busses/i2c-nforce2.c
259
data->block[0] = len;
drivers/i2c/busses/i2c-opal.c
154
req.buffer_ra = cpu_to_be64(__pa(&data->block[1]));
drivers/i2c/busses/i2c-opal.c
155
req.size = cpu_to_be32(data->block[0]);
drivers/i2c/busses/i2c-pasemi-core.c
319
len = min_t(u8, data->block[0], I2C_SMBUS_BLOCK_MAX);
drivers/i2c/busses/i2c-pasemi-core.c
322
TXFIFO_WR(smbus, data->block[i]);
drivers/i2c/busses/i2c-pasemi-core.c
323
TXFIFO_WR(smbus, data->block[len] | MTXFIFO_STOP);
drivers/i2c/busses/i2c-pasemi-core.c
336
len = min_t(u8, data->block[0], I2C_SMBUS_BLOCK_MAX - 1);
drivers/i2c/busses/i2c-pasemi-core.c
342
TXFIFO_WR(smbus, data->block[i]);
drivers/i2c/busses/i2c-pasemi-core.c
390
data->block[0] = len;
drivers/i2c/busses/i2c-pasemi-core.c
397
data->block[i] = rd & MRXFIFO_DATA_M;
drivers/i2c/busses/i2c-piix4.c
642
len = data->block[0];
drivers/i2c/busses/i2c-piix4.c
648
outb_p(data->block[i], SMBBLKDAT);
drivers/i2c/busses/i2c-piix4.c
676
data->block[0] = inb_p(SMBHSTDAT0);
drivers/i2c/busses/i2c-piix4.c
677
if (data->block[0] == 0 || data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-piix4.c
680
for (i = 1; i <= data->block[0]; i++)
drivers/i2c/busses/i2c-piix4.c
681
data->block[i] = inb_p(SMBBLKDAT);
drivers/i2c/busses/i2c-powermac.c
81
buf = data->block;
drivers/i2c/busses/i2c-powermac.c
82
len = data->block[0] + 1;
drivers/i2c/busses/i2c-powermac.c
85
buf = &data->block[1];
drivers/i2c/busses/i2c-powermac.c
86
len = data->block[0];
drivers/i2c/busses/i2c-rtl9300.c
331
xfer.data = &data->block[0];
drivers/i2c/busses/i2c-rtl9300.c
332
xfer.data_len = data->block[0] + 1;
drivers/i2c/busses/i2c-rtl9300.c
336
xfer.data = &data->block[1];
drivers/i2c/busses/i2c-rtl9300.c
337
xfer.data_len = data->block[0];
drivers/i2c/busses/i2c-scmi.c
139
len = data->block[0];
drivers/i2c/busses/i2c-scmi.c
146
mt_params[4].buffer.pointer = data->block + 1;
drivers/i2c/busses/i2c-scmi.c
253
data->block[0] = len;
drivers/i2c/busses/i2c-scmi.c
254
memcpy(data->block + 1, obj->buffer.pointer, len);
drivers/i2c/busses/i2c-sis630.c
236
len = data->block[0];
drivers/i2c/busses/i2c-sis630.c
244
"set data 0x%02x\n", data->block[i]);
drivers/i2c/busses/i2c-sis630.c
246
sis630_write(SMB_BYTE + (i - 1) % 8, data->block[i]);
drivers/i2c/busses/i2c-sis630.c
280
data->block[0] = len = 0;
drivers/i2c/busses/i2c-sis630.c
293
data->block[0] = sis630_read(SMB_COUNT);
drivers/i2c/busses/i2c-sis630.c
296
if (data->block[0] > 32)
drivers/i2c/busses/i2c-sis630.c
297
data->block[0] = 32;
drivers/i2c/busses/i2c-sis630.c
300
"block data read len=0x%x\n", data->block[0]);
drivers/i2c/busses/i2c-sis630.c
302
for (i = 0; i < 8 && len < data->block[0]; i++, len++) {
drivers/i2c/busses/i2c-sis630.c
305
data->block[len + 1] = sis630_read(SMB_BYTE +
drivers/i2c/busses/i2c-sis630.c
314
} while (len < data->block[0]);
drivers/i2c/busses/i2c-stm32f7.c
1047
if (data->block[0] > I2C_SMBUS_BLOCK_MAX ||
drivers/i2c/busses/i2c-stm32f7.c
1048
!data->block[0]) {
drivers/i2c/busses/i2c-stm32f7.c
1050
data->block[0]);
drivers/i2c/busses/i2c-stm32f7.c
1053
f7_msg->count = data->block[0] + 2;
drivers/i2c/busses/i2c-stm32f7.c
1055
f7_msg->smbus_buf[i] = data->block[i - 1];
drivers/i2c/busses/i2c-stm32f7.c
1068
if (data->block[0] > I2C_SMBUS_BLOCK_MAX - 1) {
drivers/i2c/busses/i2c-stm32f7.c
1070
data->block[0]);
drivers/i2c/busses/i2c-stm32f7.c
1073
f7_msg->count = data->block[0] + 2;
drivers/i2c/busses/i2c-stm32f7.c
1075
f7_msg->smbus_buf[i] = data->block[i - 1];
drivers/i2c/busses/i2c-stm32f7.c
1863
data->block[i] = f7_msg->smbus_buf[i];
drivers/i2c/busses/i2c-viapro.c
230
outb_p(data->block[0], SMBHSTDAT0);
drivers/i2c/busses/i2c-viapro.c
235
u8 len = data->block[0];
drivers/i2c/busses/i2c-viapro.c
241
outb_p(data->block[i], SMBBLKDAT);
drivers/i2c/busses/i2c-viapro.c
273
data->block[0] = inb_p(SMBHSTDAT0);
drivers/i2c/busses/i2c-viapro.c
274
if (data->block[0] > I2C_SMBUS_BLOCK_MAX)
drivers/i2c/busses/i2c-viapro.c
275
data->block[0] = I2C_SMBUS_BLOCK_MAX;
drivers/i2c/busses/i2c-viapro.c
277
for (i = 1; i <= data->block[0]; i++)
drivers/i2c/busses/i2c-viapro.c
278
data->block[i] = inb_p(SMBBLKDAT);
drivers/i2c/busses/i2c-xgene-slimpro.c
386
&data->block[0]);
drivers/i2c/busses/i2c-xgene-slimpro.c
392
data->block[0] + 1,
drivers/i2c/busses/i2c-xgene-slimpro.c
393
&data->block[0]);
drivers/i2c/busses/i2c-xgene-slimpro.c
404
&data->block[1]);
drivers/i2c/busses/i2c-xgene-slimpro.c
409
data->block[0],
drivers/i2c/busses/i2c-xgene-slimpro.c
410
&data->block[1]);
drivers/i2c/busses/scx200_acb.c
302
len = data->block[0];
drivers/i2c/busses/scx200_acb.c
305
buffer = &data->block[1];
drivers/i2c/i2c-core-base.c
2372
raw_id.block[0] = 3;
drivers/i2c/i2c-core-base.c
2379
id->manufacturer_id = (raw_id.block[1] << 4) | (raw_id.block[2] >> 4);
drivers/i2c/i2c-core-base.c
2380
id->part_id = ((raw_id.block[2] & 0xf) << 5) | (raw_id.block[3] >> 3);
drivers/i2c/i2c-core-base.c
2381
id->die_revision = raw_id.block[3] & 0x7;
drivers/i2c/i2c-core-smbus.c
237
memcpy(values, &data.block[1], data.block[0]);
drivers/i2c/i2c-core-smbus.c
238
return data.block[0];
drivers/i2c/i2c-core-smbus.c
259
data.block[0] = length;
drivers/i2c/i2c-core-smbus.c
260
memcpy(&data.block[1], values, length);
drivers/i2c/i2c-core-smbus.c
276
data.block[0] = length;
drivers/i2c/i2c-core-smbus.c
283
memcpy(values, &data.block[1], data.block[0]);
drivers/i2c/i2c-core-smbus.c
284
return data.block[0];
drivers/i2c/i2c-core-smbus.c
295
data.block[0] = length;
drivers/i2c/i2c-core-smbus.c
296
memcpy(data.block + 1, values, length);
drivers/i2c/i2c-core-smbus.c
403
msg[0].len = data->block[0] + 2;
drivers/i2c/i2c-core-smbus.c
407
data->block[0]);
drivers/i2c/i2c-core-smbus.c
412
memcpy(msg[0].buf + 1, data->block, msg[0].len - 1);
drivers/i2c/i2c-core-smbus.c
418
if (data->block[0] > I2C_SMBUS_BLOCK_MAX) {
drivers/i2c/i2c-core-smbus.c
421
data->block[0]);
drivers/i2c/i2c-core-smbus.c
425
msg[0].len = data->block[0] + 2;
drivers/i2c/i2c-core-smbus.c
427
memcpy(msg[0].buf + 1, data->block, msg[0].len - 1);
drivers/i2c/i2c-core-smbus.c
435
if (data->block[0] > I2C_SMBUS_BLOCK_MAX) {
drivers/i2c/i2c-core-smbus.c
438
data->block[0]);
drivers/i2c/i2c-core-smbus.c
443
msg[1].len = data->block[0];
drivers/i2c/i2c-core-smbus.c
446
msg[0].len = data->block[0] + 1;
drivers/i2c/i2c-core-smbus.c
449
memcpy(msg[0].buf + 1, data->block + 1, data->block[0]);
drivers/i2c/i2c-core-smbus.c
499
memcpy(data->block + 1, msg[1].buf, data->block[0]);
drivers/i2c/i2c-core-smbus.c
510
memcpy(data->block, msg[1].buf, msg[1].buf[0] + 1);
drivers/i2c/i2c-dev.c
373
datasize = sizeof(data->block);
drivers/i2c/i2c-dev.c
387
temp.block[0] = I2C_SMBUS_BLOCK_MAX;
drivers/i2c/i2c-stub.c
217
if (data->block[0] > 256 - command) /* Avoid overrun */
drivers/i2c/i2c-stub.c
218
data->block[0] = 256 - command;
drivers/i2c/i2c-stub.c
219
len = data->block[0];
drivers/i2c/i2c-stub.c
223
chip->words[command + i] |= data->block[1 + i];
drivers/i2c/i2c-stub.c
230
data->block[1 + i] =
drivers/i2c/i2c-stub.c
248
len = data->block[0];
drivers/i2c/i2c-stub.c
265
b->block[i] = data->block[i + 1];
drivers/i2c/i2c-stub.c
267
chip->words[command] = (b->block[0] << 8) | b->len;
drivers/i2c/i2c-stub.c
279
data->block[0] = len;
drivers/i2c/i2c-stub.c
281
data->block[i + 1] = b->block[i];
drivers/i2c/i2c-stub.c
66
u8 block[I2C_SMBUS_BLOCK_MAX];
drivers/iio/buffer/industrialio-buffer-dma.c
100
struct iio_dma_buffer_block *block = container_of(kref,
drivers/iio/buffer/industrialio-buffer-dma.c
102
struct iio_dma_buffer_queue *queue = block->queue;
drivers/iio/buffer/industrialio-buffer-dma.c
104
WARN_ON(block->fileio && block->state != IIO_BLOCK_STATE_DEAD);
drivers/iio/buffer/industrialio-buffer-dma.c
106
if (block->fileio) {
drivers/iio/buffer/industrialio-buffer-dma.c
107
dma_free_coherent(queue->dev, PAGE_ALIGN(block->size),
drivers/iio/buffer/industrialio-buffer-dma.c
108
block->vaddr, block->phys_addr);
drivers/iio/buffer/industrialio-buffer-dma.c
114
kfree(block);
drivers/iio/buffer/industrialio-buffer-dma.c
117
static void iio_buffer_block_get(struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
119
kref_get(&block->kref);
drivers/iio/buffer/industrialio-buffer-dma.c
122
static void iio_buffer_block_put(struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
124
kref_put(&block->kref, iio_buffer_block_release);
drivers/iio/buffer/industrialio-buffer-dma.c
136
struct iio_dma_buffer_block *block, *_block;
drivers/iio/buffer/industrialio-buffer-dma.c
142
list_for_each_entry_safe(block, _block, &block_list, head)
drivers/iio/buffer/industrialio-buffer-dma.c
143
iio_buffer_block_release(&block->kref);
drivers/iio/buffer/industrialio-buffer-dma.c
149
struct iio_dma_buffer_block *block;
drivers/iio/buffer/industrialio-buffer-dma.c
151
block = container_of(kref, struct iio_dma_buffer_block, kref);
drivers/iio/buffer/industrialio-buffer-dma.c
154
list_add_tail(&block->head, &iio_dma_buffer_dead_blocks);
drivers/iio/buffer/industrialio-buffer-dma.c
162
static void iio_buffer_block_put_atomic(struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
164
kref_put(&block->kref, iio_buffer_block_release_atomic);
drivers/iio/buffer/industrialio-buffer-dma.c
176
struct iio_dma_buffer_block *block __free(kfree) =
drivers/iio/buffer/industrialio-buffer-dma.c
177
kzalloc_obj(*block);
drivers/iio/buffer/industrialio-buffer-dma.c
178
if (!block)
drivers/iio/buffer/industrialio-buffer-dma.c
182
block->vaddr = dma_alloc_coherent(queue->dev, PAGE_ALIGN(size),
drivers/iio/buffer/industrialio-buffer-dma.c
183
&block->phys_addr, GFP_KERNEL);
drivers/iio/buffer/industrialio-buffer-dma.c
184
if (!block->vaddr)
drivers/iio/buffer/industrialio-buffer-dma.c
188
block->fileio = fileio;
drivers/iio/buffer/industrialio-buffer-dma.c
189
block->size = size;
drivers/iio/buffer/industrialio-buffer-dma.c
190
block->state = IIO_BLOCK_STATE_DONE;
drivers/iio/buffer/industrialio-buffer-dma.c
191
block->queue = queue;
drivers/iio/buffer/industrialio-buffer-dma.c
192
INIT_LIST_HEAD(&block->head);
drivers/iio/buffer/industrialio-buffer-dma.c
193
kref_init(&block->kref);
drivers/iio/buffer/industrialio-buffer-dma.c
200
return_ptr(block);
drivers/iio/buffer/industrialio-buffer-dma.c
203
static void _iio_dma_buffer_block_done(struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
205
if (block->state != IIO_BLOCK_STATE_DEAD)
drivers/iio/buffer/industrialio-buffer-dma.c
206
block->state = IIO_BLOCK_STATE_DONE;
drivers/iio/buffer/industrialio-buffer-dma.c
228
void iio_dma_buffer_block_done(struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
230
struct iio_dma_buffer_queue *queue = block->queue;
drivers/iio/buffer/industrialio-buffer-dma.c
236
_iio_dma_buffer_block_done(block);
drivers/iio/buffer/industrialio-buffer-dma.c
238
if (!block->fileio)
drivers/iio/buffer/industrialio-buffer-dma.c
239
iio_buffer_signal_dmabuf_done(block->fence, 0);
drivers/iio/buffer/industrialio-buffer-dma.c
241
iio_buffer_block_put_atomic(block);
drivers/iio/buffer/industrialio-buffer-dma.c
260
struct iio_dma_buffer_block *block, *_block;
drivers/iio/buffer/industrialio-buffer-dma.c
266
list_for_each_entry_safe(block, _block, list, head) {
drivers/iio/buffer/industrialio-buffer-dma.c
267
list_del(&block->head);
drivers/iio/buffer/industrialio-buffer-dma.c
268
block->bytes_used = 0;
drivers/iio/buffer/industrialio-buffer-dma.c
269
_iio_dma_buffer_block_done(block);
drivers/iio/buffer/industrialio-buffer-dma.c
271
if (!block->fileio)
drivers/iio/buffer/industrialio-buffer-dma.c
272
iio_buffer_signal_dmabuf_done(block->fence,
drivers/iio/buffer/industrialio-buffer-dma.c
274
iio_buffer_block_put_atomic(block);
drivers/iio/buffer/industrialio-buffer-dma.c
286
static bool iio_dma_block_reusable(struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
293
switch (block->state) {
drivers/iio/buffer/industrialio-buffer-dma.c
322
struct iio_dma_buffer_block *block;
drivers/iio/buffer/industrialio-buffer-dma.c
352
block = queue->fileio.blocks[i];
drivers/iio/buffer/industrialio-buffer-dma.c
355
if (block && (!iio_dma_block_reusable(block) || !try_reuse))
drivers/iio/buffer/industrialio-buffer-dma.c
356
block->state = IIO_BLOCK_STATE_DEAD;
drivers/iio/buffer/industrialio-buffer-dma.c
370
block = queue->fileio.blocks[i];
drivers/iio/buffer/industrialio-buffer-dma.c
371
if (block->state == IIO_BLOCK_STATE_DEAD) {
drivers/iio/buffer/industrialio-buffer-dma.c
373
iio_buffer_block_put(block);
drivers/iio/buffer/industrialio-buffer-dma.c
374
block = NULL;
drivers/iio/buffer/industrialio-buffer-dma.c
376
block->size = size;
drivers/iio/buffer/industrialio-buffer-dma.c
379
block = NULL;
drivers/iio/buffer/industrialio-buffer-dma.c
382
if (!block) {
drivers/iio/buffer/industrialio-buffer-dma.c
383
block = iio_dma_buffer_alloc_block(queue, size, true);
drivers/iio/buffer/industrialio-buffer-dma.c
384
if (!block)
drivers/iio/buffer/industrialio-buffer-dma.c
387
queue->fileio.blocks[i] = block;
drivers/iio/buffer/industrialio-buffer-dma.c
395
block->bytes_used = block->size;
drivers/iio/buffer/industrialio-buffer-dma.c
403
block->state = IIO_BLOCK_STATE_QUEUED;
drivers/iio/buffer/industrialio-buffer-dma.c
404
list_add_tail(&block->head, &queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
406
block->state = IIO_BLOCK_STATE_DONE;
drivers/iio/buffer/industrialio-buffer-dma.c
438
struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
450
block->state = IIO_BLOCK_STATE_ACTIVE;
drivers/iio/buffer/industrialio-buffer-dma.c
451
iio_buffer_block_get(block);
drivers/iio/buffer/industrialio-buffer-dma.c
453
ret = queue->ops->submit(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
455
if (!block->fileio)
drivers/iio/buffer/industrialio-buffer-dma.c
456
iio_buffer_signal_dmabuf_done(block->fence, ret);
drivers/iio/buffer/industrialio-buffer-dma.c
468
iio_buffer_block_put(block);
drivers/iio/buffer/industrialio-buffer-dma.c
485
struct iio_dma_buffer_block *block, *_block;
drivers/iio/buffer/industrialio-buffer-dma.c
489
list_for_each_entry_safe(block, _block, &queue->incoming, head) {
drivers/iio/buffer/industrialio-buffer-dma.c
490
list_del(&block->head);
drivers/iio/buffer/industrialio-buffer-dma.c
491
iio_dma_buffer_submit_block(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
521
struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
523
if (block->state == IIO_BLOCK_STATE_DEAD) {
drivers/iio/buffer/industrialio-buffer-dma.c
524
iio_buffer_block_put(block);
drivers/iio/buffer/industrialio-buffer-dma.c
526
iio_dma_buffer_submit_block(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
528
block->state = IIO_BLOCK_STATE_QUEUED;
drivers/iio/buffer/industrialio-buffer-dma.c
529
list_add_tail(&block->head, &queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
536
struct iio_dma_buffer_block *block;
drivers/iio/buffer/industrialio-buffer-dma.c
542
block = queue->fileio.blocks[idx];
drivers/iio/buffer/industrialio-buffer-dma.c
544
if (block->state != IIO_BLOCK_STATE_DONE)
drivers/iio/buffer/industrialio-buffer-dma.c
550
return block;
drivers/iio/buffer/industrialio-buffer-dma.c
557
struct iio_dma_buffer_block *block;
drivers/iio/buffer/industrialio-buffer-dma.c
567
block = iio_dma_buffer_dequeue(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
568
if (!block)
drivers/iio/buffer/industrialio-buffer-dma.c
572
queue->fileio.active_block = block;
drivers/iio/buffer/industrialio-buffer-dma.c
574
block = queue->fileio.active_block;
drivers/iio/buffer/industrialio-buffer-dma.c
578
if (n > block->bytes_used - queue->fileio.pos)
drivers/iio/buffer/industrialio-buffer-dma.c
579
n = block->bytes_used - queue->fileio.pos;
drivers/iio/buffer/industrialio-buffer-dma.c
580
addr = block->vaddr + queue->fileio.pos;
drivers/iio/buffer/industrialio-buffer-dma.c
591
if (queue->fileio.pos == block->bytes_used) {
drivers/iio/buffer/industrialio-buffer-dma.c
593
iio_dma_buffer_enqueue(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
643
struct iio_dma_buffer_block *block;
drivers/iio/buffer/industrialio-buffer-dma.c
661
block = queue->fileio.blocks[i];
drivers/iio/buffer/industrialio-buffer-dma.c
663
if (block != queue->fileio.active_block && block->state == IIO_BLOCK_STATE_DONE)
drivers/iio/buffer/industrialio-buffer-dma.c
664
data_available += block->size;
drivers/iio/buffer/industrialio-buffer-dma.c
676
struct iio_dma_buffer_block *block;
drivers/iio/buffer/industrialio-buffer-dma.c
687
block = iio_dma_buffer_alloc_block(queue, attach->dmabuf->size, false);
drivers/iio/buffer/industrialio-buffer-dma.c
688
if (!block)
drivers/iio/buffer/industrialio-buffer-dma.c
694
return block;
drivers/iio/buffer/industrialio-buffer-dma.c
699
struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
701
block->state = IIO_BLOCK_STATE_DEAD;
drivers/iio/buffer/industrialio-buffer-dma.c
702
iio_buffer_block_put_atomic(block);
drivers/iio/buffer/industrialio-buffer-dma.c
706
static int iio_dma_can_enqueue_block(struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dma.c
708
struct iio_dma_buffer_queue *queue = block->queue;
drivers/iio/buffer/industrialio-buffer-dma.c
714
switch (block->state) {
drivers/iio/buffer/industrialio-buffer-dma.c
728
struct iio_dma_buffer_block *block,
drivers/iio/buffer/industrialio-buffer-dma.c
741
ret = iio_dma_can_enqueue_block(block);
drivers/iio/buffer/industrialio-buffer-dma.c
745
block->bytes_used = size;
drivers/iio/buffer/industrialio-buffer-dma.c
746
block->cyclic = cyclic;
drivers/iio/buffer/industrialio-buffer-dma.c
747
block->sg_table = sgt;
drivers/iio/buffer/industrialio-buffer-dma.c
748
block->fence = fence;
drivers/iio/buffer/industrialio-buffer-dma.c
750
iio_dma_buffer_enqueue(queue, block);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
107
max_size = min(block->size, dmaengine_buffer->max_size);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
111
block->bytes_used = max_size;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
113
if (!block->bytes_used || block->bytes_used > max_size)
drivers/iio/buffer/industrialio-buffer-dmaengine.c
117
block->phys_addr,
drivers/iio/buffer/industrialio-buffer-dmaengine.c
118
block->bytes_used,
drivers/iio/buffer/industrialio-buffer-dmaengine.c
126
desc->callback_param = block;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
133
list_add_tail(&block->head, &dmaengine_buffer->active);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
51
struct iio_dma_buffer_block *block = data;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
53
scoped_guard(spinlock_irqsave, &block->queue->list_lock)
drivers/iio/buffer/industrialio-buffer-dmaengine.c
54
list_del(&block->head);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
55
block->bytes_used -= result->residue;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
56
iio_dma_buffer_block_done(block);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
60
struct iio_dma_buffer_block *block)
drivers/iio/buffer/industrialio-buffer-dmaengine.c
74
max_size = min(block->size, dmaengine_buffer->max_size);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
82
if (block->sg_table) {
drivers/iio/buffer/industrialio-buffer-dmaengine.c
83
sgl = block->sg_table->sgl;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
84
nents = sg_nents_for_len(sgl, block->bytes_used);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
92
len_total = block->bytes_used;
drivers/iio/chemical/sunrise_co2.c
60
data.block[0] = val_size;
drivers/iio/chemical/sunrise_co2.c
81
memcpy(val_buf, &data.block[1], data.block[0]);
drivers/iio/chemical/sunrise_co2.c
97
data.block[0] = count;
drivers/iio/chemical/sunrise_co2.c
98
memcpy(&data.block[1], (u8 *)val_buf + 1, count);
drivers/iio/industrialio-buffer.c
1570
buffer->access->detach_dmabuf(buffer, priv->block);
drivers/iio/industrialio-buffer.c
1717
priv->block = buffer->access->attach_dmabuf(buffer, attach);
drivers/iio/industrialio-buffer.c
1718
if (IS_ERR(priv->block)) {
drivers/iio/industrialio-buffer.c
1719
err = PTR_ERR(priv->block);
drivers/iio/industrialio-buffer.c
1916
ret = buffer->access->enqueue_dmabuf(buffer, priv->block, &fence->base,
drivers/iio/industrialio-buffer.c
46
struct iio_dma_buffer_block *block;
drivers/infiniband/hw/bng_re/bng_fw.h
137
u8 block;
drivers/infiniband/hw/bng_re/bng_fw.h
149
u32 req_sz, u32 res_sz, u8 block)
drivers/infiniband/hw/bng_re/bng_fw.h
156
msg->block = block;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
788
bool block = true;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
791
block = !(flags & RDMA_DESTROY_AH_SLEEPABLE);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
792
rc = bnxt_qplib_destroy_ah(&rdev->qplib_res, &ah->qplib_ah, block);
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
316
if (msg->block)
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
507
if (msg->block)
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
564
if (!msg->block) {
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
248
u8 block;
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
253
u32 req_sz, u32 res_sz, u8 block)
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
260
msg->block = block;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
410
bool block)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
446
sizeof(resp), block);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
456
bool block)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
472
sizeof(resp), block);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
556
bool block)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
570
sizeof(resp), block);
drivers/infiniband/hw/bnxt_re/qplib_sp.h
338
bool block);
drivers/infiniband/hw/bnxt_re/qplib_sp.h
340
bool block);
drivers/infiniband/hw/bnxt_re/qplib_sp.h
344
bool block);
drivers/infiniband/hw/mlx4/mad.c
334
int block, u32 change_bitmap)
drivers/infiniband/hw/mlx4/mad.c
352
[ix] == i + 32 * block) {
drivers/infiniband/hw/mlx4/mad.c
94
int block, u32 change_bitmap);
drivers/input/mouse/cyapa_gen3.c
647
u16 block, const u8 *data)
drivers/input/mouse/cyapa_gen3.c
659
put_unaligned_be16(block, &write_block_cmd.block_num);
drivers/input/mouse/cyapa_gen3.c
706
size_t block = start_block + i;
drivers/input/mouse/cyapa_gen3.c
710
error = cyapa_gen3_write_fw_block(cyapa, block, data);
drivers/isdn/hardware/mISDN/hfcmulti.c
2600
fifo_irq(struct hfc_multi *hc, int block)
drivers/isdn/hardware/mISDN/hfcmulti.c
2607
r_irq_fifo_bl = HFC_inb_nodebug(hc, R_IRQ_FIFO_BL0 + block);
drivers/isdn/hardware/mISDN/hfcmulti.c
2610
ch = (block << 2) + (j >> 1);
drivers/md/bcache/debug.c
84
unsigned int block = ((void *) i - (void *) ondisk) /
drivers/md/bcache/debug.c
87
pr_err("*** on disk block %u:\n", block);
drivers/md/bcache/debug.c
88
bch_dump_bset(&b->keys, i, block);
drivers/md/dm-bufio.c
1364
static inline sector_t block_to_sector(struct dm_bufio_client *c, sector_t block)
drivers/md/dm-bufio.c
1369
sector = block << c->sectors_per_block_bits;
drivers/md/dm-bufio.c
1371
sector = block * (c->block_size >> SECTOR_SHIFT);
drivers/md/dm-bufio.c
1386
sector = block_to_sector(b->c, b->block);
drivers/md/dm-bufio.c
1675
b->block = -1;
drivers/md/dm-bufio.c
1775
struct buffer_tree *tree, sector_t block,
drivers/md/dm-bufio.c
1795
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
1806
b->block = block;
drivers/md/dm-bufio.c
1871
static void *new_read(struct dm_bufio_client *c, sector_t block,
drivers/md/dm-bufio.c
1887
tree = cache_get_tree(&c->cache, block);
drivers/md/dm-bufio.c
1888
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
1913
b = __bufio_new(c, tree, block, nf, &need_submit, &write_list);
drivers/md/dm-bufio.c
1946
void *dm_bufio_get(struct dm_bufio_client *c, sector_t block,
drivers/md/dm-bufio.c
1949
return new_read(c, block, NF_GET, bp, IOPRIO_DEFAULT);
drivers/md/dm-bufio.c
1953
static void *__dm_bufio_read(struct dm_bufio_client *c, sector_t block,
drivers/md/dm-bufio.c
1959
return new_read(c, block, NF_READ, bp, ioprio);
drivers/md/dm-bufio.c
1962
void *dm_bufio_read(struct dm_bufio_client *c, sector_t block,
drivers/md/dm-bufio.c
1965
return __dm_bufio_read(c, block, bp, IOPRIO_DEFAULT);
drivers/md/dm-bufio.c
1969
void *dm_bufio_read_with_ioprio(struct dm_bufio_client *c, sector_t block,
drivers/md/dm-bufio.c
1972
return __dm_bufio_read(c, block, bp, ioprio);
drivers/md/dm-bufio.c
1976
void *dm_bufio_new(struct dm_bufio_client *c, sector_t block,
drivers/md/dm-bufio.c
1982
return new_read(c, block, NF_FRESH, bp, IOPRIO_DEFAULT);
drivers/md/dm-bufio.c
1987
sector_t block, unsigned int n_blocks,
drivers/md/dm-bufio.c
1999
for (; n_blocks--; block++) {
drivers/md/dm-bufio.c
2004
tree = cache_get_tree(&c->cache, block);
drivers/md/dm-bufio.c
2005
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
2013
b = __bufio_new(c, tree, block, NF_PREFETCH, &need_submit,
drivers/md/dm-bufio.c
2042
void dm_bufio_prefetch(struct dm_bufio_client *c, sector_t block, unsigned int n_blocks)
drivers/md/dm-bufio.c
2044
return __dm_bufio_prefetch(c, block, n_blocks, IOPRIO_DEFAULT);
drivers/md/dm-bufio.c
2048
void dm_bufio_prefetch_with_ioprio(struct dm_bufio_client *c, sector_t block,
drivers/md/dm-bufio.c
2051
return __dm_bufio_prefetch(c, block, n_blocks, ioprio);
drivers/md/dm-bufio.c
2058
struct buffer_tree *tree = cache_get_tree(&c->cache, b->block);
drivers/md/dm-bufio.c
2100
cache_mark(&c->cache, cache_get_tree(&c->cache, b->block), b,
drivers/md/dm-bufio.c
2180
tree = cache_get_tree(&c->cache, b->block);
drivers/md/dm-bufio.c
2230
int dm_bufio_issue_discard(struct dm_bufio_client *c, sector_t block, sector_t count)
drivers/md/dm-bufio.c
2240
.sector = block_to_sector(c, block),
drivers/md/dm-bufio.c
2251
static void forget_buffer(struct dm_bufio_client *c, sector_t block)
drivers/md/dm-bufio.c
2253
struct buffer_tree *tree = cache_get_tree(&c->cache, block);
drivers/md/dm-bufio.c
2256
b = cache_get(&c->cache, tree, block);
drivers/md/dm-bufio.c
2275
void dm_bufio_forget(struct dm_bufio_client *c, sector_t block)
drivers/md/dm-bufio.c
2278
forget_buffer(c, block);
drivers/md/dm-bufio.c
2288
void dm_bufio_forget_buffers(struct dm_bufio_client *c, sector_t block, sector_t n_blocks)
drivers/md/dm-bufio.c
2291
cache_remove_range(&c->cache, block, block + n_blocks, idle, __free_buffer_wake);
drivers/md/dm-bufio.c
2332
return b->block;
drivers/md/dm-bufio.c
2361
(unsigned long long)b->block, atomic_read(&b->hold_count), b->list_mode);
drivers/md/dm-bufio.c
406
sector_t block)
drivers/md/dm-bufio.c
597
static struct dm_buffer *__cache_get(const struct rb_root *root, sector_t block)
drivers/md/dm-bufio.c
605
if (b->block == block)
drivers/md/dm-bufio.c
608
n = block < b->block ? n->rb_left : n->rb_right;
drivers/md/dm-bufio.c
621
struct buffer_tree *tree, sector_t block)
drivers/md/dm-bufio.c
627
b = __cache_get(&tree->root, block);
drivers/md/dm-bufio.c
683
rb_erase(&b->node, &cache_get_tree(bc, b->block)->root);
drivers/md/dm-bufio.c
828
if (found->block == b->block)
drivers/md/dm-bufio.c
832
new = b->block < found->block ?
drivers/md/dm-bufio.c
894
static struct dm_buffer *__find_next(struct rb_root *root, sector_t block)
drivers/md/dm-bufio.c
903
if (b->block == block)
drivers/md/dm-bufio.c
906
if (block <= b->block) {
drivers/md/dm-bufio.c
928
if (!b || (b->block >= end))
drivers/md/dm-cache-metadata.c
734
static __le64 pack_value(dm_oblock_t block, unsigned int flags)
drivers/md/dm-cache-metadata.c
736
uint64_t value = from_oblock(block);
drivers/md/dm-cache-metadata.c
743
static void unpack_value(__le64 value_le, dm_oblock_t *block, unsigned int *flags)
drivers/md/dm-cache-metadata.c
748
*block = to_oblock(b);
drivers/md/dm-cache-policy-smq.c
102
static struct entry *to_entry(struct entry_space *es, unsigned int block)
drivers/md/dm-cache-policy-smq.c
104
if (block == INDEXER_NULL)
drivers/md/dm-cache-policy-smq.c
107
return __get_entry(es, block);
drivers/md/dm-cache-policy-smq.c
86
static struct entry *__get_entry(struct entry_space *es, unsigned int block)
drivers/md/dm-cache-policy-smq.c
90
e = es->begin + block;
drivers/md/dm-cache-target.c
1070
static bool optimisable_bio(struct cache *cache, struct bio *bio, dm_oblock_t block)
drivers/md/dm-cache-target.c
1073
(is_discarded_oblock(cache, block) || bio_writes_complete_block(cache, bio));
drivers/md/dm-cache-target.c
1618
static int map_bio(struct cache *cache, struct bio *bio, dm_oblock_t block,
drivers/md/dm-cache-target.c
1627
rb = bio_detain_shared(cache, block, bio);
drivers/md/dm-cache-target.c
1641
if (optimisable_bio(cache, bio, block)) {
drivers/md/dm-cache-target.c
1644
r = policy_lookup_with_work(cache->policy, block, &cblock, data_dir, true, &op);
drivers/md/dm-cache-target.c
1659
r = policy_lookup(cache->policy, block, &cblock, data_dir, false, &background_queued);
drivers/md/dm-cache-target.c
1680
remap_to_origin_clear_discard(cache, bio, block);
drivers/md/dm-cache-target.c
1703
invalidate_start(cache, cblock, block, bio);
drivers/md/dm-cache-target.c
1705
remap_to_origin_clear_discard(cache, bio, block);
drivers/md/dm-cache-target.c
1709
remap_to_origin_and_cache(cache, bio, block, cblock);
drivers/md/dm-cache-target.c
1712
remap_to_cache_dirty(cache, bio, block, cblock);
drivers/md/dm-cache-target.c
2650
dm_oblock_t block = get_bio_block(cache, bio);
drivers/md/dm-cache-target.c
2653
if (unlikely(from_oblock(block) >= from_oblock(cache->origin_blocks))) {
drivers/md/dm-cache-target.c
2669
r = map_bio(cache, bio, block, &commit_needed);
drivers/md/dm-cache-target.c
733
sector_t block = from_cblock(cblock);
drivers/md/dm-cache-target.c
738
(block * cache->sectors_per_block) +
drivers/md/dm-cache-target.c
742
(block << cache->sectors_per_block_shift) |
drivers/md/dm-dust.c
105
static int dust_add_block(struct dust_device *dd, unsigned long long block,
drivers/md/dm-dust.c
119
bblock->bb = block;
drivers/md/dm-dust.c
124
__func__, block);
drivers/md/dm-dust.c
134
__func__, block, wr_fail_cnt);
drivers/md/dm-dust.c
141
static int dust_query_block(struct dust_device *dd, unsigned long long block, char *result,
drivers/md/dm-dust.c
149
bblock = dust_rb_search(&dd->badblocklist, block);
drivers/md/dm-dust.c
151
DMEMIT("%s: block %llu found in badblocklist", __func__, block);
drivers/md/dm-dust.c
153
DMEMIT("%s: block %llu not found in badblocklist", __func__, block);
drivers/md/dm-dust.c
421
unsigned long long tmp, block;
drivers/md/dm-dust.c
464
block = tmp;
drivers/md/dm-dust.c
466
if (block > size) {
drivers/md/dm-dust.c
472
r = dust_add_block(dd, block, 0);
drivers/md/dm-dust.c
474
r = dust_remove_block(dd, block);
drivers/md/dm-dust.c
476
r = dust_query_block(dd, block, result, maxlen, &sz);
drivers/md/dm-dust.c
487
block = tmp;
drivers/md/dm-dust.c
494
if (block > size) {
drivers/md/dm-dust.c
500
r = dust_add_block(dd, block, wr_fail_cnt);
drivers/md/dm-dust.c
78
static int dust_remove_block(struct dust_device *dd, unsigned long long block)
drivers/md/dm-dust.c
84
bblock = dust_rb_search(&dd->badblocklist, block);
drivers/md/dm-dust.c
89
__func__, block);
drivers/md/dm-dust.c
98
DMINFO("%s: badblock removed at block %llu", __func__, block);
drivers/md/dm-ebs-target.c
115
block++;
drivers/md/dm-ebs-target.c
145
sector_t block, blocks, sector = bio->bi_iter.bi_sector;
drivers/md/dm-ebs-target.c
147
block = __sector_to_block(ec, sector);
drivers/md/dm-ebs-target.c
155
block++;
drivers/md/dm-ebs-target.c
163
return blocks ? dm_bufio_issue_discard(ec->bufio, block, blocks) : 0;
drivers/md/dm-ebs-target.c
73
sector_t block = __sector_to_block(ec, iter->bi_sector);
drivers/md/dm-ebs-target.c
87
ba = dm_bufio_read(ec->bufio, block, &b);
drivers/md/dm-ebs-target.c
89
ba = dm_bufio_new(ec->bufio, block, &b);
drivers/md/dm-era-target.c
106
static bool writeset_marked(struct writeset *ws, dm_block_t block)
drivers/md/dm-era-target.c
108
return test_bit(block, ws->bits);
drivers/md/dm-era-target.c
112
struct writeset_metadata *m, dm_block_t block,
drivers/md/dm-era-target.c
122
r = dm_bitset_test_bit(info, m->root, block, &m->root, result);
drivers/md/dm-era-target.c
137
struct writeset *ws, uint32_t block)
drivers/md/dm-era-target.c
141
if (!test_bit(block, ws->bits)) {
drivers/md/dm-era-target.c
142
r = dm_bitset_set_bit(info, ws->md.root, block, &ws->md.root);
drivers/md/dm-era-target.c
1563
dm_block_t block = get_block(era, bio);
drivers/md/dm-era-target.c
1577
!metadata_current_marked(era->md, block)) {
drivers/md/dm-era-target.c
970
static bool metadata_current_marked(struct era_metadata *md, dm_block_t block)
drivers/md/dm-era-target.c
977
r = writeset_marked(ws, block);
drivers/md/dm-kcopyd.c
903
int kcopyd_cancel(struct kcopyd_job *job, int block)
drivers/md/dm-log-writes.c
135
struct pending_block *block;
drivers/md/dm-log-writes.c
199
struct pending_block *block)
drivers/md/dm-log-writes.c
203
for (i = 0; i < block->vec_cnt; i++) {
drivers/md/dm-log-writes.c
204
if (block->vecs[i].bv_page)
drivers/md/dm-log-writes.c
205
__free_page(block->vecs[i].bv_page);
drivers/md/dm-log-writes.c
207
kfree(block->data);
drivers/md/dm-log-writes.c
208
kfree(block);
drivers/md/dm-log-writes.c
319
struct pending_block *block, sector_t sector)
drivers/md/dm-log-writes.c
326
entry.sector = cpu_to_le64(block->sector);
drivers/md/dm-log-writes.c
327
entry.nr_sectors = cpu_to_le64(block->nr_sectors);
drivers/md/dm-log-writes.c
328
entry.flags = cpu_to_le64(block->flags);
drivers/md/dm-log-writes.c
329
entry.data_len = cpu_to_le64(block->datalen);
drivers/md/dm-log-writes.c
331
metadatalen = (block->flags & LOG_MARK_FLAG) ? block->datalen : 0;
drivers/md/dm-log-writes.c
332
if (write_metadata(lc, &entry, sizeof(entry), block->data,
drivers/md/dm-log-writes.c
334
free_pending_block(lc, block);
drivers/md/dm-log-writes.c
340
if (block->datalen && metadatalen == 0) {
drivers/md/dm-log-writes.c
341
if (write_inline_data(lc, &entry, sizeof(entry), block->data,
drivers/md/dm-log-writes.c
342
block->datalen, sector)) {
drivers/md/dm-log-writes.c
343
free_pending_block(lc, block);
drivers/md/dm-log-writes.c
350
if (!block->vec_cnt)
drivers/md/dm-log-writes.c
354
bio = bio_alloc(lc->logdev->bdev, bio_max_segs(block->vec_cnt),
drivers/md/dm-log-writes.c
361
for (i = 0; i < block->vec_cnt; i++) {
drivers/md/dm-log-writes.c
366
ret = bio_add_page(bio, block->vecs[i].bv_page,
drivers/md/dm-log-writes.c
367
block->vecs[i].bv_len, 0);
drivers/md/dm-log-writes.c
368
if (ret != block->vecs[i].bv_len) {
drivers/md/dm-log-writes.c
372
bio_max_segs(block->vec_cnt - i),
drivers/md/dm-log-writes.c
379
ret = bio_add_page(bio, block->vecs[i].bv_page,
drivers/md/dm-log-writes.c
380
block->vecs[i].bv_len, 0);
drivers/md/dm-log-writes.c
381
if (ret != block->vecs[i].bv_len) {
drivers/md/dm-log-writes.c
387
sector += block->vecs[i].bv_len >> SECTOR_SHIFT;
drivers/md/dm-log-writes.c
391
kfree(block->data);
drivers/md/dm-log-writes.c
392
kfree(block);
drivers/md/dm-log-writes.c
396
free_pending_block(lc, block);
drivers/md/dm-log-writes.c
439
struct pending_block *block = NULL;
drivers/md/dm-log-writes.c
444
block = list_first_entry(&lc->logging_blocks,
drivers/md/dm-log-writes.c
446
list_del_init(&block->list);
drivers/md/dm-log-writes.c
451
if (!(block->flags & LOG_DISCARD_FLAG))
drivers/md/dm-log-writes.c
452
lc->next_sector += dev_to_bio_sectors(lc, block->nr_sectors);
drivers/md/dm-log-writes.c
470
super = (block->flags & (LOG_FUA_FLAG | LOG_MARK_FLAG));
drivers/md/dm-log-writes.c
477
if (block) {
drivers/md/dm-log-writes.c
479
ret = log_one_block(lc, block, sector);
drivers/md/dm-log-writes.c
488
free_pending_block(lc, block);
drivers/md/dm-log-writes.c
587
struct pending_block *block;
drivers/md/dm-log-writes.c
590
block = kzalloc_obj(struct pending_block);
drivers/md/dm-log-writes.c
591
if (!block) {
drivers/md/dm-log-writes.c
596
block->data = kstrndup(data, maxsize - 1, GFP_KERNEL);
drivers/md/dm-log-writes.c
597
if (!block->data) {
drivers/md/dm-log-writes.c
599
kfree(block);
drivers/md/dm-log-writes.c
603
block->datalen = strlen(block->data);
drivers/md/dm-log-writes.c
604
block->flags |= LOG_MARK_FLAG;
drivers/md/dm-log-writes.c
606
list_add_tail(&block->list, &lc->logging_blocks);
drivers/md/dm-log-writes.c
648
struct pending_block *block;
drivers/md/dm-log-writes.c
658
pb->block = NULL;
drivers/md/dm-log-writes.c
681
alloc_size = struct_size(block, vecs, bio_segments(bio));
drivers/md/dm-log-writes.c
683
block = kzalloc(alloc_size, GFP_NOIO);
drivers/md/dm-log-writes.c
684
if (!block) {
drivers/md/dm-log-writes.c
691
INIT_LIST_HEAD(&block->list);
drivers/md/dm-log-writes.c
692
pb->block = block;
drivers/md/dm-log-writes.c
696
block->flags |= LOG_FLUSH_FLAG;
drivers/md/dm-log-writes.c
698
block->flags |= LOG_FUA_FLAG;
drivers/md/dm-log-writes.c
700
block->flags |= LOG_DISCARD_FLAG;
drivers/md/dm-log-writes.c
702
block->flags |= LOG_METADATA_FLAG;
drivers/md/dm-log-writes.c
704
block->sector = bio_to_dev_sectors(lc, bio->bi_iter.bi_sector);
drivers/md/dm-log-writes.c
705
block->nr_sectors = bio_to_dev_sectors(lc, bio_sectors(bio));
drivers/md/dm-log-writes.c
719
list_splice_init(&lc->unflushed_blocks, &block->list);
drivers/md/dm-log-writes.c
740
free_pending_block(lc, block);
drivers/md/dm-log-writes.c
750
block->vecs[i].bv_page = page;
drivers/md/dm-log-writes.c
751
block->vecs[i].bv_len = bv.bv_len;
drivers/md/dm-log-writes.c
752
block->vec_cnt++;
drivers/md/dm-log-writes.c
759
list_splice_init(&lc->unflushed_blocks, &block->list);
drivers/md/dm-log-writes.c
773
if (bio_data_dir(bio) == WRITE && pb->block) {
drivers/md/dm-log-writes.c
774
struct pending_block *block = pb->block;
drivers/md/dm-log-writes.c
778
if (block->flags & LOG_FLUSH_FLAG) {
drivers/md/dm-log-writes.c
779
list_splice_tail_init(&block->list, &lc->logging_blocks);
drivers/md/dm-log-writes.c
780
list_add_tail(&block->list, &lc->logging_blocks);
drivers/md/dm-log-writes.c
782
} else if (block->flags & LOG_FUA_FLAG) {
drivers/md/dm-log-writes.c
783
list_add_tail(&block->list, &lc->logging_blocks);
drivers/md/dm-log-writes.c
786
list_add_tail(&block->list, &lc->unflushed_blocks);
drivers/md/dm-log.c
675
static int core_in_sync(struct dm_dirty_log *log, region_t region, int block)
drivers/md/dm-thin-metadata.c
1542
result->block = exception_block;
drivers/md/dm-thin-metadata.c
1546
static int __find_block(struct dm_thin_device *td, dm_block_t block,
drivers/md/dm-thin-metadata.c
1552
dm_block_t keys[2] = { td->id, block };
drivers/md/dm-thin-metadata.c
1567
int dm_thin_find_block(struct dm_thin_device *td, dm_block_t block,
drivers/md/dm-thin-metadata.c
1579
r = __find_block(td, block, can_issue_io, result);
drivers/md/dm-thin-metadata.c
1585
static int __find_next_mapped_block(struct dm_thin_device *td, dm_block_t block,
drivers/md/dm-thin-metadata.c
1592
dm_block_t keys[2] = { td->id, block };
drivers/md/dm-thin-metadata.c
1621
*pool_begin = lookup.block;
drivers/md/dm-thin-metadata.c
1635
if ((lookup.block != pool_end) ||
drivers/md/dm-thin-metadata.c
1665
static int __insert(struct dm_thin_device *td, dm_block_t block,
drivers/md/dm-thin-metadata.c
1671
dm_block_t keys[2] = { td->id, block };
drivers/md/dm-thin-metadata.c
1688
int dm_thin_insert_block(struct dm_thin_device *td, dm_block_t block,
drivers/md/dm-thin-metadata.c
1695
r = __insert(td, block, data_block);
drivers/md/dm-thin-metadata.h
137
dm_block_t block;
drivers/md/dm-thin-metadata.h
147
int dm_thin_find_block(struct dm_thin_device *td, dm_block_t block,
drivers/md/dm-thin-metadata.h
167
int dm_thin_insert_block(struct dm_thin_device *td, dm_block_t block,
drivers/md/dm-thin.c
1799
static void break_sharing(struct thin_c *tc, struct bio *bio, dm_block_t block,
drivers/md/dm-thin.c
1811
schedule_internal_copy(tc, block, lookup_result->block,
drivers/md/dm-thin.c
1849
dm_block_t block)
drivers/md/dm-thin.c
1865
remap_and_issue(tc, bio, block);
drivers/md/dm-thin.c
1869
dm_block_t block,
drivers/md/dm-thin.c
1881
build_data_key(tc->td, lookup_result->block, &key);
drivers/md/dm-thin.c
1888
break_sharing(tc, bio, block, &key, lookup_result, data_cell);
drivers/md/dm-thin.c
1895
remap_and_issue(tc, bio, lookup_result->block);
drivers/md/dm-thin.c
1897
remap_and_issue_shared_cell(tc, data_cell, lookup_result->block);
drivers/md/dm-thin.c
1898
remap_and_issue_shared_cell(tc, virt_cell, lookup_result->block);
drivers/md/dm-thin.c
1902
static void provision_block(struct thin_c *tc, struct bio *bio, dm_block_t block,
drivers/md/dm-thin.c
1934
schedule_external_copy(tc, block, data_block, cell, bio);
drivers/md/dm-thin.c
1936
schedule_zero(tc, block, data_block, cell, bio);
drivers/md/dm-thin.c
1956
dm_block_t block = get_bio_block(tc, bio);
drivers/md/dm-thin.c
1964
r = dm_thin_find_block(tc->td, block, 1, &lookup_result);
drivers/md/dm-thin.c
1968
process_shared_bio(tc, bio, block, &lookup_result, cell);
drivers/md/dm-thin.c
1971
remap_and_issue(tc, bio, lookup_result.block);
drivers/md/dm-thin.c
1972
inc_remap_and_issue_cell(tc, cell, lookup_result.block);
drivers/md/dm-thin.c
1994
provision_block(tc, bio, block, cell);
drivers/md/dm-thin.c
2009
dm_block_t block = get_bio_block(tc, bio);
drivers/md/dm-thin.c
2017
build_virtual_key(tc->td, block, &key);
drivers/md/dm-thin.c
2029
dm_block_t block = get_bio_block(tc, bio);
drivers/md/dm-thin.c
2032
r = dm_thin_find_block(tc->td, block, 1, &lookup_result);
drivers/md/dm-thin.c
2041
remap_and_issue(tc, bio, lookup_result.block);
drivers/md/dm-thin.c
2043
inc_remap_and_issue_cell(tc, cell, lookup_result.block);
drivers/md/dm-thin.c
2723
dm_block_t block = get_bio_block(tc, bio);
drivers/md/dm-thin.c
2751
build_virtual_key(tc->td, block, &key);
drivers/md/dm-thin.c
2755
r = dm_thin_find_block(td, block, 0, &result);
drivers/md/dm-thin.c
2781
build_data_key(tc->td, result.block, &key);
drivers/md/dm-thin.c
2791
remap(tc, bio, result.block);
drivers/md/dm-thin.c
714
static void remap(struct thin_c *tc, struct bio *bio, dm_block_t block)
drivers/md/dm-thin.c
722
(block << pool->sectors_per_block_shift) |
drivers/md/dm-thin.c
725
bio->bi_iter.bi_sector = (block * pool->sectors_per_block) +
drivers/md/dm-thin.c
787
dm_block_t block)
drivers/md/dm-thin.c
789
remap(tc, bio, block);
drivers/md/dm-thin.c
929
dm_block_t block)
drivers/md/dm-thin.c
950
remap_and_issue(info.tc, bio, block);
drivers/md/dm-vdo/data-vio.c
1429
struct compressed_block *block = data_vio->compression.block;
drivers/md/dm-vdo/data-vio.c
1430
int result = vdo_get_compressed_block_fragment(mapping_state, block,
drivers/md/dm-vdo/data-vio.c
1438
size = LZ4_decompress_safe((block->data + fragment_offset), buffer,
drivers/md/dm-vdo/data-vio.c
1559
result = vio_reset_bio(vio, (char *) data_vio->compression.block,
drivers/md/dm-vdo/data-vio.c
1766
data_vio->compression.block->data, VDO_BLOCK_SIZE,
drivers/md/dm-vdo/data-vio.c
536
memset(&data_vio->compression, 0, offsetof(struct compression_state, block));
drivers/md/dm-vdo/data-vio.c
796
&data_vio->compression.block);
drivers/md/dm-vdo/data-vio.c
827
vdo_free(vdo_forget(data_vio->compression.block));
drivers/md/dm-vdo/data-vio.h
140
struct compressed_block *block;
drivers/md/dm-vdo/dedupe.c
1198
(char *) agent->compression.block :
drivers/md/dm-vdo/encodings.c
300
int __must_check vdo_parse_geometry_block(u8 *block, struct volume_geometry *geometry)
drivers/md/dm-vdo/encodings.c
307
if (memcmp(block, VDO_GEOMETRY_MAGIC_NUMBER, VDO_GEOMETRY_MAGIC_NUMBER_SIZE) != 0)
drivers/md/dm-vdo/encodings.c
311
vdo_decode_header(block, &offset, &header);
drivers/md/dm-vdo/encodings.c
322
decode_volume_geometry(block, &offset, geometry, header.version.major_version);
drivers/md/dm-vdo/encodings.c
330
checksum = vdo_crc32(block, offset);
drivers/md/dm-vdo/encodings.c
331
decode_u32_le(block, &offset, &saved_checksum);
drivers/md/dm-vdo/encodings.c
772
struct slab_journal_entry vdo_decode_slab_journal_entry(struct packed_slab_journal_block *block,
drivers/md/dm-vdo/encodings.c
776
vdo_unpack_slab_journal_entry(&block->payload.entries[entry_count]);
drivers/md/dm-vdo/encodings.c
778
if (block->header.has_block_map_increments &&
drivers/md/dm-vdo/encodings.c
779
((block->payload.full_entries.entry_types[entry_count / 8] &
drivers/md/dm-vdo/encodings.h
1218
vdo_decode_slab_journal_entry(struct packed_slab_journal_block *block,
drivers/md/dm-vdo/encodings.h
806
int __must_check vdo_parse_geometry_block(unsigned char *block,
drivers/md/dm-vdo/errors.c
106
for (block = registered_errors.blocks;
drivers/md/dm-vdo/errors.c
107
block < registered_errors.blocks + registered_errors.count;
drivers/md/dm-vdo/errors.c
108
block++) {
drivers/md/dm-vdo/errors.c
109
if ((errnum >= block->base) && (errnum < block->last)) {
drivers/md/dm-vdo/errors.c
110
*info_ptr = block->infos + (errnum - block->base);
drivers/md/dm-vdo/errors.c
111
return block->name;
drivers/md/dm-vdo/errors.c
112
} else if ((errnum >= block->last) && (errnum < block->max)) {
drivers/md/dm-vdo/errors.c
114
return block->name;
drivers/md/dm-vdo/errors.c
275
struct error_block *block;
drivers/md/dm-vdo/errors.c
294
for (block = registered_errors.blocks;
drivers/md/dm-vdo/errors.c
295
block < registered_errors.blocks + registered_errors.count;
drivers/md/dm-vdo/errors.c
296
block++) {
drivers/md/dm-vdo/errors.c
297
if (strcmp(block_name, block->name) == 0)
drivers/md/dm-vdo/errors.c
301
if ((first_error < block->max) && (next_free_error > block->base))
drivers/md/dm-vdo/errors.c
99
struct error_block *block;
drivers/md/dm-vdo/memory-alloc.c
118
static void add_vmalloc_block(struct vmalloc_block_info *block)
drivers/md/dm-vdo/memory-alloc.c
123
block->next = memory_stats.vmalloc_list;
drivers/md/dm-vdo/memory-alloc.c
124
memory_stats.vmalloc_list = block;
drivers/md/dm-vdo/memory-alloc.c
126
memory_stats.vmalloc_bytes += block->size;
drivers/md/dm-vdo/memory-alloc.c
133
struct vmalloc_block_info *block;
drivers/md/dm-vdo/memory-alloc.c
139
(block = *block_ptr) != NULL;
drivers/md/dm-vdo/memory-alloc.c
140
block_ptr = &block->next) {
drivers/md/dm-vdo/memory-alloc.c
141
if (block->ptr == ptr) {
drivers/md/dm-vdo/memory-alloc.c
142
*block_ptr = block->next;
drivers/md/dm-vdo/memory-alloc.c
144
memory_stats.vmalloc_bytes -= block->size;
drivers/md/dm-vdo/memory-alloc.c
150
if (block != NULL)
drivers/md/dm-vdo/memory-alloc.c
151
vdo_free(block);
drivers/md/dm-vdo/memory-alloc.c
246
struct vmalloc_block_info *block;
drivers/md/dm-vdo/memory-alloc.c
248
if (vdo_allocate(1, struct vmalloc_block_info, __func__, &block) == VDO_SUCCESS) {
drivers/md/dm-vdo/memory-alloc.c
274
vdo_free(block);
drivers/md/dm-vdo/memory-alloc.c
276
block->ptr = p;
drivers/md/dm-vdo/memory-alloc.c
277
block->size = PAGE_ALIGN(size);
drivers/md/dm-vdo/memory-alloc.c
278
add_vmalloc_block(block);
drivers/md/dm-vdo/packer.c
367
static void initialize_compressed_block(struct compressed_block *block, u16 size)
drivers/md/dm-vdo/packer.c
375
block->header.version = vdo_pack_version_number(COMPRESSED_BLOCK_1_0);
drivers/md/dm-vdo/packer.c
376
block->header.sizes[0] = __cpu_to_le16(size);
drivers/md/dm-vdo/packer.c
393
struct compressed_block *block)
drivers/md/dm-vdo/packer.c
396
char *fragment = to_pack->block->data;
drivers/md/dm-vdo/packer.c
401
block->header.sizes[slot] = __cpu_to_le16(to_pack->size);
drivers/md/dm-vdo/packer.c
402
memcpy(&block->data[offset], fragment, to_pack->size);
drivers/md/dm-vdo/packer.c
430
struct compressed_block *block;
drivers/md/dm-vdo/packer.c
440
block = compression->block;
drivers/md/dm-vdo/packer.c
441
initialize_compressed_block(block, compression->size);
drivers/md/dm-vdo/packer.c
445
offset = pack_fragment(compression, client, offset, slot++, block);
drivers/md/dm-vdo/packer.c
458
memset(&block->header.sizes[slot], 0,
drivers/md/dm-vdo/packer.c
468
result = vio_reset_bio(&agent->vio, (char *) block, compressed_write_end_io,
drivers/md/dm-vdo/packer.c
47
struct compressed_block *block,
drivers/md/dm-vdo/packer.c
59
version = vdo_unpack_version_number(block->header.version);
drivers/md/dm-vdo/packer.c
67
compressed_size = __le16_to_cpu(block->header.sizes[slot]);
drivers/md/dm-vdo/packer.c
69
offset += __le16_to_cpu(block->header.sizes[i]);
drivers/md/dm-vdo/packer.h
98
struct compressed_block *block,
drivers/md/dm-vdo/recovery-journal.c
1046
struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
1048
if (!block->committing)
drivers/md/dm-vdo/recovery-journal.c
1049
vdo_waitq_enqueue_waiter(&journal->pending_writes, &block->write_waiter);
drivers/md/dm-vdo/recovery-journal.c
1060
static void release_journal_block_reference(struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
1062
vdo_release_recovery_journal_block_reference(block->journal,
drivers/md/dm-vdo/recovery-journal.c
1063
block->sequence_number,
drivers/md/dm-vdo/recovery-journal.c
1091
struct recovery_journal_block *block = context;
drivers/md/dm-vdo/recovery-journal.c
1092
struct recovery_journal *journal = block->journal;
drivers/md/dm-vdo/recovery-journal.c
1096
.sequence_number = block->sequence_number,
drivers/md/dm-vdo/recovery-journal.c
1097
.entry_count = block->entry_count,
drivers/md/dm-vdo/recovery-journal.c
1103
if (!vdo_waitq_has_waiters(&block->entry_waiters))
drivers/md/dm-vdo/recovery-journal.c
1106
vdo_waitq_enqueue_waiter(&block->entry_waiters, &data_vio->waiter);
drivers/md/dm-vdo/recovery-journal.c
1107
block->entry_count++;
drivers/md/dm-vdo/recovery-journal.c
1108
block->uncommitted_entry_count++;
drivers/md/dm-vdo/recovery-journal.c
1111
if (is_block_full(block)) {
drivers/md/dm-vdo/recovery-journal.c
1116
schedule_block_write(journal, block);
drivers/md/dm-vdo/recovery-journal.c
1147
static void recycle_journal_block(struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
1149
struct recovery_journal *journal = block->journal;
drivers/md/dm-vdo/recovery-journal.c
1152
list_move_tail(&block->list_node, &journal->free_tail_blocks);
drivers/md/dm-vdo/recovery-journal.c
1155
for (i = block->entry_count; i < journal->entries_per_block; i++)
drivers/md/dm-vdo/recovery-journal.c
1156
release_journal_block_reference(block);
drivers/md/dm-vdo/recovery-journal.c
1162
if (block->entry_count > 0)
drivers/md/dm-vdo/recovery-journal.c
1163
release_journal_block_reference(block);
drivers/md/dm-vdo/recovery-journal.c
1165
if (block == journal->active_block)
drivers/md/dm-vdo/recovery-journal.c
1217
struct recovery_journal_block *block;
drivers/md/dm-vdo/recovery-journal.c
1219
list_for_each_entry(block, &journal->active_tail_blocks, list_node) {
drivers/md/dm-vdo/recovery-journal.c
1220
if (block->committing)
drivers/md/dm-vdo/recovery-journal.c
1223
vdo_waitq_notify_all_waiters(&block->commit_waiters,
drivers/md/dm-vdo/recovery-journal.c
1226
vdo_waitq_notify_all_waiters(&block->entry_waiters,
drivers/md/dm-vdo/recovery-journal.c
1229
} else if (is_block_dirty(block) || !is_block_full(block)) {
drivers/md/dm-vdo/recovery-journal.c
1242
struct recovery_journal_block *block, *tmp;
drivers/md/dm-vdo/recovery-journal.c
1244
list_for_each_entry_safe(block, tmp, &journal->active_tail_blocks, list_node) {
drivers/md/dm-vdo/recovery-journal.c
1245
if (block->committing) {
drivers/md/dm-vdo/recovery-journal.c
1251
(is_block_dirty(block) || !is_block_full(block))) {
drivers/md/dm-vdo/recovery-journal.c
1259
recycle_journal_block(block);
drivers/md/dm-vdo/recovery-journal.c
1272
struct recovery_journal_block *block = completion->parent;
drivers/md/dm-vdo/recovery-journal.c
1273
struct recovery_journal *journal = block->journal;
drivers/md/dm-vdo/recovery-journal.c
1280
journal->events.entries.committed += block->entries_in_commit;
drivers/md/dm-vdo/recovery-journal.c
1281
block->uncommitted_entry_count -= block->entries_in_commit;
drivers/md/dm-vdo/recovery-journal.c
1282
block->entries_in_commit = 0;
drivers/md/dm-vdo/recovery-journal.c
1283
block->committing = false;
drivers/md/dm-vdo/recovery-journal.c
1286
if (block->sequence_number > journal->last_write_acknowledged)
drivers/md/dm-vdo/recovery-journal.c
1287
journal->last_write_acknowledged = block->sequence_number;
drivers/md/dm-vdo/recovery-journal.c
1290
VDO_ASSERT_LOG_ONLY((block->sequence_number >= last_active_block->sequence_number),
drivers/md/dm-vdo/recovery-journal.c
1299
if (is_block_dirty(block) && is_block_full(block))
drivers/md/dm-vdo/recovery-journal.c
1300
schedule_block_write(journal, block);
drivers/md/dm-vdo/recovery-journal.c
1310
struct recovery_journal_block *block = completion->parent;
drivers/md/dm-vdo/recovery-journal.c
1311
struct recovery_journal *journal = block->journal;
drivers/md/dm-vdo/recovery-journal.c
1316
(unsigned long long) block->sequence_number);
drivers/md/dm-vdo/recovery-journal.c
1324
struct recovery_journal_block *block = vio->completion.parent;
drivers/md/dm-vdo/recovery-journal.c
1325
struct recovery_journal *journal = block->journal;
drivers/md/dm-vdo/recovery-journal.c
1334
static void add_queued_recovery_entries(struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
1336
while (vdo_waitq_has_waiters(&block->entry_waiters)) {
drivers/md/dm-vdo/recovery-journal.c
1338
vdo_waiter_as_data_vio(vdo_waitq_dequeue_waiter(&block->entry_waiters));
drivers/md/dm-vdo/recovery-journal.c
1343
if (block->sector->entry_count == RECOVERY_JOURNAL_ENTRIES_PER_SECTOR)
drivers/md/dm-vdo/recovery-journal.c
1344
set_active_sector(block,
drivers/md/dm-vdo/recovery-journal.c
1345
(char *) block->sector + VDO_SECTOR_SIZE);
drivers/md/dm-vdo/recovery-journal.c
1348
packed_entry = &block->sector->entries[block->sector->entry_count++];
drivers/md/dm-vdo/recovery-journal.c
1362
data_vio->recovery_sequence_number = block->sequence_number;
drivers/md/dm-vdo/recovery-journal.c
1365
vdo_waitq_enqueue_waiter(&block->commit_waiters, &data_vio->waiter);
drivers/md/dm-vdo/recovery-journal.c
1378
struct recovery_journal_block *block =
drivers/md/dm-vdo/recovery-journal.c
1380
struct recovery_journal *journal = block->journal;
drivers/md/dm-vdo/recovery-journal.c
1381
struct packed_journal_header *header = get_block_header(block);
drivers/md/dm-vdo/recovery-journal.c
1383
if (block->committing || !vdo_waitq_has_waiters(&block->entry_waiters) ||
drivers/md/dm-vdo/recovery-journal.c
1387
block->entries_in_commit = vdo_waitq_num_waiters(&block->entry_waiters);
drivers/md/dm-vdo/recovery-journal.c
1388
add_queued_recovery_entries(block);
drivers/md/dm-vdo/recovery-journal.c
1392
journal->events.entries.written += block->entries_in_commit;
drivers/md/dm-vdo/recovery-journal.c
1396
header->entry_count = __cpu_to_le16(block->entry_count);
drivers/md/dm-vdo/recovery-journal.c
1398
block->committing = true;
drivers/md/dm-vdo/recovery-journal.c
1405
vdo_submit_metadata_vio(&block->vio, journal->origin + block->block_number,
drivers/md/dm-vdo/recovery-journal.c
1724
static void dump_recovery_block(const struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
1727
(unsigned long long) block->sequence_number, block->entry_count,
drivers/md/dm-vdo/recovery-journal.c
1728
(block->committing ? "committing" : "waiting"),
drivers/md/dm-vdo/recovery-journal.c
1729
vdo_waitq_num_waiters(&block->entry_waiters),
drivers/md/dm-vdo/recovery-journal.c
1730
vdo_waitq_num_waiters(&block->commit_waiters));
drivers/md/dm-vdo/recovery-journal.c
1740
const struct recovery_journal_block *block;
drivers/md/dm-vdo/recovery-journal.c
1764
list_for_each_entry(block, &journal->active_tail_blocks, list_node)
drivers/md/dm-vdo/recovery-journal.c
1765
dump_recovery_block(block);
drivers/md/dm-vdo/recovery-journal.c
202
struct recovery_journal_block *block;
drivers/md/dm-vdo/recovery-journal.c
207
block = list_last_entry(&journal->free_tail_blocks,
drivers/md/dm-vdo/recovery-journal.c
209
list_del_init(&block->list_node);
drivers/md/dm-vdo/recovery-journal.c
210
return block;
drivers/md/dm-vdo/recovery-journal.c
222
static inline bool __must_check is_block_dirty(const struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
224
return (block->uncommitted_entry_count > 0);
drivers/md/dm-vdo/recovery-journal.c
233
static inline bool __must_check is_block_empty(const struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
235
return (block->entry_count == 0);
drivers/md/dm-vdo/recovery-journal.c
244
static inline bool __must_check is_block_full(const struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
246
return ((block == NULL) || (block->journal->entries_per_block == block->entry_count));
drivers/md/dm-vdo/recovery-journal.c
282
struct recovery_journal_block *block = get_journal_block(&journal->active_tail_blocks);
drivers/md/dm-vdo/recovery-journal.c
288
return ((block != NULL) &&
drivers/md/dm-vdo/recovery-journal.c
289
(vdo_waitq_has_waiters(&block->entry_waiters) ||
drivers/md/dm-vdo/recovery-journal.c
290
vdo_waitq_has_waiters(&block->commit_waiters)));
drivers/md/dm-vdo/recovery-journal.c
294
static void recycle_journal_block(struct recovery_journal_block *block);
drivers/md/dm-vdo/recovery-journal.c
659
struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
680
VIO_PRIORITY_HIGH, block, 1, data, &block->vio);
drivers/md/dm-vdo/recovery-journal.c
686
list_add_tail(&block->list_node, &journal->free_tail_blocks);
drivers/md/dm-vdo/recovery-journal.c
687
block->journal = journal;
drivers/md/dm-vdo/recovery-journal.c
740
struct recovery_journal_block *block = &journal->blocks[i];
drivers/md/dm-vdo/recovery-journal.c
742
result = initialize_recovery_block(vdo, journal, block);
drivers/md/dm-vdo/recovery-journal.c
813
struct recovery_journal_block *block = &journal->blocks[i];
drivers/md/dm-vdo/recovery-journal.c
815
vdo_free(vdo_forget(block->vio.data));
drivers/md/dm-vdo/recovery-journal.c
816
free_vio_components(&block->vio);
drivers/md/dm-vdo/recovery-journal.c
919
get_block_header(const struct recovery_journal_block *block)
drivers/md/dm-vdo/recovery-journal.c
921
return (struct packed_journal_header *) block->vio.data;
drivers/md/dm-vdo/recovery-journal.c
929
static void set_active_sector(struct recovery_journal_block *block, void *sector)
drivers/md/dm-vdo/recovery-journal.c
931
block->sector = sector;
drivers/md/dm-vdo/recovery-journal.c
932
block->sector->check_byte = get_block_header(block)->check_byte;
drivers/md/dm-vdo/recovery-journal.c
933
block->sector->recovery_count = block->journal->recovery_count;
drivers/md/dm-vdo/recovery-journal.c
934
block->sector->entry_count = 0;
drivers/md/dm-vdo/recovery-journal.c
947
struct recovery_journal_block *block;
drivers/md/dm-vdo/recovery-journal.c
949
block = journal->active_block = pop_free_list(journal);
drivers/md/dm-vdo/recovery-journal.c
950
if (block == NULL)
drivers/md/dm-vdo/recovery-journal.c
953
list_move_tail(&block->list_node, &journal->active_tail_blocks);
drivers/md/dm-vdo/recovery-journal.c
966
header = get_block_header(block);
drivers/md/dm-vdo/recovery-journal.c
967
memset(block->vio.data, 0x0, VDO_BLOCK_SIZE);
drivers/md/dm-vdo/recovery-journal.c
968
block->sequence_number = journal->tail;
drivers/md/dm-vdo/recovery-journal.c
969
block->entry_count = 0;
drivers/md/dm-vdo/recovery-journal.c
970
block->uncommitted_entry_count = 0;
drivers/md/dm-vdo/recovery-journal.c
971
block->block_number = vdo_get_recovery_journal_block_number(journal,
drivers/md/dm-vdo/recovery-journal.c
975
set_active_sector(block, vdo_get_journal_block_sector(header, 1));
drivers/md/dm-vdo/slab-depot.c
1070
struct reference_block *block = completion->parent;
drivers/md/dm-vdo/slab-depot.c
1071
struct vdo_slab *slab = block->slab;
drivers/md/dm-vdo/slab-depot.c
1078
block->slab_journal_lock_to_release, -1);
drivers/md/dm-vdo/slab-depot.c
1085
block->is_writing = false;
drivers/md/dm-vdo/slab-depot.c
1093
if (block->is_dirty) {
drivers/md/dm-vdo/slab-depot.c
1094
vdo_waitq_enqueue_waiter(&block->slab->dirty_blocks, &block->waiter);
drivers/md/dm-vdo/slab-depot.c
1125
static vdo_refcount_t * __must_check get_reference_counters_for_block(struct reference_block *block)
drivers/md/dm-vdo/slab-depot.c
1127
size_t block_index = block - block->slab->reference_blocks;
drivers/md/dm-vdo/slab-depot.c
1129
return &block->slab->counters[block_index * COUNTS_PER_BLOCK];
drivers/md/dm-vdo/slab-depot.c
1137
static void pack_reference_block(struct reference_block *block, void *buffer)
drivers/md/dm-vdo/slab-depot.c
1140
vdo_refcount_t *counters = get_reference_counters_for_block(block);
drivers/md/dm-vdo/slab-depot.c
1144
vdo_pack_journal_point(&block->slab->slab_journal_point, &commit_point);
drivers/md/dm-vdo/slab-depot.c
1156
struct reference_block *block = vio->completion.parent;
drivers/md/dm-vdo/slab-depot.c
1157
thread_id_t thread_id = block->slab->allocator->thread_id;
drivers/md/dm-vdo/slab-depot.c
1191
struct reference_block *block = container_of(waiter, struct reference_block,
drivers/md/dm-vdo/slab-depot.c
1194
pack_reference_block(block, pooled->vio.data);
drivers/md/dm-vdo/slab-depot.c
1195
block_offset = (block - block->slab->reference_blocks);
drivers/md/dm-vdo/slab-depot.c
1196
pbn = (block->slab->ref_counts_origin + block_offset);
drivers/md/dm-vdo/slab-depot.c
1197
block->slab_journal_lock_to_release = block->slab_journal_lock;
drivers/md/dm-vdo/slab-depot.c
1198
completion->parent = block;
drivers/md/dm-vdo/slab-depot.c
1205
block->is_dirty = false;
drivers/md/dm-vdo/slab-depot.c
1212
WRITE_ONCE(block->slab->allocator->ref_counts_statistics.blocks_written,
drivers/md/dm-vdo/slab-depot.c
1213
block->slab->allocator->ref_counts_statistics.blocks_written + 1);
drivers/md/dm-vdo/slab-depot.c
1267
static void dirty_block(struct reference_block *block)
drivers/md/dm-vdo/slab-depot.c
1269
if (block->is_dirty)
drivers/md/dm-vdo/slab-depot.c
1272
block->is_dirty = true;
drivers/md/dm-vdo/slab-depot.c
1273
if (!block->is_writing)
drivers/md/dm-vdo/slab-depot.c
1274
vdo_waitq_enqueue_waiter(&block->slab->dirty_blocks, &block->waiter);
drivers/md/dm-vdo/slab-depot.c
1423
static int increment_for_data(struct vdo_slab *slab, struct reference_block *block,
drivers/md/dm-vdo/slab-depot.c
1432
block->allocated_count++;
drivers/md/dm-vdo/slab-depot.c
1470
static int decrement_for_data(struct vdo_slab *slab, struct reference_block *block,
drivers/md/dm-vdo/slab-depot.c
1500
block->allocated_count--;
drivers/md/dm-vdo/slab-depot.c
1533
static int increment_for_block_map(struct vdo_slab *slab, struct reference_block *block,
drivers/md/dm-vdo/slab-depot.c
1548
block->allocated_count++;
drivers/md/dm-vdo/slab-depot.c
1593
static int update_reference_count(struct vdo_slab *slab, struct reference_block *block,
drivers/md/dm-vdo/slab-depot.c
1605
result = decrement_for_data(slab, block, block_number, old_status,
drivers/md/dm-vdo/slab-depot.c
1613
result = increment_for_data(slab, block, block_number, old_status,
drivers/md/dm-vdo/slab-depot.c
1616
result = increment_for_block_map(slab, block, block_number, old_status,
drivers/md/dm-vdo/slab-depot.c
1636
struct reference_block *block;
drivers/md/dm-vdo/slab-depot.c
1646
block = get_reference_block(slab, block_number);
drivers/md/dm-vdo/slab-depot.c
1647
result = update_reference_count(slab, block, block_number, slab_journal_point,
drivers/md/dm-vdo/slab-depot.c
1653
if (block->is_dirty && (block->slab_journal_lock > 0)) {
drivers/md/dm-vdo/slab-depot.c
1675
block->slab_journal_lock = slab_journal_point->sequence_number;
drivers/md/dm-vdo/slab-depot.c
1677
block->slab_journal_lock = 0;
drivers/md/dm-vdo/slab-depot.c
1679
dirty_block(block);
drivers/md/dm-vdo/slab-depot.c
1898
cursor->block = cursor->first_block;
drivers/md/dm-vdo/slab-depot.c
1920
if (cursor->block == cursor->last_block) {
drivers/md/dm-vdo/slab-depot.c
1926
cursor->block++;
drivers/md/dm-vdo/slab-depot.c
1929
if (cursor->block == cursor->last_block) {
drivers/md/dm-vdo/slab-depot.c
1953
struct reference_block *block;
drivers/md/dm-vdo/slab-depot.c
1964
block = get_reference_block(slab, block_number);
drivers/md/dm-vdo/slab-depot.c
1965
result = update_reference_count(slab, block, block_number, NULL,
drivers/md/dm-vdo/slab-depot.c
1970
dirty_block(block);
drivers/md/dm-vdo/slab-depot.c
1990
struct reference_block *block = get_reference_block(slab, entry.sbn);
drivers/md/dm-vdo/slab-depot.c
1997
if (!vdo_before_journal_point(&block->commit_points[sector], entry_point)) {
drivers/md/dm-vdo/slab-depot.c
2003
result = update_reference_count(slab, block, entry.sbn, entry_point,
drivers/md/dm-vdo/slab-depot.c
2008
dirty_block(block);
drivers/md/dm-vdo/slab-depot.c
2112
return ((slab->search_cursor.block->allocated_count < COUNTS_PER_BLOCK) &&
drivers/md/dm-vdo/slab-depot.c
2151
struct reference_block *block = get_reference_block(slab, block_number);
drivers/md/dm-vdo/slab-depot.c
2160
block->allocated_count++;
drivers/md/dm-vdo/slab-depot.c
2273
struct reference_block *block)
drivers/md/dm-vdo/slab-depot.c
2276
struct vdo_slab *slab = block->slab;
drivers/md/dm-vdo/slab-depot.c
2277
vdo_refcount_t *counters = get_reference_counters_for_block(block);
drivers/md/dm-vdo/slab-depot.c
2282
vdo_unpack_journal_point(&sector->commit_point, &block->commit_points[i]);
drivers/md/dm-vdo/slab-depot.c
2287
&block->commit_points[i]))
drivers/md/dm-vdo/slab-depot.c
2288
slab->slab_journal_point = block->commit_points[i];
drivers/md/dm-vdo/slab-depot.c
2291
!journal_points_equal(block->commit_points[0],
drivers/md/dm-vdo/slab-depot.c
2292
block->commit_points[i])) {
drivers/md/dm-vdo/slab-depot.c
2293
size_t block_index = block - block->slab->reference_blocks;
drivers/md/dm-vdo/slab-depot.c
2296
i, block_index, block->slab->slab_number);
drivers/md/dm-vdo/slab-depot.c
2300
block->allocated_count = count_valid_references(counters);
drivers/md/dm-vdo/slab-depot.c
2311
struct reference_block *block = completion->parent;
drivers/md/dm-vdo/slab-depot.c
2312
struct vdo_slab *slab = block->slab;
drivers/md/dm-vdo/slab-depot.c
2317
for (i = 0; i < block_count; i++, block++, data += VDO_BLOCK_SIZE) {
drivers/md/dm-vdo/slab-depot.c
2320
unpack_reference_block(packed, block);
drivers/md/dm-vdo/slab-depot.c
2321
slab->free_blocks -= block->allocated_count;
drivers/md/dm-vdo/slab-depot.c
2332
struct reference_block *block = vio->completion.parent;
drivers/md/dm-vdo/slab-depot.c
2335
block->slab->allocator->thread_id);
drivers/md/dm-vdo/slab-depot.c
2348
struct reference_block *block =
drivers/md/dm-vdo/slab-depot.c
2350
u32 block_offset = block - block->slab->reference_blocks;
drivers/md/dm-vdo/slab-depot.c
2351
u32 max_block_count = block->slab->reference_block_count - block_offset;
drivers/md/dm-vdo/slab-depot.c
2354
vio->completion.parent = block;
drivers/md/dm-vdo/slab-depot.c
2355
vdo_submit_metadata_vio_with_size(vio, block->slab->ref_counts_origin + block_offset,
drivers/md/dm-vdo/slab-depot.c
2499
struct packed_slab_journal_block *block = (struct packed_slab_journal_block *) vio->data;
drivers/md/dm-vdo/slab-depot.c
2502
vdo_unpack_slab_journal_block_header(&block->header, &header);
drivers/md/dm-vdo/slab-depot.c
265
static void finish_updating_slab_summary_block(struct slab_summary_block *block)
drivers/md/dm-vdo/slab-depot.c
267
notify_summary_waiters(block->allocator, &block->current_update_waiters);
drivers/md/dm-vdo/slab-depot.c
268
block->writing = false;
drivers/md/dm-vdo/slab-depot.c
269
block->allocator->summary_write_count--;
drivers/md/dm-vdo/slab-depot.c
270
if (vdo_waitq_has_waiters(&block->next_update_waiters))
drivers/md/dm-vdo/slab-depot.c
271
launch_write(block);
drivers/md/dm-vdo/slab-depot.c
273
check_summary_drain_complete(block->allocator);
drivers/md/dm-vdo/slab-depot.c
282
struct slab_summary_block *block =
drivers/md/dm-vdo/slab-depot.c
285
atomic64_inc(&block->allocator->depot->summary_statistics.blocks_written);
drivers/md/dm-vdo/slab-depot.c
286
finish_updating_slab_summary_block(block);
drivers/md/dm-vdo/slab-depot.c
2861
static int apply_block_entries(struct packed_slab_journal_block *block,
drivers/md/dm-vdo/slab-depot.c
2874
vdo_decode_slab_journal_entry(block, entry_point.entry_count);
drivers/md/dm-vdo/slab-depot.c
2932
struct packed_slab_journal_block *block =
drivers/md/dm-vdo/slab-depot.c
2936
vdo_unpack_slab_journal_block_header(&block->header, &header);
drivers/md/dm-vdo/slab-depot.c
295
struct slab_summary_block *block =
drivers/md/dm-vdo/slab-depot.c
2951
result = apply_block_entries(block, header.entry_count, sequence, slab);
drivers/md/dm-vdo/slab-depot.c
300
finish_updating_slab_summary_block(block);
drivers/md/dm-vdo/slab-depot.c
306
struct slab_summary_block *block =
drivers/md/dm-vdo/slab-depot.c
309
continue_vio_after_io(vio, finish_update, block->allocator->thread_id);
drivers/md/dm-vdo/slab-depot.c
316
static void launch_write(struct slab_summary_block *block)
drivers/md/dm-vdo/slab-depot.c
318
struct block_allocator *allocator = block->allocator;
drivers/md/dm-vdo/slab-depot.c
322
if (block->writing)
drivers/md/dm-vdo/slab-depot.c
326
vdo_waitq_transfer_all_waiters(&block->next_update_waiters,
drivers/md/dm-vdo/slab-depot.c
327
&block->current_update_waiters);
drivers/md/dm-vdo/slab-depot.c
328
block->writing = true;
drivers/md/dm-vdo/slab-depot.c
331
finish_updating_slab_summary_block(block);
drivers/md/dm-vdo/slab-depot.c
335
memcpy(block->outgoing_entries, block->entries, VDO_BLOCK_SIZE);
drivers/md/dm-vdo/slab-depot.c
347
block->index);
drivers/md/dm-vdo/slab-depot.c
348
vdo_submit_metadata_vio(&block->vio, pbn, write_slab_summary_endio,
drivers/md/dm-vdo/slab-depot.c
368
struct slab_summary_block *block = &allocator->summary_blocks[index];
drivers/md/dm-vdo/slab-depot.c
372
if (vdo_is_read_only(block->vio.completion.vdo)) {
drivers/md/dm-vdo/slab-depot.c
3729
vdo_free(vdo_forget(slab->journal.block));
drivers/md/dm-vdo/slab-depot.c
3748
(char **) &journal->block);
drivers/md/dm-vdo/slab-depot.c
392
vdo_waitq_enqueue_waiter(&block->next_update_waiters, waiter);
drivers/md/dm-vdo/slab-depot.c
393
launch_write(block);
drivers/md/dm-vdo/slab-depot.c
4045
struct slab_summary_block *block = &allocator->summary_blocks[index];
drivers/md/dm-vdo/slab-depot.c
4048
result = vdo_allocate(VDO_BLOCK_SIZE, char, __func__, &block->outgoing_entries);
drivers/md/dm-vdo/slab-depot.c
4054
block->outgoing_entries, &block->vio);
drivers/md/dm-vdo/slab-depot.c
4058
block->allocator = allocator;
drivers/md/dm-vdo/slab-depot.c
4059
block->entries = &allocator->summary_entries[VDO_SLAB_SUMMARY_ENTRIES_PER_BLOCK * index];
drivers/md/dm-vdo/slab-depot.c
4060
block->index = index;
drivers/md/dm-vdo/slab-depot.c
664
sequence_number_t block;
drivers/md/dm-vdo/slab-depot.c
672
for (block = 1; block <= journal->size; block++) {
drivers/md/dm-vdo/slab-depot.c
673
VDO_ASSERT_LOG_ONLY((get_lock(journal, block)->count == 0),
drivers/md/dm-vdo/slab-depot.c
675
(unsigned long long) block);
drivers/md/dm-vdo/slab-depot.c
683
const struct packed_slab_journal_block *block =
drivers/md/dm-vdo/slab-depot.c
686
return __le64_to_cpu(block->header.sequence_number);
drivers/md/dm-vdo/slab-depot.c
757
vdo_pack_slab_journal_block_header(header, &journal->block->header);
drivers/md/dm-vdo/slab-depot.c
760
memcpy(pooled->vio.data, journal->block, VDO_BLOCK_SIZE);
drivers/md/dm-vdo/slab-depot.c
907
struct packed_slab_journal_block *block = journal->block;
drivers/md/dm-vdo/slab-depot.c
933
encode_slab_journal_entry(&journal->tail_header, &block->payload,
drivers/md/dm-vdo/slab-depot.h
136
struct packed_slab_journal_block *block;
drivers/md/dm-vdo/slab-depot.h
185
struct reference_block *block;
drivers/md/dm-vdo/vdo.c
269
char *block;
drivers/md/dm-vdo/vdo.c
272
result = vdo_allocate(VDO_BLOCK_SIZE, u8, __func__, &block);
drivers/md/dm-vdo/vdo.c
277
block, &vio);
drivers/md/dm-vdo/vdo.c
279
vdo_free(block);
drivers/md/dm-vdo/vdo.c
288
result = vio_reset_bio(vio, block, NULL, REQ_OP_READ,
drivers/md/dm-vdo/vdo.c
292
vdo_free(block);
drivers/md/dm-vdo/vdo.c
302
vdo_free(block);
drivers/md/dm-vdo/vdo.c
306
result = vdo_parse_geometry_block((u8 *) block, &vdo->geometry);
drivers/md/dm-vdo/vdo.c
307
vdo_free(block);
drivers/md/dm-verity-fec.c
107
u8 *par, *block;
drivers/md/dm-verity-fec.c
120
block = fec_buffer_rs_block(v, fio, n, i);
drivers/md/dm-verity-fec.c
124
res = decode_rs8(fio->rs, block, par_buf, v->fec->rsn,
drivers/md/dm-verity-fec.c
132
fio->output[block_offset] = block[byte_index];
drivers/md/dm-verity-fec.c
201
u64 block, ileaved;
drivers/md/dm-verity-fec.c
227
block = ileaved >> v->data_dev_block_bits;
drivers/md/dm-verity-fec.c
230
if (block >= v->data_blocks) {
drivers/md/dm-verity-fec.c
231
block -= v->data_blocks;
drivers/md/dm-verity-fec.c
237
if (unlikely(block >= v->fec->hash_blocks))
drivers/md/dm-verity-fec.c
240
block += v->hash_start;
drivers/md/dm-verity-fec.c
244
bbuf = dm_bufio_read_with_ioprio(bufio, block, &buf, bio->bi_ioprio);
drivers/md/dm-verity-fec.c
249
(unsigned long long)block, PTR_ERR(bbuf));
drivers/md/dm-verity-fec.c
260
verity_hash_for_block(v, io, block, want_digest,
drivers/md/dm-verity-fec.c
389
sector_t block, u8 *dest)
drivers/md/dm-verity-fec.c
408
block = block - v->hash_start + v->data_blocks;
drivers/md/dm-verity-fec.c
420
offset = block << v->data_dev_block_bits;
drivers/md/dm-verity-fec.c
44
u64 position, block, rem;
drivers/md/dm-verity-fec.c
52
block = div64_u64_rem(position, v->fec->io_size, &rem);
drivers/md/dm-verity-fec.c
55
res = dm_bufio_read_with_ioprio(v->fec->bufio, block, buf, ioprio);
drivers/md/dm-verity-fec.c
59
(unsigned long long)block, PTR_ERR(res));
drivers/md/dm-verity-fec.h
117
sector_t block, u8 *dest)
drivers/md/dm-verity-fec.h
77
sector_t block, u8 *dest);
drivers/md/dm-verity-target.c
112
static sector_t verity_position_at_level(struct dm_verity *v, sector_t block,
drivers/md/dm-verity-target.c
115
return block >> (level * v->hash_per_block_bits);
drivers/md/dm-verity-target.c
155
static void verity_hash_at_level(struct dm_verity *v, sector_t block, int level,
drivers/md/dm-verity-target.c
158
sector_t position = verity_position_at_level(v, block, level);
drivers/md/dm-verity-target.c
177
unsigned long long block)
drivers/md/dm-verity-target.c
204
type_str, block);
drivers/md/dm-verity-target.c
212
DM_VERITY_ENV_VAR_NAME, type, block);
drivers/md/dm-verity-target.c
241
sector_t block, int level, bool skip_unverified,
drivers/md/dm-verity-target.c
252
verity_hash_at_level(v, block, level, &hash_block, &offset);
drivers/md/dm-verity-target.c
321
block, 0);
drivers/md/dm-verity-target.c
342
sector_t block, u8 *digest, bool *is_zero)
drivers/md/dm-verity-target.c
354
r = verity_verify_level(v, io, block, 0, true, digest);
drivers/md/dm-verity-target.c
362
r = verity_verify_level(v, io, block, i, false, digest);
drivers/md/dm-verity-target.c
421
struct pending_block *block)
drivers/md/dm-verity-target.c
423
const u8 *want_digest = block->want_digest;
drivers/md/dm-verity-target.c
424
sector_t blkno = block->blkno;
drivers/md/dm-verity-target.c
425
u8 *data = block->data;
drivers/md/dm-verity-target.c
489
struct pending_block *block = &io->pending_blocks[i];
drivers/md/dm-verity-target.c
491
if (likely(memcmp(block->real_digest, block->want_digest,
drivers/md/dm-verity-target.c
494
set_bit(block->blkno, v->validated_blocks);
drivers/md/dm-verity-target.c
496
r = verity_handle_data_hash_mismatch(v, io, bio, block);
drivers/md/dm-verity-target.c
533
sector_t blkno = io->block + b;
drivers/md/dm-verity-target.c
534
struct pending_block *block;
drivers/md/dm-verity-target.c
543
block = &io->pending_blocks[io->num_pending];
drivers/md/dm-verity-target.c
545
r = verity_hash_for_block(v, io, blkno, block->want_digest,
drivers/md/dm-verity-target.c
574
block->data = data;
drivers/md/dm-verity-target.c
575
block->blkno = blkno;
drivers/md/dm-verity-target.c
68
sector_t block;
drivers/md/dm-verity-target.c
721
verity_hash_at_level(v, pw->block, i, &hash_block_start, NULL);
drivers/md/dm-verity-target.c
722
verity_hash_at_level(v, pw->block + pw->n_blocks - 1, i, &hash_block_end, NULL);
drivers/md/dm-verity-target.c
751
sector_t block = io->block;
drivers/md/dm-verity-target.c
756
while (n_blocks && test_bit(block, v->validated_blocks)) {
drivers/md/dm-verity-target.c
757
block++;
drivers/md/dm-verity-target.c
760
while (n_blocks && test_bit(block + n_blocks - 1,
drivers/md/dm-verity-target.c
775
pw->block = block;
drivers/md/dm-verity-target.c
811
io->block = bio->bi_iter.bi_sector >> (v->data_dev_block_bits - SECTOR_SHIFT);
drivers/md/dm-verity.h
102
sector_t block;
drivers/md/dm-verity.h
141
sector_t block, u8 *digest, bool *is_zero);
drivers/md/dm-writecache.c
607
uint64_t block, int flags)
drivers/md/dm-writecache.c
617
if (read_original_sector(wc, e) == block)
drivers/md/dm-writecache.c
620
node = (read_original_sector(wc, e) >= block ?
drivers/md/dm-writecache.c
625
if (read_original_sector(wc, e) >= block)
drivers/md/dm-writecache.c
647
if (read_original_sector(wc, e2) != block)
drivers/md/dm-zoned-metadata.c
1096
set, sb->dev->bdev, sb->block);
drivers/md/dm-zoned-metadata.c
1099
sb->block, sb->mblk->page);
drivers/md/dm-zoned-metadata.c
1123
zmd->sb[1].block = zmd->sb[0].block + zone_nr_blocks;
drivers/md/dm-zoned-metadata.c
1131
zmd->sb[1].block += zone_nr_blocks;
drivers/md/dm-zoned-metadata.c
1183
zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone);
drivers/md/dm-zoned-metadata.c
1185
zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone);
drivers/md/dm-zoned-metadata.c
1194
zmd->sb[src_set].block + i, page);
drivers/md/dm-zoned-metadata.c
1198
zmd->sb[dst_set].block + i, page);
drivers/md/dm-zoned-metadata.c
1235
zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone);
drivers/md/dm-zoned-metadata.c
1254
zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone);
drivers/md/dm-zoned-metadata.c
1318
sb->block = 0;
drivers/md/dm-zoned-metadata.c
133
sector_t block;
drivers/md/dm-zoned-metadata.c
536
sector_t block = zmd->sb[zmd->mblk_primary].block + mblk_no;
drivers/md/dm-zoned-metadata.c
572
bio->bi_iter.bi_sector = dmz_blk2sect(block);
drivers/md/dm-zoned-metadata.c
712
sector_t block = zmd->sb[set].block + mblk->no;
drivers/md/dm-zoned-metadata.c
723
bio->bi_iter.bi_sector = dmz_blk2sect(block);
drivers/md/dm-zoned-metadata.c
736
sector_t block, struct page *page)
drivers/md/dm-zoned-metadata.c
749
bio->bi_iter.bi_sector = dmz_blk2sect(block);
drivers/md/dm-zoned-metadata.c
800
ret = dmz_rdwr_block(dev, REQ_OP_WRITE, zmd->sb[set].block,
drivers/md/dm-zoned-reclaim.c
122
sector_t block = 0, end_block;
drivers/md/dm-zoned-reclaim.c
139
while (block < end_block) {
drivers/md/dm-zoned-reclaim.c
149
ret = dmz_first_valid_block(zmd, src_zone, &block);
drivers/md/dm-zoned-reclaim.c
160
ret = dmz_reclaim_align_wp(zrc, dst_zone, block);
drivers/md/dm-zoned-reclaim.c
166
src.sector = dmz_blk2sect(src_zone_block + block);
drivers/md/dm-zoned-reclaim.c
170
dst.sector = dmz_blk2sect(dst_zone_block + block);
drivers/md/dm-zoned-reclaim.c
184
block += nr_blocks;
drivers/md/dm-zoned-reclaim.c
186
dst_zone->wp_block = block;
drivers/md/dm-zoned-reclaim.c
60
sector_t block)
drivers/md/dm-zoned-reclaim.c
68
if (wp_block == block)
drivers/md/dm-zoned-reclaim.c
71
if (wp_block > block)
drivers/md/dm-zoned-reclaim.c
78
nr_blocks = block - wp_block;
drivers/md/dm-zoned-reclaim.c
86
(unsigned long long)block, nr_blocks, ret);
drivers/md/dm-zoned-reclaim.c
91
zone->wp_block = block;
drivers/md/dm-zoned-target.c
361
sector_t block = dmz_bio_block(bio);
drivers/md/dm-zoned-target.c
363
sector_t chunk_block = dmz_chunk_block(zmd, block);
drivers/md/dm.h
246
static inline unsigned int dm_hash_locks_index(sector_t block,
drivers/md/dm.h
249
sector_t h1 = (block * DM_HASH_LOCKS_MULT) >> DM_HASH_LOCKS_SHIFT;
drivers/md/md-bitmap.c
1148
static void md_bitmap_file_set_bit(struct bitmap *bitmap, sector_t block)
drivers/md/md-bitmap.c
1153
unsigned long chunk = block >> bitmap->counts.chunkshift;
drivers/md/md-bitmap.c
1179
static void md_bitmap_file_clear_bit(struct bitmap *bitmap, sector_t block)
drivers/md/md-bitmap.c
1184
unsigned long chunk = block >> bitmap->counts.chunkshift;
drivers/md/md-bitmap.c
1209
static int md_bitmap_file_test_bit(struct bitmap *bitmap, sector_t block)
drivers/md/md-bitmap.c
1214
unsigned long chunk = block >> bitmap->counts.chunkshift;
drivers/md/md-bitmap.c
1569
sector_t block = (sector_t)j << counts->chunkshift;
drivers/md/md-bitmap.c
1580
bmc = md_bitmap_get_counter(counts, block, &blocks, 0);
drivers/md/md-bitmap.c
1588
md_bitmap_count_page(counts, block, -1);
drivers/md/md-bitmap.c
1589
md_bitmap_file_clear_bit(bitmap, block);
drivers/md/md-bitmap.c
1592
md_bitmap_set_pending(counts, block);
drivers/md/md-bitmap.c
2289
sector_t block, lo = 0, hi = 0;
drivers/md/md-bitmap.c
2301
block = (sector_t)j << counts->chunkshift;
drivers/md/md-bitmap.c
2302
if (md_bitmap_file_test_bit(bitmap, block)) {
drivers/md/md-bitmap.c
2304
lo = block;
drivers/md/md-bitmap.c
2305
hi = block;
drivers/md/md-bitmap.c
2306
md_bitmap_file_clear_bit(bitmap, block);
drivers/md/md-bitmap.c
2307
md_bitmap_set_memory_bits(mddev->bitmap, block, 1);
drivers/md/md-bitmap.c
2308
md_bitmap_file_set_bit(mddev->bitmap, block);
drivers/md/md-bitmap.c
2381
sector_t block;
drivers/md/md-bitmap.c
2502
for (block = 0; block < blocks; ) {
drivers/md/md-bitmap.c
2506
bmc_old = md_bitmap_get_counter(&old_counts, block, &old_blocks, 0);
drivers/md/md-bitmap.c
2510
bmc_new = md_bitmap_get_counter(&bitmap->counts, block, &new_blocks, 1);
drivers/md/md-bitmap.c
2514
sector_t end = block + new_blocks;
drivers/md/md-bitmap.c
2515
sector_t start = block >> chunkshift;
drivers/md/md-bitmap.c
2519
md_bitmap_file_set_bit(bitmap, block);
drivers/md/md-bitmap.c
2523
md_bitmap_count_page(&bitmap->counts, block, 1);
drivers/md/md-bitmap.c
2524
md_bitmap_set_pending(&bitmap->counts, block);
drivers/md/md-bitmap.c
2531
block += old_blocks;
drivers/md/md-bitmap.c
2544
while (block < (chunks << chunkshift)) {
drivers/md/md-bitmap.c
2546
bmc = md_bitmap_get_counter(&bitmap->counts, block, &new_blocks, 1);
drivers/md/md-bitmap.c
2553
md_bitmap_count_page(&bitmap->counts, block, 1);
drivers/md/md-bitmap.c
2554
md_bitmap_set_pending(&bitmap->counts, block);
drivers/md/md-bitmap.c
2557
block += new_blocks;
drivers/md/md-bitmap.c
561
sector_t block, blk_cur;
drivers/md/md-bitmap.c
575
block = blk_cur;
drivers/md/md-bitmap.c
580
ret = bmap(inode, &block);
drivers/md/md-bitmap.c
581
if (ret || !block) {
drivers/md/md-bitmap.c
587
bh->b_blocknr = block;
drivers/md/md-llbitmap.c
367
unsigned int block)
drivers/md/md-llbitmap.c
373
for (pos = block * io_size; pos < (block + 1) * io_size; pos++) {
drivers/md/md-llbitmap.c
390
int block = offset / io_size;
drivers/md/md-llbitmap.c
402
set_bit(block, pctl->dirty);
drivers/md/md-llbitmap.c
413
if (test_and_set_bit(block, pctl->dirty)) {
drivers/md/md-llbitmap.c
414
llbitmap_infect_dirty_bits(llbitmap, pctl, block);
drivers/md/md-llbitmap.c
418
for (pos = block * io_size; pos < (block + 1) * io_size; pos++) {
drivers/md/md-llbitmap.c
423
llbitmap_infect_dirty_bits(llbitmap, pctl, block);
drivers/md/md-llbitmap.c
484
int block;
drivers/md/md-llbitmap.c
486
for (block = 0; block < llbitmap->blocks_per_page; block++) {
drivers/md/md-llbitmap.c
489
if (!test_and_clear_bit(block, pctl->dirty))
drivers/md/md-llbitmap.c
501
block * bit_sector;
drivers/md/md-llbitmap.c
504
block * llbitmap->io_size);
drivers/md/persistent-data/dm-array.c
154
struct dm_block **block, struct array_block **ab)
drivers/md/persistent-data/dm-array.c
158
r = dm_tm_new_block(info->btree_info.tm, &array_validator, block);
drivers/md/persistent-data/dm-array.c
162
(*ab) = dm_block_data(*block);
drivers/md/persistent-data/dm-array.c
219
struct dm_block **block, struct array_block **ab)
drivers/md/persistent-data/dm-array.c
223
r = dm_tm_read_lock(info->btree_info.tm, b, &array_validator, block);
drivers/md/persistent-data/dm-array.c
227
*ab = dm_block_data(*block);
drivers/md/persistent-data/dm-array.c
234
static void unlock_ablock(struct dm_array_info *info, struct dm_block *block)
drivers/md/persistent-data/dm-array.c
236
dm_tm_unlock(info->btree_info.tm, block);
drivers/md/persistent-data/dm-array.c
252
unsigned int index, struct dm_block **block,
drivers/md/persistent-data/dm-array.c
263
return get_ablock(info, le64_to_cpu(block_le), block, ab);
drivers/md/persistent-data/dm-array.c
270
struct dm_block *block, dm_block_t *root)
drivers/md/persistent-data/dm-array.c
272
__le64 block_le = cpu_to_le64(dm_block_location(block));
drivers/md/persistent-data/dm-array.c
281
struct dm_block **block, struct array_block **ab)
drivers/md/persistent-data/dm-array.c
285
&array_validator, block, &inc);
drivers/md/persistent-data/dm-array.c
289
*ab = dm_block_data(*block);
drivers/md/persistent-data/dm-array.c
301
struct dm_block *block, dm_block_t b,
drivers/md/persistent-data/dm-array.c
306
if (dm_block_location(block) != b) {
drivers/md/persistent-data/dm-array.c
314
r = insert_ablock(info, index, block, root);
drivers/md/persistent-data/dm-array.c
326
unsigned int index, struct dm_block **block,
drivers/md/persistent-data/dm-array.c
339
r = __shadow_ablock(info, b, block, ab);
drivers/md/persistent-data/dm-array.c
343
return __reinsert_ablock(info, index, *block, b, root);
drivers/md/persistent-data/dm-array.c
355
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
358
r = alloc_ablock(info, size_of_block, max_entries, &block, &ab);
drivers/md/persistent-data/dm-array.c
363
r = insert_ablock(info, block_index, block, root);
drivers/md/persistent-data/dm-array.c
364
unlock_ablock(info, block);
drivers/md/persistent-data/dm-array.c
468
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
490
resize->new_nr_full_blocks, &block, &ab);
drivers/md/persistent-data/dm-array.c
495
unlock_ablock(resize->info, block);
drivers/md/persistent-data/dm-array.c
507
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
511
resize->old_nr_full_blocks, &block, &ab);
drivers/md/persistent-data/dm-array.c
516
unlock_ablock(resize->info, block);
drivers/md/persistent-data/dm-array.c
591
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
610
r = get_ablock(info, b, &block, &ab);
drivers/md/persistent-data/dm-array.c
618
unlock_ablock(info, block);
drivers/md/persistent-data/dm-array.c
735
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
748
r = alloc_ablock(info, size_of_block, max_entries, &block, &ab);
drivers/md/persistent-data/dm-array.c
756
unlock_ablock(info, block);
drivers/md/persistent-data/dm-array.c
760
r = insert_ablock(info, block_index, block, root);
drivers/md/persistent-data/dm-array.c
761
unlock_ablock(info, block);
drivers/md/persistent-data/dm-array.c
782
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
790
r = lookup_ablock(info, root, index / max_entries, &block, &ab);
drivers/md/persistent-data/dm-array.c
801
unlock_ablock(info, block);
drivers/md/persistent-data/dm-array.c
810
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
821
r = shadow_ablock(info, &root, index / max_entries, &block, &ab);
drivers/md/persistent-data/dm-array.c
843
unlock_ablock(info, block);
drivers/md/persistent-data/dm-array.c
873
struct dm_block *block;
drivers/md/persistent-data/dm-array.c
877
r = get_ablock(wi->info, le64_to_cpu(block_le), &block, &ab);
drivers/md/persistent-data/dm-array.c
891
unlock_ablock(wi->info, block);
drivers/md/persistent-data/dm-array.c
917
if (c->block)
drivers/md/persistent-data/dm-array.c
918
unlock_ablock(c->info, c->block);
drivers/md/persistent-data/dm-array.c
928
r = get_ablock(c->info, le64_to_cpu(value_le), &c->block, &c->ab);
drivers/md/persistent-data/dm-array.c
939
c->block = NULL;
drivers/md/persistent-data/dm-array.c
963
if (c->block)
drivers/md/persistent-data/dm-array.c
964
unlock_ablock(c->info, c->block);
drivers/md/persistent-data/dm-array.c
974
if (!c->block)
drivers/md/persistent-data/dm-array.h
200
struct dm_block *block;
drivers/md/persistent-data/dm-btree-remove.c
157
struct dm_block *block;
drivers/md/persistent-data/dm-btree-remove.c
172
&result->block, &inc);
drivers/md/persistent-data/dm-btree-remove.c
176
result->n = dm_block_data(result->block);
drivers/md/persistent-data/dm-btree-remove.c
182
cpu_to_le64(dm_block_location(result->block));
drivers/md/persistent-data/dm-btree-remove.c
189
dm_tm_unlock(info->tm, c->block);
drivers/md/persistent-data/dm-btree-remove.c
264
dm_tm_dec(info->tm, dm_block_location(r->block));
drivers/md/persistent-data/dm-btree-remove.c
344
dm_tm_dec(info->tm, dm_block_location(c->block));
drivers/md/persistent-data/dm-btree-spine.c
166
struct dm_block *block;
drivers/md/persistent-data/dm-btree-spine.c
169
block = s->nodes[s->count - 1];
drivers/md/persistent-data/dm-btree-spine.c
171
return dm_block_data(block);
drivers/md/persistent-data/dm-btree.c
1241
dm_block_t block = root;
drivers/md/persistent-data/dm-btree.c
1250
r = btree_insert_raw(&spine, block, &le64_type, keys[level], &index);
drivers/md/persistent-data/dm-btree.c
1274
block = value64(n, index);
drivers/md/persistent-data/dm-btree.c
1277
r = btree_insert_raw(&spine, block, &info->value_type,
drivers/md/persistent-data/dm-btree.c
1340
static int find_key(struct ro_spine *s, dm_block_t block, bool find_highest,
drivers/md/persistent-data/dm-btree.c
1347
r = ro_step(s, block);
drivers/md/persistent-data/dm-btree.c
1365
block = value64(ro_node(s), i);
drivers/md/persistent-data/dm-btree.c
1367
block = value64(ro_node(s), 0);
drivers/md/persistent-data/dm-btree.c
1373
*next_block = block;
drivers/md/persistent-data/dm-btree.c
1421
static int walk_node(struct dm_btree_info *info, dm_block_t block,
drivers/md/persistent-data/dm-btree.c
1431
r = bn_read_lock(info, block, &node);
drivers/md/persistent-data/dm-btree.c
343
static int btree_lookup_raw(struct ro_spine *s, dm_block_t block, uint64_t key,
drivers/md/persistent-data/dm-btree.c
351
r = ro_step(s, block);
drivers/md/persistent-data/dm-btree.c
363
block = value64(ro_node(s), i);
drivers/md/persistent-data/dm-btree.c
977
struct dm_block *block;
drivers/md/persistent-data/dm-btree.c
980
r = bn_read_lock(info, b, &block);
drivers/md/persistent-data/dm-btree.c
984
node = dm_block_data(block);
drivers/md/persistent-data/dm-btree.c
988
unlock_block(info, block);
drivers/md/persistent-data/dm-space-map-common.c
1000
memcpy(&ll->mi_le, dm_block_data(block), sizeof(ll->mi_le));
drivers/md/persistent-data/dm-space-map-common.c
1001
dm_tm_unlock(ll->tm, block);
drivers/md/persistent-data/dm-space-map-common.c
993
struct dm_block *block;
drivers/md/persistent-data/dm-space-map-common.c
996
&index_validator, &block);
drivers/md/raid5-cache.c
707
struct r5l_meta_block *block;
drivers/md/raid5-cache.c
715
block = page_address(io->meta_page);
drivers/md/raid5-cache.c
716
block->meta_size = cpu_to_le32(io->meta_offset);
drivers/md/raid5-cache.c
717
crc = crc32c(log->uuid_checksum, block, PAGE_SIZE);
drivers/md/raid5-cache.c
718
block->checksum = cpu_to_le32(crc);
drivers/md/raid5-cache.c
765
struct r5l_meta_block *block;
drivers/md/raid5-cache.c
779
block = page_address(io->meta_page);
drivers/md/raid5-cache.c
780
clear_page(block);
drivers/md/raid5-cache.c
781
block->magic = cpu_to_le32(R5LOG_MAGIC);
drivers/md/raid5-cache.c
782
block->version = R5LOG_VERSION;
drivers/md/raid5-cache.c
783
block->seq = cpu_to_le64(log->seq);
drivers/md/raid5-cache.c
784
block->position = cpu_to_le64(log->log_start);
drivers/media/cec/core/cec-adap.c
1609
static void cec_claim_log_addrs(struct cec_adapter *adap, bool block)
drivers/media/cec/core/cec-adap.c
1626
} else if (block) {
drivers/media/cec/core/cec-adap.c
1697
void __cec_s_phys_addr(struct cec_adapter *adap, u16 phys_addr, bool block)
drivers/media/cec/core/cec-adap.c
1729
cec_claim_log_addrs(adap, block);
drivers/media/cec/core/cec-adap.c
1732
void cec_s_phys_addr(struct cec_adapter *adap, u16 phys_addr, bool block)
drivers/media/cec/core/cec-adap.c
1738
__cec_s_phys_addr(adap, phys_addr, block);
drivers/media/cec/core/cec-adap.c
1785
struct cec_log_addrs *log_addrs, bool block)
drivers/media/cec/core/cec-adap.c
1943
cec_claim_log_addrs(adap, block);
drivers/media/cec/core/cec-adap.c
1948
struct cec_log_addrs *log_addrs, bool block)
drivers/media/cec/core/cec-adap.c
1953
err = __cec_s_log_addrs(adap, log_addrs, block);
drivers/media/cec/core/cec-adap.c
781
struct cec_fh *fh, bool block)
drivers/media/cec/core/cec-adap.c
828
!block ? ", nb" : "");
drivers/media/cec/core/cec-adap.c
831
__func__, msg->len, msg->msg, !block ? " (nb)" : "");
drivers/media/cec/core/cec-adap.c
936
data->blocking = block;
drivers/media/cec/core/cec-adap.c
960
if (!block)
drivers/media/cec/core/cec-adap.c
994
bool block)
drivers/media/cec/core/cec-adap.c
999
ret = cec_transmit_msg_fh(adap, msg, NULL, block);
drivers/media/cec/core/cec-api.c
122
bool block, __u16 __user *parg)
drivers/media/cec/core/cec-api.c
139
__cec_s_phys_addr(adap, phys_addr, block);
drivers/media/cec/core/cec-api.c
168
bool block, struct cec_log_addrs __user *parg)
drivers/media/cec/core/cec-api.c
184
err = __cec_s_log_addrs(adap, &log_addrs, block);
drivers/media/cec/core/cec-api.c
212
bool block, struct cec_msg __user *parg)
drivers/media/cec/core/cec-api.c
230
err = cec_transmit_msg_fh(adap, &msg, fh, block);
drivers/media/cec/core/cec-api.c
240
static int cec_receive_msg(struct cec_fh *fh, struct cec_msg *msg, bool block)
drivers/media/cec/core/cec-api.c
268
if (!block)
drivers/media/cec/core/cec-api.c
291
bool block, struct cec_msg __user *parg)
drivers/media/cec/core/cec-api.c
299
err = cec_receive_msg(fh, &msg, block);
drivers/media/cec/core/cec-api.c
309
bool block, struct cec_event __user *parg)
drivers/media/cec/core/cec-api.c
318
while (!fh->total_queued_events && block) {
drivers/media/cec/core/cec-api.c
508
bool block = !(filp->f_flags & O_NONBLOCK);
drivers/media/cec/core/cec-api.c
522
return cec_adap_s_phys_addr(adap, fh, block, parg);
drivers/media/cec/core/cec-api.c
528
return cec_adap_s_log_addrs(adap, fh, block, parg);
drivers/media/cec/core/cec-api.c
534
return cec_transmit(adap, fh, block, parg);
drivers/media/cec/core/cec-api.c
537
return cec_receive(adap, fh, block, parg);
drivers/media/cec/core/cec-api.c
540
return cec_dqevent(adap, fh, block, parg);
drivers/media/cec/core/cec-priv.h
49
void __cec_s_phys_addr(struct cec_adapter *adap, u16 phys_addr, bool block);
drivers/media/cec/core/cec-priv.h
51
struct cec_log_addrs *log_addrs, bool block);
drivers/media/cec/core/cec-priv.h
53
struct cec_fh *fh, bool block);
drivers/media/dvb-frontends/dib9000.c
666
u16 *block;
drivers/media/dvb-frontends/dib9000.c
669
block = state->platform.risc.message_cache[i];
drivers/media/dvb-frontends/dib9000.c
670
if (*block == 0) {
drivers/media/dvb-frontends/dib9000.c
671
size = dib9000_mbx_read(state, block, 1, attr);
drivers/media/dvb-frontends/dib9000.c
675
switch (*block >> 8) {
drivers/media/dvb-frontends/dib9000.c
677
dib9000_risc_debug_buf(state, block + 1, size); /* debug-messages are going to be printed right away */
drivers/media/dvb-frontends/dib9000.c
678
*block = 0; /* free the block */
drivers/media/dvb-frontends/dib9000.c
682
dib9000_risc_data_process(state, block + 1, size);
drivers/media/dvb-frontends/dib9000.c
683
*block = 0;
drivers/media/dvb-frontends/dib9000.c
731
u16 *block;
drivers/media/dvb-frontends/dib9000.c
738
block = state->platform.risc.message_cache[i];
drivers/media/dvb-frontends/dib9000.c
739
if ((*block >> 8) == id) {
drivers/media/dvb-frontends/dib9000.c
740
*size = (*block & 0xff) - 1;
drivers/media/dvb-frontends/dib9000.c
741
memcpy(msg, block + 1, (*size) * 2);
drivers/media/dvb-frontends/dib9000.c
742
*block = 0; /* free the block */
drivers/media/i2c/ccs/ccs-data.c
118
ccs_data_parse_format_version(const struct __ccs_data_block *block)
drivers/media/i2c/ccs/ccs-data.c
120
return block->id >> CCS_DATA_BLOCK_HEADER_ID_VERSION_SHIFT;
drivers/media/i2c/ccs/ccs-data.c
123
static u8 ccs_data_parse_block_id(const struct __ccs_data_block *block,
drivers/media/i2c/ccs/ccs-data.c
127
return block->id;
drivers/media/i2c/ccs/ccs-data.c
129
return block->id & ((1 << CCS_DATA_BLOCK_HEADER_ID_VERSION_SHIFT) - 1);
drivers/media/i2c/ccs/ccs-data.c
172
static int ccs_data_block_parse_header(const struct __ccs_data_block *block,
drivers/media/i2c/ccs/ccs-data.c
183
if (!is_contained(block, endp))
drivers/media/i2c/ccs/ccs-data.c
186
rval = ccs_data_parse_length_specifier(&block->length, &hlen, &plen,
drivers/media/i2c/ccs/ccs-data.c
191
block_id = ccs_data_parse_block_id(block, is_first);
drivers/media/i2c/ccs/ccs-data.c
198
if (!has_headroom(&block->length, hlen + plen, endp))
drivers/media/i2c/ccs/ccs-data.c
205
*payload = (void *)&block->length + hlen;
drivers/media/i2c/ccs/ccs-data.c
208
*next_block = (void *)&block->length + hlen + plen;
drivers/media/i2c/ccs/ccs-data.c
810
const struct __ccs_data_block *block = data;
drivers/media/i2c/ccs/ccs-data.c
816
version = ccs_data_parse_format_version(block);
drivers/media/i2c/ccs/ccs-data.c
828
while (block < endp) {
drivers/media/i2c/ccs/ccs-data.c
833
rval = ccs_data_block_parse_header(block, is_first, &block_id,
drivers/media/i2c/ccs/ccs-data.c
928
block = next_block;
drivers/media/i2c/ds90ub953.c
156
static int ub953_select_ind_reg_block(struct ub953_data *priv, u8 block)
drivers/media/i2c/ds90ub953.c
161
if (priv->current_indirect_target == block)
drivers/media/i2c/ds90ub953.c
164
ret = regmap_write(priv->regmap, UB953_REG_IND_ACC_CTL, block << 2);
drivers/media/i2c/ds90ub953.c
167
__func__, block, ret);
drivers/media/i2c/ds90ub953.c
171
priv->current_indirect_target = block;
drivers/media/i2c/ds90ub953.c
177
static int ub953_read_ind(struct ub953_data *priv, u8 block, u8 reg, u8 *val,
drivers/media/i2c/ds90ub953.c
188
ret = ub953_select_ind_reg_block(priv, block);
drivers/media/i2c/ds90ub953.c
196
block, reg, ret);
drivers/media/i2c/ds90ub953.c
204
block, reg, ret);
drivers/media/i2c/ds90ub953.c
220
static int ub953_write_ind(struct ub953_data *priv, u8 block, u8 reg, u8 val,
drivers/media/i2c/ds90ub953.c
230
ret = ub953_select_ind_reg_block(priv, block);
drivers/media/i2c/ds90ub953.c
238
block, reg, ret);
drivers/media/i2c/ds90ub953.c
246
block, reg, ret);
drivers/media/i2c/ds90ub960.c
1093
static int ub960_select_ind_reg_block(struct ub960_data *priv, u8 block)
drivers/media/i2c/ds90ub960.c
1100
if (priv->reg_current.indirect_target == block)
drivers/media/i2c/ds90ub960.c
1103
ret = regmap_write(priv->regmap, UB960_SR_IND_ACC_CTL, block << 2);
drivers/media/i2c/ds90ub960.c
1106
__func__, block, ret);
drivers/media/i2c/ds90ub960.c
1110
priv->reg_current.indirect_target = block;
drivers/media/i2c/ds90ub960.c
1115
static int ub960_read_ind(struct ub960_data *priv, u8 block, u8 reg, u8 *val,
drivers/media/i2c/ds90ub960.c
1127
ret = ub960_select_ind_reg_block(priv, block);
drivers/media/i2c/ds90ub960.c
1135
block, reg, ret);
drivers/media/i2c/ds90ub960.c
1143
block, reg, ret);
drivers/media/i2c/ds90ub960.c
1158
static int ub960_write_ind(struct ub960_data *priv, u8 block, u8 reg, u8 val,
drivers/media/i2c/ds90ub960.c
1169
ret = ub960_select_ind_reg_block(priv, block);
drivers/media/i2c/ds90ub960.c
1177
block, reg, ret);
drivers/media/i2c/ds90ub960.c
1185
block, reg, ret);
drivers/media/i2c/ds90ub960.c
1198
static int ub960_ind_update_bits(struct ub960_data *priv, u8 block, u8 reg,
drivers/media/i2c/ds90ub960.c
1209
ret = ub960_select_ind_reg_block(priv, block);
drivers/media/i2c/ds90ub960.c
1217
block, reg, ret);
drivers/media/i2c/ds90ub960.c
1226
block, reg, ret);
drivers/media/i2c/thp7312.c
1581
u8 block;
drivers/media/i2c/thp7312.c
1584
for (block = 0; block < 3; block++) {
drivers/media/i2c/thp7312.c
1585
const u8 jedec_se[] = { SPINOR_OP_SE, block, 0x00, 0x00 };
drivers/media/pci/zoran/videocodec.h
269
long (*put_image)(struct videocodec *codec, int tr_type, int block,
drivers/media/pci/zoran/videocodec.h
271
long (*get_image)(struct videocodec *codec, int tr_type, int block,
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
152
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
154
const struct c3_isp_params_awb_config *awb_cfg = &block->awb_cfg;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
156
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
205
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
269
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
271
const struct c3_isp_params_ae_config *ae_cfg = &block->ae_cfg;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
273
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
303
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
327
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
329
const struct c3_isp_params_af_config *af_cfg = &block->af_cfg;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
331
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
351
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
358
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
360
const struct c3_isp_params_pst_gamma *gm = &block->pst_gamma;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
364
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
393
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
401
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
403
const struct c3_isp_params_ccm *ccm = &block->ccm;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
405
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
442
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
450
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
452
const struct c3_isp_params_csc *csc = &block->csc;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
454
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
491
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
499
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
501
const struct c3_isp_params_blc *blc = &block->blc;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
503
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
517
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
534
#define C3_ISP_PARAMS_BLOCK_INFO(block, data) \
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
535
[C3_ISP_PARAMS_BLOCK_ ## block] = { \
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
54
const union c3_isp_params_block *block);
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
564
const union c3_isp_params_block *block;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
567
block = (const union c3_isp_params_block *)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
570
block_handler = c3_isp_params_handlers[block->header.type];
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
571
block_handler(params->isp, block);
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
573
block_offset += block->header.size;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
62
const union c3_isp_params_block *block)
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
64
const struct c3_isp_params_awb_gains *awb_gains = &block->awb_gains;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
66
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_DISABLE) {
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
89
if (block->header.flags & C3_ISP_PARAMS_BLOCK_FL_ENABLE)
drivers/media/platform/amphion/vpu.h
119
bool block;
drivers/media/platform/amphion/vpu_mbox.c
37
if (mbox->block) {
drivers/media/platform/amphion/vpu_mbox.c
59
core->tx_type.block = true;
drivers/media/platform/amphion/vpu_mbox.c
62
core->tx_data.block = false;
drivers/media/platform/amphion/vpu_mbox.c
65
core->rx.block = true;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
113
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
120
if (block.header->type == MALI_C55_PARAM_BLOCK_AEXP_HIST) {
drivers/media/platform/arm/mali-c55/mali-c55-params.c
130
params = block.aexp_hist;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
132
if (block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE) {
drivers/media/platform/arm/mali-c55/mali-c55-params.c
164
if (block.header->type == MALI_C55_PARAM_BLOCK_AEXP_HIST)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
172
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
177
params = block.aexp_weights;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
179
if (block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
182
base = block.header->type == MALI_C55_PARAM_BLOCK_AEXP_HIST_WEIGHTS ?
drivers/media/platform/arm/mali-c55/mali-c55-params.c
218
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
223
dgain = block.digital_gain;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
229
gain = block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE ?
drivers/media/platform/arm/mali-c55/mali-c55-params.c
238
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
243
gains = block.awb_gains;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
250
u32 addr1 = block.header->type == MALI_C55_PARAM_BLOCK_AWB_GAINS ?
drivers/media/platform/arm/mali-c55/mali-c55-params.c
253
u32 addr2 = block.header->type == MALI_C55_PARAM_BLOCK_AWB_GAINS ?
drivers/media/platform/arm/mali-c55/mali-c55-params.c
258
if (block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE) {
drivers/media/platform/arm/mali-c55/mali-c55-params.c
281
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
285
params = block.awb_config;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
287
if (block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE) {
drivers/media/platform/arm/mali-c55/mali-c55-params.c
336
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
342
params = block.shading_config;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
344
if (block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE) {
drivers/media/platform/arm/mali-c55/mali-c55-params.c
382
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
386
params = block.shading_selection;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
388
if (block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
66
union mali_c55_params_block block);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
72
union mali_c55_params_block block)
drivers/media/platform/arm/mali-c55/mali-c55-params.c
720
union mali_c55_params_block block;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
723
block.data = &config->data[block_offset];
drivers/media/platform/arm/mali-c55/mali-c55-params.c
726
handler = mali_c55_params_handlers[block.header->type];
drivers/media/platform/arm/mali-c55/mali-c55-params.c
727
handler(mali_c55, block);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
729
block_offset += block.header->size;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
77
p = block.sensor_offs;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
79
if (block.header->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE) {
drivers/media/platform/qcom/venus/hfi.c
415
int hfi_session_flush(struct venus_inst *inst, u32 type, bool block)
drivers/media/platform/qcom/venus/hfi.c
429
if (block) {
drivers/media/platform/qcom/venus/hfi.h
164
int hfi_session_flush(struct venus_inst *inst, u32 type, bool block);
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
488
unsigned int block, offset, reg;
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
490
block = interface->index - RKCIF_MIPI_BASE;
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
492
if (WARN_ON_ONCE(block > RKCIF_MIPI_MAX - RKCIF_MIPI_BASE) ||
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
496
offset = rkcif->match_data->mipi->blocks[block].offset;
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
508
unsigned int block, id, offset, reg;
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
510
block = stream->interface->index - RKCIF_MIPI_BASE;
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
513
if (WARN_ON_ONCE(block > RKCIF_MIPI_MAX - RKCIF_MIPI_BASE) ||
drivers/media/platform/rockchip/rkcif/rkcif-capture-mipi.c
518
offset = rkcif->match_data->mipi->blocks[block].offset;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1690
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1692
const struct rkisp1_ext_params_bls_config *bls = &block->bls;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1710
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1712
const struct rkisp1_ext_params_dpcc_config *dpcc = &block->dpcc;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1730
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1732
const struct rkisp1_ext_params_sdg_config *sdg = &block->sdg;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1750
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1752
const struct rkisp1_ext_params_lsc_config *lsc = &block->lsc;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1770
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1772
const struct rkisp1_ext_params_awb_gain_config *awbg = &block->awbg;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1790
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1792
const struct rkisp1_ext_params_flt_config *flt = &block->flt;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1810
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1812
const struct rkisp1_ext_params_bdm_config *bdm = &block->bdm;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1830
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1832
const struct rkisp1_ext_params_ctk_config *ctk = &block->ctk;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1848
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1850
const struct rkisp1_ext_params_goc_config *goc = &block->goc;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1870
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1872
const struct rkisp1_ext_params_dpf_config *dpf = &block->dpf;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1890
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1892
const struct rkisp1_ext_params_dpf_strength_config *dpfs = &block->dpfs;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1899
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1901
const struct rkisp1_ext_params_cproc_config *cproc = &block->cproc;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1919
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1921
const struct rkisp1_ext_params_ie_config *ie = &block->ie;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1937
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1939
const struct rkisp1_ext_params_awb_meas_config *awbm = &block->awbm;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1957
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1959
const struct rkisp1_ext_params_hst_config *hst = &block->hst;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1975
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1977
const struct rkisp1_ext_params_aec_config *aec = &block->aec;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1995
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
1997
const struct rkisp1_ext_params_afc_config *afc = &block->afc;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2014
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2017
&block->compand_bls;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2034
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2037
&block->compand_curve;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2054
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2057
&block->compand_curve;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2074
const union rkisp1_ext_params_config *block)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2076
const struct rkisp1_ext_params_wdr_config *wdr = &block->wdr;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2190
#define RKISP1_PARAMS_BLOCK_INFO(block, data) \
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2191
[RKISP1_EXT_PARAMS_BLOCK_TYPE_ ## block] = { \
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2235
const union rkisp1_ext_params_config *block;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2237
block = (const union rkisp1_ext_params_config *)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2239
block_offset += block->header.size;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2245
block_handler = &rkisp1_ext_params_handlers[block->header.type];
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2253
block_handler->handler(params, block);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2255
if (block->header.flags & RKISP1_EXT_PARAMS_FL_BLOCK_DISABLE)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2256
params->enabled_blocks &= ~BIT(block->header.type);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2257
else if (block->header.flags & RKISP1_EXT_PARAMS_FL_BLOCK_ENABLE)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2258
params->enabled_blocks |= BIT(block->header.type);
drivers/media/platform/rockchip/rkvdec/rkvdec.c
1580
u8 block;
drivers/media/platform/rockchip/rkvdec/rkvdec.c
1587
block = j / 16;
drivers/media/platform/rockchip/rkvdec/rkvdec.c
1589
block_offset = (block & 1) * 32 + (block & 2) * 2;
drivers/media/platform/ti/omap3isp/isppreview.c
257
const __u32 *block = cfa->table[order[i]];
drivers/media/platform/ti/omap3isp/isppreview.c
260
isp_reg_writel(isp, block[j], OMAP3_ISP_IOMEM_PREV,
drivers/media/test-drivers/vicodec/codec-fwht.c
121
s16 block[8 * 8 + 16];
drivers/media/test-drivers/vicodec/codec-fwht.c
122
s16 *wp = block;
drivers/media/test-drivers/vicodec/codec-fwht.c
161
wp = block;
drivers/media/test-drivers/vicodec/codec-fwht.c
248
static void noinline_for_stack fwht(const u8 *block, s16 *output_block,
drivers/media/test-drivers/vicodec/codec-fwht.c
254
const u8 *tmp = block;
drivers/media/test-drivers/vicodec/codec-fwht.c
383
fwht16(const s16 *block, s16 *output_block, int stride, int intra)
drivers/media/test-drivers/vicodec/codec-fwht.c
387
const s16 *tmp = block;
drivers/media/test-drivers/vicodec/codec-fwht.c
467
ifwht(const s16 *block, s16 *output_block, int intra)
drivers/media/test-drivers/vicodec/codec-fwht.c
475
const s16 *tmp = block;
drivers/media/test-drivers/vicodec/codec-fwht.c
57
s16 block[8 * 8];
drivers/media/test-drivers/vicodec/codec-fwht.c
58
s16 *wp = block;
drivers/media/test-drivers/vicodec/codec-fwht.c
75
for (i = 63; i >= 0 && !block[zigzag[i]]; i--)
drivers/media/test-drivers/vicodec/codec-fwht.c
89
while ((tmp = block[zigzag[i]]) == 0 && cnt < 14) {
drivers/media/test-drivers/vivid/vivid-radio-rx.c
100
rds.block |= V4L2_RDS_BLOCK_CORRECTED;
drivers/media/test-drivers/vivid/vivid-radio-rx.c
103
rds.block |= V4L2_RDS_BLOCK_INVALID;
drivers/media/test-drivers/vivid/vivid-radio-rx.c
106
rds.block |= V4L2_RDS_BLOCK_ERROR;
drivers/media/test-drivers/vivid/vivid-radio-tx.c
84
if ((rds.block & V4L2_RDS_BLOCK_MSK) == V4L2_RDS_BLOCK_INVALID ||
drivers/media/test-drivers/vivid/vivid-radio-tx.c
85
(rds.block & V4L2_RDS_BLOCK_ERROR))
drivers/media/test-drivers/vivid/vivid-radio-tx.c
87
rds.block &= V4L2_RDS_BLOCK_MSK;
drivers/media/test-drivers/vivid/vivid-rds-gen.c
108
data[2].block = V4L2_RDS_BLOCK_C | (V4L2_RDS_BLOCK_C << 3);
drivers/media/test-drivers/vivid/vivid-rds-gen.c
118
data[2].block = V4L2_RDS_BLOCK_C_ALT | (V4L2_RDS_BLOCK_C_ALT << 3);
drivers/media/test-drivers/vivid/vivid-rds-gen.c
55
data[0].block = V4L2_RDS_BLOCK_A | (V4L2_RDS_BLOCK_A << 3);
drivers/media/test-drivers/vivid/vivid-rds-gen.c
58
data[1].block = V4L2_RDS_BLOCK_B | (V4L2_RDS_BLOCK_B << 3);
drivers/media/test-drivers/vivid/vivid-rds-gen.c
59
data[3].block = V4L2_RDS_BLOCK_D | (V4L2_RDS_BLOCK_D << 3);
drivers/media/test-drivers/vivid/vivid-rds-gen.c
71
data[2].block = V4L2_RDS_BLOCK_C_ALT | (V4L2_RDS_BLOCK_C_ALT << 3);
drivers/media/test-drivers/vivid/vivid-rds-gen.c
82
data[2].block = V4L2_RDS_BLOCK_C | (V4L2_RDS_BLOCK_C << 3);
drivers/media/usb/pwc/pwc-dec23.c
521
const unsigned int *block;
drivers/media/usb/pwc/pwc-dec23.c
560
block = pdec->table_subblock[rows];
drivers/media/usb/pwc/pwc-dec23.c
562
pdec->temp_colors[i] += block[MulIdx[offset1][i]];
drivers/media/usb/pwc/pwc-dec23.c
574
block = pdec->table_subblock[rows];
drivers/media/usb/pwc/pwc-dec23.c
576
pdec->temp_colors[i] += block[MulIdx[offset1][i]];
drivers/media/v4l2-core/v4l2-isp.c
107
info = &type_info[block->type];
drivers/media/v4l2-core/v4l2-isp.c
108
if (block->size != info->size &&
drivers/media/v4l2-core/v4l2-isp.c
109
(!(block->flags & V4L2_ISP_PARAMS_FL_BLOCK_DISABLE) ||
drivers/media/v4l2-core/v4l2-isp.c
110
block->size != sizeof(*block))) {
drivers/media/v4l2-core/v4l2-isp.c
113
block->size, info->size, block_offset);
drivers/media/v4l2-core/v4l2-isp.c
117
block_offset += block->size;
drivers/media/v4l2-core/v4l2-isp.c
118
buffer_size -= block->size;
drivers/media/v4l2-core/v4l2-isp.c
75
const struct v4l2_isp_params_block_header *block;
drivers/media/v4l2-core/v4l2-isp.c
77
block = (const struct v4l2_isp_params_block_header *)
drivers/media/v4l2-core/v4l2-isp.c
80
if (block->type >= num_block_types) {
drivers/media/v4l2-core/v4l2-isp.c
83
block->type, block_offset);
drivers/media/v4l2-core/v4l2-isp.c
87
if (block->size > buffer_size) {
drivers/media/v4l2-core/v4l2-isp.c
93
if ((block->flags & (V4L2_ISP_PARAMS_FL_BLOCK_ENABLE |
drivers/media/v4l2-core/v4l2-isp.c
98
block->flags, block_offset);
drivers/mfd/qcom-pm8xxx.c
124
static int pm8xxx_irq_block_handler(struct pm_irq_chip *chip, int block)
drivers/mfd/qcom-pm8xxx.c
129
ret = pm8xxx_read_block_irq(chip, block, &bits);
drivers/mfd/qcom-pm8xxx.c
131
pr_err("Failed reading %d block ret=%d", block, ret);
drivers/mfd/qcom-pm8xxx.c
135
pr_err("block bit set in master but no irqs: %d", block);
drivers/mfd/qcom-pm8xxx.c
142
pmirq = block * 8 + i;
drivers/mfd/qcom-pm8xxx.c
197
int master, int block)
drivers/mfd/qcom-pm8xxx.c
203
PM8821_SSBI_ADDR_IRQ_ROOT(master, block), &bits);
drivers/mfd/qcom-pm8xxx.c
205
pr_err("Reading block %d failed ret=%d", block, ret);
drivers/mfd/qcom-pm8xxx.c
210
block += (master * PM8821_BLOCKS_PER_MASTER) - 1;
drivers/mfd/qcom-pm8xxx.c
215
pmirq = block * 8 + i;
drivers/mfd/qcom-pm8xxx.c
224
int block;
drivers/mfd/qcom-pm8xxx.c
226
for (block = 1; block < 8; block++)
drivers/mfd/qcom-pm8xxx.c
227
if (master_val & BIT(block))
drivers/mfd/qcom-pm8xxx.c
228
pm8821_irq_block_handler(chip, master, block);
drivers/mfd/qcom-pm8xxx.c
268
u8 block, config;
drivers/mfd/qcom-pm8xxx.c
270
block = pmirq / 8;
drivers/mfd/qcom-pm8xxx.c
273
pm8xxx_config_irq(chip, block, config);
drivers/mfd/qcom-pm8xxx.c
280
u8 block, config;
drivers/mfd/qcom-pm8xxx.c
282
block = pmirq / 8;
drivers/mfd/qcom-pm8xxx.c
285
pm8xxx_config_irq(chip, block, config);
drivers/mfd/qcom-pm8xxx.c
293
u8 block, config;
drivers/mfd/qcom-pm8xxx.c
295
block = pmirq / 8;
drivers/mfd/qcom-pm8xxx.c
315
return pm8xxx_config_irq(chip, block, config);
drivers/mfd/qcom-pm8xxx.c
327
u8 block;
drivers/mfd/qcom-pm8xxx.c
333
block = pmirq / 8;
drivers/mfd/qcom-pm8xxx.c
337
rc = regmap_write(chip->regmap, SSBI_REG_ADDR_IRQ_BLK_SEL, block);
drivers/mfd/qcom-pm8xxx.c
339
pr_err("Failed Selecting Block %d rc=%d\n", block, rc);
drivers/mfd/qcom-pm8xxx.c
403
u8 block, master;
drivers/mfd/qcom-pm8xxx.c
406
block = pmirq / 8;
drivers/mfd/qcom-pm8xxx.c
407
master = block / PM8821_BLOCKS_PER_MASTER;
drivers/mfd/qcom-pm8xxx.c
409
block %= PM8821_BLOCKS_PER_MASTER;
drivers/mfd/qcom-pm8xxx.c
412
PM8821_SSBI_ADDR_IRQ_MASK(master, block),
drivers/mfd/qcom-pm8xxx.c
420
PM8821_SSBI_ADDR_IRQ_CLEAR(master, block),
drivers/mfd/qcom-pm8xxx.c
431
u8 block, master;
drivers/mfd/qcom-pm8xxx.c
433
block = pmirq / 8;
drivers/mfd/qcom-pm8xxx.c
434
master = block / PM8821_BLOCKS_PER_MASTER;
drivers/mfd/qcom-pm8xxx.c
436
block %= PM8821_BLOCKS_PER_MASTER;
drivers/mfd/qcom-pm8xxx.c
439
PM8821_SSBI_ADDR_IRQ_MASK(master, block),
drivers/mfd/qcom-pm8xxx.c
452
u8 block, irq_bit, master;
drivers/mfd/qcom-pm8xxx.c
455
block = pmirq / 8;
drivers/mfd/qcom-pm8xxx.c
456
master = block / PM8821_BLOCKS_PER_MASTER;
drivers/mfd/qcom-pm8xxx.c
458
block %= PM8821_BLOCKS_PER_MASTER;
drivers/mfd/qcom-pm8xxx.c
461
PM8821_SSBI_ADDR_IRQ_RT_STATUS(master, block), &bits);
drivers/mfd/si476x-cmd.c
946
report->rds[V4L2_RDS_BLOCK_A].block = V4L2_RDS_BLOCK_A;
drivers/mfd/si476x-cmd.c
950
report->rds[V4L2_RDS_BLOCK_B].block = V4L2_RDS_BLOCK_B;
drivers/mfd/si476x-cmd.c
954
report->rds[V4L2_RDS_BLOCK_C].block = V4L2_RDS_BLOCK_C;
drivers/mfd/si476x-cmd.c
958
report->rds[V4L2_RDS_BLOCK_D].block = V4L2_RDS_BLOCK_D;
drivers/mfd/stmpe.c
1014
static int stmpe24xx_get_altfunc(struct stmpe *stmpe, enum stmpe_block block)
drivers/mfd/stmpe.c
1016
switch (block) {
drivers/mfd/stmpe.c
1309
struct stmpe_variant_block *block = &variant->blocks[i];
drivers/mfd/stmpe.c
1311
if (!(platform_blocks & block->block))
drivers/mfd/stmpe.c
1314
for (j = 0; j < block->cell->num_resources; j++) {
drivers/mfd/stmpe.c
1316
(struct resource *) &block->cell->resources[j];
drivers/mfd/stmpe.c
1320
res->start = res->end = block->irq + j;
drivers/mfd/stmpe.c
1323
platform_blocks &= ~block->block;
drivers/mfd/stmpe.c
1324
ret = stmpe_add_device(stmpe, block->cell);
drivers/mfd/stmpe.c
260
int stmpe_set_altfunc(struct stmpe *stmpe, u32 pins, enum stmpe_block block)
drivers/mfd/stmpe.c
284
af = variant->get_altfunc(stmpe, block);
drivers/mfd/stmpe.c
397
.block = STMPE_BLOCK_GPIO,
drivers/mfd/stmpe.c
404
.block = STMPE_BLOCK_GPIO,
drivers/mfd/stmpe.c
513
.block = STMPE_BLOCK_GPIO,
drivers/mfd/stmpe.c
518
.block = STMPE_BLOCK_TOUCHSCREEN,
drivers/mfd/stmpe.c
523
.block = STMPE_BLOCK_ADC,
drivers/mfd/stmpe.c
574
static int stmpe811_get_altfunc(struct stmpe *stmpe, enum stmpe_block block)
drivers/mfd/stmpe.c
577
return block != STMPE_BLOCK_TOUCHSCREEN;
drivers/mfd/stmpe.c
638
.block = STMPE_BLOCK_GPIO,
drivers/mfd/stmpe.c
701
.block = STMPE_BLOCK_GPIO,
drivers/mfd/stmpe.c
706
.block = STMPE_BLOCK_KEYPAD,
drivers/mfd/stmpe.c
711
.block = STMPE_BLOCK_PWM,
drivers/mfd/stmpe.c
800
static int stmpe1601_get_altfunc(struct stmpe *stmpe, enum stmpe_block block)
drivers/mfd/stmpe.c
802
switch (block) {
drivers/mfd/stmpe.c
868
.block = STMPE_BLOCK_GPIO,
drivers/mfd/stmpe.c
873
.block = STMPE_BLOCK_KEYPAD,
drivers/mfd/stmpe.c
985
.block = STMPE_BLOCK_GPIO,
drivers/mfd/stmpe.c
990
.block = STMPE_BLOCK_KEYPAD,
drivers/mfd/stmpe.c
995
.block = STMPE_BLOCK_PWM,
drivers/mfd/stmpe.h
43
enum stmpe_block block;
drivers/mfd/stmpe.h
74
int (*get_altfunc)(struct stmpe *stmpe, enum stmpe_block block);
drivers/mfd/tc3589x.c
287
unsigned int blocks = tc3589x->pdata->block;
drivers/mfd/tc3589x.c
345
pdata->block |= TC3589x_BLOCK_GPIO;
drivers/mfd/tc3589x.c
347
pdata->block |= TC3589x_BLOCK_KEYPAD;
drivers/misc/cb710/debug.c
22
static inline int allow_reg_read(unsigned block, unsigned offset, unsigned bits)
drivers/misc/cb710/debug.c
26
return ((allow[block] >> offset) & mask) == mask;
drivers/misc/mei/client.c
1600
int mei_cl_notify_get(struct mei_cl *cl, bool block, bool *notify_ev)
drivers/misc/mei/client.c
1623
if (!block)
drivers/misc/mei/client.h
261
int mei_cl_notify_get(struct mei_cl *cl, bool block, bool *notify_ev);
drivers/misc/mei/main.c
646
bool block = (file->f_flags & O_NONBLOCK) == 0;
drivers/misc/mei/main.c
649
rets = mei_cl_notify_get(cl, block, &notify_ev);
drivers/misc/sram-exec.c
22
int sram_check_protect_exec(struct sram_dev *sram, struct sram_reserve *block,
drivers/misc/sram-exec.c
26
unsigned long end = base + block->size;
drivers/misc/sram.c
107
virt_base = devm_ioremap_resource(sram->dev, &block->res);
drivers/misc/sram.c
109
virt_base = devm_ioremap_resource_wc(sram->dev, &block->res);
drivers/misc/sram.c
112
dev_err(sram->dev, "could not map SRAM at %pr\n", &block->res);
drivers/misc/sram.c
118
part->base = sram->virt_base + block->start;
drivers/misc/sram.c
121
if (block->pool) {
drivers/misc/sram.c
122
ret = sram_add_pool(sram, block, start, part);
drivers/misc/sram.c
126
if (block->export) {
drivers/misc/sram.c
127
ret = sram_add_export(sram, block, start, part);
drivers/misc/sram.c
131
if (block->protect_exec) {
drivers/misc/sram.c
132
ret = sram_check_protect_exec(sram, block, part);
drivers/misc/sram.c
136
ret = sram_add_pool(sram, block, start, part);
drivers/misc/sram.c
179
struct sram_reserve *rblocks, *block;
drivers/misc/sram.c
198
block = &rblocks[0];
drivers/misc/sram.c
218
block->start = child_res.start - res->start;
drivers/misc/sram.c
219
block->size = resource_size(&child_res);
drivers/misc/sram.c
220
block->res = child_res;
drivers/misc/sram.c
221
list_add_tail(&block->list, &reserve_list);
drivers/misc/sram.c
223
block->export = of_property_read_bool(child, "export");
drivers/misc/sram.c
224
block->pool = of_property_read_bool(child, "pool");
drivers/misc/sram.c
225
block->protect_exec = of_property_read_bool(child, "protect-exec");
drivers/misc/sram.c
227
if ((block->export || block->pool || block->protect_exec) &&
drivers/misc/sram.c
228
block->size) {
drivers/misc/sram.c
240
block->label = devm_kasprintf(sram->dev, GFP_KERNEL,
drivers/misc/sram.c
243
block->label = devm_kstrdup(sram->dev,
drivers/misc/sram.c
245
if (!block->label) {
drivers/misc/sram.c
251
block->export ? "exported " : "", block->label,
drivers/misc/sram.c
252
block->start, block->start + block->size);
drivers/misc/sram.c
255
block->start, block->start + block->size);
drivers/misc/sram.c
258
block++;
drivers/misc/sram.c
280
list_for_each_entry(block, &reserve_list, list) {
drivers/misc/sram.c
282
if (block->start < cur_start) {
drivers/misc/sram.c
285
block->start, cur_start);
drivers/misc/sram.c
291
if ((block->export || block->pool || block->protect_exec) &&
drivers/misc/sram.c
292
block->size) {
drivers/misc/sram.c
293
ret = sram_add_partition(sram, block,
drivers/misc/sram.c
294
res->start + block->start);
drivers/misc/sram.c
302
if (block->start == cur_start) {
drivers/misc/sram.c
303
cur_start = block->start + block->size;
drivers/misc/sram.c
312
cur_size = block->start - cur_start;
drivers/misc/sram.c
328
cur_start = block->start + block->size;
drivers/misc/sram.c
57
static int sram_add_pool(struct sram_dev *sram, struct sram_reserve *block,
drivers/misc/sram.c
63
NUMA_NO_NODE, block->label);
drivers/misc/sram.c
68
block->size, NUMA_NO_NODE);
drivers/misc/sram.c
77
static int sram_add_export(struct sram_dev *sram, struct sram_reserve *block,
drivers/misc/sram.c
90
part->battr.size = block->size;
drivers/misc/sram.c
95
static int sram_add_partition(struct sram_dev *sram, struct sram_reserve *block,
drivers/misc/sram.h
47
int sram_check_protect_exec(struct sram_dev *sram, struct sram_reserve *block,
drivers/misc/sram.h
52
struct sram_reserve *block,
drivers/mmc/host/bcm2835.c
1022
bool block, busy, data;
drivers/mmc/host/bcm2835.c
1026
block = host->irq_block;
drivers/mmc/host/bcm2835.c
1037
if (block)
drivers/mtd/chips/cfi_cmdset_0001.c
2498
int block, status, i;
drivers/mtd/chips/cfi_cmdset_0001.c
2507
for (block = 0; block < region->numblocks; block++){
drivers/mtd/chips/cfi_cmdset_0001.c
2509
adr = region->offset + block * len;
drivers/mtd/chips/cfi_cmdset_0001.c
2514
set_bit(block, region->lockmap);
drivers/mtd/chips/cfi_cmdset_0001.c
2516
clear_bit(block, region->lockmap);
drivers/mtd/chips/cfi_cmdset_0001.c
2599
int block, i;
drivers/mtd/chips/cfi_cmdset_0001.c
2608
for_each_clear_bit(block, region->lockmap, region->numblocks) {
drivers/mtd/chips/cfi_cmdset_0001.c
2610
adr = region->offset + block * len;
drivers/mtd/devices/docg3.c
986
int block = DOC_LAYOUT_BLOCK_BBT;
drivers/mtd/devices/docg3.c
992
ret = doc_read_page_prepare(docg3, block, block + 1,
drivers/mtd/devices/mtd_intel_dg.c
389
const u32 block = 0x10;
drivers/mtd/devices/mtd_intel_dg.c
396
iowrite32(region << 24 | block, base + NVM_ERASE_REG);
drivers/mtd/ftl.c
955
unsigned long block, char *buf)
drivers/mtd/ftl.c
957
return ftl_read((void *)dev, buf, block, 1);
drivers/mtd/ftl.c
961
unsigned long block, char *buf)
drivers/mtd/ftl.c
963
return ftl_write((void *)dev, buf, block, 1);
drivers/mtd/inftlcore.c
242
int block, silly;
drivers/mtd/inftlcore.c
267
for (block = 0; block < inftl->EraseSize/SECTORSIZE; block ++) {
drivers/mtd/inftlcore.c
268
if ((BlockMap[block] != BLOCK_NIL) ||
drivers/mtd/inftlcore.c
269
BlockDeleted[block])
drivers/mtd/inftlcore.c
273
+ (block * SECTORSIZE), 16, &retlen,
drivers/mtd/inftlcore.c
284
BlockMap[block] = thisEUN;
drivers/mtd/inftlcore.c
287
BlockDeleted[block] = 1;
drivers/mtd/inftlcore.c
292
block, thisEUN, status);
drivers/mtd/inftlcore.c
313
for (block = 0; block < inftl->EraseSize/SECTORSIZE ; block++) {
drivers/mtd/inftlcore.c
321
if (BlockMap[block] == targetEUN || (pendingblock ==
drivers/mtd/inftlcore.c
322
(thisVUC * (inftl->EraseSize / SECTORSIZE) + block))) {
drivers/mtd/inftlcore.c
330
if (BlockMap[block] == BLOCK_NIL)
drivers/mtd/inftlcore.c
334
(inftl->EraseSize * BlockMap[block]) + (block * SECTORSIZE),
drivers/mtd/inftlcore.c
340
(inftl->EraseSize * BlockMap[block]) + (block * SECTORSIZE),
drivers/mtd/inftlcore.c
351
(block * SECTORSIZE), SECTORSIZE, &retlen,
drivers/mtd/inftlcore.c
460
static inline u16 INFTL_findwriteunit(struct INFTLrecord *inftl, unsigned block)
drivers/mtd/inftlcore.c
462
unsigned int thisVUC = block / (inftl->EraseSize / SECTORSIZE);
drivers/mtd/inftlcore.c
464
unsigned long blockofs = (block * SECTORSIZE) & (inftl->EraseSize -1);
drivers/mtd/inftlcore.c
473
inftl, block);
drivers/mtd/inftlcore.c
491
block , writeEUN, status);
drivers/mtd/inftlcore.c
538
thisEUN = INFTL_makefreeblock(inftl, block);
drivers/mtd/inftlcore.c
632
int block, silly;
drivers/mtd/inftlcore.c
655
for (block = 0; block < inftl->EraseSize/SECTORSIZE; block++) {
drivers/mtd/inftlcore.c
656
if (BlockUsed[block] || BlockDeleted[block])
drivers/mtd/inftlcore.c
660
+ (block * SECTORSIZE), 8 , &retlen,
drivers/mtd/inftlcore.c
671
BlockUsed[block] = 1;
drivers/mtd/inftlcore.c
674
BlockDeleted[block] = 1;
drivers/mtd/inftlcore.c
679
block, thisEUN, status);
drivers/mtd/inftlcore.c
692
for (block = 0; block < inftl->EraseSize/SECTORSIZE; block++)
drivers/mtd/inftlcore.c
693
if (BlockUsed[block])
drivers/mtd/inftlcore.c
746
static int INFTL_deleteblock(struct INFTLrecord *inftl, unsigned block)
drivers/mtd/inftlcore.c
748
unsigned int thisEUN = inftl->VUtable[block / (inftl->EraseSize / SECTORSIZE)];
drivers/mtd/inftlcore.c
749
unsigned long blockofs = (block * SECTORSIZE) & (inftl->EraseSize - 1);
drivers/mtd/inftlcore.c
757
"block=%d)\n", inftl, block);
drivers/mtd/inftlcore.c
778
block, thisEUN, status);
drivers/mtd/inftlcore.c
785
block / (inftl->EraseSize / SECTORSIZE));
drivers/mtd/inftlcore.c
800
INFTL_trydeletechain(inftl, block / (inftl->EraseSize / SECTORSIZE));
drivers/mtd/inftlcore.c
805
static int inftl_writeblock(struct mtd_blktrans_dev *mbd, unsigned long block,
drivers/mtd/inftlcore.c
810
unsigned long blockofs = (block * SECTORSIZE) & (inftl->EraseSize - 1);
drivers/mtd/inftlcore.c
816
"buffer=%p)\n", inftl, block, buffer);
drivers/mtd/inftlcore.c
824
writeEUN = INFTL_findwriteunit(inftl, block);
drivers/mtd/inftlcore.c
847
INFTL_deleteblock(inftl, block);
drivers/mtd/inftlcore.c
853
static int inftl_readblock(struct mtd_blktrans_dev *mbd, unsigned long block,
drivers/mtd/inftlcore.c
857
unsigned int thisEUN = inftl->VUtable[block / (inftl->EraseSize / SECTORSIZE)];
drivers/mtd/inftlcore.c
858
unsigned long blockofs = (block * SECTORSIZE) & (inftl->EraseSize - 1);
drivers/mtd/inftlcore.c
866
"buffer=%p)\n", inftl, block, buffer);
drivers/mtd/inftlcore.c
887
block, thisEUN, status);
drivers/mtd/inftlcore.c
894
block / (inftl->EraseSize / SECTORSIZE));
drivers/mtd/inftlmount.c
107
mtd_read(mtd, block * inftl->EraseSize + 4096, SECTORSIZE,
drivers/mtd/inftlmount.c
172
block >>= mh->BlockMultiplierBits;
drivers/mtd/inftlmount.c
280
inftl->PUtable[block] = BLOCK_RESERVED;
drivers/mtd/inftlmount.c
294
inftl->MediaUnit = block;
drivers/mtd/inftlmount.c
35
unsigned int i, block;
drivers/mtd/inftlmount.c
361
int INFTL_formatblock(struct INFTLrecord *inftl, int block)
drivers/mtd/inftlmount.c
369
pr_debug("INFTL: INFTL_formatblock(inftl=%p,block=%d)\n", inftl, block);
drivers/mtd/inftlmount.c
377
instr->addr = block * inftl->EraseSize;
drivers/mtd/inftlmount.c
389
block);
drivers/mtd/inftlmount.c
408
instr->addr = block * inftl->EraseSize + SECTORSIZE * 2;
drivers/mtd/inftlmount.c
429
unsigned int block = first_block, block1;
drivers/mtd/inftlmount.c
435
block1 = inftl->PUtable[block];
drivers/mtd/inftlmount.c
437
printk(KERN_WARNING "INFTL: formatting block %d\n", block);
drivers/mtd/inftlmount.c
438
if (INFTL_formatblock(inftl, block) < 0) {
drivers/mtd/inftlmount.c
442
inftl->PUtable[block] = BLOCK_RESERVED;
drivers/mtd/inftlmount.c
444
inftl->PUtable[block] = BLOCK_FREE;
drivers/mtd/inftlmount.c
448
block = block1;
drivers/mtd/inftlmount.c
450
if (block == BLOCK_NIL || block >= inftl->lastEUN)
drivers/mtd/inftlmount.c
502
int logical, block, i;
drivers/mtd/inftlmount.c
509
block = s->VUtable[logical];
drivers/mtd/inftlmount.c
510
if (block >= s->nb_blocks)
drivers/mtd/inftlmount.c
512
pr_debug(" LOGICAL %d --> %d ", logical, block);
drivers/mtd/inftlmount.c
514
if (s->PUtable[block] == BLOCK_NIL)
drivers/mtd/inftlmount.c
516
block = s->PUtable[block];
drivers/mtd/inftlmount.c
517
pr_debug("%d ", block);
drivers/mtd/inftlmount.c
529
unsigned int block, first_block, prev_block, last_block;
drivers/mtd/inftlmount.c
55
for (block = 0; block < inftl->nb_blocks; block++) {
drivers/mtd/inftlmount.c
550
logical_block = block = BLOCK_NIL;
drivers/mtd/inftlmount.c
572
block = first_block;
drivers/mtd/inftlmount.c
577
(s->PUtable[block] != BLOCK_NOTEXPLORED)) {
drivers/mtd/inftlmount.c
582
if (inftl_read_oob(mtd, block * s->EraseSize + 8,
drivers/mtd/inftlmount.c
584
inftl_read_oob(mtd, block * s->EraseSize +
drivers/mtd/inftlmount.c
595
ANACtable[block] = h0.ANAC;
drivers/mtd/inftlmount.c
602
if (s->PUtable[block] != BLOCK_NOTEXPLORED) {
drivers/mtd/inftlmount.c
606
s->PUtable[last_block] = block;
drivers/mtd/inftlmount.c
615
"mark 0x%x?\n", block, first_block,
drivers/mtd/inftlmount.c
62
ret = mtd_read(mtd, block * inftl->EraseSize, SECTORSIZE,
drivers/mtd/inftlmount.c
629
s->PUtable[block] = BLOCK_FREE;
drivers/mtd/inftlmount.c
640
block, first_block);
drivers/mtd/inftlmount.c
661
s->PUtable[block] = BLOCK_NIL;
drivers/mtd/inftlmount.c
663
s->PUtable[last_block] = block;
drivers/mtd/inftlmount.c
664
last_block = block;
drivers/mtd/inftlmount.c
665
block = prev_block;
drivers/mtd/inftlmount.c
668
if (block == BLOCK_NIL)
drivers/mtd/inftlmount.c
672
if (block > s->lastEUN) {
drivers/mtd/inftlmount.c
674
"block %d in chain %d?\n", block,
drivers/mtd/inftlmount.c
705
block = s->VUtable[logical_block];
drivers/mtd/inftlmount.c
709
if (block >= BLOCK_RESERVED)
drivers/mtd/inftlmount.c
712
ANAC = ANACtable[block];
drivers/mtd/inftlmount.c
714
if (s->PUtable[block] == BLOCK_NIL)
drivers/mtd/inftlmount.c
716
if (s->PUtable[block] > s->lastEUN) {
drivers/mtd/inftlmount.c
719
s->PUtable[block], logical_block);
drivers/mtd/inftlmount.c
72
block * inftl->EraseSize,
drivers/mtd/inftlmount.c
720
s->PUtable[block] = BLOCK_NIL;
drivers/mtd/inftlmount.c
723
if (ANACtable[block] != ANAC) {
drivers/mtd/inftlmount.c
729
s->VUtable[logical_block] = block;
drivers/mtd/inftlmount.c
735
last_block = block;
drivers/mtd/inftlmount.c
736
block = s->PUtable[block];
drivers/mtd/inftlmount.c
758
for (block = s->firstEUN; block <= s->lastEUN; block++) {
drivers/mtd/inftlmount.c
759
if (s->PUtable[block] == BLOCK_NOTEXPLORED) {
drivers/mtd/inftlmount.c
761
block);
drivers/mtd/inftlmount.c
762
if (INFTL_formatblock(s, block) < 0)
drivers/mtd/inftlmount.c
763
s->PUtable[block] = BLOCK_RESERVED;
drivers/mtd/inftlmount.c
765
s->PUtable[block] = BLOCK_FREE;
drivers/mtd/inftlmount.c
767
if (s->PUtable[block] == BLOCK_FREE) {
drivers/mtd/inftlmount.c
770
s->LastFreeEUN = block;
drivers/mtd/inftlmount.c
89
block * inftl->EraseSize + SECTORSIZE + 8,
drivers/mtd/inftlmount.c
94
"(err %d)\n", block * inftl->EraseSize,
drivers/mtd/maps/vmu-flash.c
18
unsigned int block; /* Which block was cached */
drivers/mtd/maps/vmu-flash.c
211
pcache->block = num;
drivers/mtd/maps/vmu-flash.c
380
(pcache->block == vblock->num)) {
drivers/mtd/mtd_blkdevs.c
51
unsigned long block, nsect;
drivers/mtd/mtd_blkdevs.c
54
block = blk_rq_pos(req) << 9 >> tr->blkshift;
drivers/mtd/mtd_blkdevs.c
63
if (tr->discard(dev, block, nsect))
drivers/mtd/mtd_blkdevs.c
68
for (; nsect > 0; nsect--, block++, buf += tr->blksize) {
drivers/mtd/mtd_blkdevs.c
69
if (tr->readsect(dev, block, buf)) {
drivers/mtd/mtd_blkdevs.c
87
for (; nsect > 0; nsect--, block++, buf += tr->blksize) {
drivers/mtd/mtd_blkdevs.c
88
if (tr->writesect(dev, block, buf)) {
drivers/mtd/mtdblock.c
231
unsigned long block, char *buf)
drivers/mtd/mtdblock.c
234
return do_cached_read(mtdblk, block<<9, 512, buf);
drivers/mtd/mtdblock.c
238
unsigned long block, char *buf)
drivers/mtd/mtdblock.c
250
return do_cached_write(mtdblk, block<<9, 512, buf);
drivers/mtd/mtdblock_ro.c
16
unsigned long block, char *buf)
drivers/mtd/mtdblock_ro.c
21
err = mtd_read(dev->mtd, (block * 512), 512, &retlen, buf);
drivers/mtd/mtdblock_ro.c
28
unsigned long block, char *buf)
drivers/mtd/mtdblock_ro.c
32
if (mtd_write(dev->mtd, (block * 512), 512, &retlen, buf))
drivers/mtd/mtdswap.c
563
unsigned int *block)
drivers/mtd/mtdswap.c
593
*block = (d->curr_write - d->eb_data) * d->pages_per_eblk +
drivers/mtd/mtdswap.c
597
d->revmap[*block] = page;
drivers/mtd/mtdswap.c
734
unsigned int i, block, eblk_base, newblock;
drivers/mtd/mtdswap.c
744
block = eblk_base + i;
drivers/mtd/mtdswap.c
745
if (d->revmap[block] == PAGE_UNDEF)
drivers/mtd/mtdswap.c
748
ret = mtdswap_move_block(d, block, &newblock);
drivers/mtd/nand/onenand/onenand_base.c
224
static int onenand_block_address(struct onenand_chip *this, int block)
drivers/mtd/nand/onenand/onenand_base.c
227
if (block & this->density_mask)
drivers/mtd/nand/onenand/onenand_base.c
228
return ONENAND_DDP_CHIP1 | (block ^ this->density_mask);
drivers/mtd/nand/onenand/onenand_base.c
230
return block;
drivers/mtd/nand/onenand/onenand_base.c
241
static int onenand_bufferram_address(struct onenand_chip *this, int block)
drivers/mtd/nand/onenand/onenand_base.c
244
if (block & this->density_mask)
drivers/mtd/nand/onenand/onenand_base.c
2454
int block;
drivers/mtd/nand/onenand/onenand_base.c
2457
block = onenand_block(this, ofs);
drivers/mtd/nand/onenand/onenand_base.c
2459
bbm->bbt[block >> 2] |= 0x01 << ((block & 0x03) << 1);
drivers/mtd/nand/onenand/onenand_base.c
2508
int start, end, block, value, status;
drivers/mtd/nand/onenand/onenand_base.c
2546
for (block = start; block < end + 1; block++) {
drivers/mtd/nand/onenand/onenand_base.c
2548
value = onenand_block_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
2551
value = onenand_bufferram_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
2554
this->write_word(block, this->base + ONENAND_REG_START_BLOCK_ADDRESS);
drivers/mtd/nand/onenand/onenand_base.c
2570
__func__, block, status);
drivers/mtd/nand/onenand/onenand_base.c
2620
unsigned int value, block, status;
drivers/mtd/nand/onenand/onenand_base.c
2624
for (block = 0; block < end; block++) {
drivers/mtd/nand/onenand/onenand_base.c
2626
value = onenand_block_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
2629
value = onenand_bufferram_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
2632
this->write_word(block, this->base + ONENAND_REG_START_BLOCK_ADDRESS);
drivers/mtd/nand/onenand/onenand_base.c
2638
__func__, block, status);
drivers/mtd/nand/onenand/onenand_base.c
2704
int value, block, page;
drivers/mtd/nand/onenand/onenand_base.c
2709
block = (int) (addr >> this->erase_shift);
drivers/mtd/nand/onenand/onenand_base.c
2714
block = (int) (addr >> this->erase_shift);
drivers/mtd/nand/onenand/onenand_base.c
2719
block &= ~1;
drivers/mtd/nand/onenand/onenand_base.c
2722
block++;
drivers/mtd/nand/onenand/onenand_base.c
2729
if (block != -1) {
drivers/mtd/nand/onenand/onenand_base.c
2731
value = onenand_block_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
2785
int block, value, status;
drivers/mtd/nand/onenand/onenand_base.c
2804
block = (int) (to >> this->erase_shift);
drivers/mtd/nand/onenand/onenand_base.c
2810
value = onenand_block_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
2819
value = onenand_bufferram_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
334
static loff_t flexonenand_addr(struct onenand_chip *this, int block)
drivers/mtd/nand/onenand/onenand_base.c
339
if (ONENAND_IS_DDP(this) && block >= this->density_mask) {
drivers/mtd/nand/onenand/onenand_base.c
340
block -= this->density_mask;
drivers/mtd/nand/onenand/onenand_base.c
346
ofs += (loff_t)block << (this->erase_shift - 1);
drivers/mtd/nand/onenand/onenand_base.c
347
if (block > (boundary + 1))
drivers/mtd/nand/onenand/onenand_base.c
348
ofs += (loff_t)(block - boundary - 1) << (this->erase_shift - 1);
drivers/mtd/nand/onenand/onenand_base.c
352
loff_t onenand_addr(struct onenand_chip *this, int block)
drivers/mtd/nand/onenand/onenand_base.c
3521
int block;
drivers/mtd/nand/onenand/onenand_base.c
3533
for (block = start; block <= end; block++) {
drivers/mtd/nand/onenand/onenand_base.c
3534
addr = flexonenand_addr(this, block);
drivers/mtd/nand/onenand/onenand_base.c
355
return (loff_t)block << this->erase_shift;
drivers/mtd/nand/onenand/onenand_base.c
3552
__func__, block);
drivers/mtd/nand/onenand/onenand_base.c
356
return flexonenand_addr(this, block);
drivers/mtd/nand/onenand/onenand_base.c
401
int value, block, page;
drivers/mtd/nand/onenand/onenand_base.c
409
block = -1;
drivers/mtd/nand/onenand/onenand_base.c
415
block = addr * this->density_mask;
drivers/mtd/nand/onenand/onenand_base.c
424
block = onenand_block(this, addr);
drivers/mtd/nand/onenand/onenand_base.c
430
block = addr * this->density_mask;
drivers/mtd/nand/onenand/onenand_base.c
435
block = onenand_block(this, addr);
drivers/mtd/nand/onenand/onenand_base.c
437
page = (int) (addr - onenand_addr(this, block))>>\
drivers/mtd/nand/onenand/onenand_base.c
443
block &= ~1;
drivers/mtd/nand/onenand/onenand_base.c
446
block++;
drivers/mtd/nand/onenand/onenand_base.c
456
value = onenand_bufferram_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
469
if (block != -1) {
drivers/mtd/nand/onenand/onenand_base.c
471
value = onenand_block_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
475
value = onenand_bufferram_address(this, block);
drivers/mtd/nand/onenand/onenand_base.c
876
int blockpage, block, page;
drivers/mtd/nand/onenand/onenand_base.c
879
block = (int) (addr >> this->erase_shift) & ~1;
drivers/mtd/nand/onenand/onenand_base.c
882
block++;
drivers/mtd/nand/onenand/onenand_base.c
884
blockpage = (block << 7) | page;
drivers/mtd/nand/onenand/onenand_base.c
923
int block = onenand_block(this, addr);
drivers/mtd/nand/onenand/onenand_base.c
924
int value = onenand_bufferram_address(this, block);
drivers/mtd/nand/onenand/onenand_bbt.c
148
int block;
drivers/mtd/nand/onenand/onenand_bbt.c
152
block = (int) (onenand_block(this, offs) << 1);
drivers/mtd/nand/onenand/onenand_bbt.c
153
res = (bbm->bbt[block >> 3] >> (block & 0x06)) & 0x03;
drivers/mtd/nand/onenand/onenand_bbt.c
156
(unsigned int) offs, block >> 1, res);
drivers/mtd/nand/onenand/onenand_samsung.c
731
unsigned int block, end;
drivers/mtd/nand/onenand/onenand_samsung.c
735
for (block = 0; block < end; block++) {
drivers/mtd/nand/onenand/onenand_samsung.c
736
unsigned int mem_addr = onenand->mem_addr(block, 0, 0);
drivers/mtd/nand/onenand/onenand_samsung.c
740
dev_err(dev, "block %d is write-protected!\n", block);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2114
unsigned int block;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2139
for (block = 0; block < search_area_size_in_blocks; block++) {
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2141
dev_dbg(dev, "\tErasing block 0x%x\n", block);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2142
status = nand_erase_op(chip, block);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2176
unsigned int block;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2205
for (block = 0; block < block_count; block++) {
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2210
chipnr = block >> (chip->chip_shift - chip->phys_erase_shift);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2211
page = block << (chip->phys_erase_shift - chip->page_shift);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2212
byte = block << chip->phys_erase_shift;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
2229
dev_dbg(dev, "Transcribing mark in block %u\n", block);
drivers/mtd/nand/raw/nand_bbt.c
1133
int i, j, chips, block, nrblocks, update;
drivers/mtd/nand/raw/nand_bbt.c
1150
block = td->pages[i] >> (this->bbt_erase_shift - this->page_shift);
drivers/mtd/nand/raw/nand_bbt.c
1151
oldval = bbt_get_entry(this, block);
drivers/mtd/nand/raw/nand_bbt.c
1152
bbt_mark_entry(this, block, BBT_BLOCK_RESERVED);
drivers/mtd/nand/raw/nand_bbt.c
1155
nand_update_bbt(this, (loff_t)block <<
drivers/mtd/nand/raw/nand_bbt.c
1161
block = ((i + 1) * nrblocks) - td->maxblocks;
drivers/mtd/nand/raw/nand_bbt.c
1163
block = i * nrblocks;
drivers/mtd/nand/raw/nand_bbt.c
1165
oldval = bbt_get_entry(this, block);
drivers/mtd/nand/raw/nand_bbt.c
1166
bbt_mark_entry(this, block, BBT_BLOCK_RESERVED);
drivers/mtd/nand/raw/nand_bbt.c
1169
block++;
drivers/mtd/nand/raw/nand_bbt.c
1177
nand_update_bbt(this, (loff_t)(block - 1) <<
drivers/mtd/nand/raw/nand_bbt.c
1435
int block;
drivers/mtd/nand/raw/nand_bbt.c
1437
block = (int)(offs >> this->bbt_erase_shift);
drivers/mtd/nand/raw/nand_bbt.c
1438
return bbt_get_entry(this, block) == BBT_BLOCK_RESERVED;
drivers/mtd/nand/raw/nand_bbt.c
1449
int block, res;
drivers/mtd/nand/raw/nand_bbt.c
1451
block = (int)(offs >> this->bbt_erase_shift);
drivers/mtd/nand/raw/nand_bbt.c
1452
res = bbt_get_entry(this, block);
drivers/mtd/nand/raw/nand_bbt.c
1455
(unsigned int)offs, block, res);
drivers/mtd/nand/raw/nand_bbt.c
1478
int block, ret = 0;
drivers/mtd/nand/raw/nand_bbt.c
1480
block = (int)(offs >> this->bbt_erase_shift);
drivers/mtd/nand/raw/nand_bbt.c
1483
bbt_mark_entry(this, block, BBT_BLOCK_WORN);
drivers/mtd/nand/raw/nand_bbt.c
558
int startblock, block, dir;
drivers/mtd/nand/raw/nand_bbt.c
586
for (block = 0; block < td->maxblocks; block++) {
drivers/mtd/nand/raw/nand_bbt.c
588
int actblock = startblock + dir * block;
drivers/mtd/nand/raw/nand_bbt.c
685
int block = startblock + dir * i;
drivers/mtd/nand/raw/nand_bbt.c
688
switch (bbt_get_entry(this, block)) {
drivers/mtd/nand/raw/nand_bbt.c
694
page = block << (this->bbt_erase_shift - this->page_shift);
drivers/mtd/nand/raw/nand_bbt.c
698
return block;
drivers/mtd/nand/raw/nand_bbt.c
718
int chip, int block)
drivers/mtd/nand/raw/nand_bbt.c
723
bbt_mark_entry(this, block, BBT_BLOCK_WORN);
drivers/mtd/nand/raw/nand_bbt.c
725
to = (loff_t)block << this->bbt_erase_shift;
drivers/mtd/nand/raw/nand_bbt.c
729
res, block);
drivers/mtd/nand/raw/nand_bbt.c
76
static inline uint8_t bbt_get_entry(struct nand_chip *chip, int block)
drivers/mtd/nand/raw/nand_bbt.c
78
uint8_t entry = chip->bbt[block >> BBT_ENTRY_SHIFT];
drivers/mtd/nand/raw/nand_bbt.c
784
int block;
drivers/mtd/nand/raw/nand_bbt.c
786
block = get_bbt_block(this, td, md, chip);
drivers/mtd/nand/raw/nand_bbt.c
787
if (block < 0) {
drivers/mtd/nand/raw/nand_bbt.c
789
res = block;
drivers/mtd/nand/raw/nand_bbt.c
79
entry >>= (block & BBT_ENTRY_MASK) * 2;
drivers/mtd/nand/raw/nand_bbt.c
797
page = block << (this->bbt_erase_shift - this->page_shift);
drivers/mtd/nand/raw/nand_bbt.c
83
static inline void bbt_mark_entry(struct nand_chip *chip, int block,
drivers/mtd/nand/raw/nand_bbt.c
86
uint8_t msk = (mark & BBT_ENTRY_MASK) << ((block & BBT_ENTRY_MASK) * 2);
drivers/mtd/nand/raw/nand_bbt.c
87
chip->bbt[block >> BBT_ENTRY_SHIFT] |= msk;
drivers/mtd/nand/raw/nand_bbt.c
895
mark_bbt_block_bad(this, td, chip, block);
drivers/mtd/nand/raw/nand_bbt.c
905
mark_bbt_block_bad(this, td, chip, block);
drivers/mtd/nftlcore.c
257
int block;
drivers/mtd/nftlcore.c
284
for (block = 0; block < nftl->EraseSize / 512; block ++) {
drivers/mtd/nftlcore.c
286
(block * 512), 16 , &retlen,
drivers/mtd/nftlcore.c
288
if (block == 2) {
drivers/mtd/nftlcore.c
301
BlockLastState[block] = status;
drivers/mtd/nftlcore.c
305
BlockFreeFound[block] = 1;
drivers/mtd/nftlcore.c
309
if (!BlockFreeFound[block])
drivers/mtd/nftlcore.c
310
BlockMap[block] = thisEUN;
drivers/mtd/nftlcore.c
315
thisVUC, block);
drivers/mtd/nftlcore.c
318
if (!BlockFreeFound[block])
drivers/mtd/nftlcore.c
319
BlockMap[block] = BLOCK_NIL;
drivers/mtd/nftlcore.c
324
thisVUC, block);
drivers/mtd/nftlcore.c
331
block, thisEUN, status);
drivers/mtd/nftlcore.c
352
for (block = 0; block < nftl->EraseSize / 512 ; block++) {
drivers/mtd/nftlcore.c
353
if (BlockLastState[block] != SECTOR_FREE &&
drivers/mtd/nftlcore.c
354
BlockMap[block] != BLOCK_NIL &&
drivers/mtd/nftlcore.c
355
BlockMap[block] != targetEUN) {
drivers/mtd/nftlcore.c
359
thisVUC, block, BlockLastState[block],
drivers/mtd/nftlcore.c
360
BlockMap[block],
drivers/mtd/nftlcore.c
361
BlockMap[block]== targetEUN ? "==" : "!=",
drivers/mtd/nftlcore.c
410
for (block = 0; block < nftl->EraseSize / 512 ; block++) {
drivers/mtd/nftlcore.c
412
if (BlockMap[block] == targetEUN ||
drivers/mtd/nftlcore.c
413
(pendingblock == (thisVUC * (nftl->EraseSize / 512) + block))) {
drivers/mtd/nftlcore.c
419
if (BlockMap[block] == BLOCK_NIL)
drivers/mtd/nftlcore.c
422
NFTL_move_block(mtd, (nftl->EraseSize * BlockMap[block]) + (block * 512),
drivers/mtd/nftlcore.c
423
(nftl->EraseSize * targetEUN) + (block * 512));
drivers/mtd/nftlcore.c
521
static inline u16 NFTL_findwriteunit(struct NFTLrecord *nftl, unsigned block)
drivers/mtd/nftlcore.c
524
u16 thisVUC = block / (nftl->EraseSize / 512);
drivers/mtd/nftlcore.c
527
unsigned long blockofs = (block * 512) & (nftl->EraseSize -1);
drivers/mtd/nftlcore.c
555
block , writeEUN, le16_to_cpu(bci.Status));
drivers/mtd/nftlcore.c
672
static int nftl_writeblock(struct mtd_blktrans_dev *mbd, unsigned long block,
drivers/mtd/nftlcore.c
677
unsigned long blockofs = (block * 512) & (nftl->EraseSize - 1);
drivers/mtd/nftlcore.c
681
writeEUN = NFTL_findwriteunit(nftl, block);
drivers/mtd/nftlcore.c
699
static int nftl_readblock(struct mtd_blktrans_dev *mbd, unsigned long block,
drivers/mtd/nftlcore.c
705
u16 thisEUN = nftl->EUNtable[block / (nftl->EraseSize / 512)];
drivers/mtd/nftlcore.c
706
unsigned long blockofs = (block * 512) & (nftl->EraseSize - 1);
drivers/mtd/nftlcore.c
737
block, thisEUN, status);
drivers/mtd/nftlcore.c
743
block / (nftl->EraseSize / 512));
drivers/mtd/nftlmount.c
100
ret = mtd->read(mtd, block * nftl->EraseSize, SECTORSIZE,
drivers/mtd/nftlmount.c
104
block * nftl->EraseSize, nftl->mbd.mtd->index, ret);
drivers/mtd/nftlmount.c
111
block * nftl->EraseSize, nftl->mbd.mtd->index);
drivers/mtd/nftlmount.c
123
nftl->MediaUnit * nftl->EraseSize, block * nftl->EraseSize);
drivers/mtd/nftlmount.c
132
nftl->SpareMediaUnit = block;
drivers/mtd/nftlmount.c
135
nftl->ReplUnitTable[block] = BLOCK_RESERVED;
drivers/mtd/nftlmount.c
211
nftl->ReplUnitTable[block] = BLOCK_RESERVED;
drivers/mtd/nftlmount.c
216
The new DiskOnChip driver already scanned the bad block table. Just query it.
drivers/mtd/nftlmount.c
220
block * nftl->EraseSize + i +
drivers/mtd/nftlmount.c
240
nftl->MediaUnit = block;
drivers/mtd/nftlmount.c
28
unsigned int block, boot_record_count = 0;
drivers/mtd/nftlmount.c
302
int NFTL_formatblock(struct NFTLrecord *nftl, int block)
drivers/mtd/nftlmount.c
311
if (nftl_read_oob(mtd, block * nftl->EraseSize + SECTORSIZE + 8,
drivers/mtd/nftlmount.c
326
instr->addr = block * nftl->EraseSize;
drivers/mtd/nftlmount.c
329
printk("Error while formatting block %d\n", block);
drivers/mtd/nftlmount.c
349
if (nftl_write_oob(mtd, block * nftl->EraseSize + SECTORSIZE +
drivers/mtd/nftlmount.c
372
unsigned int block, i, status;
drivers/mtd/nftlmount.c
378
block = first_block;
drivers/mtd/nftlmount.c
382
block * nftl->EraseSize + i * SECTORSIZE,
drivers/mtd/nftlmount.c
393
check_free_sectors(nftl, block * nftl->EraseSize + i * SECTORSIZE,
drivers/mtd/nftlmount.c
397
i, block);
drivers/mtd/nftlmount.c
402
nftl_write_oob(mtd, block *
drivers/mtd/nftlmount.c
414
block = nftl->ReplUnitTable[block];
drivers/mtd/nftlmount.c
415
if (!(block == BLOCK_NIL || block < nftl->nb_blocks))
drivers/mtd/nftlmount.c
416
printk("incorrect ReplUnitTable[] : %d\n", block);
drivers/mtd/nftlmount.c
417
if (block == BLOCK_NIL || block >= nftl->nb_blocks)
drivers/mtd/nftlmount.c
425
unsigned int length = 0, block = first_block;
drivers/mtd/nftlmount.c
436
block = nftl->ReplUnitTable[block];
drivers/mtd/nftlmount.c
437
if (!(block == BLOCK_NIL || block < nftl->nb_blocks))
drivers/mtd/nftlmount.c
438
printk("incorrect ReplUnitTable[] : %d\n", block);
drivers/mtd/nftlmount.c
439
if (block == BLOCK_NIL || block >= nftl->nb_blocks)
drivers/mtd/nftlmount.c
457
unsigned int block = first_block, block1;
drivers/mtd/nftlmount.c
462
block1 = nftl->ReplUnitTable[block];
drivers/mtd/nftlmount.c
464
printk("Formatting block %d\n", block);
drivers/mtd/nftlmount.c
465
if (NFTL_formatblock(nftl, block) < 0) {
drivers/mtd/nftlmount.c
467
nftl->ReplUnitTable[block] = BLOCK_RESERVED;
drivers/mtd/nftlmount.c
469
nftl->ReplUnitTable[block] = BLOCK_FREE;
drivers/mtd/nftlmount.c
473
block = block1;
drivers/mtd/nftlmount.c
475
if (!(block == BLOCK_NIL || block < nftl->nb_blocks))
drivers/mtd/nftlmount.c
476
printk("incorrect ReplUnitTable[] : %d\n", block);
drivers/mtd/nftlmount.c
477
if (block == BLOCK_NIL || block >= nftl->nb_blocks)
drivers/mtd/nftlmount.c
48
for (block = 0; block < nftl->nb_blocks; block++) {
drivers/mtd/nftlmount.c
488
static int check_and_mark_free_block(struct NFTLrecord *nftl, int block)
drivers/mtd/nftlmount.c
496
if (nftl_read_oob(mtd, block * nftl->EraseSize + SECTORSIZE + 8, 8,
drivers/mtd/nftlmount.c
504
if (check_free_sectors (nftl, block * nftl->EraseSize, nftl->EraseSize, 1) != 0)
drivers/mtd/nftlmount.c
512
block * nftl->EraseSize + SECTORSIZE + 8, 8,
drivers/mtd/nftlmount.c
520
if (check_free_sectors (nftl, block * nftl->EraseSize + i,
drivers/mtd/nftlmount.c
524
if (nftl_read_oob(mtd, block * nftl->EraseSize + i,
drivers/mtd/nftlmount.c
53
ret = mtd_read(mtd, block * nftl->EraseSize, SECTORSIZE,
drivers/mtd/nftlmount.c
549
static int get_fold_mark(struct NFTLrecord *nftl, unsigned int block)
drivers/mtd/nftlmount.c
555
if (nftl_read_oob(mtd, block * nftl->EraseSize + 2 * SECTORSIZE + 8,
drivers/mtd/nftlmount.c
566
unsigned int block, first_block, is_first_block;
drivers/mtd/nftlmount.c
589
block = first_block;
drivers/mtd/nftlmount.c
596
block * s->EraseSize + 8, 8,
drivers/mtd/nftlmount.c
599
block * s->EraseSize +
drivers/mtd/nftlmount.c
602
s->ReplUnitTable[block] = BLOCK_NIL;
drivers/mtd/nftlmount.c
618
if (check_and_mark_free_block(s, block) < 0) {
drivers/mtd/nftlmount.c
62
block * nftl->EraseSize, nftl->mbd.mtd->index, ret);
drivers/mtd/nftlmount.c
620
printk("Formatting block %d\n", block);
drivers/mtd/nftlmount.c
621
if (NFTL_formatblock(s, block) < 0) {
drivers/mtd/nftlmount.c
623
s->ReplUnitTable[block] = BLOCK_RESERVED;
drivers/mtd/nftlmount.c
625
s->ReplUnitTable[block] = BLOCK_FREE;
drivers/mtd/nftlmount.c
629
s->ReplUnitTable[block] = BLOCK_FREE;
drivers/mtd/nftlmount.c
637
block, first_block);
drivers/mtd/nftlmount.c
638
s->ReplUnitTable[block] = BLOCK_NIL;
drivers/mtd/nftlmount.c
655
block, logical_block, first_logical_block);
drivers/mtd/nftlmount.c
664
if (get_fold_mark(s, block) != FOLD_MARK_IN_PROGRESS ||
drivers/mtd/nftlmount.c
667
block);
drivers/mtd/nftlmount.c
673
block);
drivers/mtd/nftlmount.c
680
s->ReplUnitTable[block] = BLOCK_NIL;
drivers/mtd/nftlmount.c
684
block, rep_block);
drivers/mtd/nftlmount.c
686
s->ReplUnitTable[block] = BLOCK_NIL;
drivers/mtd/nftlmount.c
700
s->ReplUnitTable[block] = rep_block;
drivers/mtd/nftlmount.c
704
block, rep_block);
drivers/mtd/nftlmount.c
707
s->ReplUnitTable[block] = BLOCK_NIL;
drivers/mtd/nftlmount.c
712
s->ReplUnitTable[block] = rep_block;
drivers/mtd/nftlmount.c
713
block = rep_block;
drivers/mtd/nftlmount.c
73
block * nftl->EraseSize, nftl->mbd.mtd->index);
drivers/mtd/nftlmount.c
767
for (block = 0; block < s->nb_blocks; block++) {
drivers/mtd/nftlmount.c
768
if (s->ReplUnitTable[block] == BLOCK_NOTEXPLORED) {
drivers/mtd/nftlmount.c
769
printk("Unreferenced block %d, formatting it\n", block);
drivers/mtd/nftlmount.c
770
if (NFTL_formatblock(s, block) < 0)
drivers/mtd/nftlmount.c
771
s->ReplUnitTable[block] = BLOCK_RESERVED;
drivers/mtd/nftlmount.c
773
s->ReplUnitTable[block] = BLOCK_FREE;
drivers/mtd/nftlmount.c
775
if (s->ReplUnitTable[block] == BLOCK_FREE) {
drivers/mtd/nftlmount.c
777
s->LastFreeEUN = block;
drivers/mtd/nftlmount.c
79
ret = nftl_read_oob(mtd, block * nftl->EraseSize +
drivers/mtd/nftlmount.c
84
block * nftl->EraseSize, nftl->mbd.mtd->index, ret);
drivers/mtd/nftlmount.c
94
block * nftl->EraseSize, nftl->mbd.mtd->index,
drivers/mtd/rfd_ftl.c
101
block->state = BLOCK_UNUSED;
drivers/mtd/rfd_ftl.c
105
block->state = BLOCK_OK;
drivers/mtd/rfd_ftl.c
116
block->free_sectors++;
drivers/mtd/rfd_ftl.c
139
part->sector_map[entry] = block->offset +
drivers/mtd/rfd_ftl.c
142
block->used_sectors++;
drivers/mtd/rfd_ftl.c
145
if (block->free_sectors == part->data_sectors_per_block)
drivers/mtd/rfd_ftl.c
188
part->blocks = kzalloc_objs(struct block, part->total_blocks);
drivers/mtd/rfd_ftl.c
267
static int erase_block(struct partition *part, int block)
drivers/mtd/rfd_ftl.c
276
erase->addr = part->blocks[block].offset;
drivers/mtd/rfd_ftl.c
279
part->blocks[block].state = BLOCK_ERASING;
drivers/mtd/rfd_ftl.c
280
part->blocks[block].free_sectors = 0;
drivers/mtd/rfd_ftl.c
287
part->blocks[block].state = BLOCK_FAILED;
drivers/mtd/rfd_ftl.c
288
part->blocks[block].free_sectors = 0;
drivers/mtd/rfd_ftl.c
289
part->blocks[block].used_sectors = 0;
drivers/mtd/rfd_ftl.c
294
part->blocks[block].state = BLOCK_ERASED;
drivers/mtd/rfd_ftl.c
295
part->blocks[block].free_sectors = part->data_sectors_per_block;
drivers/mtd/rfd_ftl.c
296
part->blocks[block].used_sectors = 0;
drivers/mtd/rfd_ftl.c
297
part->blocks[block].erases++;
drivers/mtd/rfd_ftl.c
299
rc = mtd_write(part->mbd.mtd, part->blocks[block].offset,
drivers/mtd/rfd_ftl.c
306
part->mbd.mtd->name, part->blocks[block].offset);
drivers/mtd/rfd_ftl.c
307
part->blocks[block].state = BLOCK_FAILED;
drivers/mtd/rfd_ftl.c
309
part->blocks[block].state = BLOCK_OK;
drivers/mtd/rfd_ftl.c
408
int block, best_block, score, old_sector_block;
drivers/mtd/rfd_ftl.c
421
for (block=0; block<part->total_blocks; block++) {
drivers/mtd/rfd_ftl.c
424
if (block == part->reserved_block)
drivers/mtd/rfd_ftl.c
432
if (part->blocks[block].free_sectors)
drivers/mtd/rfd_ftl.c
435
this_score = part->blocks[block].used_sectors;
drivers/mtd/rfd_ftl.c
437
if (block == old_sector_block)
drivers/mtd/rfd_ftl.c
441
if (part->blocks[block].used_sectors ==
drivers/mtd/rfd_ftl.c
446
this_score += part->blocks[block].erases;
drivers/mtd/rfd_ftl.c
449
best_block = block;
drivers/mtd/rfd_ftl.c
480
int block, stop;
drivers/mtd/rfd_ftl.c
482
block = part->current_block == -1 ?
drivers/mtd/rfd_ftl.c
484
stop = block;
drivers/mtd/rfd_ftl.c
487
if (part->blocks[block].free_sectors &&
drivers/mtd/rfd_ftl.c
488
block != part->reserved_block)
drivers/mtd/rfd_ftl.c
489
return block;
drivers/mtd/rfd_ftl.c
491
if (part->blocks[block].state == BLOCK_UNUSED)
drivers/mtd/rfd_ftl.c
492
erase_block(part, block);
drivers/mtd/rfd_ftl.c
494
if (++block >= part->total_blocks)
drivers/mtd/rfd_ftl.c
495
block = 0;
drivers/mtd/rfd_ftl.c
497
} while (block != stop);
drivers/mtd/rfd_ftl.c
504
int rc, block;
drivers/mtd/rfd_ftl.c
507
block = find_free_block(part);
drivers/mtd/rfd_ftl.c
509
if (block == -1) {
drivers/mtd/rfd_ftl.c
515
block = find_free_block(part);
drivers/mtd/rfd_ftl.c
518
if (block == -1) {
drivers/mtd/rfd_ftl.c
524
rc = mtd_read(part->mbd.mtd, part->blocks[block].offset,
drivers/mtd/rfd_ftl.c
534
part->blocks[block].offset);
drivers/mtd/rfd_ftl.c
538
part->current_block = block;
drivers/mtd/rfd_ftl.c
546
int block, offset, rc;
drivers/mtd/rfd_ftl.c
551
block = old_addr / part->block_size;
drivers/mtd/rfd_ftl.c
555
addr = part->blocks[block].offset +
drivers/mtd/rfd_ftl.c
568
if (block == part->current_block)
drivers/mtd/rfd_ftl.c
571
part->blocks[block].used_sectors--;
drivers/mtd/rfd_ftl.c
573
if (!part->blocks[block].used_sectors &&
drivers/mtd/rfd_ftl.c
574
!part->blocks[block].free_sectors)
drivers/mtd/rfd_ftl.c
575
rc = erase_block(part, block);
drivers/mtd/rfd_ftl.c
581
static int find_free_sector(const struct partition *part, const struct block *block)
drivers/mtd/rfd_ftl.c
585
i = stop = part->data_sectors_per_block - block->free_sectors;
drivers/mtd/rfd_ftl.c
603
struct block *block;
drivers/mtd/rfd_ftl.c
618
block = &part->blocks[part->current_block];
drivers/mtd/rfd_ftl.c
620
i = find_free_sector(part, block);
drivers/mtd/rfd_ftl.c
628
block->offset;
drivers/mtd/rfd_ftl.c
647
addr = block->offset + (HEADER_MAP_OFFSET + i) * sizeof(u16);
drivers/mtd/rfd_ftl.c
659
block->used_sectors++;
drivers/mtd/rfd_ftl.c
660
block->free_sectors--;
drivers/mtd/rfd_ftl.c
88
struct block *blocks;
drivers/mtd/rfd_ftl.c
95
struct block *block = &part->blocks[block_no];
drivers/mtd/rfd_ftl.c
98
block->offset = part->block_size * block_no;
drivers/mtd/sm_ftl.c
1017
int zone_num, block, boffset;
drivers/mtd/sm_ftl.c
1019
sm_break_offset(ftl, sect_no << 9, &zone_num, &block, &boffset);
drivers/mtd/sm_ftl.c
1030
if (ftl->cache_zone == zone_num && ftl->cache_block == block) {
drivers/mtd/sm_ftl.c
1037
block = zone->lba_to_phys_table[block];
drivers/mtd/sm_ftl.c
1039
if (block == -1) {
drivers/mtd/sm_ftl.c
1044
if (sm_read_sector(ftl, zone_num, block, boffset, buf, NULL)) {
drivers/mtd/sm_ftl.c
1062
int error = 0, zone_num, block, boffset;
drivers/mtd/sm_ftl.c
1065
sm_break_offset(ftl, sec_no << 9, &zone_num, &block, &boffset);
drivers/mtd/sm_ftl.c
1078
if (ftl->cache_block != block || ftl->cache_zone != zone_num) {
drivers/mtd/sm_ftl.c
1084
ftl->cache_block = block;
drivers/mtd/sm_ftl.c
190
static loff_t sm_mkoffset(struct sm_ftl *ftl, int zone, int block, int boffset)
drivers/mtd/sm_ftl.c
194
WARN_ON(block >= ftl->zone_size);
drivers/mtd/sm_ftl.c
197
if (block == -1)
drivers/mtd/sm_ftl.c
200
return (zone * SM_MAX_ZONE_SIZE + block) * ftl->block_size + boffset;
drivers/mtd/sm_ftl.c
205
int *zone, int *block, int *boffset)
drivers/mtd/sm_ftl.c
209
*block = do_div(offset, ftl->max_lba);
drivers/mtd/sm_ftl.c
236
int zone, int block, int boffset,
drivers/mtd/sm_ftl.c
246
if (block == -1) {
drivers/mtd/sm_ftl.c
268
if (zone == 0 && block == ftl->cis_block && boffset ==
drivers/mtd/sm_ftl.c
280
ret = mtd_read_oob(mtd, sm_mkoffset(ftl, zone, block, boffset), &ops);
drivers/mtd/sm_ftl.c
285
block, zone, ret);
drivers/mtd/sm_ftl.c
303
" as bad" , block, zone);
drivers/mtd/sm_ftl.c
312
block, zone);
drivers/mtd/sm_ftl.c
321
int zone, int block, int boffset,
drivers/mtd/sm_ftl.c
330
if (zone == 0 && (block == ftl->cis_block || block == 0)) {
drivers/mtd/sm_ftl.c
345
ret = mtd_write_oob(mtd, sm_mkoffset(ftl, zone, block, boffset), &ops);
drivers/mtd/sm_ftl.c
351
block, zone, ret);
drivers/mtd/sm_ftl.c
368
int zone, int block, int lba,
drivers/mtd/sm_ftl.c
406
if (!sm_write_sector(ftl, zone, block, boffset,
drivers/mtd/sm_ftl.c
419
if (sm_erase_block(ftl, zone, block, 0))
drivers/mtd/sm_ftl.c
425
sm_mark_block_bad(ftl, zone, block);
drivers/mtd/sm_ftl.c
434
static void sm_mark_block_bad(struct sm_ftl *ftl, int zone, int block)
drivers/mtd/sm_ftl.c
448
sm_printk("marking block %d of zone %d as bad", block, zone);
drivers/mtd/sm_ftl.c
455
sm_write_sector(ftl, zone, block, boffset, NULL, &oob);
drivers/mtd/sm_ftl.c
462
static int sm_erase_block(struct sm_ftl *ftl, int zone_num, uint16_t block,
drivers/mtd/sm_ftl.c
469
erase.addr = sm_mkoffset(ftl, zone_num, block, 0);
drivers/mtd/sm_ftl.c
477
if (zone_num == 0 && (block == ftl->cis_block || block == 0)) {
drivers/mtd/sm_ftl.c
484
block, zone_num);
drivers/mtd/sm_ftl.c
490
(const unsigned char *)&block, sizeof(block));
drivers/mtd/sm_ftl.c
494
sm_mark_block_bad(ftl, zone_num, block);
drivers/mtd/sm_ftl.c
499
static int sm_check_block(struct sm_ftl *ftl, int zone, int block)
drivers/mtd/sm_ftl.c
516
if (sm_read_sector(ftl, zone, block, boffset, NULL, &oob))
drivers/mtd/sm_ftl.c
531
sm_erase_block(ftl, zone, block, 1);
drivers/mtd/sm_ftl.c
681
int block, boffset;
drivers/mtd/sm_ftl.c
686
for (block = 0 ; block < ftl->zone_size - ftl->max_lba ; block++) {
drivers/mtd/sm_ftl.c
688
if (sm_read_sector(ftl, 0, block, 0, NULL, &oob))
drivers/mtd/sm_ftl.c
704
if (sm_read_sector(ftl, 0, block, boffset, NULL, &oob))
drivers/mtd/sm_ftl.c
715
ftl->cis_block = block;
drivers/mtd/sm_ftl.c
728
block * ftl->block_size +
drivers/mtd/sm_ftl.c
754
uint16_t block;
drivers/mtd/sm_ftl.c
776
for (block = 0 ; block < ftl->zone_size ; block++) {
drivers/mtd/sm_ftl.c
779
if (zone_num == 0 && block <= ftl->cis_block)
drivers/mtd/sm_ftl.c
783
if (sm_read_sector(ftl, zone_num, block, 0, NULL, &oob)) {
drivers/mtd/sm_ftl.c
794
(unsigned char *)&block, 2);
drivers/mtd/sm_ftl.c
804
dbg("PH %04d <-> <marked bad>", block);
drivers/mtd/sm_ftl.c
816
dbg("PH %04d <-> LBA %04d(bad)", block, lba);
drivers/mtd/sm_ftl.c
825
dbg_verbose("PH %04d <-> LBA %04d", block, lba);
drivers/mtd/sm_ftl.c
826
zone->lba_to_phys_table[lba] = block;
drivers/mtd/sm_ftl.c
832
lba, zone->lba_to_phys_table[lba], block, zone_num);
drivers/mtd/sm_ftl.c
835
if (sm_check_block(ftl, zone_num, block))
drivers/mtd/sm_ftl.c
841
zone->lba_to_phys_table[lba] = block;
drivers/mtd/sm_ftl.c
850
sm_erase_block(ftl, zone_num, block, 1);
drivers/mtd/sm_ftl.c
870
(unsigned char *)&block, 2);
drivers/mtd/sm_ftl.c
872
kfifo_in(&zone->free_sectors, (const unsigned char *)&block, 2);
drivers/mtd/sm_ftl.h
83
static int sm_erase_block(struct sm_ftl *ftl, int zone_num, uint16_t block,
drivers/mtd/sm_ftl.h
85
static void sm_mark_block_bad(struct sm_ftl *ftl, int zone_num, int block);
drivers/mtd/ubi/block.c
158
module_param_cb(block, &ubiblock_param_ops, NULL, 0);
drivers/mtd/ubi/block.c
159
MODULE_PARM_DESC(block, "Attach block devices to UBI volumes. Parameter format: block=<path|dev,num|dev,name>.\n"
drivers/net/can/kvaser_pciefd/kvaser_pciefd_core.c
264
#define KVASER_PCIEFD_GET_BLOCK_ADDR(pcie, block) \
drivers/net/can/kvaser_pciefd/kvaser_pciefd_core.c
265
((pcie)->reg_base + (pcie)->driver_data->address_offset->block)
drivers/net/dsa/b53/b53_serdes.c
25
static void b53_serdes_write_blk(struct b53_device *dev, u8 offset, u16 block,
drivers/net/dsa/b53/b53_serdes.c
28
b53_write16(dev, B53_SERDES_PAGE, B53_SERDES_BLKADDR, block);
drivers/net/dsa/b53/b53_serdes.c
32
static u16 b53_serdes_read_blk(struct b53_device *dev, u8 offset, u16 block)
drivers/net/dsa/b53/b53_serdes.c
36
b53_write16(dev, B53_SERDES_PAGE, B53_SERDES_BLKADDR, block);
drivers/net/dsa/b53/b53_serdes.c
55
u8 offset, u16 block, u16 value)
drivers/net/dsa/b53/b53_serdes.c
58
b53_serdes_write_blk(dev, offset, block, value);
drivers/net/dsa/b53/b53_serdes.c
62
u8 offset, u16 block)
drivers/net/dsa/b53/b53_serdes.c
65
return b53_serdes_read_blk(dev, offset, block);
drivers/net/dsa/mv88e6xxx/port.h
605
int mv88e6xxx_port_hidden_write(struct mv88e6xxx_chip *chip, int block,
drivers/net/dsa/mv88e6xxx/port.h
608
int mv88e6xxx_port_hidden_read(struct mv88e6xxx_chip *chip, int block, int port,
drivers/net/dsa/mv88e6xxx/port_hidden.c
18
int mv88e6xxx_port_hidden_write(struct mv88e6xxx_chip *chip, int block,
drivers/net/dsa/mv88e6xxx/port_hidden.c
31
block << MV88E6XXX_PORT_RESERVED_1A_BLOCK_SHIFT |
drivers/net/dsa/mv88e6xxx/port_hidden.c
48
int mv88e6xxx_port_hidden_read(struct mv88e6xxx_chip *chip, int block, int port,
drivers/net/dsa/mv88e6xxx/port_hidden.c
56
block << MV88E6XXX_PORT_RESERVED_1A_BLOCK_SHIFT |
drivers/net/dsa/ocelot/felix.c
157
block_vcap_es0 = &ocelot->block[VCAP_ES0];
drivers/net/dsa/ocelot/felix.c
247
block_vcap_is1 = &ocelot->block[VCAP_IS1];
drivers/net/dsa/ocelot/felix.c
248
block_vcap_is2 = &ocelot->block[VCAP_IS2];
drivers/net/dsa/ocelot/felix.c
342
block_vcap_es0 = &ocelot->block[VCAP_ES0];
drivers/net/dsa/ocelot/felix.c
398
block_vcap_is2 = &ocelot->block[VCAP_IS2];
drivers/net/dsa/rzn1_a5psw.c
329
static void a5psw_port_rx_block_set(struct a5psw *a5psw, int port, bool block)
drivers/net/dsa/rzn1_a5psw.c
332
u32 reg = block ? mask : 0;
drivers/net/dsa/vitesse-vsc73xx-core.c
455
int vsc73xx_is_addr_valid(u8 block, u8 subblock)
drivers/net/dsa/vitesse-vsc73xx-core.c
457
switch (block) {
drivers/net/dsa/vitesse-vsc73xx-core.c
494
static int vsc73xx_read(struct vsc73xx *vsc, u8 block, u8 subblock, u8 reg,
drivers/net/dsa/vitesse-vsc73xx-core.c
497
return vsc->ops->read(vsc, block, subblock, reg, val);
drivers/net/dsa/vitesse-vsc73xx-core.c
500
static int vsc73xx_write(struct vsc73xx *vsc, u8 block, u8 subblock, u8 reg,
drivers/net/dsa/vitesse-vsc73xx-core.c
503
return vsc->ops->write(vsc, block, subblock, reg, val);
drivers/net/dsa/vitesse-vsc73xx-core.c
506
static int vsc73xx_update_bits(struct vsc73xx *vsc, u8 block, u8 subblock,
drivers/net/dsa/vitesse-vsc73xx-core.c
513
ret = vsc73xx_read(vsc, block, subblock, reg, &orig);
drivers/net/dsa/vitesse-vsc73xx-core.c
518
return vsc73xx_write(vsc, block, subblock, reg, tmp);
drivers/net/dsa/vitesse-vsc73xx-platform.c
42
static u32 vsc73xx_make_addr(u8 block, u8 subblock, u8 reg)
drivers/net/dsa/vitesse-vsc73xx-platform.c
46
ret = (block & VSC73XX_CMD_PLATFORM_BLOCK_MASK)
drivers/net/dsa/vitesse-vsc73xx-platform.c
55
static int vsc73xx_platform_read(struct vsc73xx *vsc, u8 block, u8 subblock,
drivers/net/dsa/vitesse-vsc73xx-platform.c
61
if (!vsc73xx_is_addr_valid(block, subblock))
drivers/net/dsa/vitesse-vsc73xx-platform.c
64
offset = vsc73xx_make_addr(block, subblock, reg);
drivers/net/dsa/vitesse-vsc73xx-platform.c
73
static int vsc73xx_platform_write(struct vsc73xx *vsc, u8 block, u8 subblock,
drivers/net/dsa/vitesse-vsc73xx-platform.c
79
if (!vsc73xx_is_addr_valid(block, subblock))
drivers/net/dsa/vitesse-vsc73xx-platform.c
82
offset = vsc73xx_make_addr(block, subblock, reg);
drivers/net/dsa/vitesse-vsc73xx-spi.c
104
if (!vsc73xx_is_addr_valid(block, subblock))
drivers/net/dsa/vitesse-vsc73xx-spi.c
119
cmd[0] = vsc73xx_make_addr(VSC73XX_CMD_SPI_MODE_WRITE, block, subblock);
drivers/net/dsa/vitesse-vsc73xx-spi.c
40
static u8 vsc73xx_make_addr(u8 mode, u8 block, u8 subblock)
drivers/net/dsa/vitesse-vsc73xx-spi.c
45
(block & VSC73XX_CMD_SPI_BLOCK_MASK) << VSC73XX_CMD_SPI_BLOCK_SHIFT;
drivers/net/dsa/vitesse-vsc73xx-spi.c
52
static int vsc73xx_spi_read(struct vsc73xx *vsc, u8 block, u8 subblock, u8 reg,
drivers/net/dsa/vitesse-vsc73xx-spi.c
62
if (!vsc73xx_is_addr_valid(block, subblock))
drivers/net/dsa/vitesse-vsc73xx-spi.c
77
cmd[0] = vsc73xx_make_addr(VSC73XX_CMD_SPI_MODE_READ, block, subblock);
drivers/net/dsa/vitesse-vsc73xx-spi.c
94
static int vsc73xx_spi_write(struct vsc73xx *vsc, u8 block, u8 subblock, u8 reg,
drivers/net/dsa/vitesse-vsc73xx.h
70
int (*read)(struct vsc73xx *vsc, u8 block, u8 subblock, u8 reg,
drivers/net/dsa/vitesse-vsc73xx.h
72
int (*write)(struct vsc73xx *vsc, u8 block, u8 subblock, u8 reg,
drivers/net/dsa/vitesse-vsc73xx.h
91
int vsc73xx_is_addr_valid(u8 block, u8 subblock);
drivers/net/ethernet/airoha/airoha_eth.c
2689
block_cb = flow_block_cb_lookup(f->block, cb, port->dev);
drivers/net/ethernet/airoha/airoha_eth.c
2703
block_cb = flow_block_cb_lookup(f->block, cb, port->dev);
drivers/net/ethernet/broadcom/asp2/bcmasp.h
100
(((sel) * 0xa04) + ((block) * 0x200) + (off) + 0x600)
drivers/net/ethernet/broadcom/asp2/bcmasp.h
101
#define ASP_RX_FILTER_NET_MASK(sel, block, off) \
drivers/net/ethernet/broadcom/asp2/bcmasp.h
102
(((sel) * 0xa04) + ((block) * 0x200) + (off) + 0x700)
drivers/net/ethernet/broadcom/asp2/bcmasp.h
99
#define ASP_RX_FILTER_NET_PAT(sel, block, off) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
43
#define BLOCK_OPS_IDX(block, stage, end) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
44
(2*(((block)*NUM_OF_INIT_PHASES) + (stage)) + (end))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
491
#define BLOCK_OPS_IDX(block, stage, end) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
492
(2*(((block)*NUM_OF_INIT_PHASES) + (stage)) + (end))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
539
#define BLOCK_PRTY_INFO(block, en_mask, m1, m1h, m2, m3) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
541
block##_REG_##block##_PRTY_MASK, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
542
block##_REG_##block##_PRTY_STS_CLR, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
543
en_mask, {m1, m1h, m2, m3}, #block \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
546
#define BLOCK_PRTY_INFO_0(block, en_mask, m1, m1h, m2, m3) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
548
block##_REG_##block##_PRTY_MASK_0, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
549
block##_REG_##block##_PRTY_STS_CLR_0, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
550
en_mask, {m1, m1h, m2, m3}, #block"_0" \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
553
#define BLOCK_PRTY_INFO_1(block, en_mask, m1, m1h, m2, m3) \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
555
block##_REG_##block##_PRTY_MASK_1, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
556
block##_REG_##block##_PRTY_STS_CLR_1, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init.h
557
en_mask, {m1, m1h, m2, m3}, #block"_1" \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
229
static void bnx2x_init_block(struct bnx2x *bp, u32 block, u32 stage)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
232
INIT_OPS_OFFSETS(bp)[BLOCK_OPS_IDX(block, stage,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
235
INIT_OPS_OFFSETS(bp)[BLOCK_OPS_IDX(block, stage,
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1944
block_cb = flow_block_cb_lookup(f->block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
160
dpaa2_switch_acl_entry_remove(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
165
struct ethsw_core *ethsw = block->ethsw;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
186
block->acl_id, acl_entry_cfg);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
202
dpaa2_switch_acl_entry_add_to_list(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
209
if (list_empty(&block->acl_entries)) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
210
list_add(&entry->list, &block->acl_entries);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
214
list_for_each_safe(pos, n, &block->acl_entries) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
225
dpaa2_switch_acl_entry_get_by_index(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
231
list_for_each_entry(tmp, &block->acl_entries, list) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
241
dpaa2_switch_acl_entry_set_precedence(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
247
err = dpaa2_switch_acl_entry_remove(block, entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
252
return dpaa2_switch_acl_entry_add(block, entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
256
dpaa2_switch_acl_tbl_add_entry(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
263
index = dpaa2_switch_acl_entry_add_to_list(block, entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
268
precedence = DPAA2_ETHSW_PORT_MAX_ACL_ENTRIES - block->num_acl_rules - 1;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
270
tmp = dpaa2_switch_acl_entry_get_by_index(block, i);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
272
err = dpaa2_switch_acl_entry_set_precedence(block, tmp,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
282
err = dpaa2_switch_acl_entry_add(block, entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
283
block->num_acl_rules++;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
289
dpaa2_switch_acl_tbl_find_entry_by_cookie(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
294
list_for_each_entry_safe(tmp, n, &block->acl_entries, list) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
302
dpaa2_switch_acl_entry_get_index(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
308
list_for_each_entry_safe(tmp, n, &block->acl_entries, list) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
317
dpaa2_switch_mirror_find_entry_by_cookie(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
322
list_for_each_entry_safe(tmp, n, &block->mirror_entries, list) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
330
dpaa2_switch_acl_tbl_remove_entry(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
336
index = dpaa2_switch_acl_entry_get_index(block, entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
339
err = dpaa2_switch_acl_entry_remove(block, entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
343
block->num_acl_rules--;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
351
tmp = dpaa2_switch_acl_entry_get_by_index(block, i);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
352
err = dpaa2_switch_acl_entry_set_precedence(block, tmp,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
401
dpaa2_switch_block_add_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
405
unsigned long block_ports = block->ports;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
406
struct ethsw_core *ethsw = block->ethsw;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
449
list_add(&entry->list, &block->mirror_entries);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
466
dpaa2_switch_block_remove_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
470
unsigned long block_ports = block->ports;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
471
struct ethsw_core *ethsw = block->ethsw;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
486
if (list_empty(&block->mirror_entries))
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
493
dpaa2_switch_cls_flower_replace_acl(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
499
struct ethsw_core *ethsw = block->ethsw;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
503
if (dpaa2_switch_acl_tbl_is_full(block)) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
525
err = dpaa2_switch_acl_tbl_add_entry(block, acl_entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
583
dpaa2_switch_cls_flower_replace_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
588
struct ethsw_core *ethsw = block->ethsw;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
625
list_for_each_safe(pos, n, &block->mirror_entries) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
644
return dpaa2_switch_block_add_mirror(block, mirror_entry, if_id,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
648
int dpaa2_switch_cls_flower_replace(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
665
return dpaa2_switch_cls_flower_replace_acl(block, cls);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
667
return dpaa2_switch_cls_flower_replace_mirror(block, cls);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
674
int dpaa2_switch_cls_flower_destroy(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
681
acl_entry = dpaa2_switch_acl_tbl_find_entry_by_cookie(block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
684
return dpaa2_switch_acl_tbl_remove_entry(block, acl_entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
687
mirror_entry = dpaa2_switch_mirror_find_entry_by_cookie(block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
690
return dpaa2_switch_block_remove_mirror(block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
697
dpaa2_switch_cls_matchall_replace_acl(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
701
struct ethsw_core *ethsw = block->ethsw;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
706
if (dpaa2_switch_acl_tbl_is_full(block)) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
724
err = dpaa2_switch_acl_tbl_add_entry(block, acl_entry);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
737
dpaa2_switch_cls_matchall_replace_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
742
struct ethsw_core *ethsw = block->ethsw;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
773
list_for_each_safe(pos, n, &block->mirror_entries) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
790
return dpaa2_switch_block_add_mirror(block, mirror_entry, if_id,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
794
int dpaa2_switch_cls_matchall_replace(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
810
return dpaa2_switch_cls_matchall_replace_acl(block, cls);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
812
return dpaa2_switch_cls_matchall_replace_mirror(block, cls);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
819
int dpaa2_switch_block_offload_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
826
list_for_each_entry(tmp, &block->mirror_entries, list) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
837
list_for_each_entry(tmp, &block->mirror_entries, list)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
845
int dpaa2_switch_block_unoffload_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
852
list_for_each_entry(tmp, &block->mirror_entries, list) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
863
list_for_each_entry(tmp, &block->mirror_entries, list)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
870
int dpaa2_switch_cls_matchall_destroy(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
877
acl_entry = dpaa2_switch_acl_tbl_find_entry_by_cookie(block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
880
return dpaa2_switch_acl_tbl_remove_entry(block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
884
mirror_entry = dpaa2_switch_mirror_find_entry_by_cookie(block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch-flower.c
887
return dpaa2_switch_block_remove_mirror(block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1161
dpaa2_switch_setup_tc_cls_matchall(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1166
return dpaa2_switch_cls_matchall_replace(block, f);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1168
return dpaa2_switch_cls_matchall_destroy(block, f);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1192
struct dpaa2_switch_filter_block *block)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1205
block->acl_id, &acl_if_cfg);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1211
block->ports |= BIT(port_priv->idx);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1212
port_priv->filter_block = block;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1219
struct dpaa2_switch_filter_block *block)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1226
if (port_priv->filter_block != block)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1232
block->acl_id, &acl_if_cfg);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1238
block->ports &= ~BIT(port_priv->idx);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1244
struct dpaa2_switch_filter_block *block)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1252
err = dpaa2_switch_block_offload_mirror(block, port_priv);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1259
if (port_priv->filter_block == block)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1272
return dpaa2_switch_port_acl_tbl_bind(port_priv, block);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1277
struct dpaa2_switch_filter_block *block)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1286
err = dpaa2_switch_block_unoffload_mirror(block, port_priv);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1293
if (block->ports == BIT(port_priv->idx))
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1296
err = dpaa2_switch_port_acl_tbl_unbind(port_priv, block);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1300
if (block->ports == 0)
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1301
block->in_use = false;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1318
block_cb = flow_block_cb_lookup(f->block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
1367
block_cb = flow_block_cb_lookup(f->block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.h
257
int dpaa2_switch_cls_flower_replace(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.h
260
int dpaa2_switch_cls_flower_destroy(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.h
263
int dpaa2_switch_cls_matchall_replace(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.h
266
int dpaa2_switch_cls_matchall_destroy(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.h
269
int dpaa2_switch_acl_entry_add(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.h
272
int dpaa2_switch_block_offload_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.h
275
int dpaa2_switch_block_unoffload_mirror(struct dpaa2_switch_filter_block *block,
drivers/net/ethernet/google/gve/gve.h
1084
struct gve_notify_block *block)
drivers/net/ethernet/google/gve/gve.h
1086
return &priv->db_bar2[be32_to_cpu(*block->irq_db_index)];
drivers/net/ethernet/google/gve/gve.h
1240
bool gve_tx_poll(struct gve_notify_block *block, int budget);
drivers/net/ethernet/google/gve/gve.h
1241
bool gve_xdp_poll(struct gve_notify_block *block, int budget);
drivers/net/ethernet/google/gve/gve.h
1242
int gve_xsk_tx_poll(struct gve_notify_block *block, int budget);
drivers/net/ethernet/google/gve/gve.h
1254
int gve_rx_poll(struct gve_notify_block *block, int budget);
drivers/net/ethernet/google/gve/gve_dqo.h
108
struct gve_notify_block *block,
drivers/net/ethernet/google/gve/gve_dqo.h
111
gve_write_irq_doorbell_dqo(priv, block,
drivers/net/ethernet/google/gve/gve_dqo.h
40
bool gve_tx_poll_dqo(struct gve_notify_block *block, bool do_clean);
drivers/net/ethernet/google/gve/gve_dqo.h
41
bool gve_xdp_poll_dqo(struct gve_notify_block *block);
drivers/net/ethernet/google/gve/gve_dqo.h
42
bool gve_xsk_tx_poll_dqo(struct gve_notify_block *block, int budget);
drivers/net/ethernet/google/gve/gve_dqo.h
43
int gve_rx_poll_dqo(struct gve_notify_block *block, int budget);
drivers/net/ethernet/google/gve/gve_dqo.h
96
const struct gve_notify_block *block, u32 val)
drivers/net/ethernet/google/gve/gve_dqo.h
98
u32 index = be32_to_cpu(*block->irq_db_index);
drivers/net/ethernet/google/gve/gve_ethtool.c
798
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_ethtool.c
800
gve_set_itr_coalesce_usecs_dqo(priv, block,
drivers/net/ethernet/google/gve/gve_ethtool.c
808
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_ethtool.c
810
gve_set_itr_coalesce_usecs_dqo(priv, block,
drivers/net/ethernet/google/gve/gve_main.c
1896
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_main.c
1905
napi_disable_locked(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
1909
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_main.c
1916
napi_disable_locked(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
1941
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_main.c
1946
napi_enable_locked(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
1951
&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
1954
iowrite32be(0, gve_irq_doorbell(priv, block));
drivers/net/ethernet/google/gve/gve_main.c
1956
gve_set_itr_coalesce_usecs_dqo(priv, block,
drivers/net/ethernet/google/gve/gve_main.c
1965
napi_schedule(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
1969
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_main.c
1974
napi_enable_locked(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
1976
&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
1979
iowrite32be(0, gve_irq_doorbell(priv, block));
drivers/net/ethernet/google/gve/gve_main.c
1981
gve_set_itr_coalesce_usecs_dqo(priv, block,
drivers/net/ethernet/google/gve/gve_main.c
1990
napi_schedule(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
2026
struct gve_notify_block *block;
drivers/net/ethernet/google/gve/gve_main.c
2029
block = gve_get_tx_notify_block(priv, txqueue);
drivers/net/ethernet/google/gve/gve_main.c
2031
if (!block)
drivers/net/ethernet/google/gve/gve_main.c
2035
if (block->tx->last_kick_msec + MIN_TX_TIMEOUT_GAP > current_time)
drivers/net/ethernet/google/gve/gve_main.c
2039
napi_schedule(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
2040
block->tx->last_kick_msec = current_time;
drivers/net/ethernet/google/gve/gve_main.c
2046
struct gve_notify_block *block;
drivers/net/ethernet/google/gve/gve_main.c
2055
block = gve_get_tx_notify_block(priv, txqueue);
drivers/net/ethernet/google/gve/gve_main.c
2056
if (block)
drivers/net/ethernet/google/gve/gve_main.c
2057
block->tx->queue_timeout++;
drivers/net/ethernet/google/gve/gve_main.c
323
struct gve_notify_block *block = arg;
drivers/net/ethernet/google/gve/gve_main.c
324
struct gve_priv *priv = block->priv;
drivers/net/ethernet/google/gve/gve_main.c
326
iowrite32be(GVE_IRQ_MASK, gve_irq_doorbell(priv, block));
drivers/net/ethernet/google/gve/gve_main.c
327
napi_schedule_irqoff(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
333
struct gve_notify_block *block = arg;
drivers/net/ethernet/google/gve/gve_main.c
336
napi_schedule_irqoff(&block->napi);
drivers/net/ethernet/google/gve/gve_main.c
354
struct gve_notify_block *block;
drivers/net/ethernet/google/gve/gve_main.c
360
block = container_of(napi, struct gve_notify_block, napi);
drivers/net/ethernet/google/gve/gve_main.c
361
priv = block->priv;
drivers/net/ethernet/google/gve/gve_main.c
363
if (block->tx) {
drivers/net/ethernet/google/gve/gve_main.c
364
if (block->tx->q_num < priv->tx_cfg.num_queues)
drivers/net/ethernet/google/gve/gve_main.c
365
reschedule |= gve_tx_poll(block, budget);
drivers/net/ethernet/google/gve/gve_main.c
367
reschedule |= gve_xdp_poll(block, budget);
drivers/net/ethernet/google/gve/gve_main.c
373
if (block->rx) {
drivers/net/ethernet/google/gve/gve_main.c
374
work_done = gve_rx_poll(block, budget);
drivers/net/ethernet/google/gve/gve_main.c
381
gve_xsk_tx_poll(block, budget));
drivers/net/ethernet/google/gve/gve_main.c
391
irq_doorbell = gve_irq_doorbell(priv, block);
drivers/net/ethernet/google/gve/gve_main.c
399
if (block->tx)
drivers/net/ethernet/google/gve/gve_main.c
400
reschedule |= gve_tx_clean_pending(priv, block->tx);
drivers/net/ethernet/google/gve/gve_main.c
401
if (block->rx)
drivers/net/ethernet/google/gve/gve_main.c
402
reschedule |= gve_rx_work_pending(block->rx);
drivers/net/ethernet/google/gve/gve_main.c
412
struct gve_notify_block *block =
drivers/net/ethernet/google/gve/gve_main.c
414
struct gve_priv *priv = block->priv;
drivers/net/ethernet/google/gve/gve_main.c
418
if (block->tx) {
drivers/net/ethernet/google/gve/gve_main.c
419
if (block->tx->q_num < priv->tx_cfg.num_queues)
drivers/net/ethernet/google/gve/gve_main.c
420
reschedule |= gve_tx_poll_dqo(block, /*do_clean=*/true);
drivers/net/ethernet/google/gve/gve_main.c
422
reschedule |= gve_xdp_poll_dqo(block);
drivers/net/ethernet/google/gve/gve_main.c
428
if (block->rx) {
drivers/net/ethernet/google/gve/gve_main.c
429
work_done = gve_rx_poll_dqo(block, budget);
drivers/net/ethernet/google/gve/gve_main.c
435
reschedule |= gve_xsk_tx_poll_dqo(block, budget);
drivers/net/ethernet/google/gve/gve_main.c
443
if (likely(gve_is_napi_on_home_cpu(priv, block->irq)))
drivers/net/ethernet/google/gve/gve_main.c
464
gve_write_irq_doorbell_dqo(priv, block,
drivers/net/ethernet/google/gve/gve_main.c
553
struct gve_notify_block *block = &priv->ntfy_blocks[i];
drivers/net/ethernet/google/gve/gve_main.c
556
snprintf(block->name, sizeof(block->name), "gve-ntfy-blk%d@pci:%s",
drivers/net/ethernet/google/gve/gve_main.c
558
block->priv = priv;
drivers/net/ethernet/google/gve/gve_main.c
561
IRQF_NO_AUTOEN, block->name, block);
drivers/net/ethernet/google/gve/gve_main.c
567
block->irq = priv->msix_vectors[msix_idx].vector;
drivers/net/ethernet/google/gve/gve_main.c
568
irq_set_affinity_and_hint(block->irq,
drivers/net/ethernet/google/gve/gve_main.c
570
block->irq_db_index = &priv->irq_db_indices[i].index;
drivers/net/ethernet/google/gve/gve_main.c
583
struct gve_notify_block *block = &priv->ntfy_blocks[j];
drivers/net/ethernet/google/gve/gve_main.c
588
free_irq(priv->msix_vectors[msix_idx].vector, block);
drivers/net/ethernet/google/gve/gve_main.c
589
block->irq = 0;
drivers/net/ethernet/google/gve/gve_main.c
617
struct gve_notify_block *block = &priv->ntfy_blocks[i];
drivers/net/ethernet/google/gve/gve_main.c
622
free_irq(priv->msix_vectors[msix_idx].vector, block);
drivers/net/ethernet/google/gve/gve_main.c
623
block->irq = 0;
drivers/net/ethernet/google/gve/gve_rx.c
1085
int gve_rx_poll(struct gve_notify_block *block, int budget)
drivers/net/ethernet/google/gve/gve_rx.c
1087
struct gve_rx_ring *rx = block->rx;
drivers/net/ethernet/google/gve/gve_rx.c
1091
feat = block->napi.dev->features;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1004
int gve_rx_poll_dqo(struct gve_notify_block *block, int budget)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1017
napi = &block->napi;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1020
rx = block->rx;
drivers/net/ethernet/google/gve/gve_tx.c
1011
bool gve_tx_poll(struct gve_notify_block *block, int budget)
drivers/net/ethernet/google/gve/gve_tx.c
1013
struct gve_priv *priv = block->priv;
drivers/net/ethernet/google/gve/gve_tx.c
1014
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx.c
995
bool gve_xdp_poll(struct gve_notify_block *block, int budget)
drivers/net/ethernet/google/gve/gve_tx.c
997
struct gve_priv *priv = block->priv;
drivers/net/ethernet/google/gve/gve_tx.c
998
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1472
bool gve_tx_poll_dqo(struct gve_notify_block *block, bool do_clean)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1475
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1476
struct gve_priv *priv = block->priv;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1480
&block->napi);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1510
bool gve_xdp_poll_dqo(struct gve_notify_block *block)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1513
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1514
struct gve_priv *priv = block->priv;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1516
gve_clean_tx_done_dqo(priv, tx, &block->napi);
drivers/net/ethernet/google/gve/gve_utils.c
111
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_utils.c
113
netif_napi_add_locked(priv->dev, &block->napi, gve_poll);
drivers/net/ethernet/google/gve/gve_utils.c
114
netif_napi_set_irq_locked(&block->napi, block->irq);
drivers/net/ethernet/google/gve/gve_utils.c
115
enable_irq(block->irq);
drivers/net/ethernet/google/gve/gve_utils.c
120
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_utils.c
122
disable_irq(block->irq);
drivers/net/ethernet/google/gve/gve_utils.c
123
netif_napi_del_locked(&block->napi);
drivers/net/ethernet/google/gve/gve_utils.c
13
struct gve_notify_block *block =
drivers/net/ethernet/google/gve/gve_utils.c
16
return block->tx != NULL;
drivers/net/ethernet/google/gve/gve_utils.c
21
struct gve_notify_block *block =
drivers/net/ethernet/google/gve/gve_utils.c
24
block->tx = NULL;
drivers/net/ethernet/google/gve/gve_utils.c
32
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_utils.c
35
block->tx = tx;
drivers/net/ethernet/google/gve/gve_utils.c
43
struct gve_notify_block *block =
drivers/net/ethernet/google/gve/gve_utils.c
46
return block->rx != NULL;
drivers/net/ethernet/google/gve/gve_utils.c
51
struct gve_notify_block *block =
drivers/net/ethernet/google/gve/gve_utils.c
54
block->rx = NULL;
drivers/net/ethernet/google/gve/gve_utils.c
60
struct gve_notify_block *block = &priv->ntfy_blocks[ntfy_idx];
drivers/net/ethernet/google/gve/gve_utils.c
63
block->rx = rx;
drivers/net/ethernet/intel/ice/ice_fw_update.c
295
u16 block_size, u8 *block, bool last_cmd,
drivers/net/ethernet/intel/ice/ice_fw_update.c
312
err = ice_aq_update_nvm(hw, module, offset, block_size, block,
drivers/net/ethernet/intel/ice/ice_fw_update.c
414
u8 *block;
drivers/net/ethernet/intel/ice/ice_fw_update.c
424
block = kzalloc(ICE_AQ_MAX_BUF_LEN, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_fw_update.c
425
if (!block)
drivers/net/ethernet/intel/ice/ice_fw_update.c
438
memcpy(block, image + offset, block_size);
drivers/net/ethernet/intel/ice/ice_fw_update.c
441
block, last_cmd, reset_level,
drivers/net/ethernet/intel/ice/ice_fw_update.c
461
kfree(block);
drivers/net/ethernet/intel/ice/ice_fw_update.h
13
u16 block_size, u8 *block, bool last_cmd,
drivers/net/ethernet/intel/ice/ice_main.c
9609
block_cb = flow_block_cb_lookup(f->block,
drivers/net/ethernet/intel/ice/ice_ptp.c
581
err = ice_get_phy_tx_tstamp_ready(hw, tx->block, &tstamp_ready);
drivers/net/ethernet/intel/ice/ice_ptp.c
622
err = ice_read_phy_tstamp(hw, tx->block, phy_idx, &raw_tstamp);
drivers/net/ethernet/intel/ice/ice_ptp.c
748
err = ice_get_phy_tx_tstamp_ready(hw, tx->block, &tstamp_ready);
drivers/net/ethernet/intel/ice/ice_ptp.c
751
tx->block, err);
drivers/net/ethernet/intel/ice/ice_ptp.c
765
ice_clear_phy_tstamp(hw, tx->block, phy_idx);
drivers/net/ethernet/intel/ice/ice_ptp.c
868
tx->block = ICE_GET_QUAD_NUM(port);
drivers/net/ethernet/intel/ice/ice_ptp.c
889
tx->block = port;
drivers/net/ethernet/intel/ice/ice_ptp.h
119
u8 block;
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5431
int ice_read_phy_tstamp(struct ice_hw *hw, u8 block, u8 idx, u64 *tstamp)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5435
return ice_read_phy_tstamp_e810(hw, block, idx, tstamp);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5440
return ice_read_phy_tstamp_e82x(hw, block, idx, tstamp);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5442
return ice_read_ptp_tstamp_eth56g(hw, block, idx, tstamp);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5464
int ice_clear_phy_tstamp(struct ice_hw *hw, u8 block, u8 idx)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5468
return ice_clear_phy_tstamp_e810(hw, block, idx);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5470
return ice_clear_phy_tstamp_e82x(hw, block, idx);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5472
return ice_clear_ptp_tstamp_eth56g(hw, block, idx);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5584
int ice_get_phy_tx_tstamp_ready(struct ice_hw *hw, u8 block, u64 *tstamp_ready)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5588
return ice_get_phy_tx_tstamp_ready_e810(hw, block,
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5591
ice_get_phy_tx_tstamp_ready_e830(hw, block, tstamp_ready);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5594
return ice_get_phy_tx_tstamp_ready_e82x(hw, block,
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
5597
return ice_get_phy_tx_tstamp_ready_eth56g(hw, block,
drivers/net/ethernet/intel/ice/ice_ptp_hw.h
297
int ice_read_phy_tstamp(struct ice_hw *hw, u8 block, u8 idx, u64 *tstamp);
drivers/net/ethernet/intel/ice/ice_ptp_hw.h
298
int ice_clear_phy_tstamp(struct ice_hw *hw, u8 block, u8 idx);
drivers/net/ethernet/intel/ice/ice_ptp_hw.h
302
int ice_get_phy_tx_tstamp_ready(struct ice_hw *hw, u8 block, u64 *tstamp_ready);
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
241
u16 block_size, u8 *block, bool last_cmd,
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
246
return ixgbe_aci_update_nvm(hw, module, offset, block_size, block,
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
275
u8 *block;
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
281
block = kzalloc(IXGBE_ACI_MAX_BUFFER_SIZE, GFP_KERNEL);
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
282
if (!block)
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
292
memcpy(block, image + offset, block_size);
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
295
block_size, block, last_cmd,
drivers/net/ethernet/intel/ixgbe/ixgbe_fw_update.c
313
kfree(block);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1000
err = rvu_alloc_bitmap(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1009
block = &hw->block[BLKADDR_SSOW];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1010
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1012
block->lf.max = (cfg >> 56) & 0xFF;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1013
block->addr = BLKADDR_SSOW;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1014
block->type = BLKTYPE_SSOW;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1015
block->multislot = true;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1016
block->lfshift = 3;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1017
block->lookup_reg = SSOW_AF_RVU_LF_HWS_CFG_DEBUG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1018
block->pf_lfcnt_reg = RVU_PRIV_PFX_SSOW_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1019
block->vf_lfcnt_reg = RVU_PRIV_HWVFX_SSOW_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1020
block->lfcfg_reg = SSOW_PRIV_LFX_HWS_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1021
block->msixcfg_reg = SSOW_PRIV_LFX_HWS_INT_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1022
block->lfreset_reg = SSOW_AF_LF_HWS_RST;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1023
block->rvu = rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1024
sprintf(block->name, "SSOWS");
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1025
err = rvu_alloc_bitmap(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1034
block = &hw->block[BLKADDR_TIM];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1035
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1038
block->lf.max = cfg & 0xFFFF;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1039
block->addr = BLKADDR_TIM;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1040
block->type = BLKTYPE_TIM;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1041
block->multislot = true;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1042
block->lfshift = 3;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1043
block->lookup_reg = TIM_AF_RVU_LF_CFG_DEBUG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1044
block->pf_lfcnt_reg = RVU_PRIV_PFX_TIM_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1045
block->vf_lfcnt_reg = RVU_PRIV_HWVFX_TIM_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1046
block->lfcfg_reg = TIM_PRIV_LFX_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1047
block->msixcfg_reg = TIM_PRIV_LFX_INT_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1048
block->lfreset_reg = TIM_AF_LF_RST;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1049
block->rvu = rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
105
reg = rvu->afreg_base + ((block << 28) | offset);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1050
sprintf(block->name, "TIM");
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1051
err = rvu_alloc_bitmap(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1101
block = &hw->block[blkid];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1102
if (!block->lf.bmap)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1106
block->fn_map = devm_kcalloc(rvu->dev, block->lf.max,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1108
if (!block->fn_map) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1116
rvu_scan_block(rvu, block);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1312
static int rvu_lookup_rsrc(struct rvu *rvu, struct rvu_block *block,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1318
rvu_write64(rvu, block->addr, block->lookup_reg, val);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1321
while (rvu_read64(rvu, block->addr, block->lookup_reg) & (1ULL << 13))
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1324
val = rvu_read64(rvu, block->addr, block->lookup_reg);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1339
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1350
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1351
if (block->type != blktype)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1391
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1400
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1402
num_lfs = rvu_get_rsrc_mapcount(pfvf, block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1407
lf = rvu_lookup_rsrc(rvu, block, pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1416
rvu_write64(rvu, blkaddr, block->lfcfg_reg |
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1417
(lf << block->lfshift), 0x00ULL);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1420
rvu_update_rsrc_map(rvu, pfvf, block,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1424
rvu_free_rsrc(&block->lf, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1427
rvu_clear_msix_offset(rvu, pfvf, block, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1436
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1449
block = &hw->block[blkid];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1450
if (!block->lf.bmap)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1470
rvu_detach_block(rvu, pcifunc, block->type);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1559
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1571
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1572
if (!block->lf.bmap)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1577
lf = rvu_alloc_rsrc(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1582
rvu_write64(rvu, blkaddr, block->lfcfg_reg |
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1583
(lf << block->lfshift), cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1584
rvu_update_rsrc_map(rvu, pfvf, block,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1588
rvu_set_msix_offset(rvu, pfvf, block, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1600
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1604
block = &hw->block[BLKADDR_NPA];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1605
free_lfs = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1621
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1622
free_lfs = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1633
block = &hw->block[BLKADDR_SSO];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1635
if (req->sso > block->lf.max) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1638
pcifunc, req->sso, block->lf.max);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1641
mappedlfs = rvu_get_rsrc_mapcount(pfvf, block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1642
free_lfs = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1650
block = &hw->block[BLKADDR_SSOW];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1651
if (req->ssow > block->lf.max) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1654
pcifunc, req->ssow, block->lf.max);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1657
mappedlfs = rvu_get_rsrc_mapcount(pfvf, block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1658
free_lfs = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1665
block = &hw->block[BLKADDR_TIM];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1666
if (req->timlfs > block->lf.max) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1669
pcifunc, req->timlfs, block->lf.max);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1672
mappedlfs = rvu_get_rsrc_mapcount(pfvf, block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1673
free_lfs = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1684
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1685
if (req->cptlfs > block->lf.max) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1688
pcifunc, req->cptlfs, block->lf.max);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1691
mappedlfs = rvu_get_rsrc_mapcount(pfvf, block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1692
free_lfs = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1701
dev_info(rvu->dev, "Request for %s failed\n", block->name);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1841
struct rvu_block *block, int lf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1846
cfg = rvu_read64(rvu, block->addr, block->msixcfg_reg |
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1847
(lf << block->lfshift));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1857
rvu_write64(rvu, block->addr, block->msixcfg_reg |
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1858
(lf << block->lfshift), (cfg & ~0x7FFULL) | offset);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1862
pfvf->msix_lfmap[offset + vec] = MSIX_BLKLF(block->addr, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1866
struct rvu_block *block, int lf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1871
cfg = rvu_read64(rvu, block->addr, block->msixcfg_reg |
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1872
(lf << block->lfshift));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1876
rvu_write64(rvu, block->addr, block->msixcfg_reg |
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1877
(lf << block->lfshift), cfg & ~0x7FFULL);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1879
offset = rvu_get_msix_offset(rvu, pfvf, block->addr, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1902
lf = rvu_get_lf(rvu, &hw->block[BLKADDR_NPA], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1910
lf = rvu_get_lf(rvu, &hw->block[blkaddr], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1916
lf = rvu_get_lf(rvu, &hw->block[BLKADDR_SSO], pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1923
lf = rvu_get_lf(rvu, &hw->block[BLKADDR_SSOW], pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1930
lf = rvu_get_lf(rvu, &hw->block[BLKADDR_TIM], pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1937
lf = rvu_get_lf(rvu, &hw->block[BLKADDR_CPT0], pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1944
lf = rvu_get_lf(rvu, &hw->block[BLKADDR_CPT1], pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1956
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1962
block = &hw->block[BLKADDR_NPA];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1963
rsp->npa = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1965
block = &hw->block[BLKADDR_NIX0];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1966
rsp->nix = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1968
block = &hw->block[BLKADDR_NIX1];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1969
rsp->nix1 = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1971
block = &hw->block[BLKADDR_SSO];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1972
rsp->sso = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1974
block = &hw->block[BLKADDR_SSOW];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1975
rsp->ssow = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1977
block = &hw->block[BLKADDR_TIM];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1978
rsp->tim = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1980
block = &hw->block[BLKADDR_CPT0];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1981
rsp->cpt = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1983
block = &hw->block[BLKADDR_CPT1];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
1984
rsp->cpt1 = rvu_rsrc_free_count(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2113
nixlf = rvu_get_lf(rvu, &hw->block[blkaddr],
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2143
lfidx = rvu_get_lf(rvu, &hw->block[lfblkaddr], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2163
lfidx = rvu_get_lf(rvu, &hw->block[lfblkaddr], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
225
int rvu_get_lf(struct rvu *rvu, struct rvu_block *block, u16 pcifunc, u16 slot)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
231
for (lf = 0; lf < block->lf.max; lf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
232
if (block->fn_map[lf] == pcifunc) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2773
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2777
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2779
block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2783
lf = rvu_get_lf(rvu, block, pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2788
if (block->addr == BLKADDR_NIX0 || block->addr == BLKADDR_NIX1)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2789
rvu_nix_lf_teardown(rvu, pcifunc, block->addr, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2790
else if (block->addr == BLKADDR_NPA)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2792
else if ((block->addr == BLKADDR_CPT0) ||
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2793
(block->addr == BLKADDR_CPT1))
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2794
rvu_cpt_lf_teardown(rvu, pcifunc, block->addr, lf,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2797
err = rvu_lf_reset(rvu, block, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
2800
block->addr, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
30
struct rvu_block *block, int lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
32
struct rvu_block *block, int lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
342
struct rvu_block *block, u16 pcifunc,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
349
if (lf >= block->lf.max) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
352
__func__, lf, block->name, block->lf.max);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
3577
(rvu->hw->block[BLKADDR_NIX0].lf.max));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
365
block->fn_map[lf] = attach ? pcifunc : 0;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
367
switch (block->addr) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
399
reg = is_pf ? block->pf_lfcnt_reg : block->vf_lfcnt_reg;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
462
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
467
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
468
return block->implemented;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
474
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
480
block = &hw->block[blkid];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
483
block->implemented = true;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
500
int rvu_lf_reset(struct rvu *rvu, struct rvu_block *block, int lf)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
504
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
507
rvu_write64(rvu, block->addr, block->lfreset_reg, lf | BIT_ULL(12));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
508
err = rvu_poll_reg(rvu, block->addr, block->lfreset_reg, BIT_ULL(12),
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
515
struct rvu_block *block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
518
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
548
static void rvu_scan_block(struct rvu *rvu, struct rvu_block *block)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
554
for (lf = 0; lf < block->lf.max; lf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
555
cfg = rvu_read64(rvu, block->addr,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
556
block->lfcfg_reg | (lf << block->lfshift));
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
561
__set_bit(lf, block->lf.bmap);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
565
rvu_update_rsrc_map(rvu, pfvf, block,
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
569
rvu_set_msix_offset(rvu, pfvf, block, lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
721
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
732
block = &hw->block[id];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
733
kfree(block->lf.bmap);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
847
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
852
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
853
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
857
block->lf.max = cfg & 0xFFF;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
858
block->addr = blkaddr;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
859
block->type = BLKTYPE_NIX;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
860
block->lfshift = 8;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
861
block->lookup_reg = NIX_AF_RVU_LF_CFG_DEBUG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
862
block->pf_lfcnt_reg = RVU_PRIV_PFX_NIXX_CFG(blkid);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
863
block->vf_lfcnt_reg = RVU_PRIV_HWVFX_NIXX_CFG(blkid);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
864
block->lfcfg_reg = NIX_PRIV_LFX_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
865
block->msixcfg_reg = NIX_PRIV_LFX_INT_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
866
block->lfreset_reg = NIX_AF_LF_RST;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
867
block->rvu = rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
868
sprintf(block->name, "NIX%d", blkid);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
870
return rvu_alloc_bitmap(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
876
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
881
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
882
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
886
block->lf.max = cfg & 0xFF;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
887
block->addr = blkaddr;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
888
block->type = BLKTYPE_CPT;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
889
block->multislot = true;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
890
block->lfshift = 3;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
891
block->lookup_reg = CPT_AF_RVU_LF_CFG_DEBUG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
892
block->pf_lfcnt_reg = RVU_PRIV_PFX_CPTX_CFG(blkid);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
893
block->vf_lfcnt_reg = RVU_PRIV_HWVFX_CPTX_CFG(blkid);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
894
block->lfcfg_reg = CPT_PRIV_LFX_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
895
block->msixcfg_reg = CPT_PRIV_LFX_INT_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
896
block->lfreset_reg = CPT_AF_LF_RST;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
897
block->rvu = rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
898
sprintf(block->name, "CPT%d", blkid);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
899
return rvu_alloc_bitmap(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
930
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
944
block = &hw->block[BLKADDR_NPA];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
945
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
948
block->lf.max = (cfg >> 16) & 0xFFF;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
949
block->addr = BLKADDR_NPA;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
950
block->type = BLKTYPE_NPA;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
951
block->lfshift = 8;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
952
block->lookup_reg = NPA_AF_RVU_LF_CFG_DEBUG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
953
block->pf_lfcnt_reg = RVU_PRIV_PFX_NPA_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
954
block->vf_lfcnt_reg = RVU_PRIV_HWVFX_NPA_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
955
block->lfcfg_reg = NPA_PRIV_LFX_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
956
block->msixcfg_reg = NPA_PRIV_LFX_INT_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
957
block->lfreset_reg = NPA_AF_LF_RST;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
958
block->rvu = rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
959
sprintf(block->name, "NPA");
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
960
err = rvu_alloc_bitmap(&block->lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
98
int rvu_poll_reg(struct rvu *rvu, u64 block, u64 offset, u64 mask, bool zero)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
983
block = &hw->block[BLKADDR_SSO];
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
984
if (!block->implemented)
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
987
block->lf.max = cfg & 0xFFFF;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
988
block->addr = BLKADDR_SSO;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
989
block->type = BLKTYPE_SSO;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
990
block->multislot = true;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
991
block->lfshift = 3;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
992
block->lookup_reg = SSO_AF_RVU_LF_CFG_DEBUG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
993
block->pf_lfcnt_reg = RVU_PRIV_PFX_SSO_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
994
block->vf_lfcnt_reg = RVU_PRIV_HWVFX_SSO_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
995
block->lfcfg_reg = SSO_PRIV_LFX_HWGRP_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
996
block->msixcfg_reg = SSO_PRIV_LFX_HWGRP_INT_CFG;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
997
block->lfreset_reg = SSO_AF_LF_HWGRP_RST;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
998
block->rvu = rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu.c
999
sprintf(block->name, "SSO GROUP");
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
459
struct rvu_block block[BLK_COUNT]; /* Block info */
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
673
static inline void rvu_write64(struct rvu *rvu, u64 block, u64 offset, u64 val)
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
675
writeq(val, rvu->afreg_base + ((block << 28) | offset));
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
678
static inline u64 rvu_read64(struct rvu *rvu, u64 block, u64 offset)
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
680
return readq(rvu->afreg_base + ((block << 28) | offset));
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
693
static inline void rvu_bar2_sel_write64(struct rvu *rvu, u64 block, u64 offset, u64 val)
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
698
rvu_write64(rvu, block, offset, val);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
699
rvu_read64(rvu, block, offset);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
914
int rvu_get_lf(struct rvu *rvu, struct rvu_block *block, u16 pcifunc, u16 slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
915
int rvu_lf_reset(struct rvu *rvu, struct rvu_block *block, int lf);
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
917
int rvu_poll_reg(struct rvu *rvu, u64 block, u64 offset, u64 mask, bool zero);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1087
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1108
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1110
for (lf = 0; lf < block->lf.max; lf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_cgx.c
1112
if (!((block->fn_map[lf] & ~RVU_PFVF_FUNC_MASK) == (pcifunc &
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
1001
rsp->flt_eng_map[vec] = block->cpt_flt_eng_map[vec];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
1002
rsp->rcvrd_eng_map[vec] = block->cpt_rcvrd_eng_map[vec];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
1004
block->cpt_flt_eng_map[vec] = 0x0;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
1005
block->cpt_rcvrd_eng_map[vec] = 0x0;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
116
block->cpt_flt_eng_map[vec] |= BIT_ULL(i);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
120
block->cpt_rcvrd_eng_map[vec] |= BIT_ULL(i);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
145
struct rvu_block *block = ptr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
146
struct rvu *rvu = block->rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
147
int blkaddr = block->addr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
159
struct rvu_block *block = ptr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
160
struct rvu *rvu = block->rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
161
int blkaddr = block->addr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
171
static int rvu_cpt_do_register_interrupt(struct rvu_block *block, int irq_offs,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
175
struct rvu *rvu = block->rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
179
name, block);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
190
static void cpt_10k_unregister_interrupts(struct rvu_block *block, int off)
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
192
struct rvu *rvu = block->rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
193
int blkaddr = block->addr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
215
free_irq(pci_irq_vector(rvu->pdev, off + i), block);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
223
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
234
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
236
return cpt_10k_unregister_interrupts(block, offs);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
246
free_irq(pci_irq_vector(rvu->pdev, offs + i), block);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
257
static int cpt_10k_register_interrupts(struct rvu_block *block, int off)
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
260
struct rvu *rvu = block->rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
261
int blkaddr = block->addr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
284
ret = rvu_cpt_do_register_interrupt(block, off + i,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
298
ret = rvu_cpt_do_register_interrupt(block, off + rvu_intr_vec,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
305
ret = rvu_cpt_do_register_interrupt(block, off + ras_intr_vec,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
321
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
328
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
337
return cpt_10k_register_interrupts(block, offs);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
349
ret = rvu_cpt_do_register_interrupt(block, offs + i,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
356
ret = rvu_cpt_do_register_interrupt(block, offs + CPT_AF_INT_VEC_RVU,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
363
ret = rvu_cpt_do_register_interrupt(block, offs + CPT_AF_INT_VEC_RAS,
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
449
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
461
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
463
block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
486
cptlf = rvu_get_lf(rvu, block, pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
518
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
520
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
522
block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
527
cptlf = rvu_get_lf(rvu, block, pcifunc, slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
535
err = rvu_lf_reset(rvu, block, cptlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
538
block->addr, cptlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
664
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
673
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
675
cptlf = rvu_get_lf(rvu, block, pcifunc, actual_slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
701
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
715
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
717
num_lfs = rvu_get_rsrc_mapcount(pfvf, block->addr);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
723
lf = rvu_get_lf(rvu, &rvu->hw->block[blkaddr],
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
83
struct rvu_block *block = ptr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
84
struct rvu *rvu = block->rvu;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
85
int blkaddr = block->addr;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
954
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
964
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
966
cptlf = rvu_get_lf(rvu, block, pcifunc, actual_slot);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
972
ret = rvu_lf_reset(rvu, block, cptlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
975
block->addr, cptlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
986
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_cpt.c
996
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3868
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3874
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3875
if (!block->lf.bmap)
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3879
for (lf = 0; lf < block->lf.max; lf++) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3886
reg = rvu_read64(rvu, blkaddr, block->lfcfg_reg |
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
3887
(lf << block->lfshift));
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
646
static void get_lf_str_list(const struct rvu_block *block, int pcifunc,
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
649
int lf = 0, seq = 0, len = 0, prev_lf = block->lf.max;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
651
for_each_set_bit(lf, block->lf.bmap, block->lf.max) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
652
if (lf >= block->lf.max)
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
655
if (block->fn_map[lf] != pcifunc)
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
683
struct rvu_block block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
698
block = rvu->hw->block[index];
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
699
if (!strlen(block.name))
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
702
get_lf_str_list(&block, pcifunc, buf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
721
struct rvu_block block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
747
if (strlen(rvu->hw->block[index].name)) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
750
rvu->hw->block[index].name);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
781
block = rvu->hw->block[index];
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
782
if (!strlen(block.name))
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
786
get_lf_str_list(&block, pcifunc, lfs);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
940
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
944
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
946
if (lf < 0 || lf >= block->lf.max) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
948
block->lf.max - 1);
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
952
*pcifunc = block->fn_map[lf];
drivers/net/ethernet/marvell/octeontx2/af/rvu_devlink.c
1426
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_devlink.c
1432
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_devlink.c
1433
block->lf.max = ctx->val.vu16;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1046
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1055
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1056
aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1063
nixlf = rvu_get_lf(rvu, block, pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1201
rc = nix_aq_enqueue_wait(rvu, block, &inst);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
122
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
126
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
127
max += block->lf.max;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
143
*nixlf = rvu_get_lf(rvu, &hw->block[blkaddr], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1505
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1521
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1522
nixlf = rvu_get_lf(rvu, block, pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1558
err = rvu_lf_reset(rvu, block, nixlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1561
block->addr - BLKADDR_NIX0, nixlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1739
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1748
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1749
nixlf = rvu_get_lf(rvu, block, pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1769
err = rvu_lf_reset(rvu, block, nixlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
1772
block->addr - BLKADDR_NIX0, nixlf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2544
nixlf = rvu_get_lf(rvu, &hw->block[blkaddr], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2633
nixlf = rvu_get_lf(rvu, &hw->block[blkaddr], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2922
nixlf = rvu_get_lf(rvu, &hw->block[blkaddr],
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3571
nixlf = rvu_get_lf(rvu, &hw->block[blkaddr], pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4853
static int nix_aq_init(struct rvu *rvu, struct rvu_block *block)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4859
cfg = rvu_read64(rvu, block->addr, NIX_AF_CFG);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4862
rvu_write64(rvu, block->addr, NIX_AF_CFG, cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4865
rvu_write64(rvu, block->addr, NIX_AF_CFG, cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4869
cfg = rvu_read64(rvu, block->addr, NIX_AF_NDC_CFG);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4875
rvu_write64(rvu, block->addr, NIX_AF_NDC_CFG, cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4881
err = rvu_aq_alloc(rvu, &block->aq,
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4887
rvu_write64(rvu, block->addr, NIX_AF_AQ_CFG, AQ_SIZE);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4888
rvu_write64(rvu, block->addr,
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4889
NIX_AF_AQ_BASE, (u64)block->aq->inst->iova);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4923
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4927
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4965
err = nix_aq_init(rvu, block);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5130
struct rvu_block *block)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5138
rvu_aq_free(rvu, block->aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5168
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5173
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5174
rvu_nix_block_freemem(rvu, blkaddr, block);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5399
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5412
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
5413
nixlf = rvu_get_lf(rvu, block, pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6650
struct rvu_block *block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
6653
if (!block->implemented || is_cn20k(rvu->pdev))
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
966
static int nix_aq_enqueue_wait(struct rvu *rvu, struct rvu_block *block,
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
969
struct admin_queue *aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
978
reg = rvu_read64(rvu, block->addr, NIX_AF_AQ_STATUS);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
988
rvu_write64(rvu, block->addr, NIX_AF_AQ_DOOR, 1);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
15
static int npa_aq_enqueue_wait(struct rvu *rvu, struct rvu_block *block,
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
158
rc = npa_aq_enqueue_wait(rvu, block, &inst);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
18
struct admin_queue *aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
26
reg = rvu_read64(rvu, block->addr, NPA_AF_AQ_STATUS);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
334
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
351
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
352
npalf = rvu_get_lf(rvu, block, pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
357
err = rvu_lf_reset(rvu, block, npalf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
36
rvu_write64(rvu, block->addr, NPA_AF_AQ_DOOR, 1);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
429
cfg = rvu_read64(rvu, block->addr, NPA_AF_BATCH_CTL);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
440
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
450
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
451
npalf = rvu_get_lf(rvu, block, pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
456
err = rvu_lf_reset(rvu, block, npalf);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
467
static void npa_aq_ndc_config(struct rvu *rvu, struct rvu_block *block)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
475
cfg = rvu_read64(rvu, block->addr, NPA_AF_NDC_CFG);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
481
rvu_write64(rvu, block->addr, NPA_AF_NDC_CFG, cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
484
static int npa_aq_init(struct rvu *rvu, struct rvu_block *block)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
490
cfg = rvu_read64(rvu, block->addr, NPA_AF_GEN_CFG);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
493
rvu_write64(rvu, block->addr, NPA_AF_GEN_CFG, cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
496
rvu_write64(rvu, block->addr, NPA_AF_GEN_CFG, cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
499
npa_aq_ndc_config(rvu, block);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
503
cfg = rvu_read64(rvu, block->addr, NPA_AF_BATCH_CTL);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
506
rvu_write64(rvu, block->addr, NPA_AF_BATCH_CTL, cfg);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
512
err = rvu_aq_alloc(rvu, &block->aq,
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
518
rvu_write64(rvu, block->addr, NPA_AF_AQ_CFG, AQ_SIZE);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
519
rvu_write64(rvu, block->addr,
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
520
NPA_AF_AQ_BASE, (u64)block->aq->inst->iova);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
534
return npa_aq_init(rvu, &hw->block[blkaddr]);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
540
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
547
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
548
rvu_aq_free(rvu, block->aq);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
68
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
82
block = &hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
83
aq = block->aq;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npa.c
89
npalf = rvu_get_lf(rvu, block, pcifunc, 0);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
128
struct rvu_block *block;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
139
block = &rvu->hw->block[blkaddr];
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
140
max += block->lf.max;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
137
struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
144
if (!prestera_acl_chain_is_supported(chain_index, block->ingress))
drivers/net/ethernet/marvell/prestera/prestera_acl.c
152
ruleset->ingress = block->ingress;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
153
ruleset->ht_key.block = block;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
24
struct prestera_flow_block *block;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
270
struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
276
ht_key.block = block;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
284
struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
289
ruleset = __prestera_acl_ruleset_lookup(acl, block, chain_index);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
299
struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
304
ruleset = __prestera_acl_ruleset_lookup(acl, block, chain_index);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
310
return prestera_acl_ruleset_create(acl, block, chain_index);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
346
struct prestera_flow_block *block)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
351
block->ruleset_zero = ruleset;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
352
list_for_each_entry(binding, &block->binding_list, list) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
360
list_for_each_entry_continue_reverse(binding, &block->binding_list,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
363
block->ruleset_zero = NULL;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
370
struct prestera_flow_block *block)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
374
list_for_each_entry(binding, &block->binding_list, list)
drivers/net/ethernet/marvell/prestera/prestera_acl.c
376
block->ruleset_zero = NULL;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
47
struct prestera_counter_block *block;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
482
struct prestera_flow_block *block = ruleset->ht_key.block;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
509
err = prestera_acl_ruleset_block_bind(ruleset, block);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
533
struct prestera_flow_block *block = ruleset->ht_key.block;
drivers/net/ethernet/marvell/prestera/prestera_acl.c
545
prestera_acl_ruleset_block_unbind(ruleset, block);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
557
rule->re->counter.block,
drivers/net/ethernet/marvell/prestera/prestera_acl.c
621
if (e->counter.block) {
drivers/net/ethernet/marvell/prestera/prestera_acl.c
637
prestera_counter_put(sw->counter, e->counter.block, e->counter.id);
drivers/net/ethernet/marvell/prestera/prestera_acl.c
695
&e->counter.block,
drivers/net/ethernet/marvell/prestera/prestera_acl.h
182
struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_acl.h
186
struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
110
struct prestera_counter_block *block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
121
counter->block_list[i] = block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
134
counter->block_list[counter->block_list_len] = block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
143
struct prestera_counter_block *block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
146
block = prestera_counter_block_lookup_not_full(counter, client);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
147
if (block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
148
return block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
150
block = kzalloc_obj(*block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
151
if (!block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
155
&block->id, &block->offset,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
156
&block->num_counters);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
160
block->stats = kzalloc_objs(*block->stats, block->num_counters);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
161
if (!block->stats) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
166
block->counter_flag = kcalloc(block->num_counters,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
167
sizeof(*block->counter_flag),
drivers/net/ethernet/marvell/prestera/prestera_counter.c
169
if (!block->counter_flag) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
174
block->client = client;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
175
mutex_init(&block->mtx);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
176
refcount_set(&block->refcnt, 1);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
177
idr_init_base(&block->counter_idr, block->offset);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
179
err = prestera_counter_block_list_add(counter, block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
183
return block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
186
idr_destroy(&block->counter_idr);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
187
mutex_destroy(&block->mtx);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
188
kfree(block->counter_flag);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
190
kfree(block->stats);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
192
prestera_hw_counter_block_release(counter->sw, block->id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
194
kfree(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
199
struct prestera_counter_block *block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
203
if (!prestera_counter_block_decref(block))
drivers/net/ethernet/marvell/prestera/prestera_counter.c
209
counter->block_list[i]->id == block->id) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
216
WARN_ON(!idr_is_empty(&block->counter_idr));
drivers/net/ethernet/marvell/prestera/prestera_counter.c
218
prestera_hw_counter_block_release(counter->sw, block->id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
219
idr_destroy(&block->counter_idr);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
220
mutex_destroy(&block->mtx);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
221
kfree(block->stats);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
222
kfree(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
225
static int prestera_counter_get_vacant(struct prestera_counter_block *block,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
230
if (block->full)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
233
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
234
free_id = idr_alloc_cyclic(&block->counter_idr, NULL, block->offset,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
235
block->offset + block->num_counters,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
239
block->full = true;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
241
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
245
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
253
struct prestera_counter_block *block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
258
block = prestera_counter_block_get(counter, client);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
259
if (IS_ERR(block))
drivers/net/ethernet/marvell/prestera/prestera_counter.c
260
return PTR_ERR(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
262
err = prestera_counter_get_vacant(block, &id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
264
prestera_counter_block_put(counter, block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
272
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
273
if (block->is_updating)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
274
block->counter_flag[id - block->offset] = COUNTER_FLAG_INVALID;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
275
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
278
*bl = block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
284
struct prestera_counter_block *block, u32 counter_id)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
286
if (!block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
289
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
290
idr_remove(&block->counter_idr, counter_id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
291
block->full = false;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
292
prestera_counter_stats_clear(block, counter_id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
293
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
295
prestera_hw_counter_clear(counter->sw, block->id, counter_id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
296
prestera_counter_block_put(counter, block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
341
struct prestera_counter_block *block;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
348
block = prestera_counter_block_get_by_idx(counter, counter->curr_idx);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
349
if (!block) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
357
err = prestera_hw_counter_trigger(counter->sw, block->id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
361
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
362
block->is_updating = true;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
363
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
371
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
374
&block->stats[counter->total_read]);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
375
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
380
if (!done || counter->total_read < block->num_counters) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
385
for (i = 0; i < block->num_counters; i++) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
386
if (block->counter_flag[i] == COUNTER_FLAG_INVALID) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
387
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
388
block->counter_flag[i] = COUNTER_FLAG_READY;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
389
memset(&block->stats[i], 0, sizeof(*block->stats));
drivers/net/ethernet/marvell/prestera/prestera_counter.c
390
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
394
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
395
block->is_updating = false;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
396
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
406
if (block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
407
prestera_counter_block_put(counter, block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
416
struct prestera_counter_block *block,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
419
if (!block || !prestera_counter_is_ready(block, counter_id)) {
drivers/net/ethernet/marvell/prestera/prestera_counter.c
425
prestera_counter_block_lock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
426
*packets = block->stats[counter_id - block->offset].packets;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
427
*bytes = block->stats[counter_id - block->offset].bytes;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
429
prestera_counter_stats_clear(block, counter_id);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
430
prestera_counter_block_unlock(block);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
45
prestera_counter_is_ready(struct prestera_counter_block *block, u32 id)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
47
return block->counter_flag[id - block->offset] == COUNTER_FLAG_READY;
drivers/net/ethernet/marvell/prestera/prestera_counter.c
60
static void prestera_counter_block_lock(struct prestera_counter_block *block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
62
mutex_lock(&block->mtx);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
65
static void prestera_counter_block_unlock(struct prestera_counter_block *block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
67
mutex_unlock(&block->mtx);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
70
static bool prestera_counter_block_incref(struct prestera_counter_block *block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
72
return refcount_inc_not_zero(&block->refcnt);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
75
static bool prestera_counter_block_decref(struct prestera_counter_block *block)
drivers/net/ethernet/marvell/prestera/prestera_counter.c
77
return refcount_dec_and_test(&block->refcnt);
drivers/net/ethernet/marvell/prestera/prestera_counter.c
81
static void prestera_counter_stats_clear(struct prestera_counter_block *block,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
84
memset(&block->stats[counter_id - block->offset], 0,
drivers/net/ethernet/marvell/prestera/prestera_counter.c
85
sizeof(*block->stats));
drivers/net/ethernet/marvell/prestera/prestera_counter.h
22
struct prestera_counter_block **block,
drivers/net/ethernet/marvell/prestera/prestera_counter.h
25
struct prestera_counter_block *block, u32 counter_id);
drivers/net/ethernet/marvell/prestera/prestera_counter.h
27
struct prestera_counter_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
103
struct prestera_flow_block *block = cb_priv;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
105
prestera_flow_block_destroy(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
109
prestera_flow_block_is_bound(const struct prestera_flow_block *block)
drivers/net/ethernet/marvell/prestera/prestera_flow.c
111
return block->ruleset_zero;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
115
prestera_flow_block_lookup(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
120
list_for_each_entry(binding, &block->binding_list, list)
drivers/net/ethernet/marvell/prestera/prestera_flow.c
127
static int prestera_flow_block_bind(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
140
if (prestera_flow_block_is_bound(block)) {
drivers/net/ethernet/marvell/prestera/prestera_flow.c
141
err = prestera_acl_ruleset_bind(block->ruleset_zero, port);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
146
list_add(&binding->list, &block->binding_list);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
154
static int prestera_flow_block_unbind(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
159
binding = prestera_flow_block_lookup(block, port);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
16
static int prestera_flow_block_mall_cb(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
165
if (prestera_flow_block_is_bound(block))
drivers/net/ethernet/marvell/prestera/prestera_flow.c
166
prestera_acl_ruleset_unbind(block->ruleset_zero, port);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
178
struct prestera_flow_block *block;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
181
block_cb = flow_block_cb_lookup(f->block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
184
block = prestera_flow_block_create(sw, f->net, ingress);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
185
if (!block)
drivers/net/ethernet/marvell/prestera/prestera_flow.c
189
sw, block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
192
prestera_flow_block_destroy(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
196
block->block_cb = block_cb;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
199
block = flow_block_cb_priv(block_cb);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
205
return block;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
208
static void prestera_flow_block_put(struct prestera_flow_block *block)
drivers/net/ethernet/marvell/prestera/prestera_flow.c
21
return prestera_mall_replace(block, f);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
210
struct flow_block_cb *block_cb = block->block_cb;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
216
prestera_flow_block_destroy(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
223
struct prestera_flow_block *block;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
228
block = prestera_flow_block_get(sw, f, &register_block, ingress);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
229
if (IS_ERR(block))
drivers/net/ethernet/marvell/prestera/prestera_flow.c
23
prestera_mall_destroy(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
230
return PTR_ERR(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
232
block_cb = block->block_cb;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
234
err = prestera_flow_block_bind(block, port);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
244
port->ingress_flow_block = block;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
246
port->egress_flow_block = block;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
251
prestera_flow_block_put(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
260
struct prestera_flow_block *block;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
264
block_cb = flow_block_cb_lookup(f->block, prestera_flow_block_cb, sw);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
268
block = flow_block_cb_priv(block_cb);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
270
prestera_mall_destroy(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
272
err = prestera_flow_block_unbind(block, port);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
30
static int prestera_flow_block_flower_cb(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flow.c
35
return prestera_flower_replace(block, f);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
37
prestera_flower_destroy(block, f);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
40
return prestera_flower_stats(block, f);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
42
return prestera_flower_tmplt_create(block, f);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
44
prestera_flower_tmplt_destroy(block, f);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
54
struct prestera_flow_block *block = cb_priv;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
58
return prestera_flow_block_flower_cb(block, type_data);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
60
return prestera_flow_block_mall_cb(block, type_data);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
68
struct prestera_flow_block *block = cb_priv;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
70
prestera_flower_template_cleanup(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
72
WARN_ON(!list_empty(&block->template_list));
drivers/net/ethernet/marvell/prestera/prestera_flow.c
73
WARN_ON(!list_empty(&block->binding_list));
drivers/net/ethernet/marvell/prestera/prestera_flow.c
75
kfree(block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
83
struct prestera_flow_block *block;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
85
block = kzalloc_obj(*block);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
86
if (!block)
drivers/net/ethernet/marvell/prestera/prestera_flow.c
89
INIT_LIST_HEAD(&block->binding_list);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
90
INIT_LIST_HEAD(&block->template_list);
drivers/net/ethernet/marvell/prestera/prestera_flow.c
91
block->net = net;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
92
block->sw = sw;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
93
block->mall.prio_min = UINT_MAX;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
94
block->mall.prio_max = 0;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
95
block->mall.bound = false;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
96
block->ingress = ingress;
drivers/net/ethernet/marvell/prestera/prestera_flow.c
98
return block;
drivers/net/ethernet/marvell/prestera/prestera_flower.c
120
rule->re_arg.police.ingress = block->ingress;
drivers/net/ethernet/marvell/prestera/prestera_flower.c
123
err = prestera_flower_parse_goto_action(block, rule,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
141
struct prestera_flow_block *block)
drivers/net/ethernet/marvell/prestera/prestera_flower.c
163
ingress_dev = __dev_get_by_index(block->net,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
191
static int prestera_flower_parse(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
222
err = prestera_flower_parse_meta(rule, f, block);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
24
void prestera_flower_template_cleanup(struct prestera_flow_block *block)
drivers/net/ethernet/marvell/prestera/prestera_flower.c
29
list_for_each_entry_safe(template, tmp, &block->template_list, list)
drivers/net/ethernet/marvell/prestera/prestera_flower.c
34
prestera_flower_parse_goto_action(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
369
return prestera_flower_parse_actions(block, rule, &f->rule->action,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
374
static int prestera_flower_prio_check(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
381
err = prestera_mall_prio_get(block, &mall_prio_min, &mall_prio_max);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
391
if (f->common.prio <= mall_prio_max && block->ingress) {
drivers/net/ethernet/marvell/prestera/prestera_flower.c
396
if (f->common.prio >= mall_prio_min && !block->ingress) {
drivers/net/ethernet/marvell/prestera/prestera_flower.c
404
int prestera_flower_prio_get(struct prestera_flow_block *block, u32 chain_index,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
409
ruleset = prestera_acl_ruleset_lookup(block->sw->acl, block, chain_index);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
417
int prestera_flower_replace(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
421
struct prestera_acl *acl = block->sw->acl;
drivers/net/ethernet/marvell/prestera/prestera_flower.c
425
err = prestera_flower_prio_check(block, f);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
429
ruleset = prestera_acl_ruleset_get(acl, block, f->common.chain_index);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
441
err = prestera_flower_parse(block, rule, f);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
451
err = prestera_acl_rule_add(block->sw, rule);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
466
void prestera_flower_destroy(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
472
ruleset = prestera_acl_ruleset_lookup(block->sw->acl, block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
479
prestera_acl_rule_del(block->sw, rule);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
48
ruleset = prestera_acl_ruleset_get(block->sw->acl, block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
485
int prestera_flower_tmplt_create(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
494
err = prestera_flower_parse(block, &rule, f);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
505
ruleset = prestera_acl_ruleset_get(block->sw->acl, block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
528
list_add_rcu(&template->list, &block->template_list);
drivers/net/ethernet/marvell/prestera/prestera_flower.c
540
void prestera_flower_tmplt_destroy(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
545
list_for_each_entry_safe(template, tmp, &block->template_list, list)
drivers/net/ethernet/marvell/prestera/prestera_flower.c
553
int prestera_flower_stats(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
563
ruleset = prestera_acl_ruleset_lookup(block->sw->acl, block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
574
err = prestera_acl_rule_get_stats(block->sw->acl, rule, &packets,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
61
static int prestera_flower_parse_actions(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.c
83
err = prestera_acl_chain_to_client(chain_index, block->ingress,
drivers/net/ethernet/marvell/prestera/prestera_flower.h
11
int prestera_flower_replace(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.h
13
void prestera_flower_destroy(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.h
15
int prestera_flower_stats(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.h
17
int prestera_flower_tmplt_create(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.h
19
void prestera_flower_tmplt_destroy(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_flower.h
21
void prestera_flower_template_cleanup(struct prestera_flow_block *block);
drivers/net/ethernet/marvell/prestera/prestera_flower.h
22
int prestera_flower_prio_get(struct prestera_flow_block *block, u32 chain_index,
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
105
prestera_mall_prio_update(block, f);
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
107
block->mall.bound = true;
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
112
&block->binding_list, list)
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
113
prestera_span_rule_del(binding, block->ingress);
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
117
void prestera_mall_destroy(struct prestera_flow_block *block)
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
121
list_for_each_entry(binding, &block->binding_list, list)
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
122
prestera_span_rule_del(binding, block->ingress);
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
124
block->mall.prio_min = UINT_MAX;
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
125
block->mall.prio_max = 0;
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
126
block->mall.bound = false;
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
14
static int prestera_mall_prio_check(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
21
err = prestera_flower_prio_get(block, f->common.chain_index,
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
32
if (f->common.prio <= flower_prio_max && !block->ingress) {
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
36
if (f->common.prio >= flower_prio_min && block->ingress) {
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
44
int prestera_mall_prio_get(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
47
if (!block->mall.bound)
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
50
*prio_min = block->mall.prio_min;
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
51
*prio_max = block->mall.prio_max;
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
55
static void prestera_mall_prio_update(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
58
block->mall.prio_min = min(block->mall.prio_min, f->common.prio);
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
59
block->mall.prio_max = max(block->mall.prio_max, f->common.prio);
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
62
int prestera_mall_replace(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
91
err = prestera_mall_prio_check(block, f);
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
97
list_for_each_entry(binding, &block->binding_list, list) {
drivers/net/ethernet/marvell/prestera/prestera_matchall.c
98
err = prestera_span_rule_add(binding, port, block->ingress);
drivers/net/ethernet/marvell/prestera/prestera_matchall.h
11
int prestera_mall_replace(struct prestera_flow_block *block,
drivers/net/ethernet/marvell/prestera/prestera_matchall.h
13
void prestera_mall_destroy(struct prestera_flow_block *block);
drivers/net/ethernet/marvell/prestera/prestera_matchall.h
14
int prestera_mall_prio_get(struct prestera_flow_block *block,
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
632
block_cb = flow_block_cb_lookup(f->block, cb, dev);
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
646
block_cb = flow_block_cb_lookup(f->block, cb, dev);
drivers/net/ethernet/mediatek/mtk_wed.c
2715
block_cb = flow_block_cb_lookup(f->block, cb, dev);
drivers/net/ethernet/mediatek/mtk_wed.c
2738
block_cb = flow_block_cb_lookup(f->block, cb, dev);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1370
struct mlx5_cmd_prot_block *block;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1390
block = next->buf;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1391
memcpy(block->data, from, copy);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1394
block->token = token;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1403
struct mlx5_cmd_prot_block *block;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1423
block = next->buf;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1425
memcpy(to, block->data, copy);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1467
struct mlx5_cmd_prot_block *block;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1488
block = tmp->buf;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1490
block->next = cpu_to_be64(tmp->next ? tmp->next->dma : 0);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1491
block->block_num = cpu_to_be32(n - i - 1);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1492
block->token = token;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
238
static int verify_block_sig(struct mlx5_cmd_prot_block *block)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
241
int xor_len = sizeof(*block) - sizeof(block->data) - 1;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
243
if (xor8_buf(block, rsvd0_off, xor_len) != 0xff)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
246
if (xor8_buf(block, 0, sizeof(*block)) != 0xff)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
252
static void calc_block_sig(struct mlx5_cmd_prot_block *block)
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
254
int ctrl_xor_len = sizeof(*block) - sizeof(block->data) - 2;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
257
block->ctrl_sig = ~xor8_buf(block, rsvd0_off, ctrl_xor_len);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
258
block->sig = ~xor8_buf(block, 0, sizeof(*block) - 1);
drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c
102
block->version = MLX5_HV_VHCA_STATS_VERSION;
drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c
103
block->rings = priv->stats_nch;
drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c
105
if (!block->command) {
drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c
110
sagent->delay = block->command == MLX5_HV_VHCA_STATS_UPDATE_ONCE ? 0 :
drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c
111
msecs_to_jiffies(block->command * 100);
drivers/net/ethernet/mellanox/mlx5/core/en/hv_vhca_stats.c
94
struct mlx5_hv_vhca_control_block *block)
drivers/net/ethernet/mellanox/mlx5/core/en/rep/tc.c
534
block_cb = flow_block_cb_lookup(f->block, setup_cb, indr_priv);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
115
struct mlx5_hv_vhca_control_block *block)
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
125
if (!(AGENT_MASK(agent->type) & block->control))
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
128
agent->control(agent, block);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
151
struct mlx5_hv_vhca_control_block *block;
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
155
block = kzalloc_obj(*block);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
156
if (!block)
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
159
err = mlx5_hv_read_config(dev, block, sizeof(*block), 0);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
167
memset(block, 0, sizeof(*block));
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
171
if (block->capabilities != capabilities)
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
172
block->capabilities = capabilities;
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
174
if (block->control & ~capabilities)
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
177
mlx5_hv_vhca_agents_control(hv_vhca, block);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
178
block->command_ack = block->command;
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
181
mlx5_hv_write_config(dev, block, sizeof(*block), 0);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
184
kfree(block);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
255
struct mlx5_hv_vhca_control_block *block),
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.c
35
struct mlx5_hv_vhca_control_block *block);
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.h
42
struct mlx5_hv_vhca_control_block *block),
drivers/net/ethernet/mellanox/mlx5/core/lib/hv_vhca.h
83
struct mlx5_hv_vhca_control_block *block),
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
692
struct mlx5_cmd_prot_block *block;
drivers/net/ethernet/mellanox/mlx5/core/pagealloc.c
696
ret = (sizeof(lay->out) + MLX5_BLKS_FOR_RECLAIM_PAGES * sizeof(block->data) -
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1009
mlxsw_afa_block_need_split(const struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1012
struct mlxsw_afa_set *cur_set = block->cur_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1023
static char *mlxsw_afa_block_append_action_ext(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1030
if (block->finished)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1032
if (block->cur_act_index + action_size > block->afa->max_acts_per_set ||
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1033
mlxsw_afa_block_need_split(block, type)) {
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1042
set->prev = block->cur_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1043
block->cur_act_index = 0;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1044
block->cur_set->next = set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1045
block->cur_set = set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1050
block->cur_set->has_trap = true;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1053
block->cur_set->has_police = true;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1059
actions = block->cur_set->ht_key.enc_actions;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1060
oneact = actions + block->cur_act_index * MLXSW_AFA_ONE_ACTION_LEN;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1061
block->cur_act_index += action_size;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1066
static char *mlxsw_afa_block_append_action(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1069
return mlxsw_afa_block_append_action_ext(block, action_code,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1140
int mlxsw_afa_block_append_vlan_modify(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1144
char *act = mlxsw_afa_block_append_action(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1252
static char *mlxsw_afa_block_append_action_trap(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1255
return mlxsw_afa_block_append_action_ext(block, action_code,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1260
static int mlxsw_afa_block_append_drop_plain(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1263
char *act = mlxsw_afa_block_append_action_trap(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1277
mlxsw_afa_block_append_drop_with_cookie(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1287
cookie_ref = mlxsw_afa_cookie_ref_create(block, fa_cookie);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1294
act = mlxsw_afa_block_append_action_trap(block, MLXSW_AFA_TRAPWU_CODE,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1309
mlxsw_afa_cookie_ref_destroy(block, cookie_ref);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1313
int mlxsw_afa_block_append_drop(struct mlxsw_afa_block *block, bool ingress,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1318
mlxsw_afa_block_append_drop_with_cookie(block, ingress,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1320
mlxsw_afa_block_append_drop_plain(block, ingress);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1324
int mlxsw_afa_block_append_trap(struct mlxsw_afa_block *block, u16 trap_id)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1326
char *act = mlxsw_afa_block_append_action_trap(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1338
int mlxsw_afa_block_append_trap_and_forward(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1341
char *act = mlxsw_afa_block_append_action_trap(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1361
mlxsw_afa_mirror_destroy(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1365
block->afa->ops->mirror_del(block->afa->ops_priv,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1373
mlxsw_afa_mirror_destructor(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1379
mlxsw_afa_mirror_destroy(block, mirror);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1383
mlxsw_afa_mirror_create(struct mlxsw_afa_block *block, u16 local_in_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1393
err = block->afa->ops->mirror_add(block->afa->ops_priv,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1402
mlxsw_afa_resource_add(block, &mirror->resource);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1411
mlxsw_afa_block_append_allocated_mirror(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1414
char *act = mlxsw_afa_block_append_action_trap(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1427
mlxsw_afa_block_append_mirror(struct mlxsw_afa_block *block, u16 local_in_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1434
mirror = mlxsw_afa_mirror_create(block, local_in_port, out_dev,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1440
err = mlxsw_afa_block_append_allocated_mirror(block, mirror->span_id);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1449
mlxsw_afa_mirror_destroy(block, mirror);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1553
static int __mlxsw_afa_block_append_qos_dsfield(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1558
char *act = mlxsw_afa_block_append_action(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1578
int mlxsw_afa_block_append_qos_dsfield(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1582
return __mlxsw_afa_block_append_qos_dsfield(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1589
int mlxsw_afa_block_append_qos_dscp(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1592
return __mlxsw_afa_block_append_qos_dsfield(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1599
int mlxsw_afa_block_append_qos_ecn(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1602
return __mlxsw_afa_block_append_qos_dsfield(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1609
int mlxsw_afa_block_append_qos_switch_prio(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1613
char *act = mlxsw_afa_block_append_action(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1666
int mlxsw_afa_block_append_fwd(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1679
fwd_entry_ref = mlxsw_afa_fwd_entry_ref_create(block, local_port);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1686
act = mlxsw_afa_block_append_action(block, MLXSW_AFA_FORWARD_CODE,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1698
mlxsw_afa_fwd_entry_ref_destroy(block, fwd_entry_ref);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1766
int mlxsw_afa_block_append_allocated_counter(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1769
char *act = mlxsw_afa_block_append_action(block, MLXSW_AFA_POLCNT_CODE,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1779
int mlxsw_afa_block_append_counter(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1787
counter = mlxsw_afa_counter_create(block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1794
err = mlxsw_afa_block_append_allocated_counter(block, counter_index);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1804
mlxsw_afa_counter_destroy(block, counter);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1809
int mlxsw_afa_block_append_police(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1818
policer_ref = mlxsw_afa_policer_ref_create(block, fa_index,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1825
act = mlxsw_afa_block_append_action_ext(block, MLXSW_AFA_POLCNT_CODE,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1838
mlxsw_afa_policer_ref_destroy(block, policer_ref);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1875
int mlxsw_afa_block_append_fid_set(struct mlxsw_afa_block *block, u16 fid,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1878
char *act = mlxsw_afa_block_append_action(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1917
int mlxsw_afa_block_append_ignore(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1920
char *act = mlxsw_afa_block_append_action(block, MLXSW_AFA_IGNORE_CODE,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1986
int mlxsw_afa_block_append_mcrouter(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
1990
char *act = mlxsw_afa_block_append_action(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2056
int mlxsw_afa_block_append_ip(struct mlxsw_afa_block *block, bool is_dip,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2064
char *act = mlxsw_afa_block_append_action(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2110
int mlxsw_afa_block_append_l4port(struct mlxsw_afa_block *block, bool is_dport, u16 l4_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2115
char *act = mlxsw_afa_block_append_action(block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2163
static void mlxsw_afa_sampler_destroy(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2167
block->afa->ops->sampler_del(block->afa->ops_priv, sampler->local_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2172
static void mlxsw_afa_sampler_destructor(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2178
mlxsw_afa_sampler_destroy(block, sampler);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2182
mlxsw_afa_sampler_create(struct mlxsw_afa_block *block, u16 local_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2194
err = block->afa->ops->sampler_add(block->afa->ops_priv, local_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2204
mlxsw_afa_resource_add(block, &sampler->resource);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2213
mlxsw_afa_block_append_allocated_sampler(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2216
char *act = mlxsw_afa_block_append_action(block, MLXSW_AFA_SAMPLER_CODE,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2225
int mlxsw_afa_block_append_sampler(struct mlxsw_afa_block *block, u16 local_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2239
sampler = mlxsw_afa_sampler_create(block, local_port, psample_group,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2245
err = mlxsw_afa_block_append_allocated_sampler(block, sampler->span_id,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
2255
mlxsw_afa_sampler_destroy(block, sampler);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
381
void (*destructor)(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
385
static void mlxsw_afa_resource_add(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
388
list_add(&resource->list, &block->resource_list);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
396
static void mlxsw_afa_resources_destroy(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
400
list_for_each_entry_safe(resource, tmp, &block->resource_list, list) {
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
401
resource->destructor(block, resource);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
407
struct mlxsw_afa_block *block;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
409
block = kzalloc_obj(*block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
410
if (!block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
412
INIT_LIST_HEAD(&block->resource_list);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
413
block->afa = mlxsw_afa;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
416
block->first_set = mlxsw_afa_set_create(true);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
417
if (!block->first_set)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
424
block->cur_set = mlxsw_afa_set_create(false);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
425
if (!block->cur_set)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
427
block->cur_set->prev = block->first_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
428
block->first_set->next = block->cur_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
430
block->cur_set = block->first_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
433
return block;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
436
mlxsw_afa_set_destroy(block->first_set);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
438
kfree(block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
443
void mlxsw_afa_block_destroy(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
445
struct mlxsw_afa_set *set = block->first_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
450
mlxsw_afa_set_put(block->afa, set);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
453
mlxsw_afa_resources_destroy(block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
454
kfree(block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
458
int mlxsw_afa_block_commit(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
460
struct mlxsw_afa_set *set = block->cur_set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
463
block->cur_set = NULL;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
464
block->finished = true;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
473
set = mlxsw_afa_set_get(block->afa, set);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
487
block->first_set = set;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
492
char *mlxsw_afa_block_first_set(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
494
return block->first_set->ht_key.enc_actions;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
498
char *mlxsw_afa_block_cur_set(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
500
return block->cur_set->ht_key.enc_actions;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
504
u32 mlxsw_afa_block_first_kvdl_index(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
509
if (WARN_ON(!block->first_set->next))
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
511
return block->first_set->next->kvdl_index;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
515
int mlxsw_afa_block_activity_get(struct mlxsw_afa_block *block, bool *activity)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
517
u32 kvdl_index = mlxsw_afa_block_first_kvdl_index(block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
519
return block->afa->ops->kvdl_set_activity_get(block->afa->ops_priv,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
524
int mlxsw_afa_block_continue(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
526
if (block->finished)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
528
mlxsw_afa_set_goto_set(block->cur_set,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
530
block->finished = true;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
535
int mlxsw_afa_block_jump(struct mlxsw_afa_block *block, u16 group_id)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
537
if (block->finished)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
539
mlxsw_afa_set_goto_set(block->cur_set,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
541
block->finished = true;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
546
int mlxsw_afa_block_terminate(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
548
if (block->finished)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
550
mlxsw_afa_set_goto_set(block->cur_set,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
552
block->finished = true;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
630
mlxsw_afa_fwd_entry_ref_destroy(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
634
mlxsw_afa_fwd_entry_put(block->afa, fwd_entry_ref->fwd_entry);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
639
mlxsw_afa_fwd_entry_ref_destructor(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
646
mlxsw_afa_fwd_entry_ref_destroy(block, fwd_entry_ref);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
650
mlxsw_afa_fwd_entry_ref_create(struct mlxsw_afa_block *block, u16 local_port)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
659
fwd_entry = mlxsw_afa_fwd_entry_get(block->afa, local_port);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
666
mlxsw_afa_resource_add(block, &fwd_entry_ref->resource);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
680
mlxsw_afa_counter_destroy(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
684
block->afa->ops->counter_index_put(block->afa->ops_priv,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
690
mlxsw_afa_counter_destructor(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
696
mlxsw_afa_counter_destroy(block, counter);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
700
mlxsw_afa_counter_create(struct mlxsw_afa_block *block)
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
709
err = block->afa->ops->counter_index_get(block->afa->ops_priv,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
714
mlxsw_afa_resource_add(block, &counter->resource);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
823
mlxsw_afa_cookie_ref_destroy(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
827
mlxsw_afa_cookie_put(block->afa, cookie_ref->cookie);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
832
mlxsw_afa_cookie_ref_destructor(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
839
mlxsw_afa_cookie_ref_destroy(block, cookie_ref);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
843
mlxsw_afa_cookie_ref_create(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
853
cookie = mlxsw_afa_cookie_get(block->afa, fa_cookie);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
860
mlxsw_afa_resource_add(block, &cookie_ref->resource);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
949
mlxsw_afa_policer_ref_destroy(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
953
mlxsw_afa_policer_put(block->afa, policer_ref->policer);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
958
mlxsw_afa_policer_ref_destructor(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
965
mlxsw_afa_policer_ref_destroy(block, policer_ref);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
969
mlxsw_afa_policer_ref_create(struct mlxsw_afa_block *block, u32 fa_index,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
981
policer = mlxsw_afa_policer_get(block->afa, fa_index, rate_bytes_ps,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.c
990
mlxsw_afa_resource_add(block, &policer_ref->resource);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
100
int mlxsw_afa_block_append_l4port(struct mlxsw_afa_block *block, bool is_dport, u16 l4_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
102
int mlxsw_afa_block_append_police(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
106
int mlxsw_afa_block_append_sampler(struct mlxsw_afa_block *block, u16 local_port,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
47
void mlxsw_afa_block_destroy(struct mlxsw_afa_block *block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
48
int mlxsw_afa_block_commit(struct mlxsw_afa_block *block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
49
char *mlxsw_afa_block_first_set(struct mlxsw_afa_block *block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
50
char *mlxsw_afa_block_cur_set(struct mlxsw_afa_block *block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
51
u32 mlxsw_afa_block_first_kvdl_index(struct mlxsw_afa_block *block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
52
int mlxsw_afa_block_activity_get(struct mlxsw_afa_block *block, bool *activity);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
53
int mlxsw_afa_block_continue(struct mlxsw_afa_block *block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
54
int mlxsw_afa_block_jump(struct mlxsw_afa_block *block, u16 group_id);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
55
int mlxsw_afa_block_terminate(struct mlxsw_afa_block *block);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
58
int mlxsw_afa_block_append_drop(struct mlxsw_afa_block *block, bool ingress,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
61
int mlxsw_afa_block_append_trap(struct mlxsw_afa_block *block, u16 trap_id);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
62
int mlxsw_afa_block_append_trap_and_forward(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
64
int mlxsw_afa_block_append_mirror(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
69
int mlxsw_afa_block_append_fwd(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
72
int mlxsw_afa_block_append_vlan_modify(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
75
int mlxsw_afa_block_append_qos_switch_prio(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
78
int mlxsw_afa_block_append_qos_dsfield(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
81
int mlxsw_afa_block_append_qos_dscp(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
83
int mlxsw_afa_block_append_qos_ecn(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
85
int mlxsw_afa_block_append_allocated_counter(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
87
int mlxsw_afa_block_append_counter(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
90
int mlxsw_afa_block_append_fid_set(struct mlxsw_afa_block *block, u16 fid,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
92
int mlxsw_afa_block_append_ignore(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
94
int mlxsw_afa_block_append_mcrouter(struct mlxsw_afa_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_actions.h
97
int mlxsw_afa_block_append_ip(struct mlxsw_afa_block *block, bool is_dip,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
154
const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i];
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
156
for (j = 0; j < block->instances_count; j++) {
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
159
elinst = &block->instances[j];
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
382
mlxsw_afk_block_elinst_get(const struct mlxsw_afk_block *block,
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
387
for (i = 0; i < block->instances_count; i++) {
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
390
elinst = &block->instances[i];
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
403
const struct mlxsw_afk_block *block;
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
409
block = key_info->blocks[block_index];
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
411
elinst = mlxsw_afk_block_elinst_get(block, element);
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
66
const struct mlxsw_afk_block *block = &mlxsw_afk->blocks[i];
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
68
for (j = 0; j < block->instances_count; j++) {
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.c
72
elinst = &block->instances[j];
drivers/net/ethernet/mellanox/mlxsw/core_acl_flex_keys.h
190
void (*encode_block)(char *output, int block_index, char *block);
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1015
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1049
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1213
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1215
void mlxsw_sp_mall_destroy(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1217
int mlxsw_sp_mall_port_bind(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1220
void mlxsw_sp_mall_port_unbind(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1222
int mlxsw_sp_mall_prio_get(struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1227
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1230
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1233
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1236
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1239
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1242
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
900
mlxsw_sp_flow_block_mlxsw_sp(struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
902
return block->mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
906
mlxsw_sp_flow_block_rule_count(const struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
908
return block ? block->rule_count : 0;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
912
mlxsw_sp_flow_block_disable_inc(struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
914
if (block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
915
block->disable_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
919
mlxsw_sp_flow_block_disable_dec(struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
921
if (block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
922
block->disable_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
926
mlxsw_sp_flow_block_disabled(const struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
928
return block->disable_count;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
932
mlxsw_sp_flow_block_is_egress_bound(const struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
934
return block->egress_binding_count;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
938
mlxsw_sp_flow_block_is_ingress_bound(const struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
940
return block->ingress_binding_count;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
944
mlxsw_sp_flow_block_is_mixed_bound(const struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
946
return block->ingress_binding_count && block->egress_binding_count;
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
951
void mlxsw_sp_flow_block_destroy(struct mlxsw_sp_flow_block *block);
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
968
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
971
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
975
struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
979
struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
107
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
110
struct mlxsw_sp_acl_ruleset *ruleset = block->ruleset_zero;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
118
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
121
struct mlxsw_sp_acl_ruleset *ruleset = block->ruleset_zero;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
131
struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
136
block->ruleset_zero = ruleset;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
137
list_for_each_entry(binding, &block->binding_list, list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
138
err = mlxsw_sp_acl_ruleset_bind(mlxsw_sp, block, binding);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
145
list_for_each_entry_continue_reverse(binding, &block->binding_list,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
147
mlxsw_sp_acl_ruleset_unbind(mlxsw_sp, block, binding);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
148
block->ruleset_zero = NULL;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
156
struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
160
list_for_each_entry(binding, &block->binding_list, list)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
161
mlxsw_sp_acl_ruleset_unbind(mlxsw_sp, block, binding);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
162
block->ruleset_zero = NULL;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
167
struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
181
ruleset->ht_key.block = block;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
239
struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
245
ht_key.block = block;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
254
struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
264
ruleset = __mlxsw_sp_acl_ruleset_lookup(acl, block, chain_index, ops);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
272
struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
284
ruleset = __mlxsw_sp_acl_ruleset_lookup(acl, block, chain_index, ops);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
289
return mlxsw_sp_acl_ruleset_create(mlxsw_sp, block, chain_index, ops,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
450
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
457
if (!list_is_singular(&block->binding_list)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
461
binding = list_first_entry(&block->binding_list,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
50
struct mlxsw_sp_flow_block *block;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
790
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
798
if (!list_is_singular(&block->binding_list)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
802
binding = list_first_entry(&block->binding_list,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
864
struct mlxsw_sp_flow_block *block = ruleset->ht_key.block;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
882
err = mlxsw_sp_acl_ruleset_block_bind(mlxsw_sp, ruleset, block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
890
block->rule_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
891
block->ingress_blocker_rule_count += rule->rulei->ingress_bind_blocker;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
892
block->egress_blocker_rule_count += rule->rulei->egress_bind_blocker;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
908
struct mlxsw_sp_flow_block *block = ruleset->ht_key.block;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
910
block->egress_blocker_rule_count -= rule->rulei->egress_bind_blocker;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
911
block->ingress_blocker_rule_count -= rule->rulei->ingress_bind_blocker;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
912
block->rule_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl.c
918
mlxsw_sp_acl_ruleset_block_unbind(mlxsw_sp, ruleset, block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_keys.c
104
char *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_keys.c
109
memcpy(output_indexed, block, MLXSW_SP1_AFK_KEY_BLOCK_SIZE);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_keys.c
246
MLXSW_ITEM64(sp2_afk, block, value, 0x00, 0, MLXSW_SP2_AFK_BITS_PER_BLOCK);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_keys.c
303
char *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_keys.c
305
u64 block_value = mlxsw_sp2_afk_block_value_get(block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
102
mlxsw_sp_mall_port_unbind(block, mlxsw_sp_port);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
108
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
114
binding = mlxsw_sp_flow_block_lookup(block, mlxsw_sp_port, ingress);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
121
block->ingress_binding_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
123
block->egress_binding_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
125
if (mlxsw_sp_flow_block_ruleset_bound(block))
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
126
mlxsw_sp_acl_ruleset_unbind(mlxsw_sp, block, binding);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
130
mlxsw_sp_mall_port_unbind(block, mlxsw_sp_port);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
15
struct mlxsw_sp_flow_block *block;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
17
block = kzalloc_obj(*block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
18
if (!block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
20
INIT_LIST_HEAD(&block->binding_list);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
21
INIT_LIST_HEAD(&block->mall.list);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
211
block_cb = flow_block_cb_lookup(f->block, mlxsw_sp_flow_block_cb,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
22
block->mlxsw_sp = mlxsw_sp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
23
block->net = net;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
24
return block;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
261
block_cb = flow_block_cb_lookup(f->block, mlxsw_sp_flow_block_cb,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
27
void mlxsw_sp_flow_block_destroy(struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
29
WARN_ON(!list_empty(&block->binding_list));
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
30
kfree(block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
34
mlxsw_sp_flow_block_lookup(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
39
list_for_each_entry(binding, &block->binding_list, list)
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
47
mlxsw_sp_flow_block_ruleset_bound(const struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
49
return block->ruleset_zero;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
53
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
61
if (WARN_ON(mlxsw_sp_flow_block_lookup(block, mlxsw_sp_port, ingress)))
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
64
if (ingress && block->ingress_blocker_rule_count) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
69
if (!ingress && block->egress_blocker_rule_count) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
74
err = mlxsw_sp_mall_port_bind(block, mlxsw_sp_port, extack);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
86
if (mlxsw_sp_flow_block_ruleset_bound(block)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
87
err = mlxsw_sp_acl_ruleset_bind(mlxsw_sp, block, binding);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
93
block->ingress_binding_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
95
block->egress_binding_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_flow.c
96
list_add(&binding->list, &block->binding_list);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
100
if (mlxsw_sp_flow_block_is_mixed_bound(block)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
104
ingress = mlxsw_sp_flow_block_is_ingress_bound(block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
134
ruleset = mlxsw_sp_acl_ruleset_lookup(mlxsw_sp, block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
153
if (mlxsw_sp_flow_block_is_egress_bound(block)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
201
block, out_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
279
block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
306
const struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
321
ingress_dev = __dev_get_by_index(block->net,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
334
if (mlxsw_sp_port->mlxsw_sp != block->mlxsw_sp) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
349
struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
362
return mlxsw_sp_flower_parse_meta_iif(rulei, block, &match,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
568
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
59
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
599
err = mlxsw_sp_flower_parse_meta(rulei, f, block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
662
if (mlxsw_sp_flow_block_is_egress_bound(block) &&
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
707
return mlxsw_sp_flower_parse_actions(mlxsw_sp, block, rulei,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
712
static int mlxsw_sp_flower_mall_prio_check(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
715
bool ingress = mlxsw_sp_flow_block_is_ingress_bound(block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
720
err = mlxsw_sp_mall_prio_get(block, f->common.chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
741
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
749
err = mlxsw_sp_flower_mall_prio_check(block, f);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
753
ruleset = mlxsw_sp_acl_ruleset_get(mlxsw_sp, block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
767
err = mlxsw_sp_flower_parse(mlxsw_sp, block, rulei, f);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
792
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
798
ruleset = mlxsw_sp_acl_ruleset_get(mlxsw_sp, block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
814
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
826
ruleset = mlxsw_sp_acl_ruleset_get(mlxsw_sp, block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
855
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
863
err = mlxsw_sp_flower_parse(mlxsw_sp, block, &rulei, f);
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
866
ruleset = mlxsw_sp_acl_ruleset_get(mlxsw_sp, block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
876
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
881
ruleset = mlxsw_sp_acl_ruleset_get(mlxsw_sp, block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
892
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_flower.c
898
ruleset = mlxsw_sp_acl_ruleset_lookup(mlxsw_sp, block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
14
mlxsw_sp_mall_entry_find(struct mlxsw_sp_flow_block *block, unsigned long cookie)
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
18
list_for_each_entry(mall_entry, &block->mall.list, list)
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
184
static void mlxsw_sp_mall_prio_update(struct mlxsw_sp_flow_block *block)
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
188
if (list_empty(&block->mall.list))
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
190
block->mall.min_prio = UINT_MAX;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
191
block->mall.max_prio = 0;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
192
list_for_each_entry(mall_entry, &block->mall.list, list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
193
if (mall_entry->priority < block->mall.min_prio)
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
194
block->mall.min_prio = mall_entry->priority;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
195
if (mall_entry->priority > block->mall.max_prio)
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
196
block->mall.max_prio = mall_entry->priority;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
201
struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
223
if (mlxsw_sp_flow_block_is_mixed_bound(block)) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
228
err = mlxsw_sp_flower_prio_get(mlxsw_sp, block, f->common.chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
251
mall_entry->ingress = mlxsw_sp_flow_block_is_ingress_bound(block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
285
list_for_each_entry(binding, &block->binding_list, list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
292
block->rule_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
294
block->egress_blocker_rule_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
296
block->ingress_blocker_rule_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
297
list_add_tail(&mall_entry->list, &block->mall.list);
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
298
mlxsw_sp_mall_prio_update(block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
302
list_for_each_entry_continue_reverse(binding, &block->binding_list,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
310
void mlxsw_sp_mall_destroy(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
316
mall_entry = mlxsw_sp_mall_entry_find(block, f->cookie);
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
324
block->egress_blocker_rule_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
326
block->ingress_blocker_rule_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
327
block->rule_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
328
list_for_each_entry(binding, &block->binding_list, list)
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
331
mlxsw_sp_mall_prio_update(block);
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
334
int mlxsw_sp_mall_port_bind(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
341
list_for_each_entry(mall_entry, &block->mall.list, list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
350
list_for_each_entry_continue_reverse(mall_entry, &block->mall.list,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
356
void mlxsw_sp_mall_port_unbind(struct mlxsw_sp_flow_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
361
list_for_each_entry(mall_entry, &block->mall.list, list)
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
365
int mlxsw_sp_mall_prio_get(struct mlxsw_sp_flow_block *block, u32 chain_index,
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
368
if (chain_index || list_empty(&block->mall.list))
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
374
*p_min_prio = block->mall.min_prio;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
375
*p_max_prio = block->mall.max_prio;
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
1978
mlxsw_sp_qevent_mall_entry_find(struct mlxsw_sp_qevent_block *block, unsigned long cookie)
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
1982
list_for_each_entry(mall_entry, &block->mall_entry_list, list)
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
2153
mlxsw_sp_qevent_binding_lookup(struct mlxsw_sp_qevent_block *block,
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
2160
list_for_each_entry(qevent_binding, &block->binding_list, list)
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
2183
block_cb = flow_block_cb_lookup(f->block, mlxsw_sp_qevent_block_cb, mlxsw_sp);
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
2257
block_cb = flow_block_cb_lookup(f->block, mlxsw_sp_qevent_block_cb, mlxsw_sp);
drivers/net/ethernet/mscc/ocelot.c
1662
block_vcap_is2 = &ocelot->block[VCAP_IS2];
drivers/net/ethernet/mscc/ocelot.c
1707
block_vcap_is2 = &ocelot->block[VCAP_IS2];
drivers/net/ethernet/mscc/ocelot.c
519
struct ocelot_vcap_block *block_vcap_is1 = &ocelot->block[VCAP_IS1];
drivers/net/ethernet/mscc/ocelot.c
751
struct ocelot_vcap_block *block = &ocelot->block[VCAP_IS1];
drivers/net/ethernet/mscc/ocelot.c
757
list_for_each_entry(filter, &block->rules, list) {
drivers/net/ethernet/mscc/ocelot_flower.c
1005
struct ocelot_vcap_block *block;
drivers/net/ethernet/mscc/ocelot_flower.c
1025
block = &ocelot->block[block_id];
drivers/net/ethernet/mscc/ocelot_flower.c
1027
filter = ocelot_vcap_block_find_filter_by_id(block, f->cookie, true);
drivers/net/ethernet/mscc/ocelot_flower.c
148
struct ocelot_vcap_block *block;
drivers/net/ethernet/mscc/ocelot_flower.c
156
block = &ocelot->block[VCAP_IS1];
drivers/net/ethernet/mscc/ocelot_flower.c
158
list_for_each_entry(filter, &block->rules, list)
drivers/net/ethernet/mscc/ocelot_flower.c
914
filter = ocelot_vcap_block_find_filter_by_id(&ocelot->block[block_id],
drivers/net/ethernet/mscc/ocelot_flower.c
968
struct ocelot_vcap_block *block;
drivers/net/ethernet/mscc/ocelot_flower.c
982
block = &ocelot->block[block_id];
drivers/net/ethernet/mscc/ocelot_flower.c
984
filter = ocelot_vcap_block_find_filter_by_id(block, f->cookie, true);
drivers/net/ethernet/mscc/ocelot_mrp.c
41
block_vcap_is2 = &ocelot->block[VCAP_IS2];
drivers/net/ethernet/mscc/ocelot_net.c
466
block_cb = flow_block_cb_lookup(f->block, cb, priv);
drivers/net/ethernet/mscc/ocelot_vcap.c
1004
block->count++;
drivers/net/ethernet/mscc/ocelot_vcap.c
1006
list_for_each_entry(tmp, &block->rules, list) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1023
static int ocelot_vcap_block_get_filter_index(struct ocelot_vcap_block *block,
drivers/net/ethernet/mscc/ocelot_vcap.c
1029
list_for_each_entry(tmp, &block->rules, list) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1039
ocelot_vcap_block_find_filter_by_index(struct ocelot_vcap_block *block,
drivers/net/ethernet/mscc/ocelot_vcap.c
1045
list_for_each_entry(tmp, &block->rules, list) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1055
ocelot_vcap_block_find_filter_by_id(struct ocelot_vcap_block *block,
drivers/net/ethernet/mscc/ocelot_vcap.c
1060
list_for_each_entry(filter, &block->rules, list)
drivers/net/ethernet/mscc/ocelot_vcap.c
1142
struct ocelot_vcap_block *block = &ocelot->block[filter->block_id];
drivers/net/ethernet/mscc/ocelot_vcap.c
1153
for (i = 0; i < block->count; i++) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1154
tmp = ocelot_vcap_block_find_filter_by_index(block, i);
drivers/net/ethernet/mscc/ocelot_vcap.c
1167
for (i = 0; i < block->count; i++) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1168
tmp = ocelot_vcap_block_find_filter_by_index(block, i);
drivers/net/ethernet/mscc/ocelot_vcap.c
1188
struct ocelot_vcap_block *block = &ocelot->block[filter->block_id];
drivers/net/ethernet/mscc/ocelot_vcap.c
1198
ret = ocelot_vcap_filter_add_to_block(ocelot, block, filter, extack);
drivers/net/ethernet/mscc/ocelot_vcap.c
1203
index = ocelot_vcap_block_get_filter_index(block, filter);
drivers/net/ethernet/mscc/ocelot_vcap.c
1208
for (i = block->count - 1; i > index; i--) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1211
tmp = ocelot_vcap_block_find_filter_by_index(block, i);
drivers/net/ethernet/mscc/ocelot_vcap.c
1224
struct ocelot_vcap_block *block,
drivers/net/ethernet/mscc/ocelot_vcap.c
1229
list_for_each_entry_safe(tmp, n, &block->rules, list) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1237
block->count--;
drivers/net/ethernet/mscc/ocelot_vcap.c
1243
struct ocelot_vcap_block *block = &ocelot->block[filter->block_id];
drivers/net/ethernet/mscc/ocelot_vcap.c
1254
index = ocelot_vcap_block_get_filter_index(block, filter);
drivers/net/ethernet/mscc/ocelot_vcap.c
1259
ocelot_vcap_block_remove_filter(ocelot, block, filter);
drivers/net/ethernet/mscc/ocelot_vcap.c
1262
for (i = index; i < block->count; i++) {
drivers/net/ethernet/mscc/ocelot_vcap.c
1265
tmp = ocelot_vcap_block_find_filter_by_index(block, i);
drivers/net/ethernet/mscc/ocelot_vcap.c
1272
vcap_entry_set(ocelot, block->count, &del_filter);
drivers/net/ethernet/mscc/ocelot_vcap.c
1281
struct ocelot_vcap_block *block = &ocelot->block[filter->block_id];
drivers/net/ethernet/mscc/ocelot_vcap.c
1284
index = ocelot_vcap_block_get_filter_index(block, filter);
drivers/net/ethernet/mscc/ocelot_vcap.c
1297
struct ocelot_vcap_block *block = &ocelot->block[filter->block_id];
drivers/net/ethernet/mscc/ocelot_vcap.c
1301
index = ocelot_vcap_block_get_filter_index(block, filter);
drivers/net/ethernet/mscc/ocelot_vcap.c
1419
struct ocelot_vcap_block *block = &ocelot->block[i];
drivers/net/ethernet/mscc/ocelot_vcap.c
1422
INIT_LIST_HEAD(&block->rules);
drivers/net/ethernet/mscc/ocelot_vcap.c
992
struct ocelot_vcap_block *block,
drivers/net/ethernet/mscc/ocelot_vcap.c
996
struct list_head *pos = &block->rules;
drivers/net/ethernet/netronome/nfp/flower/offload.c
1811
block_cb = flow_block_cb_lookup(f->block,
drivers/net/ethernet/netronome/nfp/flower/offload.c
1930
block_cb = flow_block_cb_lookup(f->block,
drivers/net/ethernet/qlogic/qed/qed_dbg_hsi.h
1134
enum block_id block,
drivers/net/ethernet/qlogic/qed/qed_debug.c
1474
const struct dbg_block_chip *block;
drivers/net/ethernet/qlogic/qed/qed_debug.c
1476
block = qed_get_dbg_block_per_chip(p_hwfn, BLOCK_DBG);
drivers/net/ethernet/qlogic/qed/qed_debug.c
1477
reset_reg = qed_get_dbg_reset_reg(p_hwfn, block->reset_reg_id);
drivers/net/ethernet/qlogic/qed/qed_debug.c
1483
old_reset_reg_val & ~BIT(block->reset_reg_bit_offset);
drivers/net/ethernet/qlogic/qed/qed_debug.c
1506
const struct dbg_block_chip *block =
drivers/net/ethernet/qlogic/qed/qed_debug.c
1509
qed_wr(p_hwfn, p_ptt, DWORDS_TO_BYTES(block->dbg_select_reg_addr),
drivers/net/ethernet/qlogic/qed/qed_debug.c
1511
qed_wr(p_hwfn, p_ptt, DWORDS_TO_BYTES(block->dbg_dword_enable_reg_addr),
drivers/net/ethernet/qlogic/qed/qed_debug.c
1513
qed_wr(p_hwfn, p_ptt, DWORDS_TO_BYTES(block->dbg_shift_reg_addr),
drivers/net/ethernet/qlogic/qed/qed_debug.c
1515
qed_wr(p_hwfn, p_ptt, DWORDS_TO_BYTES(block->dbg_force_valid_reg_addr),
drivers/net/ethernet/qlogic/qed/qed_debug.c
1517
qed_wr(p_hwfn, p_ptt, DWORDS_TO_BYTES(block->dbg_force_frame_reg_addr),
drivers/net/ethernet/qlogic/qed/qed_debug.c
1598
const struct dbg_block *block;
drivers/net/ethernet/qlogic/qed/qed_debug.c
1601
block = get_dbg_block(p_hwfn, block_id);
drivers/net/ethernet/qlogic/qed/qed_debug.c
1604
if (block->associated_storm_letter) {
drivers/net/ethernet/qlogic/qed/qed_debug.c
1606
qed_get_id_from_letter(block->associated_storm_letter);
drivers/net/ethernet/qlogic/qed/qed_debug.c
1716
const struct dbg_block_chip *block;
drivers/net/ethernet/qlogic/qed/qed_debug.c
1718
block = qed_get_dbg_block_per_chip(p_hwfn,
drivers/net/ethernet/qlogic/qed/qed_debug.c
1722
GET_FIELD(block->flags, DBG_BLOCK_CHIP_IS_REMOVED);
drivers/net/ethernet/qlogic/qed/qed_debug.c
1724
GET_FIELD(block->flags,
drivers/net/ethernet/qlogic/qed/qed_debug.c
1727
GET_FIELD(block->flags,
drivers/net/ethernet/qlogic/qed/qed_debug.c
1731
reg_val[block->reset_reg_id] |=
drivers/net/ethernet/qlogic/qed/qed_debug.c
1732
BIT(block->reset_reg_bit_offset);
drivers/net/ethernet/qlogic/qed/qed_debug.c
2728
const struct dbg_block *block;
drivers/net/ethernet/qlogic/qed/qed_debug.c
2752
block = get_dbg_block(p_hwfn,
drivers/net/ethernet/qlogic/qed/qed_debug.c
2758
if (block->associated_storm_letter)
drivers/net/ethernet/qlogic/qed/qed_debug.c
2759
storm_letter = block->associated_storm_letter;
drivers/net/ethernet/qlogic/qed/qed_debug.c
3448
const struct dbg_block *block;
drivers/net/ethernet/qlogic/qed/qed_debug.c
3477
block = get_dbg_block(p_hwfn, (enum block_id)block_id);
drivers/net/ethernet/qlogic/qed/qed_debug.c
3481
block->name,
drivers/net/ethernet/qlogic/qed/qed_debug.c
419
#define NUM_EXTRA_DBG_LINES(block) \
drivers/net/ethernet/qlogic/qed/qed_debug.c
420
(GET_FIELD((block)->flags, DBG_BLOCK_CHIP_HAS_LATENCY_EVENTS) ? 2 : 1)
drivers/net/ethernet/qlogic/qed/qed_debug.c
421
#define NUM_DBG_LINES(block) \
drivers/net/ethernet/qlogic/qed/qed_debug.c
422
((block)->num_of_dbg_bus_lines + NUM_EXTRA_DBG_LINES(block))
drivers/net/ethernet/qlogic/qed/qed_debug.c
6468
const struct dbg_block_user *block =
drivers/net/ethernet/qlogic/qed/qed_debug.c
6472
return (const char *)block->name;
drivers/net/ethernet/sfc/tc_bindings.c
110
block_cb = flow_block_cb_lookup(tcb->block,
drivers/net/ethernet/sfc/tc_bindings.c
170
binding = efx_tc_create_binding(efx, NULL, net_dev, tcb->block);
drivers/net/ethernet/sfc/tc_bindings.c
192
block_cb = flow_block_cb_lookup(tcb->block,
drivers/net/ethernet/sfc/tc_bindings.c
20
struct flow_block *block;
drivers/net/ethernet/sfc/tc_bindings.c
60
struct net_device *otherdev, struct flow_block *block)
drivers/net/ethernet/sfc/tc_bindings.c
69
binding->block = block;
drivers/net/ethernet/sfc/tc_bindings.c
89
binding = efx_tc_create_binding(efx, efv, net_dev, tcb->block);
drivers/net/phy/mscc/mscc_macsec.c
108
bool block)
drivers/net/phy/mscc/mscc_macsec.c
114
if (block)
drivers/net/usb/r8152.c
4661
struct fw_block *block = (struct fw_block *)&fw->data[i];
drivers/net/usb/r8152.c
4664
if ((i + sizeof(*block)) > fw->size)
drivers/net/usb/r8152.c
4667
type = __le32_to_cpu(block->type);
drivers/net/usb/r8152.c
4670
if (__le32_to_cpu(block->length) != sizeof(*block))
drivers/net/usb/r8152.c
4680
if (!rtl8152_is_fw_mac_ok(tp, (struct fw_mac *)block)) {
drivers/net/usb/r8152.c
4694
if (!rtl8152_is_fw_mac_ok(tp, (struct fw_mac *)block)) {
drivers/net/usb/r8152.c
4714
if (__le32_to_cpu(block->length) != sizeof(struct fw_phy_patch_key)) {
drivers/net/usb/r8152.c
4729
if (__le32_to_cpu(block->length) != sizeof(*block)) {
drivers/net/usb/r8152.c
4750
if (!rtl8152_is_fw_phy_nc_ok(tp, (struct fw_phy_nc *)block)) {
drivers/net/usb/r8152.c
4773
if (!rtl8152_is_fw_phy_union_ok(tp, (struct fw_phy_union *)block)) {
drivers/net/usb/r8152.c
4794
if (!rtl8152_is_fw_phy_union_ok(tp, (struct fw_phy_union *)block)) {
drivers/net/usb/r8152.c
4814
if (!rtl8152_is_fw_phy_union_ok(tp, (struct fw_phy_union *)block)) {
drivers/net/usb/r8152.c
4833
if (!rtl8152_is_fw_phy_union_ok(tp, (struct fw_phy_union *)block)) {
drivers/net/usb/r8152.c
4851
if (!rtl8152_is_fw_phy_union_ok(tp, (struct fw_phy_union *)block)) {
drivers/net/usb/r8152.c
4858
if (!rtl8152_is_fw_phy_union_ok(tp, (struct fw_phy_union *)block)) {
drivers/net/usb/r8152.c
4864
if (!rtl8152_is_fw_phy_fixup_ok(tp, (struct fw_phy_fixup *)block)) {
drivers/net/usb/r8152.c
4875
if (!rtl8152_is_fw_phy_speed_up_ok(tp, (struct fw_phy_speed_up *)block)) {
drivers/net/usb/r8152.c
4898
if (!rtl8152_is_fw_phy_ver_ok(tp, (struct fw_phy_ver *)block)) {
drivers/net/usb/r8152.c
4911
i += ALIGN(__le32_to_cpu(block->length), 8);
drivers/net/usb/r8152.c
5197
struct fw_block *block = (struct fw_block *)&fw->data[i];
drivers/net/usb/r8152.c
5199
switch (__le32_to_cpu(block->type)) {
drivers/net/usb/r8152.c
5204
rtl8152_fw_mac_apply(tp, (struct fw_mac *)block);
drivers/net/usb/r8152.c
5209
key = (struct fw_phy_patch_key *)block;
drivers/net/usb/r8152.c
5220
rtl8152_fw_phy_nc_apply(tp, (struct fw_phy_nc *)block);
drivers/net/usb/r8152.c
5223
patch_phy = rtl8152_fw_phy_ver(tp, (struct fw_phy_ver *)block);
drivers/net/usb/r8152.c
5232
rtl8152_fw_phy_union_apply(tp, (struct fw_phy_union *)block);
drivers/net/usb/r8152.c
5236
rtl8152_fw_phy_fixup(tp, (struct fw_phy_fixup *)block);
drivers/net/usb/r8152.c
5239
rtl_ram_code_speed_up(tp, (struct fw_phy_speed_up *)block, !power_cut);
drivers/net/usb/r8152.c
5245
i += ALIGN(__le32_to_cpu(block->length), 8);
drivers/net/usb/usbnet.c
633
goto block;
drivers/net/usb/usbnet.c
648
block:
drivers/net/wireless/ath/ath12k/qmi.h
628
static inline void ath12k_qmi_set_event_block(struct ath12k_qmi *qmi, bool block)
drivers/net/wireless/ath/ath12k/qmi.h
632
qmi->block_event = block;
drivers/net/wireless/ath/ath5k/ath5k.h
1324
spinlock_t block; /* protects beacon */
drivers/net/wireless/ath/ath5k/base.c
2195
spin_lock_bh(&ah->block);
drivers/net/wireless/ath/ath5k/base.c
2221
spin_unlock_bh(&ah->block);
drivers/net/wireless/ath/ath5k/base.c
2247
spin_lock(&ah->block);
drivers/net/wireless/ath/ath5k/base.c
2249
spin_unlock(&ah->block);
drivers/net/wireless/ath/ath5k/base.c
2623
spin_lock_init(&ah->block);
drivers/net/wireless/ath/ath5k/mac80211-ops.c
302
spin_lock_bh(&ah->block);
drivers/net/wireless/ath/ath5k/mac80211-ops.c
304
spin_unlock_bh(&ah->block);
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3146
u8 *block,
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3158
offset = block[it];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3161
length = block[it+1];
drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
3168
memcpy(&mptr[spot], &block[it+2], length);
drivers/net/wireless/ath/wil6210/fw_inc.c
362
const struct wil_fw_data_dwrite *block = d->data;
drivers/net/wireless/ath/wil6210/fw_inc.c
365
if (size % sizeof(*block)) {
drivers/net/wireless/ath/wil6210/fw_inc.c
367
sizeof(*block), size);
drivers/net/wireless/ath/wil6210/fw_inc.c
370
n = size / sizeof(*block);
drivers/net/wireless/ath/wil6210/fw_inc.c
374
u32 m = le32_to_cpu(block[i].mask);
drivers/net/wireless/ath/wil6210/fw_inc.c
375
u32 v = le32_to_cpu(block[i].value);
drivers/net/wireless/ath/wil6210/fw_inc.c
378
if (!wil_fw_addr_check(wil, &dst, block[i].addr, 0, "address"))
drivers/net/wireless/ath/wil6210/fw_inc.c
385
le32_to_cpu(block[i].addr), y, x, v, m);
drivers/net/wireless/ath/wil6210/fw_inc.c
419
const struct wil_fw_data_gw *block = d->data;
drivers/net/wireless/ath/wil6210/fw_inc.c
427
if (size < sizeof(*d) + sizeof(*block)) {
drivers/net/wireless/ath/wil6210/fw_inc.c
432
if ((size - sizeof(*d)) % sizeof(*block)) {
drivers/net/wireless/ath/wil6210/fw_inc.c
435
sizeof(*block), size - sizeof(*d));
drivers/net/wireless/ath/wil6210/fw_inc.c
438
n = (size - sizeof(*d)) / sizeof(*block);
drivers/net/wireless/ath/wil6210/fw_inc.c
464
u32 a = le32_to_cpu(block[i].addr);
drivers/net/wireless/ath/wil6210/fw_inc.c
465
u32 v = le32_to_cpu(block[i].value);
drivers/net/wireless/ath/wil6210/fw_inc.c
483
const struct wil_fw_data_gw4 *block = d->data;
drivers/net/wireless/ath/wil6210/fw_inc.c
485
void __iomem *gwa_val[ARRAY_SIZE(block->value)];
drivers/net/wireless/ath/wil6210/fw_inc.c
491
if (size < sizeof(*d) + sizeof(*block)) {
drivers/net/wireless/ath/wil6210/fw_inc.c
496
if ((size - sizeof(*d)) % sizeof(*block)) {
drivers/net/wireless/ath/wil6210/fw_inc.c
499
sizeof(*block), size - sizeof(*d));
drivers/net/wireless/ath/wil6210/fw_inc.c
502
n = (size - sizeof(*d)) / sizeof(*block);
drivers/net/wireless/ath/wil6210/fw_inc.c
512
for (k = 0; k < ARRAY_SIZE(block->value); k++)
drivers/net/wireless/ath/wil6210/fw_inc.c
533
u32 a = le32_to_cpu(block[i].addr);
drivers/net/wireless/ath/wil6210/fw_inc.c
534
u32 v[ARRAY_SIZE(block->value)];
drivers/net/wireless/ath/wil6210/fw_inc.c
536
for (k = 0; k < ARRAY_SIZE(block->value); k++)
drivers/net/wireless/ath/wil6210/fw_inc.c
537
v[k] = le32_to_cpu(block[i].value[k]);
drivers/net/wireless/ath/wil6210/fw_inc.c
543
for (k = 0; k < ARRAY_SIZE(block->value); k++)
drivers/net/wireless/atmel/at76c50x-usb.c
1253
u8 *block;
drivers/net/wireless/atmel/at76c50x-usb.c
1268
block = kmalloc(FW_BLOCK_SIZE, GFP_KERNEL);
drivers/net/wireless/atmel/at76c50x-usb.c
1269
if (!block)
drivers/net/wireless/atmel/at76c50x-usb.c
1277
memcpy(block, buf, bsize);
drivers/net/wireless/atmel/at76c50x-usb.c
1281
ret = at76_load_ext_fw_block(udev, blockno, block, bsize);
drivers/net/wireless/atmel/at76c50x-usb.c
1300
kfree(block);
drivers/net/wireless/atmel/at76c50x-usb.c
320
void *block, int size)
drivers/net/wireless/atmel/at76c50x-usb.c
324
USB_RECIP_INTERFACE, blockno, 0, block, size,
drivers/net/wireless/atmel/at76c50x-usb.c
371
u8 *block = NULL;
drivers/net/wireless/atmel/at76c50x-usb.c
387
block = kmalloc(FW_BLOCK_SIZE, GFP_KERNEL);
drivers/net/wireless/atmel/at76c50x-usb.c
388
if (!block) {
drivers/net/wireless/atmel/at76c50x-usb.c
440
memcpy(block, buf, bsize);
drivers/net/wireless/atmel/at76c50x-usb.c
445
at76_load_int_fw_block(udev, blockno, block, bsize);
drivers/net/wireless/atmel/at76c50x-usb.c
505
kfree(block);
drivers/net/wireless/atmel/at76c50x-usb.c
577
void *block, int size)
drivers/net/wireless/atmel/at76c50x-usb.c
581
0x0802, blockno, block, size,
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
1295
memcpy(range->data, fwrt->trans->init_dram.paging[idx].block,
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
1316
memcpy(range->data, frag->block, frag->size);
drivers/net/wireless/intel/iwlwifi/fw/paging.c
173
struct iwl_fw_paging *block = &fwrt->fw_paging_db[idx];
drivers/net/wireless/intel/iwlwifi/fw/paging.c
175
int len = block->fw_paging_size;
drivers/net/wireless/intel/iwlwifi/fw/paging.c
191
} else if (block->fw_paging_size > remaining) {
drivers/net/wireless/intel/iwlwifi/fw/paging.c
199
memcpy(page_address(block->fw_paging_block),
drivers/net/wireless/intel/iwlwifi/fw/paging.c
201
block->fw_offs = image->sec[sec_idx].offset + offset;
drivers/net/wireless/intel/iwlwifi/fw/paging.c
203
block->fw_paging_phys,
drivers/net/wireless/intel/iwlwifi/fw/paging.c
204
block->fw_paging_size,
drivers/net/wireless/intel/iwlwifi/fw/paging.c
211
offset += block->fw_paging_size;
drivers/net/wireless/intel/iwlwifi/fw/paging.c
43
struct page *block;
drivers/net/wireless/intel/iwlwifi/fw/paging.c
72
block = alloc_pages(GFP_KERNEL, order);
drivers/net/wireless/intel/iwlwifi/fw/paging.c
73
if (!block) {
drivers/net/wireless/intel/iwlwifi/fw/paging.c
79
fwrt->fw_paging_db[blk_idx].fw_paging_block = block;
drivers/net/wireless/intel/iwlwifi/fw/paging.c
82
phys = dma_map_page(fwrt->trans->dev, block, 0,
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
405
dma_free_coherent(trans->dev, frag->size, frag->block,
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
409
frag->block = NULL;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
543
void *block = NULL;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
557
block = dma_alloc_coherent(fwrt->dev, pages * PAGE_SIZE,
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
560
if (block)
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
569
if (!block)
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
573
frag->block = block;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
596
memset(fw_mon->frags[i].block, 0,
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
800
if (!frags || !frags->block)
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
803
dram_info = frags->block;
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
542
void *block;
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
787
int block;
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
454
bool block;
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
470
if (data->block) {
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
482
int iwl_mld_update_emlsr_block(struct iwl_mld *mld, bool block,
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
486
.block = block,
drivers/net/wireless/intel/iwlwifi/mld/mlo.h
155
int iwl_mld_update_emlsr_block(struct iwl_mld *mld, bool block,
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
373
dram->block = iwl_pcie_ctxt_info_dma_alloc_coherent(trans, len,
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
375
if (!dram->block) {
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
381
memcpy(dram->block, pnvm_data->chunks[0].data, len0);
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
382
memcpy((u8 *)dram->block + len0, pnvm_data->chunks[1].data, len1);
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
401
desc_dram->block = iwl_pcie_ctxt_info_dma_alloc_coherent
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
405
if (!desc_dram->block) {
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
410
memset(desc_dram->block, 0, len);
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
432
addresses = desc_dram->block;
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c
52
dram->block = iwl_pcie_ctxt_info_dma_alloc_coherent(trans, len,
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c
54
if (!dram->block)
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c
58
memcpy(dram->block, data, len);
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info.c
76
dram->paging[i].block,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
879
dram->fw[i].block, dram->fw[i].physical);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
104
dma_free_coherent(trans->dev, fw_mon->size, fw_mon->block,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
107
fw_mon->block = NULL;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
116
void *block = NULL;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
122
memset(fw_mon->block, 0, fw_mon->size);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
129
block = dma_alloc_coherent(trans->dev, size, &physical,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
131
if (!block)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
140
if (WARN_ON_ONCE(!block))
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
149
fw_mon->block = block;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
1929
dram_regions->drams[i].block,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
1935
if (desc_dram->block) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
1937
desc_dram->block,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
3077
u8 *cpu_addr = (void *)trans->dbg.fw_mon.block, *curr_buf;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
3446
memcpy(fw_mon_data->data, fw_mon->block, fw_mon->size);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
3669
trans->init_dram.paging[i].block, page_len);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
42
u32 block, u32 vec, u32 iter)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
46
IWL_ERR(trans, "Host monitor block 0x%x vector 0x%x\n", block, vec);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
47
iwl_write32(trans, CSR_MONITOR_CFG_REG, (block << 8) | vec);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
119
if (!txq->block)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1306
static void iwl_trans_pcie_block_txq_ptrs(struct iwl_trans *trans, bool block)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1320
if (!block && !(WARN_ON_ONCE(!txq->block))) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1321
txq->block--;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1322
if (!txq->block) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1326
} else if (block) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1327
txq->block++;
drivers/net/wireless/marvell/libertas/cmd.c
230
int lbs_set_ps_mode(struct lbs_private *priv, u16 cmd_action, bool block)
drivers/net/wireless/marvell/libertas/cmd.c
253
if (block)
drivers/net/wireless/marvell/libertas/cmd.h
125
int lbs_set_ps_mode(struct lbs_private *priv, u16 cmd_action, bool block);
drivers/net/wireless/mediatek/mt76/npu.c
342
block_cb = flow_block_cb_lookup(f->block, cb, dev);
drivers/net/wireless/mediatek/mt76/npu.c
357
block_cb = flow_block_cb_lookup(f->block, cb, dev);
drivers/net/wireless/realtek/rtw88/mac.c
867
u32 block;
drivers/net/wireless/realtek/rtw88/mac.c
878
for (block = 0; block < block_nr; block++) {
drivers/net/wireless/realtek/rtw89/core.h
3796
enum rtw89_efuse_block block);
drivers/net/wireless/realtek/rtw89/efuse_be.c
308
enum rtw89_efuse_block block)
drivers/net/wireless/realtek/rtw89/efuse_be.c
315
efuse_block = &chip->efuse_blocks[block];
drivers/net/wireless/realtek/rtw89/efuse_be.c
323
rtw89_warn(rtwdev, "failed to dump efuse logical block %d\n", block);
drivers/net/wireless/realtek/rtw89/efuse_be.c
329
ret = rtwdev->chip->ops->read_efuse(rtwdev, log_map, block);
drivers/net/wireless/realtek/rtw89/efuse_be.c
345
enum rtw89_efuse_block block;
drivers/net/wireless/realtek/rtw89/efuse_be.c
376
block = RTW89_EFUSE_BLOCK_HCI_DIG_USB;
drivers/net/wireless/realtek/rtw89/efuse_be.c
378
block = RTW89_EFUSE_BLOCK_HCI_DIG_PCIE_SDIO;
drivers/net/wireless/realtek/rtw89/efuse_be.c
380
ret = rtw89_parse_logical_efuse_block_be(rtwdev, phy_map, phy_size, block);
drivers/net/wireless/realtek/rtw89/regd.c
407
static void __rtw89_regd_setup_policy_6ghz(struct rtw89_dev *rtwdev, bool block,
drivers/net/wireless/realtek/rtw89/regd.c
420
if (block)
drivers/net/wireless/realtek/rtw89/rtw8851b.c
610
enum rtw89_efuse_block block)
drivers/net/wireless/realtek/rtw89/rtw8852a.c
656
enum rtw89_efuse_block block)
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
242
enum rtw89_efuse_block block)
drivers/net/wireless/realtek/rtw89/rtw8852b_common.h
152
enum rtw89_efuse_block block);
drivers/net/wireless/realtek/rtw89/rtw8852b_common.h
307
enum rtw89_efuse_block block)
drivers/net/wireless/realtek/rtw89/rtw8852b_common.h
309
return rtw8852bx_info.read_efuse(rtwdev, log_map, block);
drivers/net/wireless/realtek/rtw89/rtw8852c.c
577
enum rtw89_efuse_block block)
drivers/net/wireless/realtek/rtw89/rtw8922a.c
635
enum rtw89_efuse_block block)
drivers/net/wireless/realtek/rtw89/rtw8922a.c
640
switch (block) {
drivers/net/wireless/st/cw1200/fwio.c
180
for (block = 0; block < num_blocks; block++) {
drivers/net/wireless/st/cw1200/fwio.c
53
int ret, block, num_blocks;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
386
unsigned char *block)
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
388
struct ipc_mem_lite_gen_tbl *fct = (struct ipc_mem_lite_gen_tbl *)block;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
436
u8 *block = skb->data;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
440
adgh = (struct mux_adgh *)block;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
523
u8 *block;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
527
block = (u8 *)(skb->data);
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
534
cmdh = (struct mux_cmdh *)&block[next_cmd_index];
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
593
u8 *block;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
595
block = skb->data;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
596
adbh = (struct mux_adbh *)block;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
610
adth = (struct mux_adth *)(block + adth_index);
drivers/nvdimm/btt.c
1329
if (btt_is_badblock(btt, arena, arena->freelist[lane].block))
drivers/nvdimm/btt.c
1344
new_postmap = arena->freelist[lane].block;
drivers/nvdimm/btt.c
399
arena->freelist[lane].block = ent_lba(le32_to_cpu(ent->old_map));
drivers/nvdimm/btt.c
513
u32 lba = arena->freelist[lane].block;
drivers/nvdimm/btt.c
558
arena->freelist[i].block = log_oldmap;
drivers/nvdimm/btt.h
118
u32 block;
drivers/perf/xgene_pmu.c
304
XGENE_PMU_EVENT_ATTR(csw-read-block, 0x06),
drivers/platform/chrome/cros_ec_chardev.c
122
bool fetch, bool block)
drivers/platform/chrome/cros_ec_chardev.c
128
if (!block && list_empty(&priv->events)) {
drivers/platform/chrome/cros_ec_vbc.c
72
const size_t data_sz = sizeof(params->block);
drivers/platform/chrome/cros_ec_vbc.c
84
memcpy(params->block, buf, data_sz);
drivers/platform/mellanox/mlxbf-pmc.c
1100
(offset + MLXBF_PMC_REG_SIZE <= pmc->block[blk_num].blk_size))
drivers/platform/mellanox/mlxbf-pmc.c
1255
return mlxbf_pmc_write(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1274
pmcaddr = pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1279
pmcaddr = pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1331
addr = pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1353
addr = pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1354
MLXBF_PMC_CRSPACE_PERFMON_VAL0(pmc->block[blk_num].counters) +
drivers/platform/mellanox/mlxbf-pmc.c
1365
if (cnt_num >= pmc->block[blk_num].counters)
drivers/platform/mellanox/mlxbf-pmc.c
1371
if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE)
drivers/platform/mellanox/mlxbf-pmc.c
1390
if (mlxbf_pmc_write(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1404
if (mlxbf_pmc_write(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1415
if (mlxbf_pmc_write(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1430
status = mlxbf_pmc_readl(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1438
status = mlxbf_pmc_readl(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1460
status = mlxbf_pmc_readl(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1461
MLXBF_PMC_CRSPACE_PERFMON_VAL0(pmc->block[blk_num].counters) +
drivers/platform/mellanox/mlxbf-pmc.c
1478
if (cnt_num >= pmc->block[blk_num].counters)
drivers/platform/mellanox/mlxbf-pmc.c
1484
if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE)
drivers/platform/mellanox/mlxbf-pmc.c
1489
pmc->block[blk_num].counters * MLXBF_PMC_REG_SIZE;
drivers/platform/mellanox/mlxbf-pmc.c
1497
status = mlxbf_pmc_write(pmc->block[blk_num].mmio_base + perfcfg_offset,
drivers/platform/mellanox/mlxbf-pmc.c
1504
return mlxbf_pmc_read(pmc->block[blk_num].mmio_base + perfval_offset,
drivers/platform/mellanox/mlxbf-pmc.c
1518
pmcaddr = pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1523
pmcaddr = pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1567
addr = pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1589
if (cnt_num >= pmc->block[blk_num].counters)
drivers/platform/mellanox/mlxbf-pmc.c
1595
if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE)
drivers/platform/mellanox/mlxbf-pmc.c
1600
pmc->block[blk_num].counters * MLXBF_PMC_REG_SIZE;
drivers/platform/mellanox/mlxbf-pmc.c
1608
if (mlxbf_pmc_write(pmc->block[blk_num].mmio_base + perfcfg_offset,
drivers/platform/mellanox/mlxbf-pmc.c
1613
if (mlxbf_pmc_read(pmc->block[blk_num].mmio_base + perfval_offset,
drivers/platform/mellanox/mlxbf-pmc.c
1629
if (mlxbf_pmc_readl(pmc->block[blk_num].mmio_base + offset,
drivers/platform/mellanox/mlxbf-pmc.c
1638
return mlxbf_pmc_read(pmc->block[blk_num].mmio_base + offset,
drivers/platform/mellanox/mlxbf-pmc.c
165
struct mlxbf_pmc_block_info block[MLXBF_PMC_MAX_BLOCKS];
drivers/platform/mellanox/mlxbf-pmc.c
1651
return mlxbf_pmc_write(pmc->block[blk_num].mmio_base + offset,
drivers/platform/mellanox/mlxbf-pmc.c
1656
return mlxbf_pmc_write(pmc->block[blk_num].mmio_base + offset,
drivers/platform/mellanox/mlxbf-pmc.c
1679
if ((pmc->block[blk_num].type == MLXBF_PMC_TYPE_COUNTER) ||
drivers/platform/mellanox/mlxbf-pmc.c
1680
(pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE)) {
drivers/platform/mellanox/mlxbf-pmc.c
1683
} else if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_REGISTER) {
drivers/platform/mellanox/mlxbf-pmc.c
1724
if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_COUNTER) {
drivers/platform/mellanox/mlxbf-pmc.c
1732
} else if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_REGISTER) {
drivers/platform/mellanox/mlxbf-pmc.c
1740
} else if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE) {
drivers/platform/mellanox/mlxbf-pmc.c
1867
if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE) {
drivers/platform/mellanox/mlxbf-pmc.c
1868
if (mlxbf_pmc_readl(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1869
MLXBF_PMC_CRSPACE_PERFMON_CTL(pmc->block[blk_num].counters),
drivers/platform/mellanox/mlxbf-pmc.c
1875
if (mlxbf_pmc_readl(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1904
if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE) {
drivers/platform/mellanox/mlxbf-pmc.c
1905
err = mlxbf_pmc_readl(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1906
MLXBF_PMC_CRSPACE_PERFMON_CTL(pmc->block[blk_num].counters),
drivers/platform/mellanox/mlxbf-pmc.c
1916
mlxbf_pmc_write(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1917
MLXBF_PMC_CRSPACE_PERFMON_CTL(pmc->block[blk_num].counters),
drivers/platform/mellanox/mlxbf-pmc.c
1945
if (mlxbf_pmc_readl(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1946
MLXBF_PMC_CRSPACE_PERFMON_COUNT_CLOCK(pmc->block[blk_num].counters),
drivers/platform/mellanox/mlxbf-pmc.c
1970
mlxbf_pmc_write(pmc->block[blk_num].mmio_base +
drivers/platform/mellanox/mlxbf-pmc.c
1971
MLXBF_PMC_CRSPACE_PERFMON_COUNT_CLOCK(pmc->block[blk_num].counters),
drivers/platform/mellanox/mlxbf-pmc.c
1987
attr = &pmc->block[blk_num].attr_event_list;
drivers/platform/mellanox/mlxbf-pmc.c
1995
pmc->block[blk_num].block_attr[i] = &attr->dev_attr.attr;
drivers/platform/mellanox/mlxbf-pmc.c
2000
((pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE))) {
drivers/platform/mellanox/mlxbf-pmc.c
2001
attr = &pmc->block[blk_num].attr_enable;
drivers/platform/mellanox/mlxbf-pmc.c
2011
pmc->block[blk_num].block_attr[++i] = &attr->dev_attr.attr;
drivers/platform/mellanox/mlxbf-pmc.c
2015
if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE) {
drivers/platform/mellanox/mlxbf-pmc.c
2017
attr = &pmc->block[blk_num].attr_count_clock;
drivers/platform/mellanox/mlxbf-pmc.c
2027
pmc->block[blk_num].block_attr[++i] = &attr->dev_attr.attr;
drivers/platform/mellanox/mlxbf-pmc.c
2031
pmc->block[blk_num].attr_counter = devm_kcalloc(
drivers/platform/mellanox/mlxbf-pmc.c
2032
dev, pmc->block[blk_num].counters,
drivers/platform/mellanox/mlxbf-pmc.c
2034
if (!pmc->block[blk_num].attr_counter)
drivers/platform/mellanox/mlxbf-pmc.c
2037
pmc->block[blk_num].attr_event = devm_kcalloc(
drivers/platform/mellanox/mlxbf-pmc.c
2038
dev, pmc->block[blk_num].counters,
drivers/platform/mellanox/mlxbf-pmc.c
2040
if (!pmc->block[blk_num].attr_event)
drivers/platform/mellanox/mlxbf-pmc.c
2044
for (j = 0; j < pmc->block[blk_num].counters; ++j) {
drivers/platform/mellanox/mlxbf-pmc.c
2045
attr = &pmc->block[blk_num].attr_counter[j];
drivers/platform/mellanox/mlxbf-pmc.c
2055
pmc->block[blk_num].block_attr[++i] = &attr->dev_attr.attr;
drivers/platform/mellanox/mlxbf-pmc.c
2058
attr = &pmc->block[blk_num].attr_event[j];
drivers/platform/mellanox/mlxbf-pmc.c
2068
pmc->block[blk_num].block_attr[++i] = &attr->dev_attr.attr;
drivers/platform/mellanox/mlxbf-pmc.c
2087
pmc->block[blk_num].attr_event = devm_kcalloc(
drivers/platform/mellanox/mlxbf-pmc.c
2089
if (!pmc->block[blk_num].attr_event)
drivers/platform/mellanox/mlxbf-pmc.c
2094
attr = &pmc->block[blk_num].attr_event[count];
drivers/platform/mellanox/mlxbf-pmc.c
2104
pmc->block[blk_num].block_attr[i] = &attr->dev_attr.attr;
drivers/platform/mellanox/mlxbf-pmc.c
2118
if ((pmc->block[blk_num].type == MLXBF_PMC_TYPE_COUNTER) ||
drivers/platform/mellanox/mlxbf-pmc.c
2119
(pmc->block[blk_num].type == MLXBF_PMC_TYPE_CRSPACE))
drivers/platform/mellanox/mlxbf-pmc.c
2121
else if (pmc->block[blk_num].type == MLXBF_PMC_TYPE_REGISTER)
drivers/platform/mellanox/mlxbf-pmc.c
2130
pmc->block[blk_num].block_attr_grp.attrs = pmc->block[blk_num].block_attr;
drivers/platform/mellanox/mlxbf-pmc.c
2131
pmc->block[blk_num].block_attr_grp.name = devm_kasprintf(
drivers/platform/mellanox/mlxbf-pmc.c
2133
if (!pmc->block[blk_num].block_attr_grp.name)
drivers/platform/mellanox/mlxbf-pmc.c
2135
pmc->groups[pmc->group_num] = &pmc->block[blk_num].block_attr_grp;
drivers/platform/mellanox/mlxbf-pmc.c
2239
pmc->block[i].mmio_base = (void __iomem *)info[0];
drivers/platform/mellanox/mlxbf-pmc.c
2241
pmc->block[i].mmio_base =
drivers/platform/mellanox/mlxbf-pmc.c
2244
pmc->block[i].blk_size = info[1];
drivers/platform/mellanox/mlxbf-pmc.c
2245
pmc->block[i].counters = info[2];
drivers/platform/mellanox/mlxbf-pmc.c
2246
pmc->block[i].type = info[3];
drivers/platform/mellanox/mlxbf-pmc.c
2248
if (!pmc->block[i].mmio_base)
drivers/platform/wmi/core.c
322
struct guid_block *block;
drivers/platform/wmi/core.c
330
block = &wblock->gblock;
drivers/platform/wmi/core.c
336
if (!(block->flags & ACPI_WMI_METHOD))
drivers/platform/wmi/core.c
339
if (block->instance_count <= instance)
drivers/platform/wmi/core.c
433
struct guid_block *block;
drivers/platform/wmi/core.c
442
block = &wblock->gblock;
drivers/platform/wmi/core.c
445
if (block->instance_count <= instance)
drivers/platform/wmi/core.c
449
if (block->flags & (ACPI_WMI_EVENT | ACPI_WMI_METHOD))
drivers/platform/wmi/core.c
596
struct guid_block *block = &wblock->gblock;
drivers/platform/wmi/core.c
604
if (block->instance_count <= instance)
drivers/platform/wmi/core.c
608
if (block->flags & (ACPI_WMI_EVENT | ACPI_WMI_METHOD))
drivers/platform/x86/dell/dell-laptop.c
557
int block = rfkill_blocked(rfkill);
drivers/platform/x86/dell/dell-laptop.c
559
1 | (radio << 8) | (block << 16), 0, 0, 0);
drivers/platform/x86/lenovo/wmi-capdata.c
540
} __packed * block;
drivers/platform/x86/lenovo/wmi-capdata.c
547
block = (struct cd_fan_block *)ret_obj->buffer.pointer;
drivers/platform/x86/lenovo/wmi-capdata.c
550
count = size >= sizeof(*block) ? block->nr : 0;
drivers/platform/x86/lenovo/wmi-capdata.c
551
if (size < struct_size(block, data, count * 3)) {
drivers/platform/x86/lenovo/wmi-capdata.c
554
size, struct_size(block, data, count * 3));
drivers/platform/x86/lenovo/wmi-capdata.c
578
.id = block->data[idx],
drivers/platform/x86/lenovo/wmi-capdata.c
579
.max_rpm = block->data[idx + block->nr],
drivers/platform/x86/lenovo/wmi-capdata.c
580
.min_rpm = block->data[idx + (2 * block->nr)],
drivers/power/supply/bq27xxx_battery.c
1080
u8 block;
drivers/power/supply/bq27xxx_battery.c
1087
.block = (di)->dm_regs[i].offset / BQ27XXX_DM_SZ, \
drivers/power/supply/bq27xxx_battery.c
1094
buf->block == reg->offset / BQ27XXX_DM_SZ)
drivers/power/supply/bq27xxx_battery.c
1283
ret = bq27xxx_write(di, BQ27XXX_DM_BLOCK, buf->block, true);
drivers/power/supply/bq27xxx_battery.c
1430
ret = bq27xxx_write(di, BQ27XXX_DM_BLOCK, buf->block, true);
drivers/power/supply/bq27xxx_battery.c
1494
bool same = bd.class == bt.class && bd.block == bt.block;
drivers/rapidio/devices/rio_mport_cdev.c
967
(void __user *)(uintptr_t)transaction.block,
drivers/rapidio/devices/rio_mport_cdev.c
979
if (unlikely(copy_to_user((void __user *)(uintptr_t)transaction.block,
drivers/s390/block/dasd.c
1075
#define dasd_profile_start(block, cqr, req) do {} while (0)
drivers/s390/block/dasd.c
1076
#define dasd_profile_end(block, cqr, req) do {} while (0)
drivers/s390/block/dasd.c
1340
if (((cqr->block &&
drivers/s390/block/dasd.c
1341
test_bit(DASD_FLAG_LOCK_STOLEN, &cqr->block->base->flags)) ||
drivers/s390/block/dasd.c
1368
if (cqr->block)
drivers/s390/block/dasd.c
1369
cqr->trkcount = atomic_read(&cqr->block->trkcount);
drivers/s390/block/dasd.c
151
struct dasd_block *block;
drivers/s390/block/dasd.c
153
block = kzalloc_obj(*block, GFP_ATOMIC);
drivers/s390/block/dasd.c
1530
if (device->block) {
drivers/s390/block/dasd.c
1531
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd.c
1532
if (device->block->gdp)
drivers/s390/block/dasd.c
1533
blk_mq_run_hw_queues(device->block->gdp->queue, true);
drivers/s390/block/dasd.c
154
if (!block)
drivers/s390/block/dasd.c
1545
static int dasd_ese_needs_format(struct dasd_block *block, struct irb *irb)
drivers/s390/block/dasd.c
1550
if (!block)
drivers/s390/block/dasd.c
1552
device = block->base;
drivers/s390/block/dasd.c
157
atomic_set(&block->open_count, -1);
drivers/s390/block/dasd.c
159
atomic_set(&block->tasklet_scheduled, 0);
drivers/s390/block/dasd.c
160
tasklet_init(&block->tasklet, dasd_block_tasklet,
drivers/s390/block/dasd.c
161
(unsigned long) block);
drivers/s390/block/dasd.c
162
INIT_LIST_HEAD(&block->ccw_queue);
drivers/s390/block/dasd.c
163
spin_lock_init(&block->queue_lock);
drivers/s390/block/dasd.c
164
INIT_LIST_HEAD(&block->format_list);
drivers/s390/block/dasd.c
165
spin_lock_init(&block->format_lock);
drivers/s390/block/dasd.c
166
timer_setup(&block->timer, dasd_block_timeout, 0);
drivers/s390/block/dasd.c
167
spin_lock_init(&block->profile.lock);
drivers/s390/block/dasd.c
169
return block;
drivers/s390/block/dasd.c
1693
if (dasd_ese_needs_format(cqr->block, irb)) {
drivers/s390/block/dasd.c
176
void dasd_free_block(struct dasd_block *block)
drivers/s390/block/dasd.c
178
kfree(block);
drivers/s390/block/dasd.c
1833
if (!ref_cqr->block)
drivers/s390/block/dasd.c
1839
ref_cqr->block == cqr->block) {
drivers/s390/block/dasd.c
1903
struct dasd_block *block;
drivers/s390/block/dasd.c
1908
block = cqr->block;
drivers/s390/block/dasd.c
1909
if (!block) {
drivers/s390/block/dasd.c
1912
spin_lock_bh(&block->queue_lock);
drivers/s390/block/dasd.c
1914
spin_unlock_bh(&block->queue_lock);
drivers/s390/block/dasd.c
215
struct dasd_block *block = device->block;
drivers/s390/block/dasd.c
219
if (block) {
drivers/s390/block/dasd.c
220
rc = dasd_gendisk_alloc(block);
drivers/s390/block/dasd.c
223
block->debugfs_dentry =
drivers/s390/block/dasd.c
224
debugfs_create_dir(block->gdp->disk_name,
drivers/s390/block/dasd.c
226
dasd_profile_init(&block->profile, block->debugfs_dentry);
drivers/s390/block/dasd.c
228
dasd_profile_on(&device->block->profile);
drivers/s390/block/dasd.c
2568
if (device->block)
drivers/s390/block/dasd.c
2569
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd.c
261
if (device->block) {
drivers/s390/block/dasd.c
262
dasd_profile_exit(&device->block->profile);
drivers/s390/block/dasd.c
263
debugfs_remove(device->block->debugfs_dentry);
drivers/s390/block/dasd.c
2635
struct dasd_block *block;
drivers/s390/block/dasd.c
2637
block = timer_container_of(block, t, timer);
drivers/s390/block/dasd.c
2638
spin_lock_irqsave(get_ccwdev_lock(block->base->cdev), flags);
drivers/s390/block/dasd.c
264
dasd_gendisk_free(device->block);
drivers/s390/block/dasd.c
2640
dasd_device_remove_stop_bits(block->base, DASD_STOPPED_PENDING);
drivers/s390/block/dasd.c
2641
spin_unlock_irqrestore(get_ccwdev_lock(block->base->cdev), flags);
drivers/s390/block/dasd.c
2642
dasd_schedule_block_bh(block);
drivers/s390/block/dasd.c
2643
blk_mq_run_hw_queues(block->gdp->queue, true);
drivers/s390/block/dasd.c
2649
void dasd_block_set_timer(struct dasd_block *block, int expires)
drivers/s390/block/dasd.c
265
dasd_block_clear_timer(device->block);
drivers/s390/block/dasd.c
2652
timer_delete(&block->timer);
drivers/s390/block/dasd.c
2654
mod_timer(&block->timer, jiffies + expires);
drivers/s390/block/dasd.c
2661
void dasd_block_clear_timer(struct dasd_block *block)
drivers/s390/block/dasd.c
2663
timer_delete(&block->timer);
drivers/s390/block/dasd.c
2691
dasd_profile_end(cqr->block, cqr, req);
drivers/s390/block/dasd.c
2694
status = cqr->block->base->discipline->free_cp(cqr, req);
drivers/s390/block/dasd.c
2746
static void __dasd_process_block_ccw_queue(struct dasd_block *block,
drivers/s390/block/dasd.c
2753
struct dasd_device *base = block->base;
drivers/s390/block/dasd.c
2757
list_for_each_safe(l, n, &block->ccw_queue) {
drivers/s390/block/dasd.c
2810
dasd_schedule_block_bh(cqr->block);
drivers/s390/block/dasd.c
2813
static void __dasd_block_start_head(struct dasd_block *block)
drivers/s390/block/dasd.c
2817
if (list_empty(&block->ccw_queue))
drivers/s390/block/dasd.c
2823
list_for_each_entry(cqr, &block->ccw_queue, blocklist) {
drivers/s390/block/dasd.c
2826
if (test_bit(DASD_FLAG_LOCK_STOLEN, &block->base->flags) &&
drivers/s390/block/dasd.c
2830
dasd_schedule_block_bh(block);
drivers/s390/block/dasd.c
2834
if (block->base->stopped & ~DASD_STOPPED_PENDING &&
drivers/s390/block/dasd.c
2836
!dasd_eer_enabled(block->base) && block->base->aq_mask == 0) {
drivers/s390/block/dasd.c
2839
dasd_schedule_block_bh(block);
drivers/s390/block/dasd.c
2843
if (block->base->stopped)
drivers/s390/block/dasd.c
2848
cqr->startdev = block->base;
drivers/s390/block/dasd.c
2864
struct dasd_block *block = (struct dasd_block *) data;
drivers/s390/block/dasd.c
2870
atomic_set(&block->tasklet_scheduled, 0);
drivers/s390/block/dasd.c
2872
spin_lock_irq(&block->queue_lock);
drivers/s390/block/dasd.c
2874
__dasd_process_block_ccw_queue(block, &final_queue);
drivers/s390/block/dasd.c
2875
spin_unlock_irq(&block->queue_lock);
drivers/s390/block/dasd.c
2887
spin_lock_irq(&block->queue_lock);
drivers/s390/block/dasd.c
2889
__dasd_block_start_head(block);
drivers/s390/block/dasd.c
2890
spin_unlock_irq(&block->queue_lock);
drivers/s390/block/dasd.c
2894
dasd_put_device(block->base);
drivers/s390/block/dasd.c
2924
static int _dasd_requests_to_flushqueue(struct dasd_block *block,
drivers/s390/block/dasd.c
2931
spin_lock_irqsave(&block->queue_lock, flags);
drivers/s390/block/dasd.c
2934
list_for_each_entry_safe(cqr, n, &block->ccw_queue, blocklist) {
drivers/s390/block/dasd.c
2952
spin_unlock_irqrestore(&block->queue_lock, flags);
drivers/s390/block/dasd.c
2962
static int dasd_flush_block_queue(struct dasd_block *block)
drivers/s390/block/dasd.c
2970
rc = _dasd_requests_to_flushqueue(block, &flush_queue);
drivers/s390/block/dasd.c
2978
spin_lock_bh(&block->queue_lock);
drivers/s390/block/dasd.c
2979
__dasd_process_erp(block->base, cqr);
drivers/s390/block/dasd.c
2980
spin_unlock_bh(&block->queue_lock);
drivers/s390/block/dasd.c
299
struct dasd_block *block = device->block;
drivers/s390/block/dasd.c
2998
void dasd_schedule_block_bh(struct dasd_block *block)
drivers/s390/block/dasd.c
3001
if (atomic_cmpxchg(&block->tasklet_scheduled, 0, 1) != 0)
drivers/s390/block/dasd.c
3004
dasd_get_device(block->base);
drivers/s390/block/dasd.c
3005
tasklet_hi_schedule(&block->tasklet);
drivers/s390/block/dasd.c
3021
struct dasd_block *block = hctx->queue->queuedata;
drivers/s390/block/dasd.c
3028
basedev = block->base;
drivers/s390/block/dasd.c
304
if (!block) {
drivers/s390/block/dasd.c
3067
cqr = basedev->discipline->build_cp(basedev, block, req);
drivers/s390/block/dasd.c
309
if (block->base->discipline->do_analysis != NULL)
drivers/s390/block/dasd.c
3092
spin_lock(&block->queue_lock);
drivers/s390/block/dasd.c
3093
list_add_tail(&cqr->blocklist, &block->ccw_queue);
drivers/s390/block/dasd.c
3095
dasd_profile_start(block, cqr, req);
drivers/s390/block/dasd.c
3096
dasd_schedule_block_bh(block);
drivers/s390/block/dasd.c
3097
spin_unlock(&block->queue_lock);
drivers/s390/block/dasd.c
310
rc = block->base->discipline->do_analysis(block);
drivers/s390/block/dasd.c
3114
struct dasd_block *block = req->q->queuedata;
drivers/s390/block/dasd.c
3125
device = cqr->startdev ? cqr->startdev : block->base;
drivers/s390/block/dasd.c
3134
spin_lock(&block->queue_lock);
drivers/s390/block/dasd.c
3147
&block->ccw_queue, blocklist) {
drivers/s390/block/dasd.c
315
kobject_uevent(&disk_to_dev(device->block->gdp)->kobj,
drivers/s390/block/dasd.c
3173
dasd_schedule_block_bh(block);
drivers/s390/block/dasd.c
3174
spin_unlock(&block->queue_lock);
drivers/s390/block/dasd.c
320
lim = queue_limits_start_update(block->gdp->queue);
drivers/s390/block/dasd.c
321
lim.max_dev_sectors = device->discipline->max_sectors(block);
drivers/s390/block/dasd.c
3223
atomic_inc(&base->block->open_count);
drivers/s390/block/dasd.c
323
lim.logical_block_size = block->bp_block;
drivers/s390/block/dasd.c
3260
atomic_dec(&base->block->open_count);
drivers/s390/block/dasd.c
3269
atomic_dec(&base->block->open_count);
drivers/s390/block/dasd.c
3291
base->discipline->fill_geometry(base->block, geo);
drivers/s390/block/dasd.c
3293
geo->start >>= base->block->s2b_shift;
drivers/s390/block/dasd.c
333
lim.discard_granularity = block->bp_block;
drivers/s390/block/dasd.c
336
max_bytes = USHRT_MAX * block->bp_block;
drivers/s390/block/dasd.c
339
lim.max_hw_discard_sectors = max_bytes / block->bp_block;
drivers/s390/block/dasd.c
3415
struct dasd_block *block;
drivers/s390/block/dasd.c
342
rc = queue_limits_commit_update(block->gdp->queue, &lim);
drivers/s390/block/dasd.c
3435
block = device->block;
drivers/s390/block/dasd.c
3441
if (block)
drivers/s390/block/dasd.c
3442
dasd_free_block(block);
drivers/s390/block/dasd.c
346
set_capacity(block->gdp, block->blocks << block->s2b_shift);
drivers/s390/block/dasd.c
349
rc = dasd_scan_partitions(block);
drivers/s390/block/dasd.c
3515
if (device->block)
drivers/s390/block/dasd.c
3516
dasd_free_block(device->block);
drivers/s390/block/dasd.c
3533
struct dasd_block *block;
drivers/s390/block/dasd.c
3553
if (device->block) {
drivers/s390/block/dasd.c
3554
max_count = device->block->bdev_file ? 0 : -1;
drivers/s390/block/dasd.c
3555
open_count = atomic_read(&device->block->open_count);
drivers/s390/block/dasd.c
3599
if (device->block && device->block->bdev_file)
drivers/s390/block/dasd.c
3600
bdev_mark_dead(file_bdev(device->block->bdev_file), false);
drivers/s390/block/dasd.c
3625
block = device->block;
drivers/s390/block/dasd.c
3631
if (block)
drivers/s390/block/dasd.c
3632
dasd_free_block(block);
drivers/s390/block/dasd.c
364
if (device->block)
drivers/s390/block/dasd.c
366
list_empty(&device->block->ccw_queue);
drivers/s390/block/dasd.c
3681
if (device->block) {
drivers/s390/block/dasd.c
3682
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd.c
3683
if (device->block->gdp)
drivers/s390/block/dasd.c
3684
blk_mq_run_hw_queues(device->block->gdp->queue, true);
drivers/s390/block/dasd.c
381
if (device->block) {
drivers/s390/block/dasd.c
382
struct dasd_block *block = device->block;
drivers/s390/block/dasd.c
383
rc = dasd_flush_block_queue(block);
drivers/s390/block/dasd.c
3836
if (device->block) {
drivers/s390/block/dasd.c
3837
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd.c
3838
if (device->block->gdp)
drivers/s390/block/dasd.c
3839
blk_mq_run_hw_queues(device->block->gdp->queue, true);
drivers/s390/block/dasd.c
3851
struct dasd_block *block = device->block;
drivers/s390/block/dasd.c
3856
if (!block)
drivers/s390/block/dasd.c
3860
rc = _dasd_requests_to_flushqueue(block, &requeue_queue);
drivers/s390/block/dasd.c
3868
spin_lock_bh(&block->queue_lock);
drivers/s390/block/dasd.c
3869
__dasd_process_erp(block->base, cqr);
drivers/s390/block/dasd.c
3870
spin_unlock_bh(&block->queue_lock);
drivers/s390/block/dasd.c
3878
cqr->block->base->discipline->free_cp(
drivers/s390/block/dasd.c
388
dasd_destroy_partitions(block);
drivers/s390/block/dasd.c
389
block->blocks = 0;
drivers/s390/block/dasd.c
3892
if (device->block)
drivers/s390/block/dasd.c
3893
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd.c
390
block->bp_block = 0;
drivers/s390/block/dasd.c
391
block->s2b_shift = 0;
drivers/s390/block/dasd.c
4021
if (device->block)
drivers/s390/block/dasd.c
4022
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd.c
414
if (device->block) {
drivers/s390/block/dasd.c
415
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd.c
417
kobject_uevent(&disk_to_dev(device->block->gdp)->kobj,
drivers/s390/block/dasd.c
421
disk_uevent(file_bdev(device->block->bdev_file)->bd_disk,
drivers/s390/block/dasd.c
441
if (device->block && !(device->features & DASD_FEATURE_USERAW))
drivers/s390/block/dasd.c
442
disk_uevent(file_bdev(device->block->bdev_file)->bd_disk,
drivers/s390/block/dasd.c
648
static void dasd_profile_start(struct dasd_block *block,
drivers/s390/block/dasd.c
658
if (dasd_global_profile_level || block->profile.data)
drivers/s390/block/dasd.c
659
list_for_each(l, &block->ccw_queue)
drivers/s390/block/dasd.c
671
spin_lock(&block->profile.lock);
drivers/s390/block/dasd.c
672
if (block->profile.data) {
drivers/s390/block/dasd.c
673
block->profile.data->dasd_io_nr_req[counter]++;
drivers/s390/block/dasd.c
675
block->profile.data->dasd_read_nr_req[counter]++;
drivers/s390/block/dasd.c
677
spin_unlock(&block->profile.lock);
drivers/s390/block/dasd.c
760
static void dasd_profile_end(struct dasd_block *block,
drivers/s390/block/dasd.c
773
block->profile.data ||
drivers/s390/block/dasd.c
805
cqr->startdev != block->base,
drivers/s390/block/dasd.c
815
spin_lock(&block->profile.lock);
drivers/s390/block/dasd.c
816
if (block->profile.data) {
drivers/s390/block/dasd.c
817
data = block->profile.data;
drivers/s390/block/dasd.c
822
dasd_profile_end_add_data(block->profile.data,
drivers/s390/block/dasd.c
823
cqr->startdev != block->base,
drivers/s390/block/dasd.c
831
spin_unlock(&block->profile.lock);
drivers/s390/block/dasd.c
841
cqr->startdev != block->base,
drivers/s390/block/dasd_3990_erp.c
1417
if (cqr->block &&
drivers/s390/block/dasd_3990_erp.c
1418
(cqr->block->base != cqr->startdev)) {
drivers/s390/block/dasd_3990_erp.c
1442
dev_name(&cqr->block->base->cdev->dev));
drivers/s390/block/dasd_3990_erp.c
1445
erp->startdev = cqr->block->base;
drivers/s390/block/dasd_3990_erp.c
2370
dasd_block_set_timer(device->block, (HZ << 3));
drivers/s390/block/dasd_3990_erp.c
2402
erp->block = cqr->block;
drivers/s390/block/dasd_3990_erp.c
80
if (erp->block)
drivers/s390/block/dasd_3990_erp.c
81
dasd_block_set_timer(erp->block, expires);
drivers/s390/block/dasd_alias.c
748
cqr->block = NULL;
drivers/s390/block/dasd_alias.c
768
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd_alias.c
775
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd_alias.c
780
dasd_schedule_block_bh(device->block);
drivers/s390/block/dasd_devmap.c
1047
if (!device->block || !device->block->gdp ||
drivers/s390/block/dasd_devmap.c
1053
atomic_inc(&device->block->open_count);
drivers/s390/block/dasd_devmap.c
1056
set_disk_ro(device->block->gdp, val);
drivers/s390/block/dasd_devmap.c
1057
atomic_dec(&device->block->open_count);
drivers/s390/block/dasd_devmap.c
1696
if (IS_ERR(device) || !device->block)
drivers/s390/block/dasd_devmap.c
1704
if (!device->block->gdp) {
drivers/s390/block/dasd_devmap.c
1710
blk_queue_rq_timeout(device->block->gdp->queue, val * HZ);
drivers/s390/block/dasd_diag.c
149
rc = mdsk_init_io(device, device->block->bp_block, 0, NULL);
drivers/s390/block/dasd_diag.c
225
mdsk_init_io(device, device->block->bp_block, 0, NULL);
drivers/s390/block/dasd_diag.c
321
struct dasd_block *block;
drivers/s390/block/dasd_diag.c
338
block = dasd_alloc_block();
drivers/s390/block/dasd_diag.c
339
if (IS_ERR(block)) {
drivers/s390/block/dasd_diag.c
344
return PTR_ERR(block);
drivers/s390/block/dasd_diag.c
346
device->block = block;
drivers/s390/block/dasd_diag.c
347
block->base = device;
drivers/s390/block/dasd_diag.c
444
block->blocks = (unsigned long) label->block_count;
drivers/s390/block/dasd_diag.c
446
block->blocks = end_block;
drivers/s390/block/dasd_diag.c
447
block->bp_block = bsize;
drivers/s390/block/dasd_diag.c
448
block->s2b_shift = 0; /* bits to shift 512 to get a block */
drivers/s390/block/dasd_diag.c
450
block->s2b_shift++;
drivers/s390/block/dasd_diag.c
451
rc = mdsk_init_io(device, block->bp_block, 0, NULL);
drivers/s390/block/dasd_diag.c
461
(unsigned long) block->bp_block,
drivers/s390/block/dasd_diag.c
462
(unsigned long) (block->blocks <<
drivers/s390/block/dasd_diag.c
463
block->s2b_shift) >> 1,
drivers/s390/block/dasd_diag.c
473
device->block = NULL;
drivers/s390/block/dasd_diag.c
474
dasd_free_block(block);
drivers/s390/block/dasd_diag.c
484
dasd_diag_fill_geometry(struct dasd_block *block, struct hd_geometry *geo)
drivers/s390/block/dasd_diag.c
486
if (dasd_check_blocksize(block->bp_block) != 0)
drivers/s390/block/dasd_diag.c
488
geo->cylinders = (block->blocks << block->s2b_shift) >> 10;
drivers/s390/block/dasd_diag.c
490
geo->sectors = 128 >> block->s2b_shift;
drivers/s390/block/dasd_diag.c
509
struct dasd_block *block,
drivers/s390/block/dasd_diag.c
529
blksize = block->bp_block;
drivers/s390/block/dasd_diag.c
531
first_rec = blk_rq_pos(req) >> block->s2b_shift;
drivers/s390/block/dasd_diag.c
533
(blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
drivers/s390/block/dasd_diag.c
540
count += bv.bv_len >> (block->s2b_shift + 9);
drivers/s390/block/dasd_diag.c
570
block->base->features & DASD_FEATURE_FAILFAST)
drivers/s390/block/dasd_diag.c
574
cqr->block = block;
drivers/s390/block/dasd_diag.c
625
static unsigned int dasd_diag_max_sectors(struct dasd_block *block)
drivers/s390/block/dasd_diag.c
627
return DIAG_MAX_BLOCKS << block->s2b_shift;
drivers/s390/block/dasd_eckd.c
1523
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
1623
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
1730
if (cqr->block)
drivers/s390/block/dasd_eckd.c
1731
data->base = cqr->block->base;
drivers/s390/block/dasd_eckd.c
1807
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
1903
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
2019
struct dasd_block *block;
drivers/s390/block/dasd_eckd.c
2027
block = dasd_alloc_block();
drivers/s390/block/dasd_eckd.c
2028
if (IS_ERR(block)) {
drivers/s390/block/dasd_eckd.c
2031
return PTR_ERR(block);
drivers/s390/block/dasd_eckd.c
2033
device->block = block;
drivers/s390/block/dasd_eckd.c
2034
block->base = device;
drivers/s390/block/dasd_eckd.c
2195
dasd_free_block(device->block);
drivers/s390/block/dasd_eckd.c
2196
device->block = NULL;
drivers/s390/block/dasd_eckd.c
2270
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
2318
static int dasd_eckd_start_analysis(struct dasd_block *block)
drivers/s390/block/dasd_eckd.c
2322
init_cqr = dasd_eckd_analysis_ccw(block->base);
drivers/s390/block/dasd_eckd.c
2337
static int dasd_eckd_end_analysis(struct dasd_block *block)
drivers/s390/block/dasd_eckd.c
2339
struct dasd_device *device = block->base;
drivers/s390/block/dasd_eckd.c
2357
block->bp_block = DASD_RAW_BLOCKSIZE;
drivers/s390/block/dasd_eckd.c
2359
block->s2b_shift = 3;
drivers/s390/block/dasd_eckd.c
2410
block->bp_block = count_area->dl;
drivers/s390/block/dasd_eckd.c
2412
if (block->bp_block == 0) {
drivers/s390/block/dasd_eckd.c
2417
block->s2b_shift = 0; /* bits to shift 512 to get a block */
drivers/s390/block/dasd_eckd.c
2418
for (sb = 512; sb < block->bp_block; sb = sb << 1)
drivers/s390/block/dasd_eckd.c
2419
block->s2b_shift++;
drivers/s390/block/dasd_eckd.c
2421
blk_per_trk = recs_per_track(&private->rdc_data, 0, block->bp_block);
drivers/s390/block/dasd_eckd.c
2424
block->blocks = ((unsigned long) private->real_cyl *
drivers/s390/block/dasd_eckd.c
2430
"%s\n", (block->bp_block >> 10),
drivers/s390/block/dasd_eckd.c
2433
blk_per_trk * (block->bp_block >> 9)) >> 1),
drivers/s390/block/dasd_eckd.c
2434
((blk_per_trk * block->bp_block) >> 10),
drivers/s390/block/dasd_eckd.c
2441
static int dasd_eckd_do_analysis(struct dasd_block *block)
drivers/s390/block/dasd_eckd.c
2443
struct dasd_eckd_private *private = block->base->private;
drivers/s390/block/dasd_eckd.c
2446
return dasd_eckd_start_analysis(block);
drivers/s390/block/dasd_eckd.c
2448
return dasd_eckd_end_analysis(block);
drivers/s390/block/dasd_eckd.c
2472
dasd_eckd_fill_geometry(struct dasd_block *block, struct hd_geometry *geo)
drivers/s390/block/dasd_eckd.c
2474
struct dasd_eckd_private *private = block->base->private;
drivers/s390/block/dasd_eckd.c
2476
if (dasd_check_blocksize(block->bp_block) == 0) {
drivers/s390/block/dasd_eckd.c
2478
0, block->bp_block);
drivers/s390/block/dasd_eckd.c
2807
base->block->bp_block);
drivers/s390/block/dasd_eckd.c
3115
struct dasd_block *block = cqr->block;
drivers/s390/block/dasd_eckd.c
3120
spin_lock_irqsave(&block->format_lock, flags);
drivers/s390/block/dasd_eckd.c
3121
if (cqr->trkcount != atomic_read(&block->trkcount)) {
drivers/s390/block/dasd_eckd.c
3131
list_for_each_entry(format, &block->format_list, list) {
drivers/s390/block/dasd_eckd.c
3137
list_add_tail(&to_format->list, &block->format_list);
drivers/s390/block/dasd_eckd.c
3140
spin_unlock_irqrestore(&block->format_lock, flags);
drivers/s390/block/dasd_eckd.c
3145
struct dasd_block *block)
drivers/s390/block/dasd_eckd.c
3149
spin_lock_irqsave(&block->format_lock, flags);
drivers/s390/block/dasd_eckd.c
3150
atomic_inc(&block->trkcount);
drivers/s390/block/dasd_eckd.c
3152
spin_unlock_irqrestore(&block->format_lock, flags);
drivers/s390/block/dasd_eckd.c
3164
clear_format_track(format, cqr->basedev->block);
drivers/s390/block/dasd_eckd.c
3179
struct dasd_block *block;
drivers/s390/block/dasd_eckd.c
3188
block = cqr->block;
drivers/s390/block/dasd_eckd.c
3189
base = block->base;
drivers/s390/block/dasd_eckd.c
3191
blksize = block->bp_block;
drivers/s390/block/dasd_eckd.c
3195
first_trk = blk_rq_pos(req) >> block->s2b_shift;
drivers/s390/block/dasd_eckd.c
3198
(blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
drivers/s390/block/dasd_eckd.c
3262
struct dasd_block *block;
drivers/s390/block/dasd_eckd.c
3273
base = cqr->block->base;
drivers/s390/block/dasd_eckd.c
3274
blksize = base->block->bp_block;
drivers/s390/block/dasd_eckd.c
3275
block = cqr->block;
drivers/s390/block/dasd_eckd.c
3281
first_trk = first_blk = blk_rq_pos(req) >> block->s2b_shift;
drivers/s390/block/dasd_eckd.c
3284
(blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
drivers/s390/block/dasd_eckd.c
3550
if (cqr->block && (cqr->startdev != cqr->block->base)) {
drivers/s390/block/dasd_eckd.c
3552
cqr->startdev = cqr->block->base;
drivers/s390/block/dasd_eckd.c
3553
cqr->lpm = dasd_path_get_opm(cqr->block->base);
drivers/s390/block/dasd_eckd.c
3597
if (!device->block && private->lcu &&
drivers/s390/block/dasd_eckd.c
3649
if (device->block && (sense[27] & DASD_SENSE_BIT_0) &&
drivers/s390/block/dasd_eckd.c
3748
dasd_eckd_dso_ras(struct dasd_device *device, struct dasd_block *block,
drivers/s390/block/dasd_eckd.c
3846
cqr->block = block;
drivers/s390/block/dasd_eckd.c
3875
struct dasd_block *block = device->block;
drivers/s390/block/dasd_eckd.c
3914
spin_lock_irq(&block->queue_lock);
drivers/s390/block/dasd_eckd.c
3916
spin_unlock_irq(&block->queue_lock);
drivers/s390/block/dasd_eckd.c
3927
spin_lock_irq(&block->queue_lock);
drivers/s390/block/dasd_eckd.c
3929
spin_unlock_irq(&block->queue_lock);
drivers/s390/block/dasd_eckd.c
3953
struct dasd_block *block,
drivers/s390/block/dasd_eckd.c
3979
basedev = block->base;
drivers/s390/block/dasd_eckd.c
3995
count += bv.bv_len >> (block->s2b_shift + 9);
drivers/s390/block/dasd_eckd.c
3997
cidaw += bv.bv_len >> (block->s2b_shift + 9);
drivers/s390/block/dasd_eckd.c
4120
block->base->features & DASD_FEATURE_FAILFAST)
drivers/s390/block/dasd_eckd.c
4124
cqr->block = block;
drivers/s390/block/dasd_eckd.c
4141
struct dasd_block *block,
drivers/s390/block/dasd_eckd.c
4169
basedev = block->base;
drivers/s390/block/dasd_eckd.c
4297
block->base->features & DASD_FEATURE_FAILFAST)
drivers/s390/block/dasd_eckd.c
4301
cqr->block = block;
drivers/s390/block/dasd_eckd.c
4473
struct dasd_block *block,
drivers/s390/block/dasd_eckd.c
4504
basedev = block->base;
drivers/s390/block/dasd_eckd.c
4620
block->base->features & DASD_FEATURE_FAILFAST)
drivers/s390/block/dasd_eckd.c
4625
cqr->block = block;
drivers/s390/block/dasd_eckd.c
4645
struct dasd_block *block,
drivers/s390/block/dasd_eckd.c
4661
basedev = block->base;
drivers/s390/block/dasd_eckd.c
4665
blksize = block->bp_block;
drivers/s390/block/dasd_eckd.c
4670
first_rec = first_trk = blk_rq_pos(req) >> block->s2b_shift;
drivers/s390/block/dasd_eckd.c
4673
(blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
drivers/s390/block/dasd_eckd.c
4695
cqr = dasd_eckd_build_cp_tpm_track(startdev, block, req,
drivers/s390/block/dasd_eckd.c
4706
cqr = dasd_eckd_build_cp_cmd_track(startdev, block, req,
drivers/s390/block/dasd_eckd.c
4716
cqr = dasd_eckd_build_cp_cmd_single(startdev, block, req,
drivers/s390/block/dasd_eckd.c
4725
struct dasd_block *block,
drivers/s390/block/dasd_eckd.c
4756
basedev = block->base;
drivers/s390/block/dasd_eckd.c
4861
block->base->features & DASD_FEATURE_FAILFAST)
drivers/s390/block/dasd_eckd.c
4865
cqr->block = block;
drivers/s390/block/dasd_eckd.c
4890
private = cqr->block->base->private;
drivers/s390/block/dasd_eckd.c
4891
blksize = cqr->block->bp_block;
drivers/s390/block/dasd_eckd.c
4893
recid = blk_rq_pos(req) >> cqr->block->s2b_shift;
drivers/s390/block/dasd_eckd.c
4963
struct dasd_block *block,
drivers/s390/block/dasd_eckd.c
4981
cqr = dasd_eckd_build_cp_raw(startdev, block, req);
drivers/s390/block/dasd_eckd.c
4983
cqr = dasd_eckd_build_cp(startdev, block, req);
drivers/s390/block/dasd_eckd.c
5472
dasd_eckd_ioctl(struct dasd_block *block, unsigned int cmd, void __user *argp)
drivers/s390/block/dasd_eckd.c
5474
struct dasd_device *device = block->base;
drivers/s390/block/dasd_eckd.c
5870
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
5939
if (!device->block && private->lcu->pav == HYPER_PAV)
drivers/s390/block/dasd_eckd.c
5963
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
6141
struct dasd_block *block;
drivers/s390/block/dasd_eckd.c
6171
block = primary->block;
drivers/s390/block/dasd_eckd.c
6172
primary->block = NULL;
drivers/s390/block/dasd_eckd.c
6173
secondary->block = block;
drivers/s390/block/dasd_eckd.c
6174
block->base = secondary;
drivers/s390/block/dasd_eckd.c
6179
gdp = block->gdp;
drivers/s390/block/dasd_eckd.c
6231
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
6291
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
6355
cqr->block = NULL;
drivers/s390/block/dasd_eckd.c
6835
static unsigned int dasd_eckd_max_sectors(struct dasd_block *block)
drivers/s390/block/dasd_eckd.c
6837
if (block->base->features & DASD_FEATURE_USERAW) {
drivers/s390/block/dasd_eckd.c
6845
return DASD_ECKD_MAX_BLOCKS_RAW << block->s2b_shift;
drivers/s390/block/dasd_eckd.c
6848
return DASD_ECKD_MAX_BLOCKS << block->s2b_shift;
drivers/s390/block/dasd_eckd.c
835
cqr->block = NULL;
drivers/s390/block/dasd_fba.c
118
struct dasd_block *block;
drivers/s390/block/dasd_fba.c
133
block = dasd_alloc_block();
drivers/s390/block/dasd_fba.c
134
if (IS_ERR(block)) {
drivers/s390/block/dasd_fba.c
139
return PTR_ERR(block);
drivers/s390/block/dasd_fba.c
141
device->block = block;
drivers/s390/block/dasd_fba.c
142
block->base = device;
drivers/s390/block/dasd_fba.c
150
device->block = NULL;
drivers/s390/block/dasd_fba.c
151
dasd_free_block(block);
drivers/s390/block/dasd_fba.c
182
static int dasd_fba_do_analysis(struct dasd_block *block)
drivers/s390/block/dasd_fba.c
184
struct dasd_fba_private *private = block->base->private;
drivers/s390/block/dasd_fba.c
189
DBF_DEV_EVENT(DBF_WARNING, block->base, "unknown blocksize %d",
drivers/s390/block/dasd_fba.c
193
block->blocks = private->rdc_data.blk_bdsa;
drivers/s390/block/dasd_fba.c
194
block->bp_block = private->rdc_data.blk_size;
drivers/s390/block/dasd_fba.c
195
block->s2b_shift = 0; /* bits to shift 512 to get a block */
drivers/s390/block/dasd_fba.c
197
block->s2b_shift++;
drivers/s390/block/dasd_fba.c
201
static int dasd_fba_fill_geometry(struct dasd_block *block,
drivers/s390/block/dasd_fba.c
204
if (dasd_check_blocksize(block->bp_block) != 0)
drivers/s390/block/dasd_fba.c
206
geo->cylinders = (block->blocks << block->s2b_shift) >> 10;
drivers/s390/block/dasd_fba.c
208
geo->sectors = 128 >> block->s2b_shift;
drivers/s390/block/dasd_fba.c
313
struct dasd_block *block,
drivers/s390/block/dasd_fba.c
323
unsigned int blksize = block->bp_block;
drivers/s390/block/dasd_fba.c
333
first_rec = blk_rq_pos(req) >> block->s2b_shift;
drivers/s390/block/dasd_fba.c
335
(blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
drivers/s390/block/dasd_fba.c
410
block->base->features & DASD_FEATURE_FAILFAST)
drivers/s390/block/dasd_fba.c
415
cqr->block = block;
drivers/s390/block/dasd_fba.c
426
struct dasd_block *block,
drivers/s390/block/dasd_fba.c
429
struct dasd_fba_private *private = block->base->private;
drivers/s390/block/dasd_fba.c
448
blksize = block->bp_block;
drivers/s390/block/dasd_fba.c
450
first_rec = blk_rq_pos(req) >> block->s2b_shift;
drivers/s390/block/dasd_fba.c
452
(blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
drivers/s390/block/dasd_fba.c
460
count += bv.bv_len >> (block->s2b_shift + 9);
drivers/s390/block/dasd_fba.c
488
block->bp_block, blk_rq_pos(req), blk_rq_sectors(req));
drivers/s390/block/dasd_fba.c
524
ccw->count = block->bp_block;
drivers/s390/block/dasd_fba.c
539
block->base->features & DASD_FEATURE_FAILFAST)
drivers/s390/block/dasd_fba.c
543
cqr->block = block;
drivers/s390/block/dasd_fba.c
552
struct dasd_block *block,
drivers/s390/block/dasd_fba.c
556
return dasd_fba_build_cp_discard(memdev, block, req);
drivers/s390/block/dasd_fba.c
558
return dasd_fba_build_cp_regular(memdev, block, req);
drivers/s390/block/dasd_fba.c
564
struct dasd_fba_private *private = cqr->block->base->private;
drivers/s390/block/dasd_fba.c
574
blksize = cqr->block->bp_block;
drivers/s390/block/dasd_fba.c
751
static unsigned int dasd_fba_max_sectors(struct dasd_block *block)
drivers/s390/block/dasd_fba.c
753
return DASD_FBA_MAX_BLOCKS << block->s2b_shift;
drivers/s390/block/dasd_genhd.c
100
blk_mq_free_tag_set(&block->tag_set);
drivers/s390/block/dasd_genhd.c
112
DBF_DEV_EVENT(DBF_ERR, block->base,
drivers/s390/block/dasd_genhd.c
122
block->gdp = gdp;
drivers/s390/block/dasd_genhd.c
123
set_capacity(block->gdp, 0);
drivers/s390/block/dasd_genhd.c
125
rc = device_add_disk(&base->cdev->dev, block->gdp, NULL);
drivers/s390/block/dasd_genhd.c
127
dasd_gendisk_free(block);
drivers/s390/block/dasd_genhd.c
147
void dasd_gendisk_free(struct dasd_block *block)
drivers/s390/block/dasd_genhd.c
149
if (block->gdp) {
drivers/s390/block/dasd_genhd.c
150
dasd_gd_free(block->gdp);
drivers/s390/block/dasd_genhd.c
151
block->gdp = NULL;
drivers/s390/block/dasd_genhd.c
152
blk_mq_free_tag_set(&block->tag_set);
drivers/s390/block/dasd_genhd.c
159
int dasd_scan_partitions(struct dasd_block *block)
drivers/s390/block/dasd_genhd.c
164
bdev_file = bdev_file_open_by_dev(disk_devt(block->gdp), BLK_OPEN_READ,
drivers/s390/block/dasd_genhd.c
167
DBF_DEV_EVENT(DBF_ERR, block->base,
drivers/s390/block/dasd_genhd.c
173
mutex_lock(&block->gdp->open_mutex);
drivers/s390/block/dasd_genhd.c
174
rc = bdev_disk_changed(block->gdp, false);
drivers/s390/block/dasd_genhd.c
175
mutex_unlock(&block->gdp->open_mutex);
drivers/s390/block/dasd_genhd.c
177
DBF_DEV_EVENT(DBF_ERR, block->base,
drivers/s390/block/dasd_genhd.c
189
block->bdev_file = bdev_file;
drivers/s390/block/dasd_genhd.c
197
void dasd_destroy_partitions(struct dasd_block *block)
drivers/s390/block/dasd_genhd.c
205
bdev_file = block->bdev_file;
drivers/s390/block/dasd_genhd.c
206
block->bdev_file = NULL;
drivers/s390/block/dasd_genhd.c
67
int dasd_gendisk_alloc(struct dasd_block *block)
drivers/s390/block/dasd_genhd.c
84
base = block->base;
drivers/s390/block/dasd_genhd.c
89
block->tag_set.ops = &dasd_mq_ops;
drivers/s390/block/dasd_genhd.c
90
block->tag_set.cmd_size = sizeof(struct dasd_ccw_req);
drivers/s390/block/dasd_genhd.c
91
block->tag_set.nr_hw_queues = nr_hw_queues;
drivers/s390/block/dasd_genhd.c
92
block->tag_set.queue_depth = queue_depth;
drivers/s390/block/dasd_genhd.c
93
block->tag_set.numa_node = NUMA_NO_NODE;
drivers/s390/block/dasd_genhd.c
94
rc = blk_mq_alloc_tag_set(&block->tag_set);
drivers/s390/block/dasd_genhd.c
98
gdp = blk_mq_alloc_disk(&block->tag_set, &lim, block);
drivers/s390/block/dasd_int.h
134
struct dasd_block *block; /* the originating block device */
drivers/s390/block/dasd_int.h
548
struct dasd_block *block;
drivers/s390/block/dasd_ioctl.c
112
static int dasd_ioctl_resume(struct dasd_block *block)
drivers/s390/block/dasd_ioctl.c
117
base = block->base;
drivers/s390/block/dasd_ioctl.c
127
dasd_schedule_block_bh(block);
drivers/s390/block/dasd_ioctl.c
135
static int dasd_ioctl_abortio(struct dasd_block *block)
drivers/s390/block/dasd_ioctl.c
141
base = block->base;
drivers/s390/block/dasd_ioctl.c
149
spin_lock_irqsave(&block->request_queue_lock, flags);
drivers/s390/block/dasd_ioctl.c
150
spin_lock(&block->queue_lock);
drivers/s390/block/dasd_ioctl.c
151
list_for_each_entry_safe(cqr, n, &block->ccw_queue, blocklist) {
drivers/s390/block/dasd_ioctl.c
156
spin_unlock(&block->queue_lock);
drivers/s390/block/dasd_ioctl.c
158
spin_lock(&block->queue_lock);
drivers/s390/block/dasd_ioctl.c
161
spin_unlock(&block->queue_lock);
drivers/s390/block/dasd_ioctl.c
162
spin_unlock_irqrestore(&block->request_queue_lock, flags);
drivers/s390/block/dasd_ioctl.c
164
dasd_schedule_block_bh(block);
drivers/s390/block/dasd_ioctl.c
171
static int dasd_ioctl_allowio(struct dasd_block *block)
drivers/s390/block/dasd_ioctl.c
175
base = block->base;
drivers/s390/block/dasd_ioctl.c
192
dasd_format(struct dasd_block *block, struct format_data_t *fdata)
drivers/s390/block/dasd_ioctl.c
197
base = block->base;
drivers/s390/block/dasd_ioctl.c
218
block->gdp->part0->bd_mapping->host->i_blkbits =
drivers/s390/block/dasd_ioctl.c
229
static int dasd_check_format(struct dasd_block *block,
drivers/s390/block/dasd_ioctl.c
235
base = block->base;
drivers/s390/block/dasd_ioctl.c
278
rc = dasd_format(base->block, &fdata);
drivers/s390/block/dasd_ioctl.c
311
rc = dasd_check_format(base->block, &cdata);
drivers/s390/block/dasd_ioctl.c
431
static int dasd_ioctl_reset_profile(struct dasd_block *block)
drivers/s390/block/dasd_ioctl.c
433
dasd_profile_reset(&block->profile);
drivers/s390/block/dasd_ioctl.c
440
static int dasd_ioctl_read_profile(struct dasd_block *block, void __user *argp)
drivers/s390/block/dasd_ioctl.c
449
spin_lock_bh(&block->profile.lock);
drivers/s390/block/dasd_ioctl.c
450
if (block->profile.data) {
drivers/s390/block/dasd_ioctl.c
451
data->dasd_io_reqs = block->profile.data->dasd_io_reqs;
drivers/s390/block/dasd_ioctl.c
452
data->dasd_io_sects = block->profile.data->dasd_io_sects;
drivers/s390/block/dasd_ioctl.c
453
memcpy(data->dasd_io_secs, block->profile.data->dasd_io_secs,
drivers/s390/block/dasd_ioctl.c
455
memcpy(data->dasd_io_times, block->profile.data->dasd_io_times,
drivers/s390/block/dasd_ioctl.c
457
memcpy(data->dasd_io_timps, block->profile.data->dasd_io_timps,
drivers/s390/block/dasd_ioctl.c
459
memcpy(data->dasd_io_time1, block->profile.data->dasd_io_time1,
drivers/s390/block/dasd_ioctl.c
461
memcpy(data->dasd_io_time2, block->profile.data->dasd_io_time2,
drivers/s390/block/dasd_ioctl.c
464
block->profile.data->dasd_io_time2ps,
drivers/s390/block/dasd_ioctl.c
466
memcpy(data->dasd_io_time3, block->profile.data->dasd_io_time3,
drivers/s390/block/dasd_ioctl.c
469
block->profile.data->dasd_io_nr_req,
drivers/s390/block/dasd_ioctl.c
471
spin_unlock_bh(&block->profile.lock);
drivers/s390/block/dasd_ioctl.c
473
spin_unlock_bh(&block->profile.lock);
drivers/s390/block/dasd_ioctl.c
484
static int dasd_ioctl_reset_profile(struct dasd_block *block)
drivers/s390/block/dasd_ioctl.c
489
static int dasd_ioctl_read_profile(struct dasd_block *block, void __user *argp)
drivers/s390/block/dasd_ioctl.c
498
static int __dasd_ioctl_information(struct dasd_block *block,
drivers/s390/block/dasd_ioctl.c
509
base = block->base;
drivers/s390/block/dasd_ioctl.c
533
dasd_info->open_count = atomic_read(&block->open_count);
drivers/s390/block/dasd_ioctl.c
534
if (!block->bdev_file)
drivers/s390/block/dasd_ioctl.c
542
(dasd_check_blocksize(block->bp_block)))
drivers/s390/block/dasd_ioctl.c
557
static int dasd_ioctl_information(struct dasd_block *block, void __user *argp,
drivers/s390/block/dasd_ioctl.c
567
error = __dasd_ioctl_information(block, dasd_info);
drivers/s390/block/dasd_ioctl.c
597
static int dasd_ioctl_readall_cmb(struct dasd_block *block, unsigned int cmd,
drivers/s390/block/dasd_ioctl.c
604
ret = cmf_readall(block->base->cdev, &data);
drivers/s390/block/dasd_ioctl.c
613
struct dasd_block *block;
drivers/s390/block/dasd_ioctl.c
626
block = base->block;
drivers/s390/block/dasd_ioctl.c
636
rc = dasd_ioctl_quiesce(block);
drivers/s390/block/dasd_ioctl.c
639
rc = dasd_ioctl_resume(block);
drivers/s390/block/dasd_ioctl.c
642
rc = dasd_ioctl_abortio(block);
drivers/s390/block/dasd_ioctl.c
645
rc = dasd_ioctl_allowio(block);
drivers/s390/block/dasd_ioctl.c
654
rc = dasd_ioctl_information(block, argp,
drivers/s390/block/dasd_ioctl.c
658
rc = dasd_ioctl_information(block, argp,
drivers/s390/block/dasd_ioctl.c
662
rc = dasd_ioctl_read_profile(block, argp);
drivers/s390/block/dasd_ioctl.c
665
rc = dasd_ioctl_reset_profile(block);
drivers/s390/block/dasd_ioctl.c
677
rc = dasd_ioctl_readall_cmb(block, cmd, argp);
drivers/s390/block/dasd_ioctl.c
689
rc = base->discipline->ioctl(block, cmd, argp);
drivers/s390/block/dasd_ioctl.c
721
error = __dasd_ioctl_information(base->block, info);
drivers/s390/block/dasd_ioctl.c
91
static int dasd_ioctl_quiesce(struct dasd_block *block)
drivers/s390/block/dasd_ioctl.c
96
base = block->base;
drivers/s390/block/dasd_proc.c
140
if (device->block)
drivers/s390/block/dasd_proc.c
141
rc = dasd_profile_on(&device->block->profile);
drivers/s390/block/dasd_proc.c
158
if (device->block)
drivers/s390/block/dasd_proc.c
159
dasd_profile_off(&device->block->profile);
drivers/s390/block/dasd_proc.c
173
if (device->block)
drivers/s390/block/dasd_proc.c
174
dasd_profile_reset(&device->block->profile);
drivers/s390/block/dasd_proc.c
34
struct dasd_block *block;
drivers/s390/block/dasd_proc.c
40
if (device->block)
drivers/s390/block/dasd_proc.c
41
block = device->block;
drivers/s390/block/dasd_proc.c
54
if (block->gdp)
drivers/s390/block/dasd_proc.c
56
MAJOR(disk_devt(block->gdp)),
drivers/s390/block/dasd_proc.c
57
MINOR(disk_devt(block->gdp)));
drivers/s390/block/dasd_proc.c
61
if (block->gdp)
drivers/s390/block/dasd_proc.c
62
seq_printf(m, " is %-8s", block->gdp->disk_name);
drivers/s390/block/dasd_proc.c
85
if (dasd_check_blocksize(block->bp_block))
drivers/s390/block/dasd_proc.c
90
block->bp_block, block->blocks,
drivers/s390/block/dasd_proc.c
91
((block->bp_block >> 9) *
drivers/s390/block/dasd_proc.c
92
block->blocks) >> 11);
drivers/s390/char/tape_3490.c
33
unsigned int block : 22;
drivers/s390/char/tape_3490.c
686
return block_id.cbid.block;
drivers/s390/char/tape_3490.c
709
bid->block = mt_count;
drivers/scsi/aacraid/aachba.c
1257
readcmd->block[0] = cpu_to_le32((u32)(lba&0xffffffff));
drivers/scsi/aacraid/aachba.c
1258
readcmd->block[1] = cpu_to_le32((u32)((lba&0xffffffff00000000LL)>>32));
drivers/scsi/aacraid/aachba.c
1297
readcmd->block = cpu_to_le32((u32)(lba&0xffffffff));
drivers/scsi/aacraid/aachba.c
1332
readcmd->block = cpu_to_le32((u32)(lba&0xffffffff));
drivers/scsi/aacraid/aachba.c
1388
writecmd->block[0] = cpu_to_le32((u32)(lba&0xffffffff));
drivers/scsi/aacraid/aachba.c
1389
writecmd->block[1] = cpu_to_le32((u32)((lba&0xffffffff00000000LL)>>32));
drivers/scsi/aacraid/aachba.c
1431
writecmd->block = cpu_to_le32((u32)(lba&0xffffffff));
drivers/scsi/aacraid/aachba.c
1466
writecmd->block = cpu_to_le32((u32)(lba&0xffffffff));
drivers/scsi/aacraid/aacraid.h
1854
__le32 block;
drivers/scsi/aacraid/aacraid.h
1864
__le32 block;
drivers/scsi/aacraid/aacraid.h
1880
__le32 block;
drivers/scsi/aacraid/aacraid.h
1891
__le32 block;
drivers/scsi/aacraid/aacraid.h
1905
__le32 block[2];
drivers/scsi/bfa/bfa_ioc.c
4695
bfa_trc(diag, diag->block);
drivers/scsi/bfa/bfa_ioc.c
4716
if (diag->block) {
drivers/scsi/bfa/bfa_ioc.c
4724
diag->block = 0;
drivers/scsi/bfa/bfa_ioc.c
4778
diag->block = 0;
drivers/scsi/bfa/bfa_ioc.c
5050
if (diag->block) {
drivers/scsi/bfa/bfa_ioc.c
5051
bfa_trc(diag, diag->block);
drivers/scsi/bfa/bfa_ioc.c
5054
diag->block = 1;
drivers/scsi/bfa/bfa_ioc.c
5099
if (diag->block || diag->fwping.lock) {
drivers/scsi/bfa/bfa_ioc.c
5100
bfa_trc(diag, diag->block);
drivers/scsi/bfa/bfa_ioc.c
5138
if (diag->block || diag->tsensor.lock) {
drivers/scsi/bfa/bfa_ioc.c
5139
bfa_trc(diag, diag->block);
drivers/scsi/bfa/bfa_ioc.c
5247
diag->block = 0;
drivers/scsi/bfa/bfa_ioc.h
647
u8 block;
drivers/scsi/hpsa.c
4675
u32 block;
drivers/scsi/hpsa.c
4687
block = (((cdb[1] & 0x1F) << 16) |
drivers/scsi/hpsa.c
4695
block = get_unaligned_be32(&cdb[2]);
drivers/scsi/hpsa.c
4703
cdb[2] = (u8) (block >> 24);
drivers/scsi/hpsa.c
4704
cdb[3] = (u8) (block >> 16);
drivers/scsi/hpsa.c
4705
cdb[4] = (u8) (block >> 8);
drivers/scsi/hpsa.c
4706
cdb[5] = (u8) (block);
drivers/scsi/megaraid/megaraid_sas_fp.c
1360
u64 block = io_info->ldStartBlock;
drivers/scsi/megaraid/megaraid_sas_fp.c
1389
diff0 = ABS_DIFF(block, lbInfo->last_accessed_block[pd0]);
drivers/scsi/megaraid/megaraid_sas_fp.c
1390
diff1 = ABS_DIFF(block, lbInfo->last_accessed_block[pd1]);
drivers/scsi/megaraid/megaraid_sas_fp.c
1406
lbInfo->last_accessed_block[io_info->pd_after_lb] = block + count - 1;
drivers/scsi/mpi3mr/mpi3mr_os.c
2877
u8 ublock = 0, block = 0, hide = 0, delete = 0, remove = 0;
drivers/scsi/mpi3mr/mpi3mr_os.c
2894
block = 1;
drivers/scsi/mpi3mr/mpi3mr_os.c
2924
if (block)
drivers/scsi/mpt3sas/mpt3sas_base.h
483
u8 block;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
3868
sas_device_priv_data->block = 1;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
3893
sas_device_priv_data->block = 0;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
3905
sas_device_priv_data->block = 1;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
3912
sas_device_priv_data->block = 0;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
3951
if (!sas_device_priv_data->block)
drivers/scsi/mpt3sas/mpt3sas_scsih.c
3979
sas_device_priv_data->block = 0;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4055
sas_device_priv_data->block = 1;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4073
if (!sas_device_priv_data->block)
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4111
sas_device_priv_data->block = 0;
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4178
if (sas_device_priv_data->block)
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4200
if (sas_device_priv_data->block)
drivers/scsi/mpt3sas/mpt3sas_scsih.c
4234
if (sas_device_priv_data->block)
drivers/scsi/mpt3sas/mpt3sas_scsih.c
5462
if (handle == MPT3SAS_INVALID_DEVICE_HANDLE || sas_device_priv_data->block) {
drivers/scsi/mpt3sas/mpt3sas_scsih.c
5487
sas_device_priv_data->block) {
drivers/scsi/mpt3sas/mpt3sas_scsih.c
6123
if (sas_device_priv_data->block) {
drivers/scsi/mvsas/mv_94xx.c
1069
u32 block = ioread32be(regs + MVS_SGPIO_DCTRL +
drivers/scsi/mvsas/mv_94xx.c
1085
block &= ~((0x7 << MVS_SGPIO_DCTRL_ACT_SHIFT)
drivers/scsi/mvsas/mv_94xx.c
1088
block |= LED_BLINKA_SOF << (
drivers/scsi/mvsas/mv_94xx.c
1093
block &= ~((0x3 << MVS_SGPIO_DCTRL_LOC_SHIFT)
drivers/scsi/mvsas/mv_94xx.c
1095
block |= bit << MVS_SGPIO_DCTRL_LOC_SHIFT;
drivers/scsi/mvsas/mv_94xx.c
1098
block &= ~((0x7 << MVS_SGPIO_DCTRL_ERR_SHIFT)
drivers/scsi/mvsas/mv_94xx.c
1100
block |= bit << MVS_SGPIO_DCTRL_ERR_SHIFT;
drivers/scsi/mvsas/mv_94xx.c
1104
iowrite32be(block,
drivers/scsi/qedi/qedi_main.c
1063
struct nvm_iscsi_block *block = NULL;
drivers/scsi/qedi/qedi_main.c
1090
block = qedi_get_nvram_block(qedi);
drivers/scsi/qedi/qedi_main.c
1091
if (block) {
drivers/scsi/qedi/qedi_main.c
1092
chap_en = !!(block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
1094
mchap_en = !!(block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
1110
rval = qedi_find_boot_info(qedi, iscsi, block);
drivers/scsi/qedi/qedi_main.c
2018
struct nvm_iscsi_block *block;
drivers/scsi/qedi/qedi_main.c
2021
block = &qedi->iscsi_image->iscsi_cfg.block[0];
drivers/scsi/qedi/qedi_main.c
2022
for (i = 0; i < NUM_OF_ISCSI_PF_SUPPORTED; i++, block++) {
drivers/scsi/qedi/qedi_main.c
2023
flags = ((block->id) & NVM_ISCSI_CFG_BLK_CTRL_FLAG_MASK) >>
drivers/scsi/qedi/qedi_main.c
2027
(pf == (block->id & NVM_ISCSI_CFG_BLK_MAPPED_PF_ID_MASK)
drivers/scsi/qedi/qedi_main.c
2029
return block;
drivers/scsi/qedi/qedi_main.c
2040
struct nvm_iscsi_block *block;
drivers/scsi/qedi/qedi_main.c
2043
block = qedi_get_nvram_block(qedi);
drivers/scsi/qedi/qedi_main.c
2044
if (!block)
drivers/scsi/qedi/qedi_main.c
2047
initiator = &block->initiator;
drivers/scsi/qedi/qedi_main.c
2048
ipv6_en = block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
2050
dhcp_en = block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
2132
struct nvm_iscsi_block *block;
drivers/scsi/qedi/qedi_main.c
2134
block = qedi_get_nvram_block(qedi);
drivers/scsi/qedi/qedi_main.c
2135
if (!block)
drivers/scsi/qedi/qedi_main.c
2138
initiator = &block->initiator;
drivers/scsi/qedi/qedi_main.c
2173
struct nvm_iscsi_block *block;
drivers/scsi/qedi/qedi_main.c
2177
block = qedi_get_nvram_block(qedi);
drivers/scsi/qedi/qedi_main.c
2178
if (!block)
drivers/scsi/qedi/qedi_main.c
2183
GET_FIELD2(block->id, NVM_ISCSI_CFG_BLK_MAPPED_PF_ID), idx);
drivers/scsi/qedi/qedi_main.c
2185
ctrl_flags = block->target[idx].ctrl_flags &
drivers/scsi/qedi/qedi_main.c
2194
ipv6_en = block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
2197
chap_en = block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
2199
chap_name = chap_en ? block->initiator.chap_name.byte : NULL;
drivers/scsi/qedi/qedi_main.c
2200
chap_secret = chap_en ? block->initiator.chap_password.byte : NULL;
drivers/scsi/qedi/qedi_main.c
2202
mchap_en = block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
2204
mchap_name = mchap_en ? block->target[idx].chap_name.byte : NULL;
drivers/scsi/qedi/qedi_main.c
2205
mchap_secret = mchap_en ? block->target[idx].chap_password.byte : NULL;
drivers/scsi/qedi/qedi_main.c
2210
block->target[idx].target_name.byte);
drivers/scsi/qedi/qedi_main.c
2215
block->target[idx].ipv6_addr.byte);
drivers/scsi/qedi/qedi_main.c
2218
block->target[idx].ipv4_addr.byte);
drivers/scsi/qedi/qedi_main.c
2222
GET_FIELD2(block->target[idx].generic_cont0,
drivers/scsi/qedi/qedi_main.c
2227
block->target[idx].lun.value[1],
drivers/scsi/qedi/qedi_main.c
2228
block->target[idx].lun.value[0]);
drivers/scsi/qedi/qedi_main.c
920
static void qedi_get_boot_tgt_info(struct nvm_iscsi_block *block,
drivers/scsi/qedi/qedi_main.c
925
ipv6_en = !!(block->generic.ctrl_flags &
drivers/scsi/qedi/qedi_main.c
929
block->target[index].target_name.byte);
drivers/scsi/qedi/qedi_main.c
935
block->target[index].ipv6_addr.byte);
drivers/scsi/qedi/qedi_main.c
938
block->target[index].ipv4_addr.byte);
drivers/scsi/qedi/qedi_main.c
943
struct nvm_iscsi_block *block)
drivers/scsi/qedi/qedi_main.c
955
pri_ctrl_flags = !!(block->target[0].ctrl_flags &
drivers/scsi/qedi/qedi_main.c
961
qedi_get_boot_tgt_info(block, pri_tgt, 0);
drivers/scsi/qedi/qedi_main.c
964
sec_ctrl_flags = !!(block->target[1].ctrl_flags &
drivers/scsi/qedi/qedi_main.c
972
qedi_get_boot_tgt_info(block, sec_tgt, 1);
drivers/scsi/qedi/qedi_nvm_iscsi_cfg.h
204
struct nvm_iscsi_block block[NUM_OF_ISCSI_PF_SUPPORTED]; /* 0x4 */
drivers/scsi/scsi_debug.c
4235
u64 block;
drivers/scsi/scsi_debug.c
4265
block = do_div(lba, sdebug_store_sectors);
drivers/scsi/scsi_debug.c
4273
fsp + (block * sdebug_sector_size),
drivers/scsi/scsi_debug.c
4280
if (++block >= sdebug_store_sectors)
drivers/scsi/scsi_debug.c
4281
block = 0;
drivers/scsi/scsi_debug.c
4308
u64 block, rest = 0;
drivers/scsi/scsi_debug.c
4313
block = do_div(lba, store_blks);
drivers/scsi/scsi_debug.c
4314
if (block + num > store_blks)
drivers/scsi/scsi_debug.c
4315
rest = block + num - store_blks;
drivers/scsi/scsi_debug.c
4317
res = !memcmp(fsp + (block * lb_size), arr, (num - rest) * lb_size);
drivers/scsi/scsi_debug.c
4328
memcpy(fsp + (block * lb_size), arr, (num - rest) * lb_size);
drivers/scsi/scsi_debug.c
5254
u64 block, lbaa;
drivers/scsi/scsi_debug.c
5277
block = do_div(lbaa, sdebug_store_sectors);
drivers/scsi/scsi_debug.c
5280
fs1p = fsp + (block * lb_size);
drivers/scsi/scsi_debug.c
5299
block = do_div(lbaa, sdebug_store_sectors);
drivers/scsi/scsi_debug.c
5300
memmove(fsp + (block * lb_size), fs1p, lb_size);
drivers/scsi/scsi_debug.c
5673
u64 block, rest = 0;
drivers/scsi/scsi_debug.c
5693
block = do_div(lba, sdebug_store_sectors);
drivers/scsi/scsi_debug.c
5694
if (block + nblks > sdebug_store_sectors)
drivers/scsi/scsi_debug.c
5695
rest = block + nblks - sdebug_store_sectors;
drivers/scsi/scsi_debug.c
5699
prefetch_range(fsp + (sdebug_sector_size * block),
drivers/scsi/scsi_debug.c
7125
static void block_unblock_all_queues(bool block)
drivers/scsi/scsi_debug.c
7134
if (block)
drivers/scsi/sd_zbc.c
330
sector_t block = sectors_to_logical(sdkp->device, sector);
drivers/scsi/sd_zbc.c
344
put_unaligned_be64(block, &cmd->cmnd[2]);
drivers/scsi/sr.c
361
int block = 0, this_count, s_size;
drivers/scsi/sr.c
372
"Doing sr request, block = %d\n", block));
drivers/scsi/sr.c
446
block = (unsigned int)blk_rq_pos(rq) / (s_size >> 9);
drivers/scsi/sr.c
453
put_unaligned_be32(block, &SCpnt->cmnd[2]);
drivers/scsi/st.c
3135
static int get_location(struct scsi_tape *STp, unsigned int *block, int *partition,
drivers/scsi/st.c
3163
*block = *partition = 0;
drivers/scsi/st.c
3169
*block = ((STp->buffer)->b_data[0] << 16)
drivers/scsi/st.c
3174
*block = ((STp->buffer)->b_data[4] << 24)
drivers/scsi/st.c
3184
*block, *partition);
drivers/scsi/st.c
3195
static int set_location(struct scsi_tape *STp, unsigned int block, int partition,
drivers/scsi/st.c
3211
block, partition);
drivers/scsi/st.c
3234
scmd[2] = (block >> 16);
drivers/scsi/st.c
3235
scmd[3] = (block >> 8);
drivers/scsi/st.c
3236
scmd[4] = block;
drivers/scsi/st.c
3240
scmd[3] = (block >> 24);
drivers/scsi/st.c
3241
scmd[4] = (block >> 16);
drivers/scsi/st.c
3242
scmd[5] = (block >> 8);
drivers/scsi/st.c
3243
scmd[6] = block;
drivers/scsi/st.c
3277
STps->last_block_visited != block) {
drivers/scsi/st.c
3283
if (block == 0)
drivers/scsi/st.c
3300
unsigned int block;
drivers/scsi/st.c
3302
if ((i = get_location(STp, &block, &partition, 1)) < 0)
drivers/soc/tegra/cbb/tegra234-cbb.c
279
unsigned int block = 0;
drivers/soc/tegra/cbb/tegra234-cbb.c
292
addr = base + APB_BLOCK_NUM_TMO_OFFSET + (block * 4);
drivers/soc/tegra/cbb/tegra234-cbb.c
301
sprintf(name, "%s_BLOCK%d_TMO", target, block);
drivers/soc/tegra/cbb/tegra234-cbb.c
312
block++;
drivers/soc/ti/knav_qmss_queue.c
1131
struct knav_link_ram_block *block)
drivers/soc/ti/knav_qmss_queue.c
1156
block->dma = (dma_addr_t)temp[0];
drivers/soc/ti/knav_qmss_queue.c
1157
block->virt = NULL;
drivers/soc/ti/knav_qmss_queue.c
1158
block->size = temp[1];
drivers/soc/ti/knav_qmss_queue.c
1160
block->size = temp[1];
drivers/soc/ti/knav_qmss_queue.c
1162
block->virt = dmam_alloc_coherent(kdev->dev,
drivers/soc/ti/knav_qmss_queue.c
1163
8 * block->size, &block->dma,
drivers/soc/ti/knav_qmss_queue.c
1165
if (!block->virt) {
drivers/soc/ti/knav_qmss_queue.c
1178
struct knav_link_ram_block *block;
drivers/soc/ti/knav_qmss_queue.c
1182
block = &kdev->link_rams[0];
drivers/soc/ti/knav_qmss_queue.c
1184
&block->dma, block->virt, block->size);
drivers/soc/ti/knav_qmss_queue.c
1185
writel_relaxed((u32)block->dma, &qmgr->reg_config->link_ram_base0);
drivers/soc/ti/knav_qmss_queue.c
1187
writel_relaxed(block->size,
drivers/soc/ti/knav_qmss_queue.c
1190
writel_relaxed(block->size - 1,
drivers/soc/ti/knav_qmss_queue.c
1192
block++;
drivers/soc/ti/knav_qmss_queue.c
1193
if (!block->size)
drivers/soc/ti/knav_qmss_queue.c
1197
&block->dma, block->virt, block->size);
drivers/soc/ti/knav_qmss_queue.c
1198
writel_relaxed(block->dma, &qmgr->reg_config->link_ram_base1);
drivers/spi/spi-cs42l43.c
107
while (buf < block) {
drivers/spi/spi-cs42l43.c
108
u8 *word = min(buf + sizeof(u32), block);
drivers/spi/spi-cs42l43.c
62
const u8 *block = min(buf + CS42L43_FIFO_SIZE, end);
drivers/spi/spi-cs42l43.c
64
while (buf < block) {
drivers/spi/spi-cs42l43.c
65
const u8 *word = min(buf + sizeof(u32), block);
drivers/spi/spi-cs42l43.c
99
u8 *block = min(buf + CS42L43_FIFO_SIZE, end);
drivers/staging/media/atomisp/pci/ia_css_acc_types.h
213
struct ia_css_binary_block_info block;
drivers/staging/media/atomisp/pci/runtime/ifmtr/src/ifmtr.c
454
bool block[] = { false, false, false, false };
drivers/staging/media/atomisp/pci/runtime/ifmtr/src/ifmtr.c
456
assert(N_INPUT_FORMATTER_ID <= (ARRAY_SIZE(block)));
drivers/staging/media/atomisp/pci/runtime/ifmtr/src/ifmtr.c
458
block[INPUT_FORMATTER0_ID] = (bool)config_a->block_no_reqs;
drivers/staging/media/atomisp/pci/runtime/ifmtr/src/ifmtr.c
460
block[INPUT_FORMATTER1_ID] = (bool)config_b->block_no_reqs;
drivers/staging/media/atomisp/pci/runtime/ifmtr/src/ifmtr.c
471
input_formatter_set_fifo_blocking_mode(id, block[id]);
drivers/staging/media/atomisp/pci/sh_css.c
4717
mycs->video_binary.info->sp.block.output_block_height);
drivers/staging/media/ipu3/ipu3-abi.h
1652
struct imgu_abi_binary_block_info block;
drivers/staging/media/ipu3/ipu3-css-fw.c
214
if (bi->info.isp.sp.block.block_width <= 0 ||
drivers/staging/media/ipu3/ipu3-css-fw.c
215
bi->info.isp.sp.block.block_width > BLOCK_MAX ||
drivers/staging/media/ipu3/ipu3-css-fw.c
216
bi->info.isp.sp.block.output_block_height <= 0 ||
drivers/staging/media/ipu3/ipu3-css-fw.c
217
bi->info.isp.sp.block.output_block_height > BLOCK_MAX)
drivers/staging/media/ipu3/ipu3-css-params.c
1190
bi->info.isp.sp.block.block_width *
drivers/staging/media/ipu3/ipu3-css.c
1261
bi->info.isp.sp.block.block_width *
drivers/staging/media/ipu3/ipu3-css.c
1265
bi->info.isp.sp.block.output_block_height);
drivers/staging/media/tegra-video/tegra210.c
143
#define CSI_PP_OFFSET(block) ((block) * 0x800)
drivers/staging/rtl8723bs/include/ioctl_cfg80211.h
81
bool block;
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
1637
if (adapter_wdev_data(padapter)->block == true) {
drivers/target/target_core_user.c
3191
u32 start, end, block, total_blocks_freed = 0;
drivers/target/target_core_user.c
3216
block = find_last_bit(udev->data_bitmap, end);
drivers/target/target_core_user.c
3217
if (block == udev->dbi_max) {
drivers/target/target_core_user.c
3224
} else if (block == end) {
drivers/target/target_core_user.c
3229
udev->dbi_thresh = start = block + 1;
drivers/target/target_core_user.c
3230
udev->dbi_max = block;
drivers/thunderbolt/property.c
116
dir = __tb_property_parse_dir(block, block_len, entry->value,
drivers/thunderbolt/property.c
132
parse_dwdata(property->value.data, block + entry->value,
drivers/thunderbolt/property.c
143
parse_dwdata(property->value.text, block + entry->value,
drivers/thunderbolt/property.c
161
static struct tb_property_dir *__tb_property_parse_dir(const u32 *block,
drivers/thunderbolt/property.c
177
dir->uuid = kmemdup(&block[dir_offset], sizeof(*dir->uuid),
drivers/thunderbolt/property.c
187
entries = (const struct tb_property_entry *)&block[content_offset];
drivers/thunderbolt/property.c
195
property = tb_property_parse(block, block_len, &entries[i]);
drivers/thunderbolt/property.c
222
struct tb_property_dir *tb_property_parse_dir(const u32 *block,
drivers/thunderbolt/property.c
226
(const struct tb_property_rootdir_entry *)block;
drivers/thunderbolt/property.c
233
return __tb_property_parse_dir(block, block_len, 0, rootdir->length,
drivers/thunderbolt/property.c
353
u32 *block, unsigned int start_offset, size_t block_len)
drivers/thunderbolt/property.c
38
static struct tb_property_dir *__tb_property_parse_dir(const u32 *block,
drivers/thunderbolt/property.c
419
pe = (struct tb_property_dir_entry *)&block[start_offset];
drivers/thunderbolt/property.c
425
re = (struct tb_property_rootdir_entry *)&block[start_offset];
drivers/thunderbolt/property.c
440
ret = __tb_property_format_dir(child, block, dir_end,
drivers/thunderbolt/property.c
451
format_dwdata(&block[data_offset], property->value.data,
drivers/thunderbolt/property.c
459
format_dwdata(&block[data_offset], property->value.text,
drivers/thunderbolt/property.c
494
ssize_t tb_property_format_dir(const struct tb_property_dir *dir, u32 *block,
drivers/thunderbolt/property.c
499
if (!block) {
drivers/thunderbolt/property.c
506
ret = __tb_property_format_dir(dir, block, 0, block_len);
drivers/thunderbolt/property.c
95
static struct tb_property *tb_property_parse(const u32 *block, size_t block_len,
drivers/thunderbolt/test.c
2730
u32 *block;
drivers/thunderbolt/test.c
2741
block = kunit_kzalloc(test, block_len * sizeof(u32), GFP_KERNEL);
drivers/thunderbolt/test.c
2742
KUNIT_ASSERT_NOT_NULL(test, block);
drivers/thunderbolt/test.c
2744
ret = tb_property_format_dir(dir, block, block_len);
drivers/thunderbolt/test.c
2748
KUNIT_EXPECT_EQ(test, root_directory[i], block[i]);
drivers/thunderbolt/test.c
2826
u32 *block;
drivers/thunderbolt/test.c
2842
block = kunit_kzalloc(test, sizeof(root_directory), GFP_KERNEL);
drivers/thunderbolt/test.c
2843
KUNIT_ASSERT_NOT_NULL(test, block);
drivers/thunderbolt/test.c
2845
ret = tb_property_format_dir(dst, block, ARRAY_SIZE(root_directory));
drivers/thunderbolt/test.c
2849
KUNIT_EXPECT_EQ(test, root_directory[i], block[i]);
drivers/thunderbolt/xdomain.c
1391
u32 *block = NULL;
drivers/thunderbolt/xdomain.c
1399
&block, &gen);
drivers/thunderbolt/xdomain.c
1421
dir = tb_property_parse_dir(block, ret);
drivers/thunderbolt/xdomain.c
1447
kfree(block);
drivers/thunderbolt/xdomain.c
1495
kfree(block);
drivers/thunderbolt/xdomain.c
325
u32 **block, u32 *generation)
drivers/thunderbolt/xdomain.c
400
*block = data;
drivers/thunderbolt/xdomain.c
660
u32 *block;
drivers/thunderbolt/xdomain.c
680
block = kcalloc(block_len, sizeof(*block), GFP_KERNEL);
drivers/thunderbolt/xdomain.c
681
if (!block) {
drivers/thunderbolt/xdomain.c
686
ret = tb_property_format_dir(dir, block, block_len);
drivers/thunderbolt/xdomain.c
690
kfree(block);
drivers/thunderbolt/xdomain.c
698
xd->local_property_block = block;
drivers/tty/mips_ejtag_fdc.c
803
int count, block;
drivers/tty/mips_ejtag_fdc.c
824
for (count = total; count; count -= block) {
drivers/tty/mips_ejtag_fdc.c
825
block = min(count, (int)(priv->xmit_size - dport->xmit_head));
drivers/tty/mips_ejtag_fdc.c
826
memcpy(dport->port.xmit_buf + dport->xmit_head, buf, block);
drivers/tty/mips_ejtag_fdc.c
827
dport->xmit_head += block;
drivers/tty/mips_ejtag_fdc.c
830
buf += block;
drivers/usb/gadget/function/f_midi2.c
103
unsigned char block;
drivers/usb/gadget/function/f_midi2.c
1690
jack->iJack = map->ep->blks[map->block].string_id;
drivers/usb/gadget/function/f_midi2.c
1718
jack->iJack = map->ep->blks[map->block].string_id;
drivers/usb/gadget/function/f_midi2.c
2784
map->block = blk;
drivers/usb/gadget/function/f_midi2.c
2799
map->block = blk;
drivers/usb/gadget/udc/lpc32xx_udc.c
645
int block)
drivers/usb/gadget/udc/lpc32xx_udc.c
651
if (block)
drivers/usb/serial/f81534.c
406
size_t block = 0;
drivers/usb/serial/f81534.c
460
offset = count + block * F81534_MAX_DATA_BLOCK;
drivers/usb/serial/f81534.c
465
++block;
drivers/usb/storage/datafab.c
551
unsigned long block, blocks;
drivers/usb/storage/datafab.c
601
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/datafab.c
607
block, blocks);
drivers/usb/storage/datafab.c
608
return datafab_read_data(us, info, block, blocks);
drivers/usb/storage/datafab.c
614
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/datafab.c
621
block, blocks);
drivers/usb/storage/datafab.c
622
return datafab_read_data(us, info, block, blocks);
drivers/usb/storage/datafab.c
626
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/datafab.c
632
block, blocks);
drivers/usb/storage/datafab.c
633
return datafab_write_data(us, info, block, blocks);
drivers/usb/storage/datafab.c
639
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/datafab.c
646
block, blocks);
drivers/usb/storage/datafab.c
647
return datafab_write_data(us, info, block, blocks);
drivers/usb/storage/jumpshot.c
478
unsigned long block, blocks;
drivers/usb/storage/jumpshot.c
530
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/jumpshot.c
536
block, blocks);
drivers/usb/storage/jumpshot.c
537
return jumpshot_read_data(us, info, block, blocks);
drivers/usb/storage/jumpshot.c
543
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/jumpshot.c
550
block, blocks);
drivers/usb/storage/jumpshot.c
551
return jumpshot_read_data(us, info, block, blocks);
drivers/usb/storage/jumpshot.c
555
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/jumpshot.c
561
block, blocks);
drivers/usb/storage/jumpshot.c
562
return jumpshot_write_data(us, info, block, blocks);
drivers/usb/storage/jumpshot.c
568
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/jumpshot.c
575
block, blocks);
drivers/usb/storage/jumpshot.c
576
return jumpshot_write_data(us, info, block, blocks);
drivers/usb/storage/shuttle_usbat.c
1684
unsigned long block, blocks;
drivers/usb/storage/shuttle_usbat.c
1729
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/shuttle_usbat.c
1735
block, blocks);
drivers/usb/storage/shuttle_usbat.c
1736
return usbat_flash_read_data(us, info, block, blocks);
drivers/usb/storage/shuttle_usbat.c
1743
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/shuttle_usbat.c
1750
block, blocks);
drivers/usb/storage/shuttle_usbat.c
1751
return usbat_flash_read_data(us, info, block, blocks);
drivers/usb/storage/shuttle_usbat.c
1755
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/shuttle_usbat.c
1761
block, blocks);
drivers/usb/storage/shuttle_usbat.c
1762
return usbat_flash_write_data(us, info, block, blocks);
drivers/usb/storage/shuttle_usbat.c
1769
block = ((u32)(srb->cmnd[2]) << 24) | ((u32)(srb->cmnd[3]) << 16) |
drivers/usb/storage/shuttle_usbat.c
1776
block, blocks);
drivers/usb/storage/shuttle_usbat.c
1777
return usbat_flash_write_data(us, info, block, blocks);
drivers/video/fbdev/core/fbmon.c
101
static int edid_is_serial_block(unsigned char *block)
drivers/video/fbdev/core/fbmon.c
103
if ((block[0] == 0x00) && (block[1] == 0x00) &&
drivers/video/fbdev/core/fbmon.c
104
(block[2] == 0x00) && (block[3] == 0xff) &&
drivers/video/fbdev/core/fbmon.c
105
(block[4] == 0x00))
drivers/video/fbdev/core/fbmon.c
111
static int edid_is_ascii_block(unsigned char *block)
drivers/video/fbdev/core/fbmon.c
113
if ((block[0] == 0x00) && (block[1] == 0x00) &&
drivers/video/fbdev/core/fbmon.c
114
(block[2] == 0x00) && (block[3] == 0xfe) &&
drivers/video/fbdev/core/fbmon.c
115
(block[4] == 0x00))
drivers/video/fbdev/core/fbmon.c
121
static int edid_is_limits_block(unsigned char *block)
drivers/video/fbdev/core/fbmon.c
123
if ((block[0] == 0x00) && (block[1] == 0x00) &&
drivers/video/fbdev/core/fbmon.c
124
(block[2] == 0x00) && (block[3] == 0xfd) &&
drivers/video/fbdev/core/fbmon.c
125
(block[4] == 0x00))
drivers/video/fbdev/core/fbmon.c
131
static int edid_is_monitor_block(unsigned char *block)
drivers/video/fbdev/core/fbmon.c
133
if ((block[0] == 0x00) && (block[1] == 0x00) &&
drivers/video/fbdev/core/fbmon.c
134
(block[2] == 0x00) && (block[3] == 0xfc) &&
drivers/video/fbdev/core/fbmon.c
135
(block[4] == 0x00))
drivers/video/fbdev/core/fbmon.c
141
static int edid_is_timing_block(unsigned char *block)
drivers/video/fbdev/core/fbmon.c
143
if ((block[0] != 0x00) || (block[1] != 0x00) ||
drivers/video/fbdev/core/fbmon.c
144
(block[2] != 0x00) || (block[4] != 0x00))
drivers/video/fbdev/core/fbmon.c
152
unsigned char *block = edid + ID_MANUFACTURER_NAME, manufacturer[4];
drivers/video/fbdev/core/fbmon.c
157
manufacturer[0] = ((block[0] & 0x7c) >> 2) + '@';
drivers/video/fbdev/core/fbmon.c
158
manufacturer[1] = ((block[0] & 0x03) << 3) +
drivers/video/fbdev/core/fbmon.c
159
((block[1] & 0xe0) >> 5) + '@';
drivers/video/fbdev/core/fbmon.c
160
manufacturer[2] = (block[1] & 0x1f) + '@';
drivers/video/fbdev/core/fbmon.c
162
model = block[2] + (block[3] << 8);
drivers/video/fbdev/core/fbmon.c
297
static void parse_vendor_block(unsigned char *block, struct fb_monspecs *specs)
drivers/video/fbdev/core/fbmon.c
299
specs->manufacturer[0] = ((block[0] & 0x7c) >> 2) + '@';
drivers/video/fbdev/core/fbmon.c
300
specs->manufacturer[1] = ((block[0] & 0x03) << 3) +
drivers/video/fbdev/core/fbmon.c
301
((block[1] & 0xe0) >> 5) + '@';
drivers/video/fbdev/core/fbmon.c
302
specs->manufacturer[2] = (block[1] & 0x1f) + '@';
drivers/video/fbdev/core/fbmon.c
304
specs->model = block[2] + (block[3] << 8);
drivers/video/fbdev/core/fbmon.c
305
specs->serial = block[4] + (block[5] << 8) +
drivers/video/fbdev/core/fbmon.c
306
(block[6] << 16) + (block[7] << 24);
drivers/video/fbdev/core/fbmon.c
307
specs->year = block[9] + 1990;
drivers/video/fbdev/core/fbmon.c
308
specs->week = block[8];
drivers/video/fbdev/core/fbmon.c
331
static void get_chroma(unsigned char *block, struct fb_monspecs *specs)
drivers/video/fbdev/core/fbmon.c
337
tmp = ((block[5] & (3 << 6)) >> 6) | (block[0x7] << 2);
drivers/video/fbdev/core/fbmon.c
343
tmp = ((block[5] & (3 << 4)) >> 4) | (block[0x8] << 2);
drivers/video/fbdev/core/fbmon.c
349
tmp = ((block[5] & (3 << 2)) >> 2) | (block[0x9] << 2);
drivers/video/fbdev/core/fbmon.c
355
tmp = (block[5] & 3) | (block[0xa] << 2);
drivers/video/fbdev/core/fbmon.c
361
tmp = ((block[6] & (3 << 6)) >> 6) | (block[0xb] << 2);
drivers/video/fbdev/core/fbmon.c
367
tmp = ((block[6] & (3 << 4)) >> 4) | (block[0xc] << 2);
drivers/video/fbdev/core/fbmon.c
373
tmp = ((block[6] & (3 << 2)) >> 2) | (block[0xd] << 2);
drivers/video/fbdev/core/fbmon.c
379
tmp = (block[6] & 3) | (block[0xe] << 2);
drivers/video/fbdev/core/fbmon.c
414
static int get_est_timing(unsigned char *block, struct fb_videomode *mode)
drivers/video/fbdev/core/fbmon.c
419
c = block[0];
drivers/video/fbdev/core/fbmon.c
456
c = block[1];
drivers/video/fbdev/core/fbmon.c
490
c = block[2];
drivers/video/fbdev/core/fbmon.c
499
static int get_std_timing(unsigned char *block, struct fb_videomode *mode,
drivers/video/fbdev/core/fbmon.c
505
u32 std_2byte_code = block[0] << 8 | block[1];
drivers/video/fbdev/core/fbmon.c
519
xres = (block[0] + 31) * 8;
drivers/video/fbdev/core/fbmon.c
523
ratio = (block[1] & 0xc0) >> 6;
drivers/video/fbdev/core/fbmon.c
542
refresh = (block[1] & 0x3f) + 60;
drivers/video/fbdev/core/fbmon.c
558
static int get_dst_timing(unsigned char *block, struct fb_videomode *mode,
drivers/video/fbdev/core/fbmon.c
563
for (j = 0; j < 6; j++, block += STD_TIMING_DESCRIPTION_SIZE)
drivers/video/fbdev/core/fbmon.c
564
num += get_std_timing(block, &mode[num], ver, rev, specs);
drivers/video/fbdev/core/fbmon.c
569
static void get_detailed_timing(unsigned char *block,
drivers/video/fbdev/core/fbmon.c
625
unsigned char *block;
drivers/video/fbdev/core/fbmon.c
645
block = edid + DETAILED_TIMING_DESCRIPTIONS_START;
drivers/video/fbdev/core/fbmon.c
646
for (i = 0; i < 4; i++, block+= DETAILED_TIMING_DESCRIPTION_SIZE) {
drivers/video/fbdev/core/fbmon.c
647
if (!(block[0] == 0x00 && block[1] == 0x00)) {
drivers/video/fbdev/core/fbmon.c
648
get_detailed_timing(block, &mode[num]);
drivers/video/fbdev/core/fbmon.c
658
block = edid + ESTABLISHED_TIMING_1;
drivers/video/fbdev/core/fbmon.c
659
num += get_est_timing(block, &mode[num]);
drivers/video/fbdev/core/fbmon.c
662
block = edid + STD_TIMING_DESCRIPTIONS_START;
drivers/video/fbdev/core/fbmon.c
663
for (i = 0; i < STD_TIMING; i++, block += STD_TIMING_DESCRIPTION_SIZE)
drivers/video/fbdev/core/fbmon.c
664
num += get_std_timing(block, &mode[num], ver, rev, specs);
drivers/video/fbdev/core/fbmon.c
666
block = edid + DETAILED_TIMING_DESCRIPTIONS_START;
drivers/video/fbdev/core/fbmon.c
667
for (i = 0; i < 4; i++, block+= DETAILED_TIMING_DESCRIPTION_SIZE) {
drivers/video/fbdev/core/fbmon.c
668
if (block[0] == 0x00 && block[1] == 0x00 && block[3] == 0xfa)
drivers/video/fbdev/core/fbmon.c
669
num += get_dst_timing(block + 5, &mode[num],
drivers/video/fbdev/core/fbmon.c
703
unsigned char *block;
drivers/video/fbdev/core/fbmon.c
705
block = edid + DETAILED_TIMING_DESCRIPTIONS_START;
drivers/video/fbdev/core/fbmon.c
709
for (i = 0; i < 4; i++, block += DETAILED_TIMING_DESCRIPTION_SIZE) {
drivers/video/fbdev/core/fbmon.c
710
if (edid_is_limits_block(block)) {
drivers/video/fbdev/core/fbmon.c
783
unsigned char c, *block;
drivers/video/fbdev/core/fbmon.c
785
block = edid + EDID_STRUCT_DISPLAY;
drivers/video/fbdev/core/fbmon.c
789
c = block[0] & 0x80;
drivers/video/fbdev/core/fbmon.c
796
switch ((block[0] & 0x60) >> 5) {
drivers/video/fbdev/core/fbmon.c
816
c = block[0] & 0x10;
drivers/video/fbdev/core/fbmon.c
819
c = block[0] & 0x0f;
drivers/video/fbdev/core/fbmon.c
842
specs->max_x = block[1];
drivers/video/fbdev/core/fbmon.c
843
specs->max_y = block[2];
drivers/video/fbdev/core/fbmon.c
855
c = block[3];
drivers/video/fbdev/core/fbmon.c
860
get_dpms_capabilities(block[4], specs);
drivers/video/fbdev/core/fbmon.c
862
switch ((block[4] & 0x18) >> 3) {
drivers/video/fbdev/core/fbmon.c
881
get_chroma(block, specs);
drivers/video/fbdev/core/fbmon.c
884
c = block[4] & 0x7;
drivers/video/fbdev/core/fbmon.c
902
unsigned char *block;
drivers/video/fbdev/core/fbmon.c
913
block = edid + DETAILED_TIMING_DESCRIPTIONS_START;
drivers/video/fbdev/core/fbmon.c
915
for (i = 0; i < 4; i++, block += DETAILED_TIMING_DESCRIPTION_SIZE) {
drivers/video/fbdev/core/fbmon.c
916
if (edid_is_timing_block(block)) {
drivers/video/fbdev/core/fbmon.c
944
unsigned char *block;
drivers/video/fbdev/core/fbmon.c
969
block = edid + DETAILED_TIMING_DESCRIPTIONS_START;
drivers/video/fbdev/core/fbmon.c
970
for (i = 0; i < 4; i++, block += DETAILED_TIMING_DESCRIPTION_SIZE) {
drivers/video/fbdev/core/fbmon.c
971
if (edid_is_serial_block(block)) {
drivers/video/fbdev/core/fbmon.c
972
copy_string(block, specs->serial_no);
drivers/video/fbdev/core/fbmon.c
974
} else if (edid_is_ascii_block(block)) {
drivers/video/fbdev/core/fbmon.c
975
copy_string(block, specs->ascii);
drivers/video/fbdev/core/fbmon.c
977
} else if (edid_is_monitor_block(block)) {
drivers/video/fbdev/core/fbmon.c
978
copy_string(block, specs->monitor);
drivers/video/fbdev/edid.h
107
#define H_SIZE_LO (unsigned)block[ 12 ]
drivers/video/fbdev/edid.h
108
#define V_SIZE_LO (unsigned)block[ 13 ]
drivers/video/fbdev/edid.h
110
#define H_SIZE_HI UPPER_NIBBLE( (unsigned)block[ 14 ] )
drivers/video/fbdev/edid.h
111
#define V_SIZE_HI LOWER_NIBBLE( (unsigned)block[ 14 ] )
drivers/video/fbdev/edid.h
116
#define H_BORDER (unsigned)block[ 15 ]
drivers/video/fbdev/edid.h
117
#define V_BORDER (unsigned)block[ 16 ]
drivers/video/fbdev/edid.h
119
#define FLAGS (unsigned)block[ 17 ]
drivers/video/fbdev/edid.h
127
#define V_MIN_RATE block[ 5 ]
drivers/video/fbdev/edid.h
128
#define V_MAX_RATE block[ 6 ]
drivers/video/fbdev/edid.h
129
#define H_MIN_RATE block[ 7 ]
drivers/video/fbdev/edid.h
130
#define H_MAX_RATE block[ 8 ]
drivers/video/fbdev/edid.h
131
#define MAX_PIXEL_CLOCK (((int)block[ 9 ]) * 10)
drivers/video/fbdev/edid.h
132
#define GTF_SUPPORT block[10]
drivers/video/fbdev/edid.h
72
#define PIXEL_CLOCK_LO (unsigned)block[ 0 ]
drivers/video/fbdev/edid.h
73
#define PIXEL_CLOCK_HI (unsigned)block[ 1 ]
drivers/video/fbdev/edid.h
75
#define H_ACTIVE_LO (unsigned)block[ 2 ]
drivers/video/fbdev/edid.h
76
#define H_BLANKING_LO (unsigned)block[ 3 ]
drivers/video/fbdev/edid.h
77
#define H_ACTIVE_HI UPPER_NIBBLE( (unsigned)block[ 4 ] )
drivers/video/fbdev/edid.h
79
#define H_BLANKING_HI LOWER_NIBBLE( (unsigned)block[ 4 ] )
drivers/video/fbdev/edid.h
82
#define V_ACTIVE_LO (unsigned)block[ 5 ]
drivers/video/fbdev/edid.h
83
#define V_BLANKING_LO (unsigned)block[ 6 ]
drivers/video/fbdev/edid.h
84
#define V_ACTIVE_HI UPPER_NIBBLE( (unsigned)block[ 7 ] )
drivers/video/fbdev/edid.h
86
#define V_BLANKING_HI LOWER_NIBBLE( (unsigned)block[ 7 ] )
drivers/video/fbdev/edid.h
89
#define H_SYNC_OFFSET_LO (unsigned)block[ 8 ]
drivers/video/fbdev/edid.h
90
#define H_SYNC_WIDTH_LO (unsigned)block[ 9 ]
drivers/video/fbdev/edid.h
92
#define V_SYNC_OFFSET_LO UPPER_NIBBLE( (unsigned)block[ 10 ] )
drivers/video/fbdev/edid.h
93
#define V_SYNC_WIDTH_LO LOWER_NIBBLE( (unsigned)block[ 10 ] )
drivers/video/fbdev/edid.h
95
#define V_SYNC_WIDTH_HI ((unsigned)block[ 11 ] & (1|2))
drivers/video/fbdev/edid.h
96
#define V_SYNC_OFFSET_HI (((unsigned)block[ 11 ] & (4|8)) >> 2)
drivers/video/fbdev/edid.h
98
#define H_SYNC_WIDTH_HI (((unsigned)block[ 11 ] & (16|32)) >> 4)
drivers/video/fbdev/edid.h
99
#define H_SYNC_OFFSET_HI (((unsigned)block[ 11 ] & (64|128)) >> 6)
drivers/w1/slaves/w1_ds2433.c
106
set_bit(block, data->validcrc);
drivers/w1/slaves/w1_ds2433.c
85
int block)
drivers/w1/slaves/w1_ds2433.c
88
int off = block * W1_PAGE_SIZE;
drivers/w1/slaves/w1_ds2433.c
90
if (test_bit(block, data->validcrc))
drivers/w1/slaves/w1_ds28e04.c
72
int block)
drivers/w1/slaves/w1_ds28e04.c
75
int off = block * W1_PAGE_SIZE;
drivers/w1/slaves/w1_ds28e04.c
77
if (data->validcrc & (1 << block))
drivers/w1/slaves/w1_ds28e04.c
93
data->validcrc |= (1 << block);
drivers/w1/w1_netlink.c
100
w1_reply_make_space(block, block->maxlen);
drivers/w1/w1_netlink.c
112
static void w1_netlink_setup_msg(struct w1_cb_block *block, u32 ack)
drivers/w1/w1_netlink.c
114
if (block->cn && block->cn->ack == ack) {
drivers/w1/w1_netlink.c
115
block->msg = (struct w1_netlink_msg *)(block->cn->data + block->cn->len);
drivers/w1/w1_netlink.c
118
if (block->cn)
drivers/w1/w1_netlink.c
119
block->cn = (struct cn_msg *)(block->cn->data +
drivers/w1/w1_netlink.c
120
block->cn->len);
drivers/w1/w1_netlink.c
122
block->cn = block->first_cn;
drivers/w1/w1_netlink.c
124
memcpy(block->cn, &block->request_cn, sizeof(*block->cn));
drivers/w1/w1_netlink.c
125
block->cn->len = 0;
drivers/w1/w1_netlink.c
126
block->cn->ack = ack;
drivers/w1/w1_netlink.c
127
block->msg = (struct w1_netlink_msg *)block->cn->data;
drivers/w1/w1_netlink.c
135
static void w1_netlink_queue_cmd(struct w1_cb_block *block,
drivers/w1/w1_netlink.c
139
w1_reply_make_space(block, sizeof(struct cn_msg) +
drivers/w1/w1_netlink.c
146
w1_netlink_setup_msg(block, block->request_cn.seq + 1);
drivers/w1/w1_netlink.c
147
memcpy(block->msg, block->cur_msg, sizeof(*block->msg));
drivers/w1/w1_netlink.c
148
block->cn->len += sizeof(*block->msg);
drivers/w1/w1_netlink.c
149
block->msg->len = 0;
drivers/w1/w1_netlink.c
150
block->cmd = (struct w1_netlink_cmd *)(block->msg->data);
drivers/w1/w1_netlink.c
153
if (block->cmd != cmd)
drivers/w1/w1_netlink.c
154
memcpy(block->cmd, cmd, space);
drivers/w1/w1_netlink.c
155
block->cn->len += space;
drivers/w1/w1_netlink.c
156
block->msg->len += space;
drivers/w1/w1_netlink.c
162
static void w1_netlink_queue_status(struct w1_cb_block *block,
drivers/w1/w1_netlink.c
167
w1_reply_make_space(block, space);
drivers/w1/w1_netlink.c
168
w1_netlink_setup_msg(block, block->request_cn.ack);
drivers/w1/w1_netlink.c
170
memcpy(block->msg, req_msg, sizeof(*req_msg));
drivers/w1/w1_netlink.c
171
block->cn->len += sizeof(*req_msg);
drivers/w1/w1_netlink.c
172
block->msg->len = 0;
drivers/w1/w1_netlink.c
173
block->msg->status = (u8)-error;
drivers/w1/w1_netlink.c
175
struct w1_netlink_cmd *cmd = (struct w1_netlink_cmd *)block->msg->data;
drivers/w1/w1_netlink.c
177
block->cn->len += sizeof(*cmd);
drivers/w1/w1_netlink.c
178
block->msg->len += sizeof(*cmd);
drivers/w1/w1_netlink.c
181
w1_netlink_check_send(block);
drivers/w1/w1_netlink.c
236
struct w1_cb_block *block = dev->priv;
drivers/w1/w1_netlink.c
237
struct w1_netlink_cmd *cache_cmd = block->cmd;
drivers/w1/w1_netlink.c
240
w1_reply_make_space(block, sizeof(*data));
drivers/w1/w1_netlink.c
243
if (!block->cmd) {
drivers/w1/w1_netlink.c
245
w1_netlink_queue_cmd(block, cache_cmd);
drivers/w1/w1_netlink.c
248
data = (u64 *)(block->cmd->data + block->cmd->len);
drivers/w1/w1_netlink.c
251
block->cn->len += sizeof(*data);
drivers/w1/w1_netlink.c
252
block->msg->len += sizeof(*data);
drivers/w1/w1_netlink.c
253
block->cmd->len += sizeof(*data);
drivers/w1/w1_netlink.c
41
struct w1_cb_block *block;
drivers/w1/w1_netlink.c
453
dev->priv = node->block;
drivers/w1/w1_netlink.c
456
node->block->cur_msg = node->msg;
drivers/w1/w1_netlink.c
468
w1_netlink_check_send(node->block);
drivers/w1/w1_netlink.c
470
w1_netlink_queue_status(node->block, node->msg, cmd, err);
drivers/w1/w1_netlink.c
479
w1_netlink_queue_status(node->block, node->msg, cmd, err);
drivers/w1/w1_netlink.c
495
w1_unref_block(node->block);
drivers/w1/w1_netlink.c
543
struct w1_cb_block *block = NULL;
drivers/w1/w1_netlink.c
55
static u16 w1_reply_len(struct w1_cb_block *block)
drivers/w1/w1_netlink.c
57
if (!block->cn)
drivers/w1/w1_netlink.c
59
return (u8 *)block->cn - (u8 *)block->first_cn + block->cn->len;
drivers/w1/w1_netlink.c
603
block = kzalloc(size, GFP_KERNEL);
drivers/w1/w1_netlink.c
604
if (!block) {
drivers/w1/w1_netlink.c
612
atomic_set(&block->refcnt, 1);
drivers/w1/w1_netlink.c
613
block->portid = nsp->portid;
drivers/w1/w1_netlink.c
614
block->request_cn = *cn;
drivers/w1/w1_netlink.c
615
memcpy(block->request_cn.data, cn->data, cn->len);
drivers/w1/w1_netlink.c
616
node = (struct w1_cb_node *)(block->request_cn.data + cn->len);
drivers/w1/w1_netlink.c
62
static void w1_unref_block(struct w1_cb_block *block)
drivers/w1/w1_netlink.c
627
block->maxlen = reply_size;
drivers/w1/w1_netlink.c
628
block->first_cn = (struct cn_msg *)(node + node_count);
drivers/w1/w1_netlink.c
629
memset(block->first_cn, 0, sizeof(*block->first_cn));
drivers/w1/w1_netlink.c
64
if (atomic_sub_return(1, &block->refcnt) == 0) {
drivers/w1/w1_netlink.c
65
u16 len = w1_reply_len(block);
drivers/w1/w1_netlink.c
67
cn_netlink_send_mult(block->first_cn, len,
drivers/w1/w1_netlink.c
679
atomic_inc(&block->refcnt);
drivers/w1/w1_netlink.c
68
block->portid, 0,
drivers/w1/w1_netlink.c
681
node->block = block;
drivers/w1/w1_netlink.c
682
node->msg = (struct w1_netlink_msg *)((u8 *)&block->request_cn +
drivers/w1/w1_netlink.c
71
kfree(block);
drivers/w1/w1_netlink.c
710
if (block)
drivers/w1/w1_netlink.c
711
w1_unref_block(block);
drivers/w1/w1_netlink.c
83
static void w1_reply_make_space(struct w1_cb_block *block, u16 space)
drivers/w1/w1_netlink.c
85
u16 len = w1_reply_len(block);
drivers/w1/w1_netlink.c
86
if (len + space >= block->maxlen) {
drivers/w1/w1_netlink.c
87
cn_netlink_send_mult(block->first_cn, len, block->portid,
drivers/w1/w1_netlink.c
89
block->first_cn->len = 0;
drivers/w1/w1_netlink.c
90
block->cn = NULL;
drivers/w1/w1_netlink.c
91
block->msg = NULL;
drivers/w1/w1_netlink.c
92
block->cmd = NULL;
drivers/w1/w1_netlink.c
97
static void w1_netlink_check_send(struct w1_cb_block *block)
drivers/w1/w1_netlink.c
99
if (!(block->request_cn.flags & W1_CN_BUNDLE) && block->cn)
fs/adfs/adfs.h
207
unsigned int block)
fs/adfs/adfs.h
213
block += off << ADFS_SB(sb)->s_log2sharesize;
fs/adfs/adfs.h
216
return adfs_map_lookup(sb, indaddr >> 8, block);
fs/adfs/dir.c
103
int block;
fs/adfs/dir.c
122
block = __adfs_block_map(sb, indaddr, i);
fs/adfs/dir.c
123
if (!block) {
fs/adfs/dir.c
129
dir->bhs[i] = sb_bread(sb, block);
fs/adfs/dir.c
133
indaddr, i, block);
fs/adfs/inode.c
17
adfs_get_block(struct inode *inode, sector_t block, struct buffer_head *bh,
fs/adfs/inode.c
21
if (block >= inode->i_blocks)
fs/adfs/inode.c
24
block = __adfs_block_map(inode->i_sb, ADFS_I(inode)->indaddr,
fs/adfs/inode.c
25
block);
fs/adfs/inode.c
26
if (block)
fs/adfs/inode.c
27
map_bh(bh, inode->i_sb, block);
fs/adfs/inode.c
72
static sector_t _adfs_bmap(struct address_space *mapping, sector_t block)
fs/adfs/inode.c
74
return generic_block_bmap(mapping, block, adfs_get_block);
fs/affs/affs.h
158
extern void affs_free_block(struct super_block *sb, u32 block);
fs/affs/affs.h
222
static inline bool affs_validblock(struct super_block *sb, int block)
fs/affs/affs.h
224
return(block >= AFFS_SB(sb)->s_reserved &&
fs/affs/affs.h
225
block < AFFS_SB(sb)->s_partition_size);
fs/affs/affs.h
234
affs_bread(struct super_block *sb, int block)
fs/affs/affs.h
236
pr_debug("%s: %d\n", __func__, block);
fs/affs/affs.h
237
if (affs_validblock(sb, block))
fs/affs/affs.h
238
return sb_bread(sb, block);
fs/affs/affs.h
242
affs_getblk(struct super_block *sb, int block)
fs/affs/affs.h
244
pr_debug("%s: %d\n", __func__, block);
fs/affs/affs.h
245
if (affs_validblock(sb, block))
fs/affs/affs.h
246
return sb_getblk(sb, block);
fs/affs/affs.h
250
affs_getzeroblk(struct super_block *sb, int block)
fs/affs/affs.h
253
pr_debug("%s: %d\n", __func__, block);
fs/affs/affs.h
254
if (affs_validblock(sb, block)) {
fs/affs/affs.h
255
bh = sb_getblk(sb, block);
fs/affs/affs.h
265
affs_getemptyblk(struct super_block *sb, int block)
fs/affs/affs.h
268
pr_debug("%s: %d\n", __func__, block);
fs/affs/affs.h
269
if (affs_validblock(sb, block)) {
fs/affs/affs.h
270
bh = sb_getblk(sb, block);
fs/affs/bitmap.c
102
affs_error(sb, "affs_free_block","Block %u outside partition", block);
fs/affs/bitmap.c
39
affs_free_block(struct super_block *sb, u32 block)
fs/affs/bitmap.c
47
pr_debug("%s(%u)\n", __func__, block);
fs/affs/bitmap.c
49
if (block > sbi->s_partition_size)
fs/affs/bitmap.c
52
blk = block - sbi->s_reserved;
fs/affs/bitmap.c
90
affs_warning(sb,"affs_free_block","Trying to free block %u which is already free", block);
fs/affs/file.c
298
affs_get_block(struct inode *inode, sector_t block, struct buffer_head *bh_result, int create)
fs/affs/file.c
305
(unsigned long long)block);
fs/affs/file.c
307
BUG_ON(block > (sector_t)0x7fffffffUL);
fs/affs/file.c
309
if (block >= AFFS_I(inode)->i_blkcnt) {
fs/affs/file.c
310
if (block > AFFS_I(inode)->i_blkcnt || !create)
fs/affs/file.c
318
ext = (u32)block / AFFS_SB(sb)->s_hashsize;
fs/affs/file.c
319
block -= ext * AFFS_SB(sb)->s_hashsize;
fs/affs/file.c
323
map_bh(bh_result, sb, (sector_t)be32_to_cpu(AFFS_BLOCK(sb, ext_bh, block)));
fs/affs/file.c
338
AFFS_BLOCK(sb, ext_bh, block) = cpu_to_be32(blocknr);
fs/affs/file.c
339
AFFS_HEAD(ext_bh)->block_count = cpu_to_be32(block + 1);
fs/affs/file.c
343
if (!block) {
fs/affs/file.c
360
(unsigned long long)block);
fs/affs/file.c
454
static sector_t _affs_bmap(struct address_space *mapping, sector_t block)
fs/affs/file.c
456
return generic_block_bmap(mapping,block,affs_get_block);
fs/affs/file.c
472
affs_bread_ino(struct inode *inode, int block, int create)
fs/affs/file.c
478
err = affs_get_block(inode, block, &tmp_bh, create);
fs/affs/file.c
491
affs_getzeroblk_ino(struct inode *inode, int block)
fs/affs/file.c
497
err = affs_get_block(inode, block, &tmp_bh, 1);
fs/affs/file.c
510
affs_getemptyblk_ino(struct inode *inode, int block)
fs/affs/file.c
516
err = affs_get_block(inode, block, &tmp_bh, 1);
fs/affs/inode.c
24
u32 block;
fs/affs/inode.c
295
u32 block;
fs/affs/inode.c
301
if (!(block = affs_alloc_block(dir, dir->i_ino)))
fs/affs/inode.c
304
bh = affs_getzeroblk(sb, block);
fs/affs/inode.c
312
inode->i_ino = block;
fs/affs/inode.c
335
affs_free_block(sb, block);
fs/affs/inode.c
353
u32 block = 0;
fs/affs/inode.c
369
block = affs_alloc_block(dir, dir->i_ino);
fs/affs/inode.c
37
block = inode->i_ino;
fs/affs/inode.c
370
if (!block)
fs/affs/inode.c
374
bh = affs_getzeroblk(sb, block);
fs/affs/inode.c
38
bh = affs_bread(sb, block);
fs/affs/inode.c
393
AFFS_TAIL(sb, inode_bh)->link_chain = cpu_to_be32(block);
fs/affs/inode.c
394
affs_adjust_checksum(inode_bh, block - be32_to_cpu(chain));
fs/affs/inode.c
40
affs_warning(sb, "read_inode", "Cannot read block %d", block);
fs/affs/inode.c
415
if (block)
fs/affs/inode.c
416
affs_free_block(sb, block);
fs/affs/inode.c
46
AFFS_HEAD(bh)->ptype, block);
fs/afs/dir.c
117
union afs_xdr_dir_block *block = iter_base;
fs/afs/dir.c
119
pr_warn("[%05zx] %32phN\n", progress, block);
fs/afs/dir.c
148
union afs_xdr_dir_block *block)
fs/afs/dir.c
150
if (block->hdr.magic != AFS_DIR_MAGIC) {
fs/afs/dir.c
153
progress, ntohs(block->hdr.magic));
fs/afs/dir.c
163
((u8 *)block)[AFS_DIR_BLOCK_SIZE - 1] = 0;
fs/afs/dir.c
370
union afs_xdr_dir_block *block)
fs/afs/dir.c
387
if (!(block->hdr.bitmap[slot / 8] &
fs/afs/dir.c
397
dire = &block->dirents[slot];
fs/afs/dir.c
399
(unsigned long)(block + 1) - (unsigned long)dire->u.name - 1);
fs/afs/dir.c
422
if (!(block->hdr.bitmap[xslot / 8] & (1 << (xslot % 8)))) {
fs/afs/dir_edit.c
100
block->hdr.bitmap[2] &= ~(u8)(mask >> 2 * 8);
fs/afs/dir_edit.c
101
block->hdr.bitmap[3] &= ~(u8)(mask >> 3 * 8);
fs/afs/dir_edit.c
102
block->hdr.bitmap[4] &= ~(u8)(mask >> 4 * 8);
fs/afs/dir_edit.c
103
block->hdr.bitmap[5] &= ~(u8)(mask >> 5 * 8);
fs/afs/dir_edit.c
104
block->hdr.bitmap[6] &= ~(u8)(mask >> 6 * 8);
fs/afs/dir_edit.c
105
block->hdr.bitmap[7] &= ~(u8)(mask >> 7 * 8);
fs/afs/dir_edit.c
111
static union afs_xdr_dir_block *afs_dir_get_block(struct afs_dir_iter *iter, size_t block)
fs/afs/dir_edit.c
116
size_t blpos = block * AFS_DIR_BLOCK_SIZE;
fs/afs/dir_edit.c
117
size_t blend = (block + 1) * AFS_DIR_BLOCK_SIZE, fpos = iter->fpos;
fs/afs/dir_edit.c
165
static int afs_dir_scan_block(const union afs_xdr_dir_block *block, const struct qstr *name,
fs/afs/dir_edit.c
174
bitmap = (u64)block->hdr.bitmap[0] << 0 * 8;
fs/afs/dir_edit.c
175
bitmap |= (u64)block->hdr.bitmap[1] << 1 * 8;
fs/afs/dir_edit.c
176
bitmap |= (u64)block->hdr.bitmap[2] << 2 * 8;
fs/afs/dir_edit.c
177
bitmap |= (u64)block->hdr.bitmap[3] << 3 * 8;
fs/afs/dir_edit.c
178
bitmap |= (u64)block->hdr.bitmap[4] << 4 * 8;
fs/afs/dir_edit.c
179
bitmap |= (u64)block->hdr.bitmap[5] << 5 * 8;
fs/afs/dir_edit.c
180
bitmap |= (u64)block->hdr.bitmap[6] << 6 * 8;
fs/afs/dir_edit.c
181
bitmap |= (u64)block->hdr.bitmap[7] << 7 * 8;
fs/afs/dir_edit.c
188
de = &block->dirents[d];
fs/afs/dir_edit.c
211
union afs_xdr_dir_block *block, int block_num)
fs/afs/dir_edit.c
213
memset(block, 0, sizeof(*block));
fs/afs/dir_edit.c
214
block->hdr.npages = htons(1);
fs/afs/dir_edit.c
215
block->hdr.magic = AFS_DIR_MAGIC;
fs/afs/dir_edit.c
216
block->hdr.bitmap[0] = 1;
fs/afs/dir_edit.c
219
block->hdr.bitmap[0] = 0xff;
fs/afs/dir_edit.c
220
block->hdr.bitmap[1] = 0x1f;
fs/afs/dir_edit.c
221
memset(block->meta.alloc_ctrs,
fs/afs/dir_edit.c
223
sizeof(block->meta.alloc_ctrs));
fs/afs/dir_edit.c
24
static int afs_find_contig_bits(union afs_xdr_dir_block *block, unsigned int nr_slots)
fs/afs/dir_edit.c
245
union afs_xdr_dir_block *meta, *block;
fs/afs/dir_edit.c
287
block = afs_dir_get_block(&iter, b);
fs/afs/dir_edit.c
288
if (!block)
fs/afs/dir_edit.c
298
ntohs(block->hdr.npages),
fs/afs/dir_edit.c
299
ntohs(block->hdr.magic));
fs/afs/dir_edit.c
30
bitmap = (u64)block->hdr.bitmap[0] << 0 * 8;
fs/afs/dir_edit.c
304
afs_edit_init_block(meta, block, b);
fs/afs/dir_edit.c
31
bitmap |= (u64)block->hdr.bitmap[1] << 1 * 8;
fs/afs/dir_edit.c
311
slot = afs_find_contig_bits(block, iter.nr_slots);
fs/afs/dir_edit.c
317
kunmap_local(block);
fs/afs/dir_edit.c
32
bitmap |= (u64)block->hdr.bitmap[2] << 2 * 8;
fs/afs/dir_edit.c
33
bitmap |= (u64)block->hdr.bitmap[3] << 3 * 8;
fs/afs/dir_edit.c
332
block = afs_dir_get_block(&iter, 0);
fs/afs/dir_edit.c
34
bitmap |= (u64)block->hdr.bitmap[4] << 4 * 8;
fs/afs/dir_edit.c
340
de = &block->dirents[slot];
fs/afs/dir_edit.c
35
bitmap |= (u64)block->hdr.bitmap[5] << 5 * 8;
fs/afs/dir_edit.c
350
afs_set_contig_bits(block, slot, iter.nr_slots);
fs/afs/dir_edit.c
36
bitmap |= (u64)block->hdr.bitmap[6] << 6 * 8;
fs/afs/dir_edit.c
361
kunmap_local(block);
fs/afs/dir_edit.c
37
bitmap |= (u64)block->hdr.bitmap[7] << 7 * 8;
fs/afs/dir_edit.c
376
kunmap_local(block);
fs/afs/dir_edit.c
396
union afs_xdr_dir_block *meta, *block, *pblock;
fs/afs/dir_edit.c
436
block = afs_dir_find_block(&iter, b);
fs/afs/dir_edit.c
437
if (!block)
fs/afs/dir_edit.c
443
de = &block->dirents[slot];
fs/afs/dir_edit.c
452
afs_clear_contig_bits(block, slot, iter.nr_slots);
fs/afs/dir_edit.c
461
kunmap_local(block);
fs/afs/dir_edit.c
511
kunmap_local(block);
fs/afs/dir_edit.c
517
kunmap_local(block);
fs/afs/dir_edit.c
531
union afs_xdr_dir_block *block;
fs/afs/dir_edit.c
552
block = afs_dir_get_block(&iter, b);
fs/afs/dir_edit.c
553
if (!block)
fs/afs/dir_edit.c
560
slot = afs_dir_scan_block(block, name, b);
fs/afs/dir_edit.c
564
kunmap_local(block);
fs/afs/dir_edit.c
574
de = &block->dirents[slot];
fs/afs/dir_edit.c
581
kunmap_local(block);
fs/afs/dir_edit.c
590
kunmap_local(block);
fs/afs/dir_edit.c
69
static void afs_set_contig_bits(union afs_xdr_dir_block *block,
fs/afs/dir_edit.c
77
block->hdr.bitmap[0] |= (u8)(mask >> 0 * 8);
fs/afs/dir_edit.c
78
block->hdr.bitmap[1] |= (u8)(mask >> 1 * 8);
fs/afs/dir_edit.c
79
block->hdr.bitmap[2] |= (u8)(mask >> 2 * 8);
fs/afs/dir_edit.c
80
block->hdr.bitmap[3] |= (u8)(mask >> 3 * 8);
fs/afs/dir_edit.c
81
block->hdr.bitmap[4] |= (u8)(mask >> 4 * 8);
fs/afs/dir_edit.c
82
block->hdr.bitmap[5] |= (u8)(mask >> 5 * 8);
fs/afs/dir_edit.c
83
block->hdr.bitmap[6] |= (u8)(mask >> 6 * 8);
fs/afs/dir_edit.c
84
block->hdr.bitmap[7] |= (u8)(mask >> 7 * 8);
fs/afs/dir_edit.c
90
static void afs_clear_contig_bits(union afs_xdr_dir_block *block,
fs/afs/dir_edit.c
98
block->hdr.bitmap[0] &= ~(u8)(mask >> 0 * 8);
fs/afs/dir_edit.c
99
block->hdr.bitmap[1] &= ~(u8)(mask >> 1 * 8);
fs/afs/dir_search.c
105
iter->block = kmap_local_folio(folio, blpos - fpos);
fs/afs/dir_search.c
106
return iter->block;
fs/afs/dir_search.c
138
const union afs_xdr_dir_block *block;
fs/afs/dir_search.c
152
block = afs_dir_find_block(iter, blnum);
fs/afs/dir_search.c
153
if (!block)
fs/afs/dir_search.c
155
dire = &block->dirents[slot];
fs/afs/dir_search.c
176
if (iter->block) {
fs/afs/dir_search.c
177
kunmap_local(iter->block);
fs/afs/dir_search.c
178
iter->block = NULL;
fs/afs/dir_search.c
67
union afs_xdr_dir_block *afs_dir_find_block(struct afs_dir_iter *iter, size_t block)
fs/afs/dir_search.c
72
size_t blpos = block * AFS_DIR_BLOCK_SIZE;
fs/afs/dir_search.c
73
size_t blend = (block + 1) * AFS_DIR_BLOCK_SIZE, fpos = iter->fpos;
fs/afs/dir_search.c
76
_enter("%zx,%d", block, slot);
fs/afs/dir_search.c
78
if (iter->block) {
fs/afs/dir_search.c
79
kunmap_local(iter->block);
fs/afs/dir_search.c
80
iter->block = NULL;
fs/afs/internal.h
1125
union afs_xdr_dir_block *afs_dir_find_block(struct afs_dir_iter *iter, size_t block);
fs/afs/internal.h
985
union afs_xdr_dir_block *block;
fs/befs/datastream.c
54
befs_blocknr_t block; /* block coresponding to pos */
fs/befs/datastream.c
57
block = pos >> BEFS_SB(sb)->block_shift;
fs/befs/datastream.c
59
*off = pos - (block << BEFS_SB(sb)->block_shift);
fs/befs/datastream.c
61
if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) {
fs/befs/datastream.c
63
(unsigned long)block);
fs/befs/datastream.c
70
(unsigned long)block);
fs/befs/io.c
30
befs_blocknr_t block;
fs/befs/io.c
43
block = iaddr2blockno(sb, &iaddr);
fs/befs/io.c
45
befs_debug(sb, "%s: offset = %lu", __func__, (unsigned long)block);
fs/befs/io.c
47
bh = sb_bread(sb, block);
fs/befs/io.c
51
(unsigned long)block);
fs/befs/linuxvfs.c
119
befs_bmap(struct address_space *mapping, sector_t block)
fs/befs/linuxvfs.c
121
return generic_block_bmap(mapping, block, befs_get_block);
fs/befs/linuxvfs.c
134
befs_get_block(struct inode *inode, sector_t block,
fs/befs/linuxvfs.c
144
(unsigned long)inode->i_ino, (long)block);
fs/befs/linuxvfs.c
147
"block %ld in inode %lu", (long)block,
fs/befs/linuxvfs.c
152
res = befs_fblock2brun(sb, ds, block, &run);
fs/befs/linuxvfs.c
157
(long)block);
fs/befs/linuxvfs.c
166
__func__, (unsigned long)inode->i_ino, (long)block,
fs/befs/linuxvfs.c
46
static sector_t befs_bmap(struct address_space *mapping, sector_t block);
fs/bfs/dir.c
273
int block, sblock, eblock, off, pos;
fs/bfs/dir.c
280
for (block = sblock; block <= eblock; block++) {
fs/bfs/dir.c
281
bh = sb_bread(dir->i_sb, block);
fs/bfs/dir.c
287
pos = (block - sblock) * BFS_BSIZE + off;
fs/bfs/dir.c
321
unsigned long block = 0, offset = 0;
fs/bfs/dir.c
331
while (block * BFS_BSIZE + offset < dir->i_size) {
fs/bfs/dir.c
333
bh = sb_bread(dir->i_sb, BFS_I(dir)->i_sblock + block);
fs/bfs/dir.c
335
block++;
fs/bfs/dir.c
35
int block;
fs/bfs/dir.c
351
block++;
fs/bfs/dir.c
46
block = BFS_I(dir)->i_sblock + (ctx->pos >> BFS_BSIZE_BITS);
fs/bfs/dir.c
47
bh = sb_bread(dir->i_sb, block);
fs/bfs/file.c
109
create, (unsigned long)block, phys);
fs/bfs/file.c
120
if (phys + block >= info->si_blocks) {
fs/bfs/file.c
137
create, (unsigned long)block, phys);
fs/bfs/file.c
139
phys += block;
fs/bfs/file.c
187
static sector_t bfs_bmap(struct address_space *mapping, sector_t block)
fs/bfs/file.c
189
return generic_block_bmap(mapping, block, bfs_get_block);
fs/bfs/file.c
65
static int bfs_get_block(struct inode *inode, sector_t block,
fs/bfs/file.c
74
phys = bi->i_sblock + block;
fs/bfs/file.c
78
create, (unsigned long)block, phys);
fs/bfs/file.c
90
create, (unsigned long)block, phys);
fs/bfs/inode.c
40
int block, off;
fs/bfs/inode.c
406
int block = (i - BFS_ROOT_INO) / BFS_INODES_PER_BLOCK + 1;
fs/bfs/inode.c
412
bh = sb_bread(s, block);
fs/bfs/inode.c
53
block = (ino - BFS_ROOT_INO) / BFS_INODES_PER_BLOCK + 1;
fs/bfs/inode.c
54
bh = sb_bread(inode->i_sb, block);
fs/btrfs/qgroup.c
4708
const struct btrfs_qgroup_swapped_block *block = rb_entry(node,
fs/btrfs/qgroup.c
4711
if (block->subvol_bytenr < *bytenr)
fs/btrfs/qgroup.c
4713
else if (block->subvol_bytenr > *bytenr)
fs/btrfs/qgroup.c
4745
struct btrfs_qgroup_swapped_block *block;
fs/btrfs/qgroup.c
4763
block = kmalloc_obj(*block, GFP_NOFS);
fs/btrfs/qgroup.c
4764
if (!block) {
fs/btrfs/qgroup.c
4773
block->subvol_bytenr = btrfs_node_blockptr(reloc_parent, reloc_slot);
fs/btrfs/qgroup.c
4774
block->subvol_generation = btrfs_node_ptr_generation(reloc_parent,
fs/btrfs/qgroup.c
4776
block->reloc_bytenr = btrfs_node_blockptr(subvol_parent, subvol_slot);
fs/btrfs/qgroup.c
4777
block->reloc_generation = btrfs_node_ptr_generation(subvol_parent,
fs/btrfs/qgroup.c
4779
block->last_snapshot = last_snapshot;
fs/btrfs/qgroup.c
4780
block->level = level;
fs/btrfs/qgroup.c
4788
block->trace_leaf = true;
fs/btrfs/qgroup.c
4790
block->trace_leaf = false;
fs/btrfs/qgroup.c
4791
btrfs_node_key_to_cpu(reloc_parent, &block->first_key, reloc_slot);
fs/btrfs/qgroup.c
4795
node = rb_find_add(&block->node, &blocks->blocks[level], qgroup_swapped_block_bytenr_cmp);
fs/btrfs/qgroup.c
4801
if (entry->subvol_generation != block->subvol_generation ||
fs/btrfs/qgroup.c
4802
entry->reloc_bytenr != block->reloc_bytenr ||
fs/btrfs/qgroup.c
4803
entry->reloc_generation != block->reloc_generation) {
fs/btrfs/qgroup.c
4812
kfree(block);
fs/btrfs/qgroup.c
4837
struct btrfs_qgroup_swapped_block AUTO_KFREE(block);
fs/btrfs/qgroup.c
4861
block = rb_entry(node, struct btrfs_qgroup_swapped_block, node);
fs/btrfs/qgroup.c
4864
rb_erase(&block->node, &blocks->blocks[level]);
fs/btrfs/qgroup.c
4874
check.level = block->level;
fs/btrfs/qgroup.c
4875
check.transid = block->reloc_generation;
fs/btrfs/qgroup.c
4877
memcpy(&check.first_key, &block->first_key, sizeof(check.first_key));
fs/btrfs/qgroup.c
4880
reloc_eb = read_tree_block(fs_info, block->reloc_bytenr, &check);
fs/btrfs/qgroup.c
4892
block->last_snapshot, block->trace_leaf);
fs/btrfs/relocation.c
1916
struct tree_block *block;
fs/btrfs/relocation.c
1919
block = rb_entry(rb_node, struct tree_block, rb_node);
fs/btrfs/relocation.c
1921
kfree(block);
fs/btrfs/relocation.c
2431
struct tree_block *block)
fs/btrfs/relocation.c
2434
.level = block->level,
fs/btrfs/relocation.c
2435
.owner_root = block->owner,
fs/btrfs/relocation.c
2436
.transid = block->key.offset
fs/btrfs/relocation.c
2440
eb = read_tree_block(fs_info, block->bytenr, &check);
fs/btrfs/relocation.c
2447
if (block->level == 0)
fs/btrfs/relocation.c
2448
btrfs_item_key_to_cpu(eb, &block->key, 0);
fs/btrfs/relocation.c
2450
btrfs_node_key_to_cpu(eb, &block->key, 0);
fs/btrfs/relocation.c
2452
block->key_ready = true;
fs/btrfs/relocation.c
2550
struct reloc_control *rc, struct tree_block *block,
fs/btrfs/relocation.c
2559
root = btrfs_get_fs_root(fs_info, block->owner, true);
fs/btrfs/relocation.c
2563
nr_levels = max(btrfs_header_level(root->node) - block->level, 0) + 1;
fs/btrfs/relocation.c
2571
path->lowest_level = block->level;
fs/btrfs/relocation.c
2575
ret = btrfs_search_slot(trans, root, &block->key, path, 0, 1);
fs/btrfs/relocation.c
2598
struct tree_block *block;
fs/btrfs/relocation.c
2609
rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
fs/btrfs/relocation.c
2610
if (!block->key_ready)
fs/btrfs/relocation.c
2611
btrfs_readahead_tree_block(fs_info, block->bytenr,
fs/btrfs/relocation.c
2612
block->owner, 0,
fs/btrfs/relocation.c
2613
block->level);
fs/btrfs/relocation.c
2617
rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
fs/btrfs/relocation.c
2618
if (!block->key_ready) {
fs/btrfs/relocation.c
2619
ret = get_tree_block_key(fs_info, block);
fs/btrfs/relocation.c
2626
rbtree_postorder_for_each_entry_safe(block, next, blocks, rb_node) {
fs/btrfs/relocation.c
2632
if (block->owner &&
fs/btrfs/relocation.c
2633
(!btrfs_is_fstree(block->owner) ||
fs/btrfs/relocation.c
2634
block->owner == BTRFS_DATA_RELOC_TREE_OBJECTID)) {
fs/btrfs/relocation.c
2635
ret = relocate_cowonly_block(trans, rc, block, path);
fs/btrfs/relocation.c
2641
node = build_backref_tree(trans, rc, &block->key,
fs/btrfs/relocation.c
2642
block->level, block->bytenr);
fs/btrfs/relocation.c
2648
ret = relocate_tree_block(trans, rc, node, &block->key,
fs/btrfs/relocation.c
3048
struct tree_block *block;
fs/btrfs/relocation.c
3118
block = kmalloc_obj(*block, GFP_NOFS);
fs/btrfs/relocation.c
3119
if (!block)
fs/btrfs/relocation.c
3122
block->bytenr = extent_key->objectid;
fs/btrfs/relocation.c
3123
block->key.objectid = rc->extent_root->fs_info->nodesize;
fs/btrfs/relocation.c
3124
block->key.offset = generation;
fs/btrfs/relocation.c
3125
block->level = level;
fs/btrfs/relocation.c
3126
block->key_ready = false;
fs/btrfs/relocation.c
3127
block->owner = owner;
fs/btrfs/relocation.c
3129
rb_node = rb_simple_insert(blocks, &block->simple_node);
fs/btrfs/relocation.c
3131
btrfs_backref_panic(rc->extent_root->fs_info, block->bytenr,
fs/buffer.c
1007
sector_t block = div_u64(folio_pos(folio), size);
fs/buffer.c
1015
bh->b_blocknr = block;
fs/buffer.c
1018
if (block < end_block)
fs/buffer.c
1021
block++;
fs/buffer.c
1039
static bool grow_dev_folio(struct block_device *bdev, sector_t block,
fs/buffer.c
1088
return block < end_block;
fs/buffer.c
1096
static bool grow_buffers(struct block_device *bdev, sector_t block,
fs/buffer.c
1105
if (check_mul_overflow(block, (sector_t)size, &pos) || pos > MAX_LFS_FILESIZE) {
fs/buffer.c
1107
__func__, (unsigned long long)block,
fs/buffer.c
1113
return grow_dev_folio(bdev, block, pos / PAGE_SIZE, size, gfp);
fs/buffer.c
1117
__getblk_slow(struct block_device *bdev, sector_t block,
fs/buffer.c
1131
if (!grow_buffers(bdev, block, size, gfp))
fs/buffer.c
1135
bh = __find_get_block_nonatomic(bdev, block, size);
fs/buffer.c
1137
bh = __find_get_block(bdev, block, size);
fs/buffer.c
1358
lookup_bh_lru(struct block_device *bdev, sector_t block, unsigned size)
fs/buffer.c
1372
if (bh && bh->b_blocknr == block && bh->b_bdev == bdev &&
fs/buffer.c
1398
find_get_block_common(struct block_device *bdev, sector_t block,
fs/buffer.c
1401
struct buffer_head *bh = lookup_bh_lru(bdev, block, size);
fs/buffer.c
1405
bh = __find_get_block_slow(bdev, block, atomic);
fs/buffer.c
1415
__find_get_block(struct block_device *bdev, sector_t block, unsigned size)
fs/buffer.c
1417
return find_get_block_common(bdev, block, size, true);
fs/buffer.c
1423
__find_get_block_nonatomic(struct block_device *bdev, sector_t block,
fs/buffer.c
1426
return find_get_block_common(bdev, block, size, false);
fs/buffer.c
1444
struct buffer_head *bdev_getblk(struct block_device *bdev, sector_t block,
fs/buffer.c
1450
bh = __find_get_block_nonatomic(bdev, block, size);
fs/buffer.c
1452
bh = __find_get_block(bdev, block, size);
fs/buffer.c
1458
return __getblk_slow(bdev, block, size, gfp);
fs/buffer.c
1465
void __breadahead(struct block_device *bdev, sector_t block, unsigned size)
fs/buffer.c
1467
struct buffer_head *bh = bdev_getblk(bdev, block, size,
fs/buffer.c
1499
struct buffer_head *__bread_gfp(struct block_device *bdev, sector_t block,
fs/buffer.c
1512
bh = bdev_getblk(bdev, block, size, gfp);
fs/buffer.c
1735
void clean_bdev_aliases(struct block_device *bdev, sector_t block, sector_t len)
fs/buffer.c
1740
pgoff_t index = ((loff_t)block << blkbits) / PAGE_SIZE;
fs/buffer.c
1746
end = ((loff_t)(block + len - 1) << blkbits) / PAGE_SIZE;
fs/buffer.c
1767
if (!buffer_mapped(bh) || (bh->b_blocknr < block))
fs/buffer.c
1769
if (bh->b_blocknr >= block + len)
fs/buffer.c
181
__find_get_block_slow(struct block_device *bdev, sector_t block, bool atomic)
fs/buffer.c
1837
sector_t block;
fs/buffer.c
1860
block = div_u64(folio_pos(folio), blocksize);
fs/buffer.c
1868
if (block > last_block) {
fs/buffer.c
1882
err = get_block(inode, block, bh, 1);
fs/buffer.c
1893
block++;
fs/buffer.c
193
index = ((loff_t)block << blkbits) / PAGE_SIZE;
fs/buffer.c
2040
iomap_to_bh(struct inode *inode, sector_t block, struct buffer_head *bh,
fs/buffer.c
2043
loff_t offset = (loff_t)block << inode->i_blkbits;
fs/buffer.c
2114
sector_t block;
fs/buffer.c
2125
block = div_u64(folio_pos(folio), blocksize);
fs/buffer.c
2128
block++, block_start=block_end, bh = bh->b_this_page) {
fs/buffer.c
2142
err = get_block(inode, block, bh, 1);
fs/buffer.c
2144
err = iomap_to_bh(inode, block, bh, iomap);
fs/buffer.c
224
else if (bh->b_blocknr == block) {
fs/buffer.c
242
(unsigned long long)block,
fs/buffer.c
2753
sector_t generic_block_bmap(struct address_space *mapping, sector_t block,
fs/buffer.c
2761
get_block(inode, block, &tmp, 0);
fs/ceph/inode.c
2404
u64 block = orig_pos >> CEPH_FSCRYPT_BLOCK_SHIFT;
fs/ceph/inode.c
2497
0, block);
fs/ecryptfs/mmap.c
503
static sector_t ecryptfs_bmap(struct address_space *mapping, sector_t block)
fs/ecryptfs/mmap.c
506
int ret = bmap(lower_inode, &block);
fs/ecryptfs/mmap.c
510
return block;
fs/efs/dir.c
101
block++;
fs/efs/dir.c
103
ctx->pos = (block << EFS_DIRBSIZE_BITS) | slot;
fs/efs/dir.c
28
efs_block_t block;
fs/efs/dir.c
36
block = ctx->pos >> EFS_DIRBSIZE_BITS;
fs/efs/dir.c
42
while (block < inode->i_blocks) {
fs/efs/dir.c
47
bh = sb_bread(inode->i_sb, efs_bmap(inode, block));
fs/efs/dir.c
51
__func__, block);
fs/efs/dir.c
78
__func__, block, slot, dirblock->slots-1,
fs/efs/dir.c
83
ctx->pos = (block << EFS_DIRBSIZE_BITS) | slot;
fs/efs/file.c
27
__func__, block, inode->i_blocks, inode->i_size);
fs/efs/file.c
37
int efs_bmap(struct inode *inode, efs_block_t block) {
fs/efs/file.c
39
if (block < 0) {
fs/efs/file.c
45
if (!(block < inode->i_blocks)) {
fs/efs/file.c
51
__func__, block, inode->i_blocks, inode->i_size);
fs/efs/file.c
56
return efs_map_block(inode, block);
fs/efs/inode.c
180
efs_extent_check(efs_extent *ptr, efs_block_t block, struct efs_sb_info *sb) {
fs/efs/inode.c
193
if ((block >= offset) && (block < offset+length)) {
fs/efs/inode.c
194
return(sb->fs_start + start + block - offset);
fs/efs/inode.c
200
efs_block_t efs_map_block(struct inode *inode, efs_block_t block) {
fs/efs/inode.c
214
if ((result = efs_extent_check(&in->extents[last], block, sb)))
fs/efs/inode.c
22
static sector_t _efs_bmap(struct address_space *mapping, sector_t block)
fs/efs/inode.c
231
if ((result = efs_extent_check(&in->extents[cur], block, sb))) {
fs/efs/inode.c
237
pr_err("%s() failed to map block %u (dir)\n", __func__, block);
fs/efs/inode.c
24
return generic_block_bmap(mapping,block,efs_get_block);
fs/efs/inode.c
242
__func__, block);
fs/efs/inode.c
265
cur, block);
fs/efs/inode.c
303
if ((result = efs_extent_check(&ext, block, sb))) {
fs/efs/inode.c
310
pr_err("%s() failed to map block %u (indir)\n", __func__, block);
fs/efs/inode.c
58
efs_block_t block, offset;
fs/efs/inode.c
85
block = sb->fs_start + sb->first_block +
fs/efs/inode.c
93
bh = sb_bread(inode->i_sb, block);
fs/efs/inode.c
95
pr_warn("%s() failed at block %d\n", __func__, block);
fs/efs/namei.c
25
efs_block_t block;
fs/efs/namei.c
31
for(block = 0; block < inode->i_blocks; block++) {
fs/efs/namei.c
33
bh = sb_bread(inode->i_sb, efs_bmap(inode, block));
fs/efs/namei.c
36
__func__, block);
fs/erofs/data.c
417
static sector_t erofs_bmap(struct address_space *mapping, sector_t block)
fs/erofs/data.c
419
return iomap_bmap(mapping, block, &erofs_iomap_ops);
fs/exfat/inode.c
521
static sector_t exfat_aop_bmap(struct address_space *mapping, sector_t block)
fs/exfat/inode.c
527
blocknr = generic_block_bmap(mapping, block, exfat_get_block);
fs/ext2/balloc.c
482
void ext2_free_blocks(struct inode * inode, ext2_fsblk_t block,
fs/ext2/balloc.c
497
if (!ext2_data_block_valid(sbi, block, count)) {
fs/ext2/balloc.c
500
"block = %lu, count = %lu", block, count);
fs/ext2/balloc.c
504
ext2_debug ("freeing block(s) %lu-%lu\n", block, block + count - 1);
fs/ext2/balloc.c
508
block_group = (block - le32_to_cpu(es->s_first_data_block)) /
fs/ext2/balloc.c
510
bit = (block - le32_to_cpu(es->s_first_data_block)) %
fs/ext2/balloc.c
529
if (in_range (le32_to_cpu(desc->bg_block_bitmap), block, count) ||
fs/ext2/balloc.c
530
in_range (le32_to_cpu(desc->bg_inode_bitmap), block, count) ||
fs/ext2/balloc.c
531
in_range (block, le32_to_cpu(desc->bg_inode_table),
fs/ext2/balloc.c
533
in_range (block + count - 1, le32_to_cpu(desc->bg_inode_table),
fs/ext2/balloc.c
538
block, count);
fs/ext2/balloc.c
546
"bit already cleared for block %lu", block + i);
fs/ext2/balloc.c
560
block += count;
fs/ext2/ialloc.c
171
unsigned long block;
fs/ext2/ialloc.c
184
block = le32_to_cpu(gdp->bg_inode_table) +
fs/ext2/ialloc.c
186
sb_breadahead(inode->i_sb, block);
fs/ext2/inode.c
1319
unsigned long block;
fs/ext2/inode.c
1336
block = le32_to_cpu(gdp->bg_inode_table) +
fs/ext2/inode.c
1338
if (!(bh = sb_bread(sb, block)))
fs/ext2/inode.c
1352
(unsigned long) ino, block);
fs/ext2/inode.c
330
static inline ext2_fsblk_t ext2_find_goal(struct inode *inode, long block,
fs/ext2/inode.c
341
if (block_i && (block == block_i->last_alloc_logical_block + 1)
fs/ext2/inode.c
562
long block, Indirect *where, int num, int blks)
fs/ext2/inode.c
591
block_i->last_alloc_logical_block = block + blks - 1;
fs/ext2/inode.c
952
static sector_t ext2_bmap(struct address_space *mapping, sector_t block)
fs/ext2/inode.c
954
return generic_block_bmap(mapping,block,ext2_get_block);
fs/ext2/super.c
1139
block = descriptor_loc(sb, logic_sb_block, i);
fs/ext2/super.c
1140
sbi->s_group_desc[i] = sb_bread(sb, block);
fs/ext2/super.c
885
unsigned long block;
fs/ext2/xattr.c
746
ext2_fsblk_t block = ext2_new_blocks(inode, goal,
fs/ext2/xattr.c
751
ea_idebug(inode, "creating block %lu", block);
fs/ext2/xattr.c
753
new_bh = sb_getblk(sb, block);
fs/ext2/xattr.c
755
ext2_free_blocks(inode, block, 1);
fs/ext4/balloc.c
37
ext4_fsblk_t block)
fs/ext4/balloc.c
42
group = (block -
fs/ext4/balloc.c
46
ext4_get_group_no_and_offset(sb, block, &group, NULL);
fs/ext4/balloc.c
75
ext4_fsblk_t block,
fs/ext4/balloc.c
80
actual_group = ext4_get_group_number(sb, block);
fs/ext4/ext4.h
2457
#define EXT4_DIRENT_TAIL(block, blocksize) \
fs/ext4/ext4.h
2458
((struct ext4_dir_entry_tail *)(((void *)(block)) + \
fs/ext4/ext4.h
2786
ext4_fsblk_t block);
fs/ext4/ext4.h
2985
bool ext4_fc_replay_check_excluded(struct super_block *sb, ext4_fsblk_t block);
fs/ext4/ext4.h
3012
struct buffer_head *bh, ext4_fsblk_t block,
fs/ext4/ext4.h
3019
ext4_fsblk_t block, unsigned long count);
fs/ext4/ext4.h
3022
extern void ext4_mb_mark_bb(struct super_block *sb, ext4_fsblk_t block,
fs/ext4/ext4.h
3036
int ext4_bread_batch(struct inode *inode, ext4_lblk_t block, int bh_count,
fs/ext4/ext4.h
3183
sector_t block, blk_opf_t op_flags);
fs/ext4/ext4.h
3185
sector_t block);
fs/ext4/ext4.h
3187
sector_t block);
fs/ext4/ext4.h
3193
extern void ext4_sb_breadahead_unmovable(struct super_block *sb, sector_t block);
fs/ext4/ext4.h
3242
#define ext4_error_inode_block(inode, block, err, fmt, a...) \
fs/ext4/ext4.h
3243
__ext4_error_inode((inode), __func__, __LINE__, (block), (err), \
fs/ext4/ext4.h
3246
#define EXT4_ERROR_FILE(file, block, fmt, a...) \
fs/ext4/ext4.h
3247
ext4_error_file((file), __func__, __LINE__, (block), (fmt), ## a)
fs/ext4/ext4.h
3254
#define ext4_error_inode(inode, func, line, block, fmt, ...) \
fs/ext4/ext4.h
3255
__ext4_error_inode(inode, func, line, block, 0, fmt, ##__VA_ARGS__)
fs/ext4/ext4.h
3256
#define ext4_error_inode_err(inode, func, line, block, err, fmt, ...) \
fs/ext4/ext4.h
3257
__ext4_error_inode((inode), (func), (line), (block), \
fs/ext4/ext4.h
3259
#define ext4_error_file(file, func, line, block, fmt, ...) \
fs/ext4/ext4.h
3260
__ext4_error_file(file, func, line, block, fmt, ##__VA_ARGS__)
fs/ext4/ext4.h
3275
#define ext4_grp_locked_error(sb, grp, ino, block, fmt, ...) \
fs/ext4/ext4.h
3276
__ext4_grp_locked_error(__func__, __LINE__, sb, grp, ino, block, \
fs/ext4/ext4.h
3281
#define ext4_error_inode(inode, func, line, block, fmt, ...) \
fs/ext4/ext4.h
3284
__ext4_error_inode(inode, "", 0, block, 0, " "); \
fs/ext4/ext4.h
3286
#define ext4_error_inode_err(inode, func, line, block, err, fmt, ...) \
fs/ext4/ext4.h
3289
__ext4_error_inode(inode, "", 0, block, err, " "); \
fs/ext4/ext4.h
3291
#define ext4_error_file(file, func, line, block, fmt, ...) \
fs/ext4/ext4.h
3294
__ext4_error_file(file, "", 0, block, " "); \
fs/ext4/ext4.h
3323
#define ext4_grp_locked_error(sb, grp, ino, block, fmt, ...) \
fs/ext4/ext4.h
3326
__ext4_grp_locked_error("", 0, sb, grp, ino, block, " "); \
fs/ext4/ext4.h
3676
struct inode *dir, ext4_lblk_t block,
fs/ext4/ext4_extents.h
221
ext4_fsblk_t block;
fs/ext4/ext4_extents.h
223
block = le32_to_cpu(ex->ee_start_lo);
fs/ext4/ext4_extents.h
224
block |= ((ext4_fsblk_t) le16_to_cpu(ex->ee_start_hi) << 31) << 1;
fs/ext4/ext4_extents.h
225
return block;
fs/ext4/ext4_extents.h
234
ext4_fsblk_t block;
fs/ext4/ext4_extents.h
236
block = le32_to_cpu(ix->ei_leaf_lo);
fs/ext4/ext4_extents.h
237
block |= ((ext4_fsblk_t) le16_to_cpu(ix->ei_leaf_hi) << 31) << 1;
fs/ext4/ext4_extents.h
238
return block;
fs/ext4/extents.c
217
ext4_lblk_t block)
fs/ext4/extents.c
2222
ext4_lblk_t block, ext4_lblk_t num,
fs/ext4/extents.c
2225
ext4_lblk_t next, end = block + num - 1;
fs/ext4/extents.c
2231
while (block <= end) {
fs/ext4/extents.c
2234
if (!ext4_es_lookup_extent(inode, block, &next, &es, NULL))
fs/ext4/extents.c
2257
block = next;
fs/ext4/extents.c
245
if (block > ext_block)
fs/ext4/extents.c
246
return ext_pblk + (block - ext_block);
fs/ext4/extents.c
248
return ext_pblk - (ext_block - block);
fs/ext4/extents.c
366
ext4_fsblk_t block = ext4_ext_pblock(ext);
fs/ext4/extents.c
377
return ext4_inode_block_valid(inode, block, len);
fs/ext4/extents.c
383
ext4_fsblk_t block = ext4_idx_pblock(ext_idx);
fs/ext4/extents.c
385
return ext4_inode_block_valid(inode, block, 1);
fs/ext4/extents.c
755
struct ext4_ext_path *path, ext4_lblk_t block)
fs/ext4/extents.c
761
ext_debug(inode, "binsearch for %u(idx): ", block);
fs/ext4/extents.c
771
if (block < le32_to_cpu(m->ei_block))
fs/ext4/extents.c
799
if (block < le32_to_cpu(ix->ei_block))
fs/ext4/extents.c
816
struct ext4_ext_path *path, ext4_lblk_t block)
fs/ext4/extents.c
829
ext_debug(inode, "binsearch for %u: ", block);
fs/ext4/extents.c
840
if (block < le32_to_cpu(m->ee_block))
fs/ext4/extents.c
862
if (block < le32_to_cpu(ex->ee_block))
fs/ext4/extents.c
886
ext4_find_extent(struct inode *inode, ext4_lblk_t block,
fs/ext4/extents.c
895
KUNIT_STATIC_STUB_REDIRECT(ext4_find_extent, inode, block, path, flags);
fs/ext4/extents.c
934
ext4_ext_binsearch_idx(inode, path + ppos, block);
fs/ext4/extents.c
956
ext4_ext_binsearch(inode, path + ppos, block);
fs/ext4/extents_status.c
1491
ext4_fsblk_t block;
fs/ext4/extents_status.c
1530
block = 0x7FDEADBEEFULL;
fs/ext4/extents_status.c
1533
block = ext4_es_pblock(&orig_es) +
fs/ext4/extents_status.c
1535
ext4_es_store_pblock_status(&newes, block,
fs/ext4/extents_status.c
1551
block = orig_es.es_pblk + orig_es.es_len - len2;
fs/ext4/extents_status.c
1552
ext4_es_store_pblock(es, block);
fs/ext4/extents_status.c
1602
block = es->es_pblk + orig_len - len1;
fs/ext4/extents_status.c
1603
ext4_es_store_pblock(es, block);
fs/ext4/extents_status.h
223
ext4_fsblk_t block;
fs/ext4/extents_status.h
225
block = (pb & ~ES_MASK) | (es->es_pblk & ES_MASK);
fs/ext4/extents_status.h
226
es->es_pblk = block;
fs/ext4/indirect.c
251
static ext4_fsblk_t ext4_find_goal(struct inode *inode, ext4_lblk_t block,
fs/ext4/inline.c
1259
struct inode *dir, ext4_lblk_t block,
fs/ext4/inode.c
1045
ext4_lblk_t block, int map_flags)
fs/ext4/inode.c
1050
bh = ext4_getblk(handle, inode, block, map_flags);
fs/ext4/inode.c
1065
int ext4_bread_batch(struct inode *inode, ext4_lblk_t block, int bh_count,
fs/ext4/inode.c
1071
bhs[i] = ext4_getblk(NULL, inode, block + i, 0 /* map_flags */);
fs/ext4/inode.c
1173
sector_t block;
fs/ext4/inode.c
1189
block = EXT4_PG_TO_LBLK(inode, folio->index);
fs/ext4/inode.c
1192
block++, block_start = block_end, bh = bh->b_this_page) {
fs/ext4/inode.c
1206
err = get_block(inode, block, bh, 1);
fs/ext4/inode.c
3355
static sector_t ext4_bmap(struct address_space *mapping, sector_t block)
fs/ext4/inode.c
3378
ret = iomap_bmap(mapping, block, &ext4_iomap_ops);
fs/ext4/inode.c
4670
int block;
fs/ext4/inode.c
4730
for (block = 0; block < EXT4_N_BLOCKS; block++)
fs/ext4/inode.c
4731
raw_inode->i_block[block] = ei->i_data[block];
fs/ext4/inode.c
4771
ext4_fsblk_t block;
fs/ext4/inode.c
4793
block = ext4_inode_table(sb, gdp);
fs/ext4/inode.c
4794
if ((block <= le32_to_cpu(EXT4_SB(sb)->s_es->s_first_data_block)) ||
fs/ext4/inode.c
4795
(block >= ext4_blocks_count(EXT4_SB(sb)->s_es))) {
fs/ext4/inode.c
4797
"block_group %u", block, iloc->block_group);
fs/ext4/inode.c
4800
block += (inode_offset / inodes_per_block);
fs/ext4/inode.c
4802
bh = sb_getblk(sb, block);
fs/ext4/inode.c
4874
b = block & ~((ext4_fsblk_t) ra_blks - 1);
fs/ext4/inode.c
4900
*ret_block = block;
fs/ext4/inode.c
5145
int block;
fs/ext4/inode.c
5317
for (block = 0; block < EXT4_N_BLOCKS; block++)
fs/ext4/inode.c
5318
ei->i_data[block] = raw_inode->i_block[block];
fs/ext4/inode.c
973
ext4_lblk_t block, int map_flags)
fs/ext4/inode.c
985
map.m_lblk = block;
fs/ext4/mballoc-test.c
516
ext4_fsblk_t block;
fs/ext4/mballoc-test.c
526
block = ext4_group_first_block_no(sb, goal_group) +
fs/ext4/mballoc-test.c
528
ext4_free_blocks_simple_test(inode, block, len);
fs/ext4/mballoc-test.c
73
ext4_fsblk_t block;
fs/ext4/mballoc-test.c
98
block = ext4_count_free_clusters(sb);
fs/ext4/mballoc-test.c
99
ret = percpu_counter_init(&EXT4_SB(sb)->s_freeclusters_counter, block,
fs/ext4/mballoc.c
1546
int block, pnum;
fs/ext4/mballoc.c
1557
block = group * 2;
fs/ext4/mballoc.c
1558
pnum = EXT4_LBLK_TO_PG(inode, block);
fs/ext4/mballoc.c
1567
offset_in_folio(folio, EXT4_LBLK_TO_B(inode, block));
fs/ext4/mballoc.c
1569
block++;
fs/ext4/mballoc.c
1570
pnum = EXT4_LBLK_TO_PG(inode, block);
fs/ext4/mballoc.c
1678
int block;
fs/ext4/mballoc.c
1715
block = group * 2;
fs/ext4/mballoc.c
1716
pnum = EXT4_LBLK_TO_PG(inode, block);
fs/ext4/mballoc.c
1750
EXT4_LBLK_TO_B(inode, block)));
fs/ext4/mballoc.c
1767
offset_in_folio(folio, EXT4_LBLK_TO_B(inode, block));
fs/ext4/mballoc.c
1769
block++;
fs/ext4/mballoc.c
1770
pnum = EXT4_LBLK_TO_PG(inode, block);
fs/ext4/mballoc.c
1816
offset_in_folio(folio, EXT4_LBLK_TO_B(inode, block));
fs/ext4/mballoc.c
1846
static int mb_find_order_for_block(struct ext4_buddy *e4b, int block)
fs/ext4/mballoc.c
1852
BUG_ON(block >= (1 << (e4b->bd_blkbits + 3)));
fs/ext4/mballoc.c
1856
if (!mb_test_bit(block >> order, bb)) {
fs/ext4/mballoc.c
2005
int block;
fs/ext4/mballoc.c
2025
block = mb_test_and_clear_bits(e4b->bd_bitmap, first, count);
fs/ext4/mballoc.c
2029
if (unlikely(block != -1)) {
fs/ext4/mballoc.c
2043
blocknr += EXT4_C2B(sbi, block);
fs/ext4/mballoc.c
2049
block);
fs/ext4/mballoc.c
2088
static int mb_find_extent(struct ext4_buddy *e4b, int block,
fs/ext4/mballoc.c
2099
BUG_ON(block >= max);
fs/ext4/mballoc.c
2100
if (mb_test_bit(block, buddy)) {
fs/ext4/mballoc.c
2108
order = mb_find_order_for_block(e4b, block);
fs/ext4/mballoc.c
2110
ex->fe_len = (1 << order) - (block & ((1 << order) - 1));
fs/ext4/mballoc.c
2111
ex->fe_start = block;
fs/ext4/mballoc.c
2114
block = block >> order;
fs/ext4/mballoc.c
2119
if (block + 1 >= max)
fs/ext4/mballoc.c
2122
next = (block + 1) * (1 << order);
fs/ext4/mballoc.c
2128
block = next >> order;
fs/ext4/mballoc.c
2138
block, order, needed, ex->fe_group, ex->fe_start,
fs/ext4/mballoc.c
4201
ext4_fsblk_t block;
fs/ext4/mballoc.c
4218
block = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
fs/ext4/mballoc.c
4220
if (!ext4_inode_block_valid(ac->ac_inode, block, len)) {
fs/ext4/mballoc.c
4222
"fs metadata", block, block+len);
fs/ext4/mballoc.c
4259
void ext4_mb_mark_bb(struct super_block *sb, ext4_fsblk_t block,
fs/ext4/mballoc.c
4269
ext4_get_group_no_and_offset(sb, block, &group, &blkoff);
fs/ext4/mballoc.c
4282
if (!ext4_sb_block_valid(sb, NULL, block, thisgrp_len)) {
fs/ext4/mballoc.c
4285
block, thisgrp_len);
fs/ext4/mballoc.c
4297
block += thisgrp_len;
fs/ext4/mballoc.c
5898
ext4_grpblk_t block;
fs/ext4/mballoc.c
5912
ext4_get_group_no_and_offset(sb, goal, &group, &block);
fs/ext4/mballoc.c
5921
ac->ac_o_ex.fe_start = block;
fs/ext4/mballoc.c
6177
ext4_fsblk_t goal, block;
fs/ext4/mballoc.c
6222
block = ext4_group_first_block_no(sb, group) + EXT4_C2B(sbi, i);
fs/ext4/mballoc.c
6223
ext4_mb_mark_bb(sb, block, 1, true);
fs/ext4/mballoc.c
6227
return block;
fs/ext4/mballoc.c
6241
ext4_fsblk_t block = 0;
fs/ext4/mballoc.c
6339
block = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
fs/ext4/mballoc.c
6369
trace_ext4_allocate_blocks(ar, (unsigned long long)block);
fs/ext4/mballoc.c
6371
return block;
fs/ext4/mballoc.c
6513
static void ext4_free_blocks_simple(struct inode *inode, ext4_fsblk_t block,
fs/ext4/mballoc.c
6520
ext4_get_group_no_and_offset(sb, block, &group, &blkoff);
fs/ext4/mballoc.c
6537
ext4_fsblk_t block, unsigned long count,
fs/ext4/mballoc.c
6555
!ext4_inode_block_valid(inode, block, count)) {
fs/ext4/mballoc.c
6557
"Block = %llu, count = %lu", block, count);
fs/ext4/mballoc.c
6565
ext4_get_group_no_and_offset(sb, block, &block_group, &bit);
fs/ext4/mballoc.c
6592
!ext4_inode_block_valid(inode, block, count)) {
fs/ext4/mballoc.c
6594
"Block = %llu, count = %lu", block, count);
fs/ext4/mballoc.c
6674
block += count;
fs/ext4/mballoc.c
6698
struct buffer_head *bh, ext4_fsblk_t block,
fs/ext4/mballoc.c
6708
if (block)
fs/ext4/mballoc.c
6709
BUG_ON(block != bh->b_blocknr);
fs/ext4/mballoc.c
6711
block = bh->b_blocknr;
fs/ext4/mballoc.c
6715
ext4_free_blocks_simple(inode, block, EXT4_NUM_B2C(sbi, count));
fs/ext4/mballoc.c
6722
!ext4_inode_block_valid(inode, block, count)) {
fs/ext4/mballoc.c
6724
"block = %llu, count = %lu", block, count);
fs/ext4/mballoc.c
6729
ext4_debug("freeing block %llu\n", block);
fs/ext4/mballoc.c
6730
trace_ext4_free_blocks(inode, block, count, flags);
fs/ext4/mballoc.c
6736
inode, bh, block);
fs/ext4/mballoc.c
6746
overflow = EXT4_PBLK_COFF(sbi, block);
fs/ext4/mballoc.c
6750
block += overflow;
fs/ext4/mballoc.c
6756
block -= overflow;
fs/ext4/mballoc.c
6783
block + i);
fs/ext4/mballoc.c
6784
ext4_forget(handle, is_metadata, inode, bh, block + i);
fs/ext4/mballoc.c
6788
ext4_mb_clear_bb(handle, inode, block, count, flags);
fs/ext4/mballoc.c
6801
ext4_fsblk_t block, unsigned long count)
fs/ext4/mballoc.c
6808
ext4_fsblk_t first_cluster = EXT4_B2C(sbi, block);
fs/ext4/mballoc.c
6809
ext4_fsblk_t last_cluster = EXT4_B2C(sbi, block + count - 1);
fs/ext4/mballoc.c
6813
ext4_debug("Adding block(s) %llu-%llu\n", block, block + count - 1);
fs/ext4/mballoc.c
6818
ext4_get_group_no_and_offset(sb, block, &block_group, &bit);
fs/ext4/mballoc.c
6834
if (!ext4_sb_block_valid(sb, NULL, block, count)) {
fs/ext4/mballoc.c
6837
block, count);
fs/ext4/mballoc.c
7271
void ext4_free_blocks_simple_test(struct inode *inode, ext4_fsblk_t block,
fs/ext4/mballoc.c
7274
return ext4_free_blocks_simple(inode, block, count);
fs/ext4/mballoc.h
301
ext4_fsblk_t block, unsigned long count);
fs/ext4/migrate.c
357
ext4_fsblk_t block;
fs/ext4/migrate.c
361
block = ext4_idx_pblock(ix);
fs/ext4/migrate.c
362
bh = ext4_sb_bread(inode->i_sb, block, 0);
fs/ext4/migrate.c
382
ext4_free_blocks(handle, inode, NULL, block, 1,
fs/ext4/move_extent.c
120
sector_t block;
fs/ext4/move_extent.c
137
block = folio_pos(folio) >> inode->i_blkbits;
fs/ext4/move_extent.c
151
int err = ext4_get_block(inode, block, bh, 0);
fs/ext4/move_extent.c
167
} while (block++, (bh = bh->b_this_page) != head);
fs/ext4/move_extent.c
269
sector_t block;
fs/ext4/move_extent.c
276
block = folio_pos(folio) >> inode->i_blkbits;
fs/ext4/move_extent.c
285
ret = ext4_get_block(inode, block, bh, 0);
fs/ext4/move_extent.c
288
} while (block++, (bh = bh->b_this_page) != head);
fs/ext4/namei.c
1039
struct inode *dir, ext4_lblk_t block,
fs/ext4/namei.c
1050
(unsigned long)block));
fs/ext4/namei.c
1051
bh = ext4_read_dirblock(dir, block, DIRENT_HTREE);
fs/ext4/namei.c
1079
EXT4_LBLK_TO_B(dir, block)
fs/ext4/namei.c
1158
ext4_lblk_t block;
fs/ext4/namei.c
121
#define ext4_read_dirblock(inode, block, type) \
fs/ext4/namei.c
122
__ext4_read_dirblock((inode), (block), (type), __func__, __LINE__)
fs/ext4/namei.c
1228
block = dx_get_block(frame->at);
fs/ext4/namei.c
1229
ret = htree_dirblock_to_tree(dir_file, dir, block, &hinfo,
fs/ext4/namei.c
125
ext4_lblk_t block,
fs/ext4/namei.c
134
if (block >= inode->i_size >> inode->i_blkbits) {
fs/ext4/namei.c
1347
static void dx_insert_block(struct dx_frame *frame, u32 hash, ext4_lblk_t block)
fs/ext4/namei.c
135
ext4_error_inode(inode, func, line, block,
fs/ext4/namei.c
1357
dx_set_block(new, block);
fs/ext4/namei.c
137
block, inode->i_size);
fs/ext4/namei.c
144
bh = ext4_bread(NULL, inode, block, 0);
fs/ext4/namei.c
149
inode->i_ino, (unsigned long)block,
fs/ext4/namei.c
1490
static int is_dx_internal_node(struct inode *dir, ext4_lblk_t block,
fs/ext4/namei.c
1497
if (block == 0)
fs/ext4/namei.c
1525
ext4_lblk_t start, block;
fs/ext4/namei.c
155
if (!bh && (type == INDEX || type == DIRENT_HTREE || block == 0)) {
fs/ext4/namei.c
1556
block = start = 0;
fs/ext4/namei.c
156
ext4_error_inode(inode, func, line, block,
fs/ext4/namei.c
158
(type == INDEX) ? "index" : "leaf", block);
fs/ext4/namei.c
1586
block = start;
fs/ext4/namei.c
1596
if (block < start)
fs/ext4/namei.c
1597
ra_max = start - block;
fs/ext4/namei.c
1599
ra_max = nblocks - block;
fs/ext4/namei.c
1601
retval = ext4_bread_batch(dir, block, ra_max,
fs/ext4/namei.c
1615
(unsigned long) block);
fs/ext4/namei.c
1621
!is_dx_internal_node(dir, block,
fs/ext4/namei.c
1626
"block %lu", (unsigned long)block);
fs/ext4/namei.c
1633
EXT4_LBLK_TO_B(dir, block), res_dir);
fs/ext4/namei.c
1635
EXT4_I(dir)->i_dir_start_lookup = block;
fs/ext4/namei.c
1646
if (++block >= nblocks)
fs/ext4/namei.c
1647
block = 0;
fs/ext4/namei.c
1648
} while (block != start);
fs/ext4/namei.c
1654
block = nblocks;
fs/ext4/namei.c
1656
if (block < nblocks) {
fs/ext4/namei.c
166
if (block == 0)
fs/ext4/namei.c
1715
ext4_lblk_t block;
fs/ext4/namei.c
1725
block = dx_get_block(frame->at);
fs/ext4/namei.c
1726
bh = ext4_read_dirblock(dir, block, DIRENT_HTREE);
fs/ext4/namei.c
1731
EXT4_LBLK_TO_B(dir, block), res_dir);
fs/ext4/namei.c
174
ext4_error_inode(inode, func, line, block,
fs/ext4/namei.c
193
ext4_error_inode_err(inode, func, line, block,
fs/ext4/namei.c
205
ext4_error_inode_err(inode, func, line, block,
fs/ext4/namei.c
2222
ext4_lblk_t block;
fs/ext4/namei.c
2253
bh2 = ext4_append(handle, dir, &block);
fs/ext4/namei.c
2367
ext4_lblk_t block, blocks;
fs/ext4/namei.c
238
__le32 block;
fs/ext4/namei.c
2414
for (block = 0; block < blocks; block++) {
fs/ext4/namei.c
2415
bh = ext4_read_dirblock(dir, block, DIRENT);
fs/ext4/namei.c
2417
bh = ext4_bread(handle, dir, block,
fs/ext4/namei.c
2440
bh = ext4_append(handle, dir, &block);
fs/ext4/namei.c
2966
ext4_lblk_t block = 0;
fs/ext4/namei.c
2979
dir_block = ext4_append(handle, inode, &block);
fs/ext4/namei.c
539
return le32_to_cpu(entry->block) & 0x0fffffff;
fs/ext4/namei.c
544
entry->block = cpu_to_le32(value);
fs/ext4/namei.c
55
ext4_lblk_t *block)
fs/ext4/namei.c
66
*block = inode->i_size >> inode->i_sb->s_blocksize_bits;
fs/ext4/namei.c
67
map.m_lblk = *block;
fs/ext4/namei.c
720
ext4_lblk_t block = dx_get_block(entries);
fs/ext4/namei.c
724
printk("%s%3u:%03u hash %8x/%8x ",levels?"":" ", i, block, hash, range);
fs/ext4/namei.c
725
bh = ext4_bread(NULL,dir, block, 0);
fs/ext4/namei.c
787
ext4_lblk_t block;
fs/ext4/namei.c
83
bh = ext4_bread(handle, inode, *block, EXT4_GET_BLOCKS_CREATE);
fs/ext4/namei.c
897
block = dx_get_block(at);
fs/ext4/namei.c
899
if (blocks[i] == block) {
fs/ext4/namei.c
902
blocks[level], block);
fs/ext4/namei.c
908
blocks[level] = block;
fs/ext4/namei.c
910
frame->bh = ext4_read_dirblock(dir, block, INDEX);
fs/ext4/resize.c
1296
static struct buffer_head *ext4_get_bitmap(struct super_block *sb, __u64 block)
fs/ext4/resize.c
1298
struct buffer_head *bh = sb_getblk(sb, block);
fs/ext4/resize.c
535
ext4_fsblk_t block;
fs/ext4/resize.c
571
block = start + ext4_bg_has_super(sb, group);
fs/ext4/resize.c
573
for (j = 0; j < gdblocks; j++, block++) {
fs/ext4/resize.c
576
ext4_debug("update backup group %#04llx\n", block);
fs/ext4/resize.c
581
gdb = sb_getblk(sb, block);
fs/ext4/resize.c
622
block = group_data[i].inode_table;
fs/ext4/resize.c
624
block, sbi->s_itb_per_group);
fs/ext4/resize.c
625
err = sb_issue_zeroout(sb, block, sbi->s_itb_per_group,
fs/ext4/resize.c
635
block = group_data[i].block_bitmap;
fs/ext4/resize.c
640
bh = bclean(handle, sb, block);
fs/ext4/resize.c
664
block = group_data[i].inode_bitmap;
fs/ext4/resize.c
669
bh = bclean(handle, sb, block);
fs/ext4/resize.c
687
block = start;
fs/ext4/resize.c
689
block += group_table_count[j];
fs/ext4/resize.c
690
if (block == (&group_data[i].block_bitmap)[j]) {
fs/ext4/resize.c
704
block = start;
fs/ext4/super.c
1045
unsigned long ino, ext4_fsblk_t block,
fs/ext4/super.c
1065
if (block)
fs/ext4/super.c
1067
(unsigned long long) block);
fs/ext4/super.c
1077
save_error_info(sb, EFSCORRUPTED, ino, block, function,
fs/ext4/super.c
1084
ext4_handle_error(sb, false, EFSCORRUPTED, ino, block, function, line);
fs/ext4/super.c
1211
ext4_fsblk_t block;
fs/ext4/super.c
1214
block = ext4_count_free_clusters(sbi->s_sb);
fs/ext4/super.c
1215
ext4_free_blocks_count_set(sbi->s_es, EXT4_C2B(sbi, block));
fs/ext4/super.c
1216
err = percpu_counter_init(&sbi->s_freeclusters_counter, block,
fs/ext4/super.c
231
sector_t block,
fs/ext4/super.c
237
bh = sb_getblk_gfp(sb, block, gfp);
fs/ext4/super.c
251
struct buffer_head *ext4_sb_bread(struct super_block *sb, sector_t block,
fs/ext4/super.c
257
return __ext4_sb_bread_gfp(sb, block, op_flags, gfp);
fs/ext4/super.c
261
sector_t block)
fs/ext4/super.c
266
return __ext4_sb_bread_gfp(sb, block, 0, gfp);
fs/ext4/super.c
270
sector_t block)
fs/ext4/super.c
275
return __ext4_sb_bread_gfp(sb, block, 0, gfp);
fs/ext4/super.c
278
void ext4_sb_breadahead_unmovable(struct super_block *sb, sector_t block)
fs/ext4/super.c
280
struct buffer_head *bh = bdev_getblk(sb->s_bdev, block,
fs/ext4/super.c
4891
ext4_fsblk_t block;
fs/ext4/super.c
4916
block = descriptor_loc(sb, logical_sb_block, i);
fs/ext4/super.c
4917
ext4_sb_breadahead_unmovable(sb, block);
fs/ext4/super.c
4923
block = descriptor_loc(sb, logical_sb_block, i);
fs/ext4/super.c
4924
bh = ext4_sb_bread_unmovable(sb, block);
fs/ext4/super.c
5910
static int ext4_journal_bmap(journal_t *journal, sector_t *block)
fs/ext4/super.c
5918
map.m_lblk = *block;
fs/ext4/super.c
5924
*block, ret);
fs/ext4/super.c
5928
*block = map.m_pblk;
fs/ext4/super.c
642
__u32 ino, __u64 block,
fs/ext4/super.c
656
sbi->s_last_error_block = block;
fs/ext4/super.c
663
sbi->s_first_error_block = block;
fs/ext4/super.c
691
__u32 ino, __u64 block,
fs/ext4/super.c
705
save_error_info(sb, error, ino, block, func, line);
fs/ext4/super.c
809
unsigned int line, bool force_ro, int error, __u64 block,
fs/ext4/super.c
831
ext4_handle_error(sb, force_ro, error, 0, block, function, line);
fs/ext4/super.c
835
unsigned int line, ext4_fsblk_t block, int error,
fs/ext4/super.c
849
if (block)
fs/ext4/super.c
853
block, current->comm, &vaf);
fs/ext4/super.c
865
ext4_handle_error(inode->i_sb, false, error, inode->i_ino, block,
fs/ext4/super.c
870
unsigned int line, ext4_fsblk_t block,
fs/ext4/super.c
889
if (block)
fs/ext4/super.c
894
block, current->comm, path, &vaf);
fs/ext4/super.c
905
ext4_handle_error(inode->i_sb, false, EFSCORRUPTED, inode->i_ino, block,
fs/ext4/xattr.c
1389
unsigned long block = 0;
fs/ext4/xattr.c
1399
map.m_lblk = block += ret;
fs/ext4/xattr.c
1418
block = 0;
fs/ext4/xattr.c
1423
bh = ext4_getblk(handle, ea_inode, block, 0);
fs/ext4/xattr.c
1449
block += 1;
fs/ext4/xattr.c
2116
ext4_fsblk_t goal, block;
fs/ext4/xattr.c
2123
block = ext4_new_meta_blocks(handle, inode, goal, 0,
fs/ext4/xattr.c
2129
(unsigned long long)block);
fs/ext4/xattr.c
2131
new_bh = sb_getblk(sb, block);
fs/ext4/xattr.c
2135
ext4_free_blocks(handle, inode, NULL, block, 1,
fs/f2fs/data.c
4097
static sector_t f2fs_bmap_compress(struct inode *inode, sector_t block)
fs/f2fs/data.c
4104
start_idx = round_down(block, F2FS_I(inode)->i_cluster_size);
fs/f2fs/data.c
4112
dn.ofs_in_node += block - start_idx;
fs/f2fs/data.c
4126
static sector_t f2fs_bmap(struct address_space *mapping, sector_t block)
fs/f2fs/data.c
4139
if (unlikely(block >= max_file_blocks(inode)))
fs/f2fs/data.c
4143
blknr = f2fs_bmap_compress(inode, block);
fs/f2fs/data.c
4148
map.m_lblk = block;
fs/f2fs/data.c
4157
trace_f2fs_bmap(inode, block, blknr);
fs/f2fs/dir.c
683
unsigned long bidx, block;
fs/f2fs/dir.c
717
for (block = bidx; block <= (bidx + nblock - 1); block++) {
fs/f2fs/dir.c
718
dentry_folio = f2fs_get_new_data_folio(dir, NULL, block, true);
fs/f2fs/file.c
4148
pgoff_t off, block_t block, block_t len, u32 flags)
fs/f2fs/file.c
4150
sector_t sector = SECTOR_FROM_BLOCK(block);
fs/f2fs/file.c
4165
ret = fscrypt_zeroout_range(inode, off, block, len);
fs/f2fs/node.c
723
static int get_node_path(struct inode *inode, long block,
fs/f2fs/node.c
736
if (block < direct_index) {
fs/f2fs/node.c
737
offset[n] = block;
fs/f2fs/node.c
740
block -= direct_index;
fs/f2fs/node.c
741
if (block < direct_blks) {
fs/f2fs/node.c
744
offset[n] = block;
fs/f2fs/node.c
748
block -= direct_blks;
fs/f2fs/node.c
749
if (block < direct_blks) {
fs/f2fs/node.c
752
offset[n] = block;
fs/f2fs/node.c
756
block -= direct_blks;
fs/f2fs/node.c
757
if (block < indirect_blks) {
fs/f2fs/node.c
760
offset[n++] = block / direct_blks;
fs/f2fs/node.c
762
offset[n] = block % direct_blks;
fs/f2fs/node.c
766
block -= indirect_blks;
fs/f2fs/node.c
767
if (block < indirect_blks) {
fs/f2fs/node.c
770
offset[n++] = block / direct_blks;
fs/f2fs/node.c
772
offset[n] = block % direct_blks;
fs/f2fs/node.c
776
block -= indirect_blks;
fs/f2fs/node.c
777
if (block < dindirect_blks) {
fs/f2fs/node.c
780
offset[n++] = block / indirect_blks;
fs/f2fs/node.c
783
offset[n++] = (block / direct_blks) % dptrs_per_blk;
fs/f2fs/node.c
787
offset[n] = block % direct_blks;
fs/f2fs/super.c
4499
int block;
fs/f2fs/super.c
4508
for (block = 0; block < 2; block++) {
fs/f2fs/super.c
4509
folio = read_mapping_folio(sb->s_bdev->bd_mapping, block, NULL);
fs/f2fs/super.c
4512
block + 1);
fs/f2fs/super.c
4519
err = sanity_check_raw_super(sbi, folio, block);
fs/f2fs/super.c
4522
block + 1);
fs/f2fs/super.c
4529
memcpy(super, F2FS_SUPER_BLOCK(folio, block),
fs/f2fs/super.c
4531
*valid_super_block = block;
fs/fat/inode.c
316
static sector_t _fat_bmap(struct address_space *mapping, sector_t block)
fs/fat/inode.c
322
blocknr = generic_block_bmap(mapping, block, fat_get_block_bmap);
fs/freevxfs/vxfs_bmap.c
107
vxfs_bmap_indir(struct inode *ip, long indir, int size, long block)
fs/freevxfs/vxfs_bmap.c
127
if (block < off) {
fs/freevxfs/vxfs_bmap.c
138
block - off);
fs/freevxfs/vxfs_bmap.c
143
if ((block - off) >= fs32_to_cpu(sbi, typ->vt_size))
fs/freevxfs/vxfs_bmap.c
145
pblock = fs32_to_cpu(sbi, typ->vt_block) + block - off;
fs/freevxfs/vxfs_inode.c
141
u_long block, offset;
fs/freevxfs/vxfs_inode.c
148
block = extent + ((ino * VXFS_ISIZE) / sbp->s_blocksize);
fs/freevxfs/vxfs_inode.c
150
bp = sb_bread(sbp, block);
fs/freevxfs/vxfs_inode.c
166
printk(KERN_WARNING "vxfs: unable to read block %ld\n", block);
fs/freevxfs/vxfs_olt.c
33
vxfs_oblock(struct super_block *sbp, daddr_t block, u_long bsize)
fs/freevxfs/vxfs_olt.c
36
return (block * (sbp->s_blocksize / bsize));
fs/freevxfs/vxfs_subr.c
149
vxfs_bmap(struct address_space *mapping, sector_t block)
fs/freevxfs/vxfs_subr.c
151
return generic_block_bmap(mapping, block, vxfs_getblk);
fs/freevxfs/vxfs_subr.c
72
vxfs_bread(struct inode *ip, int block)
fs/freevxfs/vxfs_subr.c
77
pblock = vxfs_bmap1(ip, block);
fs/fuse/file.c
2573
static sector_t fuse_bmap(struct address_space *mapping, sector_t block)
fs/fuse/file.c
2586
inarg.block = block;
fs/fuse/file.c
2600
return err ? 0 : outarg.block;
fs/gfs2/bmap.c
107
error = gfs2_alloc_blocks(ip, &block, &n, 0);
fs/gfs2/bmap.c
111
gfs2_trans_remove_revoke(GFS2_SB(&ip->i_inode), block, 1);
fs/gfs2/bmap.c
112
error = gfs2_dir_get_new_buffer(ip, block, &bh);
fs/gfs2/bmap.c
119
error = gfs2_unstuffer_folio(ip, dibh, block, folio);
fs/gfs2/bmap.c
132
*(__be64 *)(di + 1) = cpu_to_be64(block);
fs/gfs2/bmap.c
233
static void find_metapath(const struct gfs2_sbd *sdp, u64 block,
fs/gfs2/bmap.c
240
mp->mp_list[i] = do_div(block, sdp->sd_inptrs);
fs/gfs2/bmap.c
395
sector_t factor = 1, block = 0;
fs/gfs2/bmap.c
400
block += mp->mp_list[hgt] * factor;
fs/gfs2/bmap.c
403
return block;
fs/gfs2/bmap.c
55
u64 block, struct folio *folio)
fs/gfs2/bmap.c
78
map_bh(bh, inode->i_sb, block);
fs/gfs2/bmap.c
94
u64 block = 0;
fs/gfs2/dir.c
103
static int gfs2_dir_get_existing_buffer(struct gfs2_inode *ip, u64 block,
fs/gfs2/dir.c
109
error = gfs2_meta_read(ip->i_gl, block, DIO_WAIT, 0, &bh);
fs/gfs2/dir.c
90
int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block,
fs/gfs2/dir.c
95
bh = gfs2_meta_new(ip->i_gl, block);
fs/gfs2/dir.h
49
int gfs2_dir_get_new_buffer(struct gfs2_inode *ip, u64 block,
fs/gfs2/lops.c
508
unsigned int block = 0, blocks_submitted = 0, blocks_read = 0;
fs/gfs2/lops.c
528
for (; block < je->lblock + je->blocks; block++, dblock++) {
fs/gfs2/lops.c
531
block >> shift);
fs/gfs2/lops.c
540
if (bio && (off || block < blocks_submitted + max_blocks)) {
fs/gfs2/lops.c
558
blocks_submitted = block;
fs/gfs2/lops.c
585
while (blocks_read < block) {
fs/gfs2/quota.c
398
unsigned int block, offset;
fs/gfs2/quota.c
411
block = qd->qd_slot / sdp->sd_qc_per_block;
fs/gfs2/quota.c
415
(loff_t)block << inode->i_blkbits,
fs/gfs2/rgrp.c
1624
static u64 gfs2_next_unreserved_block(struct gfs2_rgrpd *rgd, u64 block,
fs/gfs2/rgrp.c
1636
rc = rs_cmp(block, length, rs);
fs/gfs2/rgrp.c
1646
while (rs_cmp(block, length, rs) == 0 && rs != ignore_rs) {
fs/gfs2/rgrp.c
1647
block = rs->rs_start + rs->rs_requested;
fs/gfs2/rgrp.c
1656
return block;
fs/gfs2/rgrp.c
1680
u64 block = gfs2_rbm_to_block(rbm);
fs/gfs2/rgrp.c
1698
nblock = gfs2_next_unreserved_block(rbm->rgd, block, extlen, rs);
fs/gfs2/rgrp.c
1699
if (nblock == block) {
fs/gfs2/rgrp.c
1708
u64 len = nblock - block;
fs/gfs2/rgrp.c
1844
u64 block;
fs/gfs2/rgrp.c
1860
block = gfs2_rbm_to_block(&rbm);
fs/gfs2/rgrp.c
1861
if (gfs2_rbm_from_block(&rbm, block + 1))
fs/gfs2/rgrp.c
1863
if (*last_unlinked != NO_BLOCK && block <= *last_unlinked)
fs/gfs2/rgrp.c
1865
if (block == skip)
fs/gfs2/rgrp.c
1867
*last_unlinked = block;
fs/gfs2/rgrp.c
1869
error = gfs2_glock_get(sdp, block, &gfs2_iopen_glops, CREATE, &gl);
fs/gfs2/rgrp.c
2232
u64 block;
fs/gfs2/rgrp.c
2236
block = gfs2_rbm_to_block(rbm);
fs/gfs2/rgrp.c
2239
block++;
fs/gfs2/rgrp.c
2241
ret = gfs2_rbm_from_block(&pos, block);
fs/gfs2/rgrp.c
2247
block++;
fs/gfs2/rgrp.c
2422
u64 block; /* block, within the file system scope */
fs/gfs2/rgrp.c
2448
block = gfs2_rbm_to_block(&rbm);
fs/gfs2/rgrp.c
2449
rbm.rgd->rd_last_alloc = block - rbm.rgd->rd_data0;
fs/gfs2/rgrp.c
2451
ip->i_goal = block + *nblocks - 1;
fs/gfs2/rgrp.c
2492
gfs2_trans_remove_revoke(sdp, block, *nblocks);
fs/gfs2/rgrp.c
2496
trace_gfs2_block_alloc(ip, rbm.rgd, block, *nblocks,
fs/gfs2/rgrp.c
2498
*bn = block;
fs/gfs2/rgrp.c
2659
u64 block)
fs/gfs2/rgrp.c
2676
if (rgrp_contains_block(rgd, block))
fs/gfs2/rgrp.c
2678
rgd = gfs2_blk2rgrpd(sdp, block, 1);
fs/gfs2/rgrp.c
2681
if (!rgd || !rgrp_contains_block(rgd, block))
fs/gfs2/rgrp.c
2682
rgd = gfs2_blk2rgrpd(sdp, block, 1);
fs/gfs2/rgrp.c
2687
(unsigned long long)block);
fs/gfs2/rgrp.c
277
static int gfs2_rbm_from_block(struct gfs2_rbm *rbm, u64 block)
fs/gfs2/rgrp.c
279
if (!rgrp_contains_block(rbm->rgd, block))
fs/gfs2/rgrp.c
282
rbm->offset = block - rbm->rgd->rd_data0;
fs/gfs2/rgrp.c
382
u64 block;
fs/gfs2/rgrp.c
406
block = gfs2_rbm_to_block(&rbm);
fs/gfs2/rgrp.c
407
if (gfs2_rbm_from_block(&rbm, block + chunk_size)) {
fs/gfs2/rgrp.h
66
u64 block);
fs/gfs2/rgrp.h
85
static inline int rgrp_contains_block(struct gfs2_rgrpd *rgd, u64 block)
fs/gfs2/rgrp.h
89
return first <= block && block < last;
fs/gfs2/sys.c
587
GDLM_ATTR(block, 0644, block_show, block_store);
fs/gfs2/trace_gfs2.h
335
__field( sector_t, block )
fs/gfs2/trace_gfs2.h
343
__entry->block = bd->bd_bh->b_blocknr;
fs/gfs2/trace_gfs2.h
350
(unsigned long long)__entry->block,
fs/gfs2/trace_gfs2.h
554
u64 block, unsigned len, u8 block_state),
fs/gfs2/trace_gfs2.h
556
TP_ARGS(ip, rgd, block, len, block_state),
fs/gfs2/trace_gfs2.h
572
__entry->start = block;
fs/gfs2/xattr.c
643
u64 block;
fs/gfs2/xattr.c
646
error = gfs2_alloc_blocks(ip, &block, &n, 0);
fs/gfs2/xattr.c
649
gfs2_trans_remove_revoke(sdp, block, 1);
fs/gfs2/xattr.c
650
*bhp = gfs2_meta_new(ip->i_gl, block);
fs/gfs2/xattr.c
704
u64 block;
fs/gfs2/xattr.c
708
error = gfs2_alloc_blocks(ip, &block, &n, 0);
fs/gfs2/xattr.c
711
gfs2_trans_remove_revoke(sdp, block, 1);
fs/gfs2/xattr.c
712
bh = gfs2_meta_new(ip->i_gl, block);
fs/hfs/bnode.c
347
int size, block, i, hash;
fs/hfs/bnode.c
385
block = off >> PAGE_SHIFT;
fs/hfs/bnode.c
388
page = read_mapping_page(mapping, block++, NULL);
fs/hfs/extent.c
107
return be16_to_cpu(ext->block) + be16_to_cpu(ext->count);
fs/hfs/extent.c
151
u32 cnid, u32 block, u8 type)
fs/hfs/extent.c
155
hfs_ext_build_key(fd->search_key, cnid, block, type);
fs/hfs/extent.c
169
static inline int __hfs_ext_cache_extent(struct hfs_find_data *fd, struct inode *inode, u32 block)
fs/hfs/extent.c
180
block, HFS_IS_RSRC(inode) ? HFS_FK_RSRC : HFS_FK_DATA);
fs/hfs/extent.c
191
static int hfs_ext_read_extent(struct inode *inode, u16 block)
fs/hfs/extent.c
196
if (block >= HFS_I(inode)->cached_start &&
fs/hfs/extent.c
197
block < HFS_I(inode)->cached_start + HFS_I(inode)->cached_blocks)
fs/hfs/extent.c
202
res = __hfs_ext_cache_extent(&fd, inode, block);
fs/hfs/extent.c
21
static void hfs_ext_build_key(hfs_btree_key *key, u32 cnid, u16 block, u8 type)
fs/hfs/extent.c
215
be16_to_cpu(extent[i].block),
fs/hfs/extent.c
230
start = be16_to_cpu(extent->block);
fs/hfs/extent.c
235
extent->block = cpu_to_be16(alloc_block);
fs/hfs/extent.c
26
key->ext.FABN = cpu_to_be16(block);
fs/hfs/extent.c
267
start = be16_to_cpu(extent->block);
fs/hfs/extent.c
270
extent->block = 0;
fs/hfs/extent.c
336
int hfs_get_block(struct inode *inode, sector_t block,
fs/hfs/extent.c
345
ablock = (u32)block / HFS_SB(sb)->fs_div;
fs/hfs/extent.c
347
if (block >= HFS_I(inode)->fs_blocks) {
fs/hfs/extent.c
350
if (block > HFS_I(inode)->fs_blocks)
fs/hfs/extent.c
379
(u32)block % HFS_SB(sb)->fs_div);
fs/hfs/extent.c
420
HFS_I(inode)->first_extents[0].block = cpu_to_be16(start);
fs/hfs/extent.c
466
HFS_I(inode)->cached_extents[0].block = cpu_to_be16(start);
fs/hfs/extent.c
82
return be16_to_cpu(ext->block) + off;
fs/hfs/hfs_fs.h
173
extern int hfs_get_block(struct inode *inode, sector_t block,
fs/hfs/inode.c
62
static sector_t hfs_bmap(struct address_space *mapping, sector_t block)
fs/hfs/inode.c
64
return generic_block_bmap(mapping, block, hfs_get_block);
fs/hfs/mdb.c
209
block = be16_to_cpu(mdb->drVBMSt) + part_start;
fs/hfs/mdb.c
210
off = (loff_t)block << HFS_SECTOR_SIZE_BITS;
fs/hfs/mdb.c
334
sector_t block;
fs/hfs/mdb.c
338
block = be16_to_cpu(HFS_SB(sb)->mdb->drVBMSt) + HFS_SB(sb)->part_start;
fs/hfs/mdb.c
339
off = (block << HFS_SECTOR_SIZE_BITS) & (sb->s_blocksize - 1);
fs/hfs/mdb.c
340
block >>= sb->s_blocksize_bits - HFS_SECTOR_SIZE_BITS;
fs/hfs/mdb.c
344
bh = sb_bread(sb, block);
fs/hfs/mdb.c
357
block++;
fs/hfs/mdb.c
98
unsigned int block;
fs/hfsplus/bnode.c
455
int size, block, i, hash;
fs/hfsplus/bnode.c
495
block = off >> PAGE_SHIFT;
fs/hfsplus/bnode.c
497
for (i = 0; i < tree->pages_per_bnode; block++, i++) {
fs/hfsplus/bnode.c
498
page = read_mapping_page(mapping, block, NULL);
fs/hfsplus/extents.c
158
u32 cnid, u32 block, u8 type)
fs/hfsplus/extents.c
162
hfsplus_ext_build_key(fd->search_key, cnid, block, type);
fs/hfsplus/extents.c
178
struct inode *inode, u32 block)
fs/hfsplus/extents.c
192
block, HFSPLUS_IS_RSRC(inode) ?
fs/hfsplus/extents.c
206
static int hfsplus_ext_read_extent(struct inode *inode, u32 block)
fs/hfsplus/extents.c
212
if (block >= hip->cached_start &&
fs/hfsplus/extents.c
213
block < hip->cached_start + hip->cached_blocks)
fs/hfsplus/extents.c
218
res = __hfsplus_ext_cache_extent(&fd, inode, block);
fs/hfsplus/extents.c
42
u32 block, u8 type)
fs/hfsplus/extents.c
46
key->ext.start_block = cpu_to_be32(block);
fs/hfsplus/inode.c
57
static sector_t hfsplus_bmap(struct address_space *mapping, sector_t block)
fs/hfsplus/inode.c
59
return generic_block_bmap(mapping, block, hfsplus_get_block);
fs/hpfs/file.c
228
static sector_t _hpfs_bmap(struct address_space *mapping, sector_t block)
fs/hpfs/file.c
230
return generic_block_bmap(mapping, block, hpfs_get_block);
fs/inode.c
2047
int bmap(struct inode *inode, sector_t *block)
fs/inode.c
2052
*block = inode->i_mapping->a_ops->bmap(inode->i_mapping, *block);
fs/ioctl.c
63
sector_t block;
fs/ioctl.c
75
block = ur_block;
fs/ioctl.c
76
error = bmap(inode, &block);
fs/ioctl.c
78
if (block > INT_MAX) {
fs/ioctl.c
88
ur_block = block;
fs/isofs/dir.c
100
block = ctx->pos >> bufbits;
fs/isofs/dir.c
106
bh = isofs_bread(inode, block);
fs/isofs/dir.c
125
block = ctx->pos >> bufbits;
fs/isofs/dir.c
130
block_saved = block;
fs/isofs/dir.c
139
block++;
fs/isofs/dir.c
143
bh = isofs_bread(inode, block);
fs/isofs/dir.c
155
" in block %lu of inode %lu\n", block,
fs/isofs/dir.c
89
unsigned long block, offset, block_saved, offset_saved;
fs/isofs/export.c
154
u32 block;
fs/isofs/export.c
170
return isofs_export_iget(sb, ifid->block, ifid->offset,
fs/isofs/export.c
21
unsigned long block,
fs/isofs/export.c
27
if (block == 0)
fs/isofs/export.c
29
inode = isofs_iget(sb, block, offset);
fs/isofs/inode.c
1126
static int isofs_bmap(struct inode *inode, sector_t block)
fs/isofs/inode.c
1133
error = isofs_get_block(inode, block, &dummy, 0);
fs/isofs/inode.c
1139
struct buffer_head *isofs_bread(struct inode *inode, sector_t block)
fs/isofs/inode.c
1141
sector_t blknr = isofs_bmap(inode, block);
fs/isofs/inode.c
1157
static sector_t _isofs_bmap(struct address_space *mapping, sector_t block)
fs/isofs/inode.c
1159
return generic_block_bmap(mapping,block,isofs_get_block);
fs/isofs/inode.c
1173
unsigned long block, offset, block_saved, offset_saved;
fs/isofs/inode.c
1187
block = ei->i_iget5_block;
fs/isofs/inode.c
1195
bh = sb_bread(inode->i_sb, block);
fs/isofs/inode.c
1205
++block;
fs/isofs/inode.c
1210
block_saved = block;
fs/isofs/inode.c
1224
block++;
fs/isofs/inode.c
1228
bh = sb_bread(inode->i_sb, block);
fs/isofs/inode.c
1258
printk(KERN_INFO "ISOFS: unable to read i-node block %lu\n", block);
fs/isofs/inode.c
1274
unsigned long block;
fs/isofs/inode.c
1285
block = ei->i_iget5_block;
fs/isofs/inode.c
1286
bh = sb_bread(inode->i_sb, block);
fs/isofs/inode.c
1307
bh = sb_bread(inode->i_sb, ++block);
fs/isofs/inode.c
1472
unsigned long block;
fs/isofs/inode.c
1481
return (i->i_iget5_block == d->block)
fs/isofs/inode.c
1490
i->i_iget5_block = d->block;
fs/isofs/inode.c
1500
unsigned long block,
fs/isofs/inode.c
1512
data.block = block;
fs/isofs/inode.c
1515
hashval = (block << sb->s_blocksize_bits) | offset;
fs/isofs/inode.c
550
static bool rootdir_empty(struct super_block *sb, unsigned long block)
fs/isofs/inode.c
556
bh = sb_bread(sb, block);
fs/isofs/inode.c
586
int iso_blknum, block;
fs/isofs/inode.c
631
block = iso_blknum << (ISOFS_BLOCK_BITS - s->s_blocksize_bits);
fs/isofs/inode.c
632
if (!(bh = sb_bread(s, block)))
fs/isofs/inode.c
975
__func__, s->s_id, iso_blknum, block);
fs/isofs/isofs.h
127
unsigned long block,
fs/isofs/isofs.h
132
unsigned long block,
fs/isofs/isofs.h
135
return __isofs_iget(sb, block, offset, 0);
fs/isofs/isofs.h
139
unsigned long block,
fs/isofs/isofs.h
142
return __isofs_iget(sb, block, offset, 1);
fs/isofs/isofs.h
149
static inline unsigned long isofs_get_ino(unsigned long block,
fs/isofs/isofs.h
153
return (block << (bufbits - 5)) | (offset >> 5);
fs/isofs/isofs.h
189
unsigned long *block,
fs/isofs/isofs.h
195
*block = (unsigned long)isonum_733(de->extent)
fs/isofs/namei.c
103
" in block %lu of inode %lu\n", block,
fs/isofs/namei.c
156
unsigned long block;
fs/isofs/namei.c
166
&block, &offset,
fs/isofs/namei.c
171
inode = found ? isofs_iget(dir->i_sb, block, offset) : NULL;
fs/isofs/namei.c
42
unsigned long block, f_pos, offset, block_saved, offset_saved;
fs/isofs/namei.c
51
block = 0;
fs/isofs/namei.c
59
bh = isofs_bread(dir, block);
fs/isofs/namei.c
71
block = f_pos >> bufbits;
fs/isofs/namei.c
86
block++;
fs/isofs/namei.c
90
bh = isofs_bread(dir, block);
fs/isofs/rock.c
706
unsigned long block, offset;
fs/isofs/rock.c
716
block = ei->i_iget5_block;
fs/isofs/rock.c
717
bh = sb_bread(inode->i_sb, block);
fs/jbd2/commit.c
313
unsigned long long block)
fs/jbd2/commit.c
315
tag->t_blocknr = cpu_to_be32(block & (u32)~0);
fs/jbd2/commit.c
317
tag->t_blocknr_high = cpu_to_be32((block >> 31) >> 1);
fs/jbd2/journal.c
1018
unsigned long *block)
fs/jbd2/journal.c
1028
*block = transaction->t_log_start;
fs/jbd2/journal.c
1031
*block = transaction->t_log_start;
fs/jbd2/journal.c
1034
*block = journal->j_head;
fs/jbd2/journal.c
1037
*block = journal->j_head;
fs/jbd2/journal.c
1056
int __jbd2_update_log_tail(journal_t *journal, tid_t tid, unsigned long block)
fs/jbd2/journal.c
1069
ret = jbd2_journal_update_sb_log_tail(journal, tid, block, REQ_FUA);
fs/jbd2/journal.c
1074
freed = block - journal->j_tail;
fs/jbd2/journal.c
1075
if (block < journal->j_tail)
fs/jbd2/journal.c
1078
trace_jbd2_update_log_tail(journal, tid, block, freed);
fs/jbd2/journal.c
1082
journal->j_tail_sequence, tid, block, freed);
fs/jbd2/journal.c
1086
journal->j_tail = block;
fs/jbd2/journal.c
1098
void jbd2_update_log_tail(journal_t *journal, tid_t tid, unsigned long block)
fs/jbd2/journal.c
1102
__jbd2_update_log_tail(journal, tid, block);
fs/jbd2/journal.c
1949
unsigned long block, log_offset; /* logical */
fs/jbd2/journal.c
1969
for (block = log_offset; block < journal->j_total_len; block++) {
fs/jbd2/journal.c
1970
err = jbd2_journal_bmap(journal, block, &phys_block);
fs/jbd2/journal.c
1972
pr_err("JBD2: bad block at offset %lu", block);
fs/jbd2/journal.c
1985
block--;
fs/jbd2/journal.c
1993
if (block != journal->j_total_len - 1)
fs/jbd2/journal.c
927
sector_t block = blocknr;
fs/jbd2/journal.c
930
err = journal->j_bmap(journal, &block);
fs/jbd2/journal.c
932
*retp = block;
fs/jbd2/journal.c
934
ret = bmap(journal->j_inode, &block);
fs/jbd2/journal.c
936
if (ret || !block) {
fs/jbd2/journal.c
943
*retp = block;
fs/jbd2/recovery.c
393
unsigned long long block = be32_to_cpu(tag->t_blocknr);
fs/jbd2/recovery.c
395
block |= (u64)be32_to_cpu(tag->t_blocknr_high) << 32;
fs/jbd2/recovery.c
396
return block;
fs/jbd2/revoke.c
131
static inline int hash(journal_t *journal, unsigned long long block)
fs/jbd2/revoke.c
133
return hash_64(block, journal->j_revoke->hash_shift);
fs/jffs2/erase.c
289
struct jffs2_raw_node_ref *block, *ref;
fs/jffs2/erase.c
293
block = ref = jeb->first_node;
fs/jffs2/erase.c
298
jffs2_free_refblock(block);
fs/jffs2/erase.c
299
block = ref;
fs/jfs/inode.c
326
static sector_t jfs_bmap(struct address_space *mapping, sector_t block)
fs/jfs/inode.c
328
return generic_block_bmap(mapping, block, jfs_get_block);
fs/lockd/clnt4xdr.c
382
encode_bool(xdr, args->block);
fs/lockd/clnt4xdr.c
405
encode_bool(xdr, args->block);
fs/lockd/clntlock.c
105
void nlmclnt_queue_block(struct nlm_wait *block)
fs/lockd/clntlock.c
108
list_add(&block->b_list, &nlm_blocked);
fs/lockd/clntlock.c
115
__be32 nlmclnt_dequeue_block(struct nlm_wait *block)
fs/lockd/clntlock.c
120
list_del(&block->b_list);
fs/lockd/clntlock.c
121
status = block->b_status;
fs/lockd/clntlock.c
129
int nlmclnt_wait(struct nlm_wait *block, struct nlm_rqst *req, long timeout)
fs/lockd/clntlock.c
136
if (block == NULL)
fs/lockd/clntlock.c
147
ret = wait_event_interruptible_timeout(block->b_wait,
fs/lockd/clntlock.c
148
block->b_status != nlm_lck_blocked,
fs/lockd/clntlock.c
153
if (block->b_status == nlm_lck_denied_grace_period)
fs/lockd/clntlock.c
154
block->b_status = nlm_lck_blocked;
fs/lockd/clntlock.c
165
struct nlm_wait *block;
fs/lockd/clntlock.c
173
list_for_each_entry(block, &nlm_blocked, b_list) {
fs/lockd/clntlock.c
174
struct file_lock *fl_blocked = block->b_lock;
fs/lockd/clntlock.c
186
if (!rpc_cmp_addr(nlm_addr(block->b_host), addr))
fs/lockd/clntlock.c
193
block->b_status = nlm_granted;
fs/lockd/clntlock.c
194
wake_up(&block->b_wait);
fs/lockd/clntlock.c
230
struct nlm_wait *block;
fs/lockd/clntlock.c
284
list_for_each_entry(block, &nlm_blocked, b_list) {
fs/lockd/clntlock.c
285
if (block->b_host == host) {
fs/lockd/clntlock.c
286
block->b_status = nlm_lck_denied_grace_period;
fs/lockd/clntlock.c
287
wake_up(&block->b_wait);
fs/lockd/clntlock.c
88
void nlmclnt_prepare_block(struct nlm_wait *block, struct nlm_host *host, struct file_lock *fl)
fs/lockd/clntlock.c
90
block->b_host = host;
fs/lockd/clntlock.c
91
block->b_lock = fl;
fs/lockd/clntlock.c
92
init_waitqueue_head(&block->b_wait);
fs/lockd/clntlock.c
93
block->b_status = nlm_lck_blocked;
fs/lockd/clntproc.c
187
call->a_args.block = IS_SETLKW(cmd) ? 1 : 0;
fs/lockd/clntproc.c
527
struct nlm_wait block;
fs/lockd/clntproc.c
543
nlmclnt_prepare_block(&block, host, fl);
fs/lockd/clntproc.c
556
nlmclnt_queue_block(&block);
fs/lockd/clntproc.c
569
status = nlmclnt_wait(&block, req, NLMCLNT_POLL_TIMEOUT);
fs/lockd/clntproc.c
572
if (block.b_status != nlm_lck_blocked)
fs/lockd/clntproc.c
575
b_status = nlmclnt_dequeue_block(&block);
fs/lockd/clntproc.c
583
if (!req->a_args.block)
fs/lockd/clntproc.c
585
if (nlmclnt_cancel(host, req->a_args.block, fl) == 0)
fs/lockd/clntproc.c
786
static int nlmclnt_cancel(struct nlm_host *host, int block, struct file_lock *fl)
fs/lockd/clntproc.c
800
req->a_args.block = block;
fs/lockd/clntxdr.c
383
encode_bool(xdr, args->block);
fs/lockd/clntxdr.c
406
encode_bool(xdr, args->block);
fs/lockd/svc4proc.c
147
argp->block, &argp->cookie,
fs/lockd/svclock.c
1002
nlmsvc_insert_block(block, 10 * HZ);
fs/lockd/svclock.c
1006
nlmsvc_unlink_block(block);
fs/lockd/svclock.c
1007
fl = &block->b_call->a_args.lock.fl;
fs/lockd/svclock.c
1018
nlmsvc_unlink_block(block);
fs/lockd/svclock.c
1020
nlmsvc_release_block(block);
fs/lockd/svclock.c
1028
retry_deferred_block(struct nlm_block *block)
fs/lockd/svclock.c
1030
if (!(block->b_flags & B_GOT_CALLBACK))
fs/lockd/svclock.c
1031
block->b_flags |= B_TIMED_OUT;
fs/lockd/svclock.c
1032
nlmsvc_insert_block(block, NLM_TIMEOUT);
fs/lockd/svclock.c
1033
dprintk("revisit block %p flags %d\n", block, block->b_flags);
fs/lockd/svclock.c
1034
if (block->b_deferred_req) {
fs/lockd/svclock.c
1035
block->b_deferred_req->revisit(block->b_deferred_req, 0);
fs/lockd/svclock.c
1036
block->b_deferred_req = NULL;
fs/lockd/svclock.c
1049
struct nlm_block *block;
fs/lockd/svclock.c
1053
block = list_entry(nlm_blocked.next, struct nlm_block, b_list);
fs/lockd/svclock.c
1055
if (block->b_when == NLM_NEVER)
fs/lockd/svclock.c
1057
if (time_after(block->b_when, jiffies)) {
fs/lockd/svclock.c
1058
timeout = block->b_when - jiffies;
fs/lockd/svclock.c
1064
block, block->b_when);
fs/lockd/svclock.c
1065
if (block->b_flags & B_QUEUED) {
fs/lockd/svclock.c
1067
block, block->b_granted, block->b_flags);
fs/lockd/svclock.c
1068
retry_deferred_block(block);
fs/lockd/svclock.c
1070
nlmsvc_grant_blocked(block);
fs/lockd/svclock.c
115
list_add_tail(&block->b_list, pos);
fs/lockd/svclock.c
116
block->b_when = when;
fs/lockd/svclock.c
119
static void nlmsvc_insert_block(struct nlm_block *block, unsigned long when)
fs/lockd/svclock.c
122
nlmsvc_insert_block_locked(block, when);
fs/lockd/svclock.c
130
nlmsvc_remove_block(struct nlm_block *block)
fs/lockd/svclock.c
133
if (!list_empty(&block->b_list)) {
fs/lockd/svclock.c
134
list_del_init(&block->b_list);
fs/lockd/svclock.c
136
nlmsvc_release_block(block);
fs/lockd/svclock.c
148
struct nlm_block *block;
fs/lockd/svclock.c
157
list_for_each_entry(block, &nlm_blocked, b_list) {
fs/lockd/svclock.c
158
fl = &block->b_call->a_args.lock.fl;
fs/lockd/svclock.c
160
block->b_file, fl->c.flc_pid,
fs/lockd/svclock.c
163
nlmdbg_cookie2a(&block->b_call->a_args.cookie));
fs/lockd/svclock.c
164
if (block->b_file == file && nlm_compare_locks(fl, &lock->fl)) {
fs/lockd/svclock.c
165
kref_get(&block->b_count);
fs/lockd/svclock.c
167
return block;
fs/lockd/svclock.c
190
struct nlm_block *block;
fs/lockd/svclock.c
193
list_for_each_entry(block, &nlm_blocked, b_list) {
fs/lockd/svclock.c
194
if (nlm_cookie_match(&block->b_call->a_args.cookie,cookie))
fs/lockd/svclock.c
202
dprintk("nlmsvc_find_block(%s): block=%p\n", nlmdbg_cookie2a(cookie), block);
fs/lockd/svclock.c
203
kref_get(&block->b_count);
fs/lockd/svclock.c
205
return block;
fs/lockd/svclock.c
228
struct nlm_block *block;
fs/lockd/svclock.c
236
block = kzalloc_obj(*block);
fs/lockd/svclock.c
237
if (block == NULL)
fs/lockd/svclock.c
239
kref_init(&block->b_count);
fs/lockd/svclock.c
240
INIT_LIST_HEAD(&block->b_list);
fs/lockd/svclock.c
241
INIT_LIST_HEAD(&block->b_flist);
fs/lockd/svclock.c
251
dprintk("lockd: created block %p...\n", block);
fs/lockd/svclock.c
254
block->b_daemon = rqstp->rq_server;
fs/lockd/svclock.c
255
block->b_host = host;
fs/lockd/svclock.c
256
block->b_file = file;
fs/lockd/svclock.c
260
list_add(&block->b_flist, &file->f_blocks);
fs/lockd/svclock.c
263
block->b_call = call;
fs/lockd/svclock.c
265
call->a_block = block;
fs/lockd/svclock.c
267
return block;
fs/lockd/svclock.c
270
kfree(block);
fs/lockd/svclock.c
281
static int nlmsvc_unlink_block(struct nlm_block *block)
fs/lockd/svclock.c
284
dprintk("lockd: unlinking block %p...\n", block);
fs/lockd/svclock.c
287
status = locks_delete_block(&block->b_call->a_args.lock.fl);
fs/lockd/svclock.c
288
nlmsvc_remove_block(block);
fs/lockd/svclock.c
294
struct nlm_block *block = container_of(kref, struct nlm_block, b_count);
fs/lockd/svclock.c
295
struct nlm_file *file = block->b_file;
fs/lockd/svclock.c
297
dprintk("lockd: freeing block %p...\n", block);
fs/lockd/svclock.c
300
list_del_init(&block->b_flist);
fs/lockd/svclock.c
303
nlmsvc_freegrantargs(block->b_call);
fs/lockd/svclock.c
304
nlmsvc_release_call(block->b_call);
fs/lockd/svclock.c
305
nlm_release_file(block->b_file);
fs/lockd/svclock.c
306
kfree(block);
fs/lockd/svclock.c
309
static void nlmsvc_release_block(struct nlm_block *block)
fs/lockd/svclock.c
311
if (block != NULL)
fs/lockd/svclock.c
312
kref_put_mutex(&block->b_count, nlmsvc_free_block, &block->b_file->f_mutex);
fs/lockd/svclock.c
323
struct nlm_block *block, *next;
fs/lockd/svclock.c
328
list_for_each_entry_safe(block, next, &file->f_blocks, b_flist) {
fs/lockd/svclock.c
329
if (!match(block->b_host, host))
fs/lockd/svclock.c
333
if (list_empty(&block->b_list))
fs/lockd/svclock.c
335
kref_get(&block->b_count);
fs/lockd/svclock.c
338
nlmsvc_unlink_block(block);
fs/lockd/svclock.c
339
nlmsvc_release_block(block);
fs/lockd/svclock.c
42
static void nlmsvc_release_block(struct nlm_block *block);
fs/lockd/svclock.c
43
static void nlmsvc_insert_block(struct nlm_block *block, unsigned long);
fs/lockd/svclock.c
44
static void nlmsvc_remove_block(struct nlm_block *block);
fs/lockd/svclock.c
453
nlmsvc_defer_lock_rqst(struct svc_rqst *rqstp, struct nlm_block *block)
fs/lockd/svclock.c
457
block->b_flags |= B_QUEUED;
fs/lockd/svclock.c
459
nlmsvc_insert_block(block, NLM_TIMEOUT);
fs/lockd/svclock.c
461
block->b_cache_req = &rqstp->rq_chandle;
fs/lockd/svclock.c
463
block->b_deferred_req =
fs/lockd/svclock.c
464
rqstp->rq_chandle.defer(block->b_cache_req);
fs/lockd/svclock.c
465
if (block->b_deferred_req != NULL)
fs/lockd/svclock.c
469
block, block->b_flags, ntohl(status));
fs/lockd/svclock.c
484
struct nlm_block *block = NULL;
fs/lockd/svclock.c
511
block = nlmsvc_lookup_block(file, lock);
fs/lockd/svclock.c
512
if (block == NULL) {
fs/lockd/svclock.c
513
block = nlmsvc_create_block(rqstp, host, file, lock, cookie);
fs/lockd/svclock.c
515
if (block == NULL)
fs/lockd/svclock.c
517
lock = &block->b_call->a_args.lock;
fs/lockd/svclock.c
521
if (block->b_flags & B_QUEUED) {
fs/lockd/svclock.c
523
block, block->b_flags);
fs/lockd/svclock.c
524
if (block->b_granted) {
fs/lockd/svclock.c
525
nlmsvc_unlink_block(block);
fs/lockd/svclock.c
529
if (block->b_flags & B_TIMED_OUT) {
fs/lockd/svclock.c
530
nlmsvc_unlink_block(block);
fs/lockd/svclock.c
556
!list_empty(&block->b_list)) {
fs/lockd/svclock.c
563
nlmsvc_insert_block_locked(block, NLM_NEVER);
fs/lockd/svclock.c
575
nlmsvc_remove_block(block);
fs/lockd/svclock.c
580
nlmsvc_remove_block(block);
fs/lockd/svclock.c
588
ret = nlmsvc_defer_lock_rqst(rqstp, block);
fs/lockd/svclock.c
591
nlmsvc_remove_block(block);
fs/lockd/svclock.c
595
nlmsvc_remove_block(block);
fs/lockd/svclock.c
603
nlmsvc_release_block(block);
fs/lockd/svclock.c
715
struct nlm_block *block;
fs/lockd/svclock.c
733
block = nlmsvc_lookup_block(file, lock);
fs/lockd/svclock.c
735
if (block != NULL) {
fs/lockd/svclock.c
736
struct file_lock *fl = &block->b_call->a_args.lock.fl;
fs/lockd/svclock.c
739
vfs_cancel_lock(block->b_file->f_file[mode], fl);
fs/lockd/svclock.c
740
status = nlmsvc_unlink_block(block);
fs/lockd/svclock.c
741
nlmsvc_release_block(block);
fs/lockd/svclock.c
756
nlmsvc_update_deferred_block(struct nlm_block *block, int result)
fs/lockd/svclock.c
758
block->b_flags |= B_GOT_CALLBACK;
fs/lockd/svclock.c
760
block->b_granted = 1;
fs/lockd/svclock.c
762
block->b_flags |= B_TIMED_OUT;
fs/lockd/svclock.c
767
struct nlm_block *block;
fs/lockd/svclock.c
771
list_for_each_entry(block, &nlm_blocked, b_list) {
fs/lockd/svclock.c
772
if (nlm_compare_locks(&block->b_call->a_args.lock.fl, fl)) {
fs/lockd/svclock.c
774
block, block->b_flags);
fs/lockd/svclock.c
775
if (block->b_flags & B_QUEUED) {
fs/lockd/svclock.c
776
if (block->b_flags & B_TIMED_OUT) {
fs/lockd/svclock.c
780
nlmsvc_update_deferred_block(block, result);
fs/lockd/svclock.c
782
block->b_granted = 1;
fs/lockd/svclock.c
784
nlmsvc_insert_block_locked(block, 0);
fs/lockd/svclock.c
785
svc_wake_up(block->b_daemon);
fs/lockd/svclock.c
806
struct nlm_block *block;
fs/lockd/svclock.c
810
list_for_each_entry(block, &nlm_blocked, b_list) {
fs/lockd/svclock.c
811
if (nlm_compare_locks(&block->b_call->a_args.lock.fl, fl)) {
fs/lockd/svclock.c
812
nlmsvc_insert_block_locked(block, 0);
fs/lockd/svclock.c
814
svc_wake_up(block->b_daemon);
fs/lockd/svclock.c
851
nlmsvc_grant_blocked(struct nlm_block *block)
fs/lockd/svclock.c
853
struct nlm_file *file = block->b_file;
fs/lockd/svclock.c
854
struct nlm_lock *lock = &block->b_call->a_args.lock;
fs/lockd/svclock.c
859
dprintk("lockd: grant blocked lock %p\n", block);
fs/lockd/svclock.c
861
kref_get(&block->b_count);
fs/lockd/svclock.c
864
nlmsvc_unlink_block(block);
fs/lockd/svclock.c
869
if (block->b_granted) {
fs/lockd/svclock.c
870
nlm_rebind_host(block->b_host);
fs/lockd/svclock.c
89
nlmsvc_insert_block_locked(struct nlm_block *block, unsigned long when)
fs/lockd/svclock.c
892
nlmsvc_insert_block(block, NLM_NEVER);
fs/lockd/svclock.c
893
nlmsvc_release_block(block);
fs/lockd/svclock.c
898
nlmsvc_insert_block(block, 10 * HZ);
fs/lockd/svclock.c
899
nlmsvc_release_block(block);
fs/lockd/svclock.c
906
block->b_granted = 1;
fs/lockd/svclock.c
911
nlmsvc_insert_block(block, NLM_NEVER);
fs/lockd/svclock.c
916
error = nlm_async_call(block->b_call, NLMPROC_GRANTED_MSG,
fs/lockd/svclock.c
921
nlmsvc_insert_block(block, 10 * HZ);
fs/lockd/svclock.c
935
struct nlm_block *block = call->a_block;
fs/lockd/svclock.c
94
dprintk("lockd: nlmsvc_insert_block(%p, %ld)\n", block, when);
fs/lockd/svclock.c
949
if (list_empty(&block->b_list))
fs/lockd/svclock.c
95
if (list_empty(&block->b_list)) {
fs/lockd/svclock.c
96
kref_get(&block->b_count);
fs/lockd/svclock.c
962
nlmsvc_insert_block_locked(block, timeout);
fs/lockd/svclock.c
963
svc_wake_up(block->b_daemon);
fs/lockd/svclock.c
98
list_del_init(&block->b_list);
fs/lockd/svclock.c
990
struct nlm_block *block;
fs/lockd/svclock.c
996
if (!(block = nlmsvc_find_block(cookie)))
fs/lockd/svcproc.c
168
argp->block, &argp->cookie,
fs/lockd/xdr.c
180
if (xdr_stream_decode_bool(xdr, &argp->block) < 0)
fs/lockd/xdr.c
205
if (xdr_stream_decode_bool(xdr, &argp->block) < 0)
fs/lockd/xdr4.c
174
if (xdr_stream_decode_bool(xdr, &argp->block) < 0)
fs/lockd/xdr4.c
199
if (xdr_stream_decode_bool(xdr, &argp->block) < 0)
fs/minix/bitmap.c
109
int block;
fs/minix/bitmap.c
119
block = 2 + sbi->s_imap_blocks + sbi->s_zmap_blocks +
fs/minix/bitmap.c
121
*bh = sb_bread(sb, block);
fs/minix/bitmap.c
133
int block;
fs/minix/bitmap.c
145
block = 2 + sbi->s_imap_blocks + sbi->s_zmap_blocks +
fs/minix/bitmap.c
147
*bh = sb_bread(sb, block);
fs/minix/bitmap.c
42
void minix_free_block(struct inode *inode, unsigned long block)
fs/minix/bitmap.c
50
if (block < sbi->s_firstdatazone || block >= sbi->s_nzones) {
fs/minix/bitmap.c
54
zone = block - sbi->s_firstdatazone + 1;
fs/minix/bitmap.c
65
sb->s_id, block);
fs/minix/inode.c
173
unsigned long block;
fs/minix/inode.c
189
block = minix_blocks_needed(sbi->s_ninodes, sb->s_blocksize);
fs/minix/inode.c
190
if (sbi->s_imap_blocks < block) {
fs/minix/inode.c
196
block = minix_blocks_needed(
fs/minix/inode.c
199
if (sbi->s_zmap_blocks < block) {
fs/minix/inode.c
223
unsigned long i, block;
fs/minix/inode.c
309
block=2;
fs/minix/inode.c
311
if (!(sbi->s_imap[i]=sb_bread(s, block)))
fs/minix/inode.c
313
block++;
fs/minix/inode.c
316
if (!(sbi->s_zmap[i]=sb_bread(s, block)))
fs/minix/inode.c
318
block++;
fs/minix/inode.c
434
static int minix_get_block(struct inode *inode, sector_t block,
fs/minix/inode.c
438
return V1_minix_get_block(inode, block, bh_result, create);
fs/minix/inode.c
440
return V2_minix_get_block(inode, block, bh_result, create);
fs/minix/inode.c
483
static sector_t minix_bmap(struct address_space *mapping, sector_t block)
fs/minix/inode.c
485
return generic_block_bmap(mapping,block,minix_get_block);
fs/minix/itree_common.c
152
static int get_block(struct inode * inode, sector_t block,
fs/minix/itree_common.c
160
int depth = block_to_path(inode, block, offsets);
fs/minix/itree_v1.c
25
static int block_to_path(struct inode * inode, long block, int offsets[DEPTH])
fs/minix/itree_v1.c
29
if (block < 0) {
fs/minix/itree_v1.c
31
block, inode->i_sb->s_bdev);
fs/minix/itree_v1.c
34
if ((u64)block * BLOCK_SIZE >= inode->i_sb->s_maxbytes)
fs/minix/itree_v1.c
37
if (block < 7) {
fs/minix/itree_v1.c
38
offsets[n++] = block;
fs/minix/itree_v1.c
39
} else if ((block -= 7) < 512) {
fs/minix/itree_v1.c
41
offsets[n++] = block;
fs/minix/itree_v1.c
43
block -= 512;
fs/minix/itree_v1.c
45
offsets[n++] = block>>9;
fs/minix/itree_v1.c
46
offsets[n++] = block & 511;
fs/minix/itree_v1.c
53
int V1_minix_get_block(struct inode * inode, long block,
fs/minix/itree_v1.c
56
return get_block(inode, block, bh_result, create);
fs/minix/itree_v2.c
27
static int block_to_path(struct inode * inode, long block, int offsets[DEPTH])
fs/minix/itree_v2.c
32
if (block < 0) {
fs/minix/itree_v2.c
34
block, sb->s_bdev);
fs/minix/itree_v2.c
37
if ((u64)block * (u64)sb->s_blocksize >= sb->s_maxbytes)
fs/minix/itree_v2.c
40
if (block < DIRCOUNT) {
fs/minix/itree_v2.c
41
offsets[n++] = block;
fs/minix/itree_v2.c
42
} else if ((block -= DIRCOUNT) < INDIRCOUNT(sb)) {
fs/minix/itree_v2.c
44
offsets[n++] = block;
fs/minix/itree_v2.c
45
} else if ((block -= INDIRCOUNT(sb)) < INDIRCOUNT(sb) * INDIRCOUNT(sb)) {
fs/minix/itree_v2.c
47
offsets[n++] = block / INDIRCOUNT(sb);
fs/minix/itree_v2.c
48
offsets[n++] = block % INDIRCOUNT(sb);
fs/minix/itree_v2.c
50
block -= INDIRCOUNT(sb) * INDIRCOUNT(sb);
fs/minix/itree_v2.c
52
offsets[n++] = (block / INDIRCOUNT(sb)) / INDIRCOUNT(sb);
fs/minix/itree_v2.c
53
offsets[n++] = (block / INDIRCOUNT(sb)) % INDIRCOUNT(sb);
fs/minix/itree_v2.c
54
offsets[n++] = block % INDIRCOUNT(sb);
fs/minix/itree_v2.c
61
int V2_minix_get_block(struct inode * inode, long block,
fs/minix/itree_v2.c
64
return get_block(inode, block, bh_result, create);
fs/minix/minix.h
55
void minix_free_block(struct inode *inode, unsigned long block);
fs/mpage.c
102
int block = 0;
fs/mpage.c
120
if (block == page_block) {
fs/mpage.c
127
block++;
fs/nfs/nfs4proc.c
7456
data->arg.block = 1;
fs/nfs/nfs4xdr.c
1309
static inline int nfs4_lock_type(struct file_lock *fl, int block)
fs/nfs/nfs4xdr.c
1312
return block ? NFS4_READW_LT : NFS4_READ_LT;
fs/nfs/nfs4xdr.c
1313
return block ? NFS4_WRITEW_LT : NFS4_WRITE_LT;
fs/nfs/nfs4xdr.c
1345
*p++ = cpu_to_be32(nfs4_lock_type(args->fl, args->block));
fs/nilfs2/mdt.c
170
static int nilfs_mdt_read_block(struct inode *inode, unsigned long block,
fs/nilfs2/mdt.c
178
err = nilfs_mdt_submit_block(inode, block, REQ_OP_READ, &first_bh);
fs/nilfs2/mdt.c
186
blkoff = block + 1;
fs/nilfs2/mdt.c
207
inode->i_ino, block);
fs/nilfs2/mdt.c
30
nilfs_mdt_insert_new_block(struct inode *inode, unsigned long block,
fs/nilfs2/mdt.c
328
int nilfs_mdt_delete_block(struct inode *inode, unsigned long block)
fs/nilfs2/mdt.c
333
err = nilfs_bmap_delete(ii->i_bmap, block);
fs/nilfs2/mdt.c
336
nilfs_mdt_forget_block(inode, block);
fs/nilfs2/mdt.c
354
int nilfs_mdt_forget_block(struct inode *inode, unsigned long block)
fs/nilfs2/mdt.c
356
pgoff_t index = block >> (PAGE_SHIFT - inode->i_blkbits);
fs/nilfs2/mdt.c
372
bh = get_nth_bh(bh, block - first_block);
fs/nilfs2/mdt.c
45
ret = nilfs_bmap_insert(ii->i_bmap, block, (unsigned long)bh);
fs/nilfs2/mdt.c
64
trace_nilfs2_mdt_insert_new_block(inode, inode->i_ino, block);
fs/nilfs2/mdt.c
69
static int nilfs_mdt_create_block(struct inode *inode, unsigned long block,
fs/nilfs2/mdt.c
83
bh = nilfs_grab_buffer(inode, inode->i_mapping, block, 0);
fs/nilfs2/mdt.c
95
err = nilfs_mdt_insert_new_block(inode, block, bh, init_block);
fs/nilfs2/page.c
29
unsigned long block, pgoff_t index, int blkbits,
fs/nilfs2/page.c
40
bh = get_nth_bh(bh, block - first_block);
fs/ntfs3/fsntfs.c
1036
struct buffer_head *ntfs_bread(struct super_block *sb, sector_t block)
fs/ntfs3/fsntfs.c
1041
if (unlikely(block >= sbi->volume.blocks)) {
fs/ntfs3/fsntfs.c
1044
(u64)block << sb->s_blocksize_bits);
fs/ntfs3/fsntfs.c
1048
bh = sb_bread_unmovable(sb, block);
fs/ntfs3/fsntfs.c
1053
(u64)block << sb->s_blocksize_bits);
fs/ntfs3/fsntfs.c
1062
sector_t block = lbo >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
1070
for (; bytes; block += 1, off = 0, op = blocksize) {
fs/ntfs3/fsntfs.c
1075
bh = __bread(bdev, block, blocksize);
fs/ntfs3/fsntfs.c
1078
(u64)block);
fs/ntfs3/fsntfs.c
1082
bh = __getblk(bdev, block, blocksize);
fs/ntfs3/fsntfs.c
1107
(u64)block, err);
fs/ntfs3/fsntfs.c
1235
sector_t block = lbo >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
1252
bh = ntfs_bread(sb, block);
fs/ntfs3/fsntfs.c
1277
block += 1;
fs/ntfs3/fsntfs.c
1362
sector_t block = lbo >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
1378
bh = sb_getblk(sb, block);
fs/ntfs3/fsntfs.c
1391
bh = ntfs_bread(sb, block);
fs/ntfs3/fsntfs.c
1405
block += 1;
fs/ntfs3/index.c
376
sector_t blen, block;
fs/ntfs3/index.c
413
block = (sector_t)lcn * sbi->blocks_per_cluster;
fs/ntfs3/index.c
416
bh = ntfs_bread(sb, block + blk);
fs/ntfs3/inode.c
557
static sector_t ntfs_bmap(struct address_space *mapping, sector_t block)
fs/ntfs3/inode.c
578
return iomap_bmap(mapping, block, &ntfs_iomap_ops);
fs/ntfs3/ntfs_fs.h
630
struct buffer_head *ntfs_bread(struct super_block *sb, sector_t block);
fs/ocfs2/alloc.c
4798
u64 block;
fs/ocfs2/alloc.c
4853
block = ocfs2_clusters_to_blocks(osb->sb, bit_off);
fs/ocfs2/alloc.c
4857
status = ocfs2_insert_extent(handle, et, *logical_offset, block,
fs/ocfs2/alloc.c
7087
u64 block;
fs/ocfs2/alloc.c
7145
block = phys = ocfs2_clusters_to_blocks(inode->i_sb, bit_off);
fs/ocfs2/alloc.c
7187
ret = ocfs2_insert_extent(handle, &et, 0, block, 1, 0, NULL);
fs/ocfs2/aops.c
426
static sector_t ocfs2_bmap(struct address_space *mapping, sector_t block)
fs/ocfs2/aops.c
434
(unsigned long long)block);
fs/ocfs2/aops.c
459
err = ocfs2_extent_map_get_blocks(inode, block, &p_blkno, NULL,
fs/ocfs2/aops.c
469
(unsigned long long)block);
fs/ocfs2/buffer_head_io.c
109
bhs[i] = sb_getblk(osb->sb, block++);
fs/ocfs2/buffer_head_io.c
193
int ocfs2_read_blocks(struct ocfs2_caching_info *ci, u64 block, int nr,
fs/ocfs2/buffer_head_io.c
204
trace_ocfs2_read_blocks_begin(ci, (unsigned long long)block, nr, flags);
fs/ocfs2/buffer_head_io.c
236
bhs[i] = sb_getblk(sb, block++);
fs/ocfs2/buffer_head_io.c
396
trace_ocfs2_read_blocks_end((unsigned long long)block, nr,
fs/ocfs2/buffer_head_io.c
89
int ocfs2_read_blocks_sync(struct ocfs2_super *osb, u64 block,
fs/ocfs2/buffer_head_io.c
97
trace_ocfs2_read_blocks_sync((unsigned long long)block, nr);
fs/ocfs2/buffer_head_io.h
18
int ocfs2_read_blocks_sync(struct ocfs2_super *osb, u64 block,
fs/ocfs2/buffer_head_io.h
28
int ocfs2_read_blocks(struct ocfs2_caching_info *ci, u64 block, int nr,
fs/ocfs2/dir.c
690
unsigned long start, block, b;
fs/ocfs2/dir.c
704
block = start;
fs/ocfs2/dir.c
714
b = block;
fs/ocfs2/dir.c
721
if (b >= nblocks || (num && block == start)) {
fs/ocfs2/dir.c
735
if (ocfs2_read_dir_block(dir, block, &bh, 0)) {
fs/ocfs2/dir.c
741
block);
fs/ocfs2/dir.c
745
block << sb->s_blocksize_bits,
fs/ocfs2/dir.c
749
OCFS2_I(dir)->ip_dir_start_lookup = block;
fs/ocfs2/dir.c
758
if (++block >= nblocks)
fs/ocfs2/dir.c
759
block = 0;
fs/ocfs2/dir.c
760
} while (block != start);
fs/ocfs2/dir.c
766
block = nblocks;
fs/ocfs2/dir.c
768
if (block < nblocks) {
fs/ocfs2/ocfs2_trace.h
2686
TP_PROTO(unsigned long long block, void *ci),
fs/ocfs2/ocfs2_trace.h
2687
TP_ARGS(block, ci),
fs/ocfs2/ocfs2_trace.h
2689
__field(unsigned long long, block)
fs/ocfs2/ocfs2_trace.h
2693
__entry->block = block;
fs/ocfs2/ocfs2_trace.h
2696
TP_printk("%llu %p", __entry->block, __entry->ci)
fs/ocfs2/ocfs2_trace.h
2700
TP_PROTO(void *ci, unsigned long long block,
fs/ocfs2/ocfs2_trace.h
2702
TP_ARGS(ci, block, nr, flags),
fs/ocfs2/ocfs2_trace.h
2705
__field(unsigned long long, block)
fs/ocfs2/ocfs2_trace.h
2711
__entry->block = block;
fs/ocfs2/ocfs2_trace.h
2715
TP_printk("%p %llu %u %d", __entry->ci, __entry->block,
fs/ocfs2/suballoc.h
145
static inline u64 ocfs2_which_suballoc_group(u64 block, unsigned int bit)
fs/ocfs2/suballoc.h
147
u64 group = block - (u64) bit;
fs/ocfs2/super.c
115
int block,
fs/ocfs2/super.c
1696
int block,
fs/ocfs2/super.c
1704
*bh = sb_getblk(sb, block);
fs/ocfs2/uptodate.c
208
sector_t block)
fs/ocfs2/uptodate.c
216
if (block < item->c_block)
fs/ocfs2/uptodate.c
218
else if (block > item->c_block)
fs/ocfs2/uptodate.c
288
sector_t block)
fs/ocfs2/uptodate.c
294
(unsigned long long)block, ci->ci_num_cached);
fs/ocfs2/uptodate.c
296
ci->ci_cache.ci_array[ci->ci_num_cached] = block;
fs/ocfs2/uptodate.c
306
sector_t block = new->c_block;
fs/ocfs2/uptodate.c
313
(unsigned long long)block, ci->ci_num_cached);
fs/ocfs2/uptodate.c
320
if (block < tmp->c_block)
fs/ocfs2/uptodate.c
322
else if (block > tmp->c_block)
fs/ocfs2/uptodate.c
327
(unsigned long long) block);
fs/ocfs2/uptodate.c
385
sector_t block,
fs/ocfs2/uptodate.c
395
(unsigned long long)block, expand_tree);
fs/ocfs2/uptodate.c
402
new->c_block = block;
fs/ocfs2/uptodate.c
423
ocfs2_append_cache_array(ci, block);
fs/ocfs2/uptodate.c
556
sector_t block)
fs/ocfs2/uptodate.c
564
(unsigned long long) block, ci->ci_num_cached,
fs/ocfs2/uptodate.c
568
index = ocfs2_search_cache_array(ci, block);
fs/ocfs2/uptodate.c
572
item = ocfs2_search_cache_tree(ci, block);
fs/ocfs2/uptodate.c
590
sector_t block = bh->b_blocknr;
fs/ocfs2/uptodate.c
592
ocfs2_remove_block_from_cache(ci, block);
fs/ocfs2/uptodate.c
597
sector_t block,
fs/ocfs2/uptodate.c
603
for (i = 0; i < b_len; i++, block++)
fs/ocfs2/uptodate.c
604
ocfs2_remove_block_from_cache(ci, block);
fs/ocfs2/uptodate.h
63
sector_t block,
fs/ocfs2/xattr.c
1757
int block, block_offset;
fs/ocfs2/xattr.c
1760
block = offset >> loc->xl_inode->i_sb->s_blocksize_bits;
fs/ocfs2/xattr.c
1763
return bucket_block(bucket, block) + block_offset;
fs/ocfs2/xattr.c
2514
u64 block,
fs/ocfs2/xattr.c
2528
ret = ocfs2_read_xattr_block(inode, block, &blk_bh);
fs/ocfs2/xattr.c
5132
u64 block;
fs/ocfs2/xattr.c
5161
block = ocfs2_clusters_to_blocks(osb->sb, bit_off);
fs/ocfs2/xattr.c
5162
trace_ocfs2_add_new_xattr_cluster((unsigned long long)block, num_bits);
fs/ocfs2/xattr.c
5164
if (bucket_blkno(first) + (prev_clusters * bpc) == block &&
fs/ocfs2/xattr.c
5183
block,
fs/ocfs2/xattr.c
5193
trace_ocfs2_add_new_xattr_cluster_insert((unsigned long long)block,
fs/ocfs2/xattr.c
5195
ret = ocfs2_insert_extent(handle, &et, v_start, block,
fs/ocfs2/xattr.c
817
u64 block;
fs/ocfs2/xattr.c
844
block = ocfs2_clusters_to_blocks(inode->i_sb, phys_cpos);
fs/ocfs2/xattr.c
846
block, alloc_size);
fs/omfs/bitmap.c
100
tmp = block;
fs/omfs/bitmap.c
175
int omfs_clear_range(struct super_block *sb, u64 block, int count)
fs/omfs/bitmap.c
183
tmp = block;
fs/omfs/bitmap.c
91
int omfs_allocate_block(struct super_block *sb, u64 block)
fs/omfs/dir.c
118
u64 block;
fs/omfs/dir.c
128
block = be64_to_cpu(*entry);
fs/omfs/dir.c
141
oi->i_sibling = cpu_to_be64(block);
fs/omfs/dir.c
165
u64 block, prev;
fs/omfs/dir.c
175
block = be64_to_cpu(*entry);
fs/omfs/dir.c
177
bh2 = omfs_scan_list(dir, block, name, namelen, &prev);
fs/omfs/dir.c
34
static struct buffer_head *omfs_scan_list(struct inode *dir, u64 block,
fs/omfs/dir.c
43
while (block != ~0) {
fs/omfs/dir.c
44
bh = omfs_bread(dir->i_sb, block);
fs/omfs/dir.c
51
if (omfs_is_bad(OMFS_SB(dir->i_sb), &oi->i_head, block)) {
fs/omfs/dir.c
59
*prev_block = block;
fs/omfs/dir.c
60
block = be64_to_cpu(oi->i_sibling);
fs/omfs/dir.c
72
u64 block, dummy;
fs/omfs/dir.c
78
block = be64_to_cpu(*((__be64 *) &bh->b_data[ofs]));
fs/omfs/dir.c
81
return omfs_scan_list(dir, block, name, namelen, &dummy);
fs/omfs/file.c
192
sector_t block, int count, int *left)
fs/omfs/file.c
200
if (block >= searched &&
fs/omfs/file.c
201
block < searched + numblocks) {
fs/omfs/file.c
206
*left = numblocks - (block - searched);
fs/omfs/file.c
209
block - searched;
fs/omfs/file.c
217
static int omfs_get_block(struct inode *inode, sector_t block,
fs/omfs/file.c
253
offset = find_block(inode, entry, block, extent_count, &remain);
fs/omfs/file.c
327
static sector_t omfs_bmap(struct address_space *mapping, sector_t block)
fs/omfs/file.c
329
return generic_block_bmap(mapping, block, omfs_get_block);
fs/omfs/inode.c
26
struct buffer_head *omfs_bread(struct super_block *sb, sector_t block)
fs/omfs/inode.c
29
if (block >= sbi->s_num_blocks)
fs/omfs/inode.c
32
return sb_bread(sb, clus_to_blk(sbi, block));
fs/omfs/inode.c
342
sector_t block;
fs/omfs/inode.c
355
block = clus_to_blk(sbi, sbi->s_bitmap_ino);
fs/omfs/inode.c
356
if (block >= sbi->s_num_blocks)
fs/omfs/inode.c
361
bh = sb_bread(sb, block++);
fs/omfs/omfs.h
30
static inline sector_t clus_to_blk(struct omfs_sb_info *sbi, sector_t block)
fs/omfs/omfs.h
32
return block << sbi->s_block_shift;
fs/omfs/omfs.h
42
extern int omfs_allocate_block(struct super_block *sb, u64 block);
fs/omfs/omfs.h
45
extern int omfs_clear_range(struct super_block *sb, u64 block, int count);
fs/omfs/omfs.h
62
extern struct buffer_head *omfs_bread(struct super_block *sb, sector_t block);
fs/omfs/omfs.h
65
extern int omfs_reserve_block(struct super_block *sb, sector_t block);
fs/pstore/ram_core.c
127
uint8_t *block;
fs/pstore/ram_core.c
136
block = buffer->data + (start & ~(ecc_block_size - 1));
fs/pstore/ram_core.c
140
if (block + ecc_block_size > buffer_end)
fs/pstore/ram_core.c
141
size = buffer_end - block;
fs/pstore/ram_core.c
142
persistent_ram_encode_rs8(prz, block, size, par);
fs/pstore/ram_core.c
143
block += ecc_block_size;
fs/pstore/ram_core.c
145
} while (block < buffer->data + start + count);
fs/pstore/ram_core.c
162
uint8_t *block;
fs/pstore/ram_core.c
168
block = buffer->data;
fs/pstore/ram_core.c
170
while (block < buffer->data + buffer_size(prz)) {
fs/pstore/ram_core.c
173
if (block + size > buffer->data + prz->buffer_size)
fs/pstore/ram_core.c
174
size = buffer->data + prz->buffer_size - block;
fs/pstore/ram_core.c
175
numerr = persistent_ram_decode_rs8(prz, block, size, par);
fs/pstore/ram_core.c
177
pr_devel("error in block %p, %d\n", block, numerr);
fs/pstore/ram_core.c
180
pr_devel("uncorrectable error in block %p\n", block);
fs/pstore/ram_core.c
183
block += prz->ecc_info.block_size;
fs/qnx4/inode.c
115
block = try_extent(&xblk->xblk_xtnts[ix], &offset);
fs/qnx4/inode.c
116
if (block) {
fs/qnx4/inode.c
131
QNX4DEBUG((KERN_INFO "qnx4: mapping block %ld of inode %ld = %ld\n",iblock,inode->i_ino,block));
fs/qnx4/inode.c
132
return block;
fs/qnx4/inode.c
272
static sector_t qnx4_bmap(struct address_space *mapping, sector_t block)
fs/qnx4/inode.c
274
return generic_block_bmap(mapping,block,qnx4_get_block);
fs/qnx4/inode.c
286
int block;
fs/qnx4/inode.c
307
block = ino / QNX4_INODES_PER_BLOCK;
fs/qnx4/inode.c
309
if (!(bh = sb_bread(sb, block))) {
fs/qnx4/inode.c
93
u32 block = try_extent(&qnx4_inode->di_first_xtnt, &offset);
fs/qnx4/inode.c
95
if (block) {
fs/qnx4/namei.c
53
unsigned long block, offset, blkofs;
fs/qnx4/namei.c
58
block = offset = blkofs = 0;
fs/qnx4/namei.c
61
block = qnx4_block_map(dir, blkofs);
fs/qnx4/namei.c
62
if (block)
fs/qnx4/namei.c
63
bh = sb_bread(dir->i_sb, block);
fs/qnx4/namei.c
71
*ino = block * QNX4_INODES_PER_BLOCK +
fs/qnx4/qnx4.h
31
extern int qnx4_is_free(struct super_block *sb, long block);
fs/qnx6/inode.c
117
unsigned block = 0;
fs/qnx6/inode.c
135
block = qnx6_get_devblock(s, ei->di_block_ptr[levelptr]);
fs/qnx6/inode.c
138
bh = sb_bread(s, block);
fs/qnx6/inode.c
140
pr_err("Error reading block (%u)\n", block);
fs/qnx6/inode.c
150
block = qnx6_get_devblock(s, ptr);
fs/qnx6/inode.c
153
return block;
fs/qnx6/inode.c
485
static sector_t qnx6_bmap(struct address_space *mapping, sector_t block)
fs/qnx6/inode.c
487
return generic_block_bmap(mapping, block, qnx6_get_block);
fs/qnx6/inode.c
65
static unsigned qnx6_get_devblock(struct super_block *sb, __fs32 block)
fs/qnx6/inode.c
68
return fs32_to_cpu(sbi, block) + sbi->s_blks_off;
fs/smb/common/cifs_md4.c
123
le32_to_cpu_array(ctx->block, ARRAY_SIZE(ctx->block));
fs/smb/common/cifs_md4.c
124
md4_transform(ctx->hash, ctx->block);
fs/smb/common/cifs_md4.c
142
const u32 avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f);
fs/smb/common/cifs_md4.c
147
memcpy((char *)mctx->block + (sizeof(mctx->block) - avail),
fs/smb/common/cifs_md4.c
152
memcpy((char *)mctx->block + (sizeof(mctx->block) - avail),
fs/smb/common/cifs_md4.c
159
while (len >= sizeof(mctx->block)) {
fs/smb/common/cifs_md4.c
160
memcpy(mctx->block, data, sizeof(mctx->block));
fs/smb/common/cifs_md4.c
162
data += sizeof(mctx->block);
fs/smb/common/cifs_md4.c
163
len -= sizeof(mctx->block);
fs/smb/common/cifs_md4.c
166
memcpy(mctx->block, data, len);
fs/smb/common/cifs_md4.c
175
char *p = (char *)mctx->block + offset;
fs/smb/common/cifs_md4.c
182
p = (char *)mctx->block;
fs/smb/common/cifs_md4.c
187
mctx->block[14] = mctx->byte_count << 3;
fs/smb/common/cifs_md4.c
188
mctx->block[15] = mctx->byte_count >> 29;
fs/smb/common/cifs_md4.c
189
le32_to_cpu_array(mctx->block, (sizeof(mctx->block) -
fs/smb/common/cifs_md4.c
191
md4_transform(mctx->hash, mctx->block);
fs/smb/common/md4.h
18
u32 block[MD4_BLOCK_WORDS];
fs/squashfs/block.c
222
const sector_t block = read_start >> msblk->devblksize_log2;
fs/squashfs/block.c
226
int total_len = (block_end - block) << msblk->devblksize_log2;
fs/squashfs/block.c
235
bio->bi_iter.bi_sector = block * (msblk->devblksize >> SECTOR_SHIFT);
fs/squashfs/cache.c
103
entry->block = block;
fs/squashfs/cache.c
110
entry->length = squashfs_read_data(sb, block, length,
fs/squashfs/cache.c
161
cache->name, i, entry->block, entry->refcount, entry->error);
fs/squashfs/cache.c
165
block);
fs/squashfs/cache.c
261
entry->block = SQUASHFS_INVALID_BLK;
fs/squashfs/cache.c
336
u64 *block, int *offset, int length)
fs/squashfs/cache.c
342
TRACE("Entered squashfs_read_metadata [%llx:%x]\n", *block, *offset);
fs/squashfs/cache.c
351
entry = squashfs_cache_get(sb, msblk->block_cache, *block, 0);
fs/squashfs/cache.c
367
*block = entry->next_index;
fs/squashfs/cache.c
413
void *squashfs_read_table(struct super_block *sb, u64 block, int length)
fs/squashfs/cache.c
439
res = squashfs_read_data(sb, block, length |
fs/squashfs/cache.c
53
struct squashfs_cache *cache, u64 block, int length)
fs/squashfs/cache.c
62
if (cache->entry[i].block == block) {
fs/squashfs/dir.c
102
u64 block = squashfs_i(inode)->start + msblk->directory_table;
fs/squashfs/dir.c
108
TRACE("Entered squashfs_readdir [%llx:%x]\n", block, offset);
fs/squashfs/dir.c
145
length = get_dir_index_using_offset(inode->i_sb, &block, &offset,
fs/squashfs/dir.c
155
err = squashfs_read_metadata(inode->i_sb, &dirh, &block,
fs/squashfs/dir.c
171
err = squashfs_read_metadata(inode->i_sb, dire, &block,
fs/squashfs/dir.c
183
&block, &offset, size);
fs/squashfs/dir.c
214
ERROR("Unable to read directory block [%llx:%x]\n", block, offset);
fs/squashfs/file.c
166
long long block = 0;
fs/squashfs/file.c
191
block += SQUASHFS_COMPRESSED_SIZE_BLOCK(size);
fs/squashfs/file.c
197
return block;
fs/squashfs/file.c
330
int *offset, u64 *block)
fs/squashfs/file.c
334
int res = fill_meta_index(inode, index, start, offset, block);
fs/squashfs/file.c
337
res, index, *start, *offset, block ? *block : 0);
fs/squashfs/file.c
352
if (block)
fs/squashfs/file.c
353
*block += blks;
fs/squashfs/file.c
366
static inline int read_blocklist(struct inode *inode, int index, u64 *block)
fs/squashfs/file.c
371
return read_blocklist_ptrs(inode, index, &start, &offset, block);
fs/squashfs/file.c
482
u64 block = 0;
fs/squashfs/file.c
484
res = read_blocklist(inode, index, &block);
fs/squashfs/file.c
491
res = squashfs_readpage_block(folio, block, res, expected);
fs/squashfs/file.c
597
u64 block = 0;
fs/squashfs/file.c
623
bsize = read_blocklist(inode, start >> msblk->block_log, &block);
fs/squashfs/file.c
632
res = squashfs_read_data(inode->i_sb, block, bsize, NULL, actor);
fs/squashfs/file_cache.c
21
int squashfs_readpage_block(struct folio *folio, u64 block, int bsize, int expected)
fs/squashfs/file_cache.c
25
block, bsize);
fs/squashfs/file_cache.c
29
ERROR("Unable to read page, block %llx, size %x\n", block,
fs/squashfs/file_direct.c
22
int squashfs_readpage_block(struct folio *folio, u64 block, int bsize,
fs/squashfs/file_direct.c
76
res = squashfs_read_data(inode->i_sb, block, bsize, NULL, actor);
fs/squashfs/fragment.c
39
int block, offset, size;
fs/squashfs/fragment.c
45
block = SQUASHFS_FRAGMENT_INDEX(fragment);
fs/squashfs/fragment.c
48
start_block = le64_to_cpu(msblk->fragment_index[block]);
fs/squashfs/id.c
36
int block = SQUASHFS_ID_BLOCK(index);
fs/squashfs/id.c
45
start_block = le64_to_cpu(msblk->id_table[block]);
fs/squashfs/inode.c
111
u64 block = SQUASHFS_INODE_BLK(ino) + msblk->inode_table;
fs/squashfs/inode.c
122
err = squashfs_read_metadata(sb, sqshb_ino, &block,
fs/squashfs/inode.c
131
block = SQUASHFS_INODE_BLK(ino) + msblk->inode_table;
fs/squashfs/inode.c
142
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
178
squashfs_i(inode)->block_list_start = block;
fs/squashfs/inode.c
185
offset, squashfs_i(inode)->start, block, offset);
fs/squashfs/inode.c
194
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
238
squashfs_i(inode)->block_list_start = block;
fs/squashfs/inode.c
245
offset, squashfs_i(inode)->start, block, offset);
fs/squashfs/inode.c
251
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
275
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
288
squashfs_i(inode)->dir_idx_start = block;
fs/squashfs/inode.c
303
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
319
squashfs_i(inode)->start = block;
fs/squashfs/inode.c
326
err = squashfs_read_metadata(sb, NULL, &block,
fs/squashfs/inode.c
330
err = squashfs_read_metadata(sb, &xattr, &block,
fs/squashfs/inode.c
339
block, offset);
fs/squashfs/inode.c
347
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
370
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
394
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
412
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/namei.c
129
u64 block = squashfs_i(dir)->start + msblk->directory_table;
fs/squashfs/namei.c
134
TRACE("Entered squashfs_lookup [%llx:%x]\n", block, offset);
fs/squashfs/namei.c
147
length = get_dir_index_using_name(dir->i_sb, &block, &offset,
fs/squashfs/namei.c
156
err = squashfs_read_metadata(dir->i_sb, &dirh, &block,
fs/squashfs/namei.c
172
err = squashfs_read_metadata(dir->i_sb, dire, &block,
fs/squashfs/namei.c
184
&block, &offset, size);
fs/squashfs/squashfs.h
80
int squashfs_readpage_block(struct folio *, u64 block, int bsize, int expected);
fs/squashfs/squashfs_fs_sb.h
30
u64 block;
fs/squashfs/symlink.c
39
u64 block = squashfs_i(inode)->start;
fs/squashfs/symlink.c
47
"%llx, offset %x\n", folio->index, block, offset);
fs/squashfs/symlink.c
53
bytes = squashfs_read_metadata(sb, NULL, &block, &offset,
fs/squashfs/symlink.c
72
entry = squashfs_cache_get(sb, msblk->block_cache, block, 0);
fs/squashfs/symlink.c
88
block = entry->next_index;
fs/squashfs/xattr_id.c
32
int block = SQUASHFS_XATTR_BLOCK(index);
fs/squashfs/xattr_id.c
41
start_block = le64_to_cpu(msblk->xattr_id_table[block]);
fs/ubifs/crypto.c
39
unsigned int in_len, unsigned int *out_len, int block)
fs/ubifs/crypto.c
54
offset_in_page(p), block);
fs/ubifs/crypto.c
65
unsigned int *out_len, int block)
fs/ubifs/crypto.c
80
block);
fs/ubifs/file.c
102
unsigned int block, beyond;
fs/ubifs/file.c
114
block = folio->index << UBIFS_BLOCKS_PER_PAGE_SHIFT;
fs/ubifs/file.c
116
if (block >= beyond) {
fs/ubifs/file.c
133
if (block >= beyond) {
fs/ubifs/file.c
138
ret = read_block(inode, folio, offset, block, dn);
fs/ubifs/file.c
143
} else if (block + 1 == beyond) {
fs/ubifs/file.c
153
block += 1;
fs/ubifs/file.c
46
unsigned int block, struct ubifs_data_node *dn)
fs/ubifs/file.c
53
data_key_init(c, &key, inode->i_ino, block);
fs/ubifs/file.c
71
err = ubifs_decrypt(inode, dn, &dlen, block);
fs/ubifs/file.c
886
unsigned int block;
fs/ubifs/file.c
901
block = folio->index << UBIFS_BLOCKS_PER_PAGE_SHIFT;
fs/ubifs/file.c
904
data_key_init(c, &key, inode->i_ino, block);
fs/ubifs/file.c
911
block += 1;
fs/ubifs/file.c
94
block, inode->i_ino);
fs/ubifs/journal.c
1624
unsigned int block, struct ubifs_data_node *dn,
fs/ubifs/journal.c
1640
err = ubifs_decrypt(inode, dn, &dlen, block);
fs/ubifs/journal.c
1656
err = ubifs_encrypt(inode, dn, out_len, &data_size, block);
fs/ubifs/key.h
268
unsigned int block)
fs/ubifs/key.h
270
ubifs_assert(c, !(block & ~UBIFS_S_KEY_BLOCK_MASK));
fs/ubifs/key.h
272
key->u32[1] = block | (UBIFS_DATA_KEY << UBIFS_S_KEY_BLOCK_BITS);
fs/ubifs/tnc.c
1555
unsigned int block = key_block(c, &bu->key);
fs/ubifs/tnc.c
1622
bu->blk_cnt += (next_block - block - 1);
fs/ubifs/tnc.c
1625
block = next_block;
fs/ubifs/tnc.c
1663
block = key_block(c, &bu->key) + bu->blk_cnt;
fs/ubifs/tnc.c
1664
block &= ~(UBIFS_BLOCKS_PER_PAGE - 1);
fs/ubifs/tnc.c
1666
if (key_block(c, &bu->zbranch[bu->cnt - 1].key) < block)
fs/ubifs/tnc.c
3528
unsigned int block;
fs/ubifs/tnc.c
3535
block = (size + UBIFS_BLOCK_SIZE - 1) >> UBIFS_BLOCK_SHIFT;
fs/ubifs/tnc.c
3536
data_key_init(c, &from_key, inode->i_ino, block);
fs/ubifs/tnc.c
3563
block = key_block(c, key);
fs/ubifs/tnc.c
3566
((loff_t)block) << UBIFS_BLOCK_SHIFT);
fs/ubifs/ubifs.h
2118
int block)
fs/ubifs/ubifs.h
2126
unsigned int *out_len, int block)
fs/ubifs/ubifs.h
2135
unsigned int in_len, unsigned int *out_len, int block);
fs/ubifs/ubifs.h
2137
unsigned int *out_len, int block);
fs/udf/balloc.c
127
unsigned long block;
fs/udf/balloc.c
136
block = bloc->logicalBlockNum + offset +
fs/udf/balloc.c
140
block_group = block >> (sb->s_blocksize_bits + 3);
fs/udf/balloc.c
141
bit = block % (sb->s_blocksize << 3);
fs/udf/balloc.c
165
block += count;
fs/udf/balloc.c
181
int bit, block, block_group;
fs/udf/balloc.c
195
block = first_block + (sizeof(struct spaceBitmapDesc) << 3);
fs/udf/balloc.c
196
block_group = block >> (sb->s_blocksize_bits + 3);
fs/udf/balloc.c
203
bit = block % (sb->s_blocksize << 3);
fs/udf/balloc.c
211
block++;
fs/udf/balloc.c
228
udf_pblk_t block;
fs/udf/balloc.c
243
block = goal + (sizeof(struct spaceBitmapDesc) << 3);
fs/udf/balloc.c
244
block_group = block >> (sb->s_blocksize_bits + 3);
fs/udf/balloc.c
255
bit = block % (sb->s_blocksize << 3);
fs/udf/balloc.c
32
struct udf_bitmap *bitmap, unsigned int block,
fs/udf/balloc.c
384
epos.block = oepos.block = iinfo->i_location;
fs/udf/balloc.c
43
bh = sb_bread(sb, udf_get_lb_pblock(sb, &loc, block));
fs/udf/balloc.c
433
oepos.block = epos.block;
fs/udf/balloc.c
518
epos.block = iinfo->i_location;
fs/udf/balloc.c
591
epos.block = iinfo->i_location;
fs/udf/balloc.c
616
goal_epos.block = epos.block;
fs/udf/balloc.c
718
udf_pblk_t block;
fs/udf/balloc.c
721
block = udf_bitmap_new_block(sb,
fs/udf/balloc.c
725
block = udf_table_new_block(sb,
fs/udf/balloc.c
732
if (inode && block)
fs/udf/balloc.c
734
return block;
fs/udf/directory.c
464
sector_t block;
fs/udf/directory.c
476
block = iinfo->i_lenExtents >> iter->dir->i_blkbits;
fs/udf/directory.c
477
bh = udf_bread(iter->dir, block, 1, &err);
fs/udf/directory.c
482
err = inode_bmap(iter->dir, block, &iter->epos, &iter->eloc, &iter->elen,
fs/udf/directory.c
487
(unsigned long long)block, iter->dir->i_ino);
fs/udf/ialloc.c
35
udf_pblk_t block;
fs/udf/ialloc.c
67
block = udf_new_block(dir->i_sb, NULL,
fs/udf/ialloc.c
85
iinfo->i_location.logicalBlockNum = block;
fs/udf/inode.c
1238
struct buffer_head *udf_bread(struct inode *inode, udf_pblk_t block,
fs/udf/inode.c
1243
.lblk = block,
fs/udf/inode.c
1955
unsigned long block = udf_get_lb_pblock(sb, ino, 0);
fs/udf/inode.c
1956
struct inode *inode = iget_locked(sb, block);
fs/udf/inode.c
1981
int udf_setup_indirect_aext(struct inode *inode, udf_pblk_t block,
fs/udf/inode.c
1999
neloc.logicalBlockNum = block;
fs/udf/inode.c
2000
neloc.partitionReferenceNum = epos->block.partitionReferenceNum;
fs/udf/inode.c
2014
cpu_to_le32(epos->block.logicalBlockNum);
fs/udf/inode.c
2021
udf_new_tag(bh->b_data, TAG_IDENT_AED, ver, 1, block,
fs/udf/inode.c
2024
nepos.block = neloc;
fs/udf/inode.c
2044
udf_write_aext(inode, epos, &nepos.block,
fs/udf/inode.c
2047
__udf_add_aext(inode, epos, &nepos.block,
fs/udf/inode.c
2132
epos->block.partitionReferenceNum,
fs/udf/inode.c
2133
epos->block.logicalBlockNum, &err);
fs/udf/inode.c
2212
udf_pblk_t block;
fs/udf/inode.c
2229
epos->block = *eloc;
fs/udf/inode.c
2232
block = udf_get_lb_pblock(inode->i_sb, &epos->block, 0);
fs/udf/inode.c
2233
epos->bh = sb_bread(inode->i_sb, block);
fs/udf/inode.c
2235
udf_debug("reading block %u failed!\n", block);
fs/udf/inode.c
2369
oepos.block = epos.block;
fs/udf/inode.c
2380
udf_free_blocks(inode->i_sb, inode, &epos.block, 0, 1);
fs/udf/inode.c
2426
int inode_bmap(struct inode *inode, sector_t block, struct extent_position *pos,
fs/udf/inode.c
2431
loff_t lbcount = 0, bcount = (loff_t) block << blocksize_bits;
fs/udf/inode.c
2438
pos->block = iinfo->i_location;
fs/udf/inode.c
315
static sector_t udf_bmap(struct address_space *mapping, sector_t block)
fs/udf/inode.c
321
return generic_block_bmap(mapping, block, udf_get_block);
fs/udf/inode.c
456
static int __udf_get_block(struct inode *inode, sector_t block,
fs/udf/inode.c
461
.lblk = block,
fs/udf/inode.c
476
int udf_get_block(struct inode *inode, sector_t block,
fs/udf/inode.c
489
return __udf_get_block(inode, block, bh_result, flags);
fs/udf/inode.c
498
static int udf_get_block_wb(struct inode *inode, sector_t block,
fs/udf/inode.c
501
return __udf_get_block(inode, block, bh_result, 0);
fs/udf/inode.c
67
static int udf_get_block_wb(struct inode *inode, sector_t block,
fs/udf/inode.c
740
prev_epos.block = iinfo->i_location;
fs/udf/inode.c
762
prev_epos.block = cur_epos.block;
fs/udf/inode.c
763
cur_epos.block = next_epos.block;
fs/udf/misc.c
188
struct buffer_head *udf_read_tagged(struct super_block *sb, uint32_t block,
fs/udf/misc.c
196
if (block == 0xFFFFFFFF)
fs/udf/misc.c
199
bh = sb_bread(sb, block);
fs/udf/misc.c
202
block, location);
fs/udf/misc.c
212
block, le32_to_cpu(tag_p->tagLocation), location);
fs/udf/misc.c
220
block, checksum, tag_p->tagChecksum);
fs/udf/misc.c
228
le16_to_cpu(tag_p->descVersion), block);
fs/udf/misc.c
239
udf_debug("Crc failure block %u: crc = %u, crclen = %u\n", block,
fs/udf/namei.c
133
static int udf_expand_dir_adinicb(struct inode *inode, udf_pblk_t *block)
fs/udf/namei.c
157
*block = udf_new_block(inode->i_sb, inode,
fs/udf/namei.c
160
if (!(*block))
fs/udf/namei.c
162
newblock = udf_get_pblock(inode->i_sb, *block,
fs/udf/namei.c
181
eloc.logicalBlockNum = *block;
fs/udf/namei.c
186
epos.block = iinfo->i_location;
fs/udf/namei.c
201
iter.fi.descTag.tagLocation = cpu_to_le32(*block);
fs/udf/namei.c
227
udf_pblk_t block;
fs/udf/namei.c
262
ret = udf_expand_dir_adinicb(dir, &block);
fs/udf/namei.c
272
block = dinfo->i_location.logicalBlockNum;
fs/udf/namei.c
274
block = iter->eloc.logicalBlockNum +
fs/udf/namei.c
292
block = iter->eloc.logicalBlockNum +
fs/udf/namei.c
297
udf_new_tag((char *)(&iter->fi), TAG_IDENT_FID, 3, 1, block,
fs/udf/namei.c
300
udf_new_tag((char *)(&iter->fi), TAG_IDENT_FID, 2, 1, block,
fs/udf/namei.c
580
udf_pblk_t block;
fs/udf/namei.c
608
block = udf_new_block(sb, inode,
fs/udf/namei.c
611
if (!block)
fs/udf/namei.c
613
epos.block = iinfo->i_location;
fs/udf/namei.c
616
eloc.logicalBlockNum = block;
fs/udf/namei.c
628
block = udf_get_pblock(sb, block,
fs/udf/namei.c
631
epos.bh = sb_getblk(sb, block);
fs/udf/namei.c
931
static struct dentry *udf_nfs_get_inode(struct super_block *sb, u32 block,
fs/udf/namei.c
937
if (block == 0)
fs/udf/namei.c
940
loc.logicalBlockNum = block;
fs/udf/namei.c
962
return udf_nfs_get_inode(sb, fid->udf.block, fid->udf.partref,
fs/udf/namei.c
993
fid->udf.block = location.logicalBlockNum;
fs/udf/partition.c
101
inline uint32_t udf_get_pblock_virt20(struct super_block *sb, uint32_t block,
fs/udf/partition.c
104
return udf_get_pblock_virt15(sb, block, partition, offset);
fs/udf/partition.c
107
uint32_t udf_get_pblock_spar15(struct super_block *sb, uint32_t block,
fs/udf/partition.c
119
packet = (block + offset) & ~(sdata->s_packet_len - 1);
fs/udf/partition.c
137
((block + offset) &
fs/udf/partition.c
144
return map->s_partition_root + block + offset;
fs/udf/partition.c
25
uint32_t udf_get_pblock(struct super_block *sb, uint32_t block,
fs/udf/partition.c
275
static uint32_t udf_try_read_meta(struct inode *inode, uint32_t block,
fs/udf/partition.c
288
err = inode_bmap(inode, block, &epos, &eloc, &elen, &ext_offset, &etype);
fs/udf/partition.c
303
uint32_t udf_get_pblock_meta25(struct super_block *sb, uint32_t block,
fs/udf/partition.c
32
block, partition, offset);
fs/udf/partition.c
321
retblk = udf_try_read_meta(inode, block, partition, offset);
fs/udf/partition.c
336
retblk = udf_try_read_meta(inode, block, partition, offset);
fs/udf/partition.c
37
return map->s_partition_func(sb, block, partition, offset);
fs/udf/partition.c
39
return map->s_partition_root + block + offset;
fs/udf/partition.c
42
uint32_t udf_get_pblock_virt15(struct super_block *sb, uint32_t block,
fs/udf/partition.c
58
if (block > vdata->s_num_entries) {
fs/udf/partition.c
60
block, vdata->s_num_entries);
fs/udf/partition.c
66
vdata->s_start_offset))[block]);
fs/udf/partition.c
70
if (block >= index) {
fs/udf/partition.c
71
block -= index;
fs/udf/partition.c
72
newblock = 1 + (block / (sb->s_blocksize / sizeof(uint32_t)));
fs/udf/partition.c
73
index = block % (sb->s_blocksize / sizeof(uint32_t));
fs/udf/partition.c
76
index = vdata->s_start_offset / sizeof(uint32_t) + block;
fs/udf/partition.c
82
sb, block, partition);
fs/udf/super.c
1280
static int udf_load_partdesc(struct super_block *sb, sector_t block)
fs/udf/super.c
1291
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
1432
static int udf_load_logicalvol(struct super_block *sb, sector_t block,
fs/udf/super.c
1445
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
1741
sector_t block, sector_t lastblock,
fs/udf/super.c
1768
for (; (!done && block <= lastblock); block++) {
fs/udf/super.c
1769
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
1788
block = le32_to_cpu(vdp->nextVolDescSeqExt.extLocation);
fs/udf/super.c
1792
lastblock += block - 1;
fs/udf/super.c
1794
block--;
fs/udf/super.c
1812
curr->block = block;
fs/udf/super.c
1825
if (!data.vds[VDS_POS_PRIMARY_VOL_DESC].block) {
fs/udf/super.c
1830
ret = udf_load_pvoldesc(sb, data.vds[VDS_POS_PRIMARY_VOL_DESC].block);
fs/udf/super.c
1834
if (data.vds[VDS_POS_LOGICAL_VOL_DESC].block) {
fs/udf/super.c
1836
data.vds[VDS_POS_LOGICAL_VOL_DESC].block,
fs/udf/super.c
1844
ret = udf_load_partdesc(sb, data.part_descs_loc[i].rec.block);
fs/udf/super.c
1903
static int udf_check_anchor_block(struct super_block *sb, sector_t block,
fs/udf/super.c
1910
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
2438
udf_pblk_t block = 0, newblock;
fs/udf/super.c
2470
newblock = udf_get_lb_pblock(sb, &loc, ++block);
fs/udf/super.c
2495
epos.block = UDF_I(table)->i_location;
fs/udf/super.c
890
static int udf_load_pvoldesc(struct super_block *sb, sector_t block)
fs/udf/super.c
903
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/truncate.c
140
epos.block = iinfo->i_location;
fs/udf/truncate.c
248
udf_free_blocks(sb, NULL, &epos.block,
fs/udf/truncate.c
258
epos.block = eloc;
fs/udf/truncate.c
283
udf_free_blocks(sb, NULL, &epos.block, 0, indirect_ext_len);
fs/udf/udf_i.h
8
struct kernel_lb_addr block;
fs/udf/udfdecl.h
104
uint32_t block;
fs/udf/udfdecl.h
155
extern struct buffer_head *udf_bread(struct inode *inode, udf_pblk_t block,
fs/udf/udfdecl.h
160
extern int inode_bmap(struct inode *inode, sector_t block,
fs/udf/udfdecl.h
164
extern int udf_setup_indirect_aext(struct inode *inode, udf_pblk_t block,
fs/ufs/inode.c
1004
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
1005
if (block) {
fs/ufs/inode.c
1010
free_full_branch(inode, block, depth);
fs/ufs/inode.c
1018
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
1019
if (block) {
fs/ufs/inode.c
1024
free_data(&ctx, block, uspi->s_fpb);
fs/ufs/inode.c
1113
u64 block;
fs/ufs/inode.c
1136
block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
1137
if (!block)
fs/ufs/inode.c
1139
ubh[i] = ubh_bread(sb, block, uspi->s_bsize);
fs/ufs/inode.c
1153
block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
1154
if (block) {
fs/ufs/inode.c
1158
free_full_branch(inode, block, i - UFS_IND_BLOCK + 1);
fs/ufs/inode.c
229
unsigned block = ufs_fragstoblks(lastfrag);
fs/ufs/inode.c
239
p = ufs_get_direct_data_ptr(uspi, ufsi, block);
fs/ufs/inode.c
504
static sector_t ufs_bmap(struct address_space *mapping, sector_t block)
fs/ufs/inode.c
506
return generic_block_bmap(mapping,block,ufs_getfrag_block);
fs/ufs/inode.c
975
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
976
if (block)
fs/ufs/inode.c
977
free_full_branch(inode, block, depth);
fs/ufs/inode.c
984
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
985
if (block)
fs/ufs/inode.c
986
free_data(&ctx, block, uspi->s_fpb);
fs/xfs/libxfs/xfs_ag.c
335
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_ag.c
353
be16_add_cpu(&block->bb_numrecs, 1);
fs/xfs/libxfs/xfs_ag.c
378
be16_add_cpu(&block->bb_numrecs, 1);
fs/xfs/libxfs/xfs_ag.c
403
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_ag.c
417
rrec = XFS_RMAP_REC_ADDR(block, 1);
fs/xfs/libxfs/xfs_ag.c
424
rrec = XFS_RMAP_REC_ADDR(block, 2);
fs/xfs/libxfs/xfs_ag.c
431
rrec = XFS_RMAP_REC_ADDR(block, 3);
fs/xfs/libxfs/xfs_ag.c
439
rrec = XFS_RMAP_REC_ADDR(block, 4);
fs/xfs/libxfs/xfs_ag.c
447
rrec = XFS_RMAP_REC_ADDR(block, 5);
fs/xfs/libxfs/xfs_ag.c
452
be16_add_cpu(&block->bb_numrecs, 1);
fs/xfs/libxfs/xfs_ag.c
457
rrec = XFS_RMAP_REC_ADDR(block,
fs/xfs/libxfs/xfs_ag.c
458
be16_to_cpu(block->bb_numrecs) + 1);
fs/xfs/libxfs/xfs_ag.c
464
be16_add_cpu(&block->bb_numrecs, 1);
fs/xfs/libxfs/xfs_alloc.c
480
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_alloc.c
484
block = xfs_btree_get_block(cnt_cur, 0, &bp);
fs/xfs/libxfs/xfs_alloc.c
486
xfs_btree_get_sibling(cnt_cur, block, &ptr, XFS_BB_RIGHTSIB);
fs/xfs/libxfs/xfs_alloc_btree.c
247
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_alloc_btree.c
252
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_alloc_btree.c
270
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_alloc_btree.h
31
#define XFS_ALLOC_REC_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_alloc_btree.h
33
((char *)(block) + \
fs/xfs/libxfs/xfs_alloc_btree.h
37
#define XFS_ALLOC_KEY_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_alloc_btree.h
39
((char *)(block) + \
fs/xfs/libxfs/xfs_alloc_btree.h
43
#define XFS_ALLOC_PTR_ADDR(mp, block, index, maxrecs) \
fs/xfs/libxfs/xfs_alloc_btree.h
45
((char *)(block) + \
fs/xfs/libxfs/xfs_bmap.c
1132
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_bmap.c
1140
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_bmap.c
1143
num_recs = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_bmap.c
1147
xfs_inode_verifier_error(ip, -EFSCORRUPTED, __func__, block,
fs/xfs/libxfs/xfs_bmap.c
1148
sizeof(*block), __this_address);
fs/xfs/libxfs/xfs_bmap.c
1154
frp = xfs_bmbt_rec_addr(mp, block, 1);
fs/xfs/libxfs/xfs_bmap.c
286
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_bmap.c
295
ASSERT(be16_to_cpu(block->bb_level) > 0);
fs/xfs/libxfs/xfs_bmap.c
298
for( i = 1; i <= xfs_btree_get_numrecs(block); i++) {
fs/xfs/libxfs/xfs_bmap.c
300
keyp = xfs_bmbt_key_addr(mp, block, i);
fs/xfs/libxfs/xfs_bmap.c
312
pp = xfs_bmap_broot_ptr_addr(mp, block, i, sz);
fs/xfs/libxfs/xfs_bmap.c
314
pp = xfs_bmbt_ptr_addr(mp, block, i, dmxr);
fs/xfs/libxfs/xfs_bmap.c
316
for (j = i+1; j <= be16_to_cpu(block->bb_numrecs); j++) {
fs/xfs/libxfs/xfs_bmap.c
318
thispa = xfs_bmap_broot_ptr_addr(mp, block, j, sz);
fs/xfs/libxfs/xfs_bmap.c
320
thispa = xfs_bmbt_ptr_addr(mp, block, j, dmxr);
fs/xfs/libxfs/xfs_bmap.c
348
struct xfs_btree_block *block; /* current btree block */
fs/xfs/libxfs/xfs_bmap.c
368
block = ifp->if_broot;
fs/xfs/libxfs/xfs_bmap.c
372
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_bmap.c
374
xfs_check_block(block, mp, 1, ifp->if_broot_bytes);
fs/xfs/libxfs/xfs_bmap.c
375
pp = xfs_bmap_broot_ptr_addr(mp, block, 1, ifp->if_broot_bytes);
fs/xfs/libxfs/xfs_bmap.c
398
block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_bmap.c
407
xfs_check_block(block, mp, 0, 0);
fs/xfs/libxfs/xfs_bmap.c
408
pp = xfs_bmbt_ptr_addr(mp, block, 1, mp->m_bmap_dmxr[1]);
fs/xfs/libxfs/xfs_bmap.c
434
num_recs = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_bmap.c
440
nextbno = be64_to_cpu(block->bb_u.l.bb_rightsib);
fs/xfs/libxfs/xfs_bmap.c
448
ep = xfs_bmbt_rec_addr(mp, block, 1);
fs/xfs/libxfs/xfs_bmap.c
455
nextp = xfs_bmbt_rec_addr(mp, block, j + 1);
fs/xfs/libxfs/xfs_bmap.c
485
block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_bmap.c
640
struct xfs_btree_block *block; /* btree root block */
fs/xfs/libxfs/xfs_bmap.c
660
block = xfs_bmap_broot_realloc(ip, whichfork, 1);
fs/xfs/libxfs/xfs_bmap.c
665
xfs_bmbt_init_block(ip, block, NULL, 1, 1);
fs/xfs/libxfs/xfs_bmap.c
725
kp = xfs_bmbt_key_addr(mp, block, 1);
fs/xfs/libxfs/xfs_bmap.c
728
pp = xfs_bmbt_ptr_addr(mp, block, 1, xfs_bmbt_get_maxrecs(cur,
fs/xfs/libxfs/xfs_bmap.c
729
be16_to_cpu(block->bb_level)));
fs/xfs/libxfs/xfs_bmap.c
889
struct xfs_btree_block *block = ip->i_df.if_broot;
fs/xfs/libxfs/xfs_bmap.c
897
if (xfs_bmap_bmdr_space(block) <= xfs_inode_data_fork_size(ip))
fs/xfs/libxfs/xfs_bmap_btree.c
399
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_bmap_btree.c
403
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_bmap_btree.c
423
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_bmap_btree.h
103
((char *)block + xfs_bmbt_block_len(mp) +
fs/xfs/libxfs/xfs_bmap_btree.h
112
struct xfs_bmdr_block *block,
fs/xfs/libxfs/xfs_bmap_btree.h
116
((char *)(block + 1) +
fs/xfs/libxfs/xfs_bmap_btree.h
122
struct xfs_bmdr_block *block,
fs/xfs/libxfs/xfs_bmap_btree.h
126
((char *)(block + 1) +
fs/xfs/libxfs/xfs_bmap_btree.h
132
struct xfs_bmdr_block *block,
fs/xfs/libxfs/xfs_bmap_btree.h
137
((char *)(block + 1) +
fs/xfs/libxfs/xfs_bmap_btree.h
76
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_bmap_btree.h
80
((char *)block + xfs_bmbt_block_len(mp) +
fs/xfs/libxfs/xfs_bmap_btree.h
87
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_bmap_btree.h
91
((char *)block + xfs_bmbt_block_len(mp) +
fs/xfs/libxfs/xfs_bmap_btree.h
98
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
1011
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
1015
xfs_agblock_t left = be32_to_cpu(block->bb_u.s.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
1016
xfs_agblock_t right = be32_to_cpu(block->bb_u.s.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
1046
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
1059
block = XFS_BUF_TO_BLOCK(cur->bc_levels[lev].bp);
fs/xfs/libxfs/xfs_btree.c
1063
return xfs_btree_readahead_agblock(cur, lr, block);
fs/xfs/libxfs/xfs_btree.c
1065
return xfs_btree_readahead_fsblock(cur, lr, block);
fs/xfs/libxfs/xfs_btree.c
1067
return xfs_btree_readahead_memblock(cur, lr, block);
fs/xfs/libxfs/xfs_btree.c
1192
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
1200
ptr->l = block->bb_u.l.bb_rightsib;
fs/xfs/libxfs/xfs_btree.c
1202
ptr->l = block->bb_u.l.bb_leftsib;
fs/xfs/libxfs/xfs_btree.c
1205
ptr->s = block->bb_u.s.bb_rightsib;
fs/xfs/libxfs/xfs_btree.c
1207
ptr->s = block->bb_u.s.bb_leftsib;
fs/xfs/libxfs/xfs_btree.c
1214
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
1222
block->bb_u.l.bb_rightsib = ptr->l;
fs/xfs/libxfs/xfs_btree.c
1224
block->bb_u.l.bb_leftsib = ptr->l;
fs/xfs/libxfs/xfs_btree.c
1227
block->bb_u.s.bb_rightsib = ptr->s;
fs/xfs/libxfs/xfs_btree.c
1229
block->bb_u.s.bb_leftsib = ptr->s;
fs/xfs/libxfs/xfs_btree.c
127
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
1276
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
1282
__xfs_btree_init_block(mp, block, ops, XFS_BUF_DADDR_NULL, level,
fs/xfs/libxfs/xfs_btree.c
134
if (!uuid_equal(&block->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
136
if (block->bb_u.l.bb_blkno !=
fs/xfs/libxfs/xfs_btree.c
1361
struct xfs_btree_block **block,
fs/xfs/libxfs/xfs_btree.c
1376
*block = XFS_BUF_TO_BLOCK(*bpp);
fs/xfs/libxfs/xfs_btree.c
1389
struct xfs_btree_block **block,
fs/xfs/libxfs/xfs_btree.c
139
if (block->bb_u.l.bb_pad != cpu_to_be32(0))
fs/xfs/libxfs/xfs_btree.c
1411
*block = XFS_BUF_TO_BLOCK(*bpp);
fs/xfs/libxfs/xfs_btree.c
143
if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops))
fs/xfs/libxfs/xfs_btree.c
145
if (be16_to_cpu(block->bb_level) != level)
fs/xfs/libxfs/xfs_btree.c
147
if (be16_to_cpu(block->bb_numrecs) >
fs/xfs/libxfs/xfs_btree.c
1570
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
1571
int level = xfs_btree_get_level(block);
fs/xfs/libxfs/xfs_btree.c
161
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
1662
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
1674
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
1677
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
1683
if (++cur->bc_levels[level].ptr <= xfs_btree_get_numrecs(block))
fs/xfs/libxfs/xfs_btree.c
1687
xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB);
fs/xfs/libxfs/xfs_btree.c
169
fa = __xfs_btree_check_lblock_hdr(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
1698
block = xfs_btree_get_block(cur, lev, &bp);
fs/xfs/libxfs/xfs_btree.c
1701
error = xfs_btree_check_block(cur, block, lev, bp);
fs/xfs/libxfs/xfs_btree.c
1706
if (++cur->bc_levels[lev].ptr <= xfs_btree_get_numrecs(block))
fs/xfs/libxfs/xfs_btree.c
1731
for (block = xfs_btree_get_block(cur, lev, &bp); lev > level; ) {
fs/xfs/libxfs/xfs_btree.c
1734
ptrp = xfs_btree_ptr_addr(cur, cur->bc_levels[lev].ptr, block);
fs/xfs/libxfs/xfs_btree.c
1736
error = xfs_btree_read_buf_block(cur, ptrp, 0, &block, &bp);
fs/xfs/libxfs/xfs_btree.c
1765
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
178
if (block->bb_u.l.bb_leftsib != cpu_to_be64(NULLFSBLOCK))
fs/xfs/libxfs/xfs_btree.c
1781
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
1784
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
1790
xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_LEFTSIB);
fs/xfs/libxfs/xfs_btree.c
180
if (block->bb_u.l.bb_rightsib != cpu_to_be64(NULLFSBLOCK))
fs/xfs/libxfs/xfs_btree.c
1825
for (block = xfs_btree_get_block(cur, lev, &bp); lev > level; ) {
fs/xfs/libxfs/xfs_btree.c
1828
ptrp = xfs_btree_ptr_addr(cur, cur->bc_levels[lev].ptr, block);
fs/xfs/libxfs/xfs_btree.c
1830
error = xfs_btree_read_buf_block(cur, ptrp, 0, &block, &bp);
fs/xfs/libxfs/xfs_btree.c
1834
cur->bc_levels[lev].ptr = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_btree.c
1855
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
1865
if (be64_to_cpu(block->bb_u.l.bb_owner) != owner)
fs/xfs/libxfs/xfs_btree.c
1868
if (be32_to_cpu(block->bb_u.s.bb_owner) != owner)
fs/xfs/libxfs/xfs_btree.c
187
block->bb_u.l.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
190
block->bb_u.l.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
1944
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
1949
xfs_btree_rec_addr(cur, keyno, block));
fs/xfs/libxfs/xfs_btree.c
1953
return xfs_btree_key_addr(cur, keyno, block);
fs/xfs/libxfs/xfs_btree.c
1987
struct xfs_btree_block *block; /* current btree block */
fs/xfs/libxfs/xfs_btree.c
2003
block = NULL;
fs/xfs/libxfs/xfs_btree.c
201
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
2018
error = xfs_btree_lookup_get_block(cur, level, pp, &block);
fs/xfs/libxfs/xfs_btree.c
2036
high = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_btree.c
2042
cur->bc_mp, block,
fs/xfs/libxfs/xfs_btree.c
2043
sizeof(*block));
fs/xfs/libxfs/xfs_btree.c
2065
keyno, block, &key);
fs/xfs/libxfs/xfs_btree.c
209
fa = __xfs_btree_check_lblock_hdr(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
2095
pp = xfs_btree_ptr_addr(cur, keyno, block);
fs/xfs/libxfs/xfs_btree.c
2112
xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB);
fs/xfs/libxfs/xfs_btree.c
2114
keyno > xfs_btree_get_numrecs(block) &&
fs/xfs/libxfs/xfs_btree.c
2134
if (keyno == 0 || keyno > xfs_btree_get_numrecs(block))
fs/xfs/libxfs/xfs_btree.c
215
block->bb_u.l.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
2161
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
2170
rec = xfs_btree_rec_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
2176
for (n = 2; n <= xfs_btree_get_numrecs(block); n++) {
fs/xfs/libxfs/xfs_btree.c
2177
rec = xfs_btree_rec_addr(cur, n, block);
fs/xfs/libxfs/xfs_btree.c
218
block->bb_u.l.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
2192
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
2201
memcpy(key, xfs_btree_key_addr(cur, 1, block),
fs/xfs/libxfs/xfs_btree.c
2204
max_hkey = xfs_btree_high_key_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
2205
for (n = 2; n <= xfs_btree_get_numrecs(block); n++) {
fs/xfs/libxfs/xfs_btree.c
2206
hkey = xfs_btree_high_key_addr(cur, n, block);
fs/xfs/libxfs/xfs_btree.c
2214
memcpy(key, xfs_btree_key_addr(cur, 1, block),
fs/xfs/libxfs/xfs_btree.c
2223
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
2226
if (be16_to_cpu(block->bb_level) == 0)
fs/xfs/libxfs/xfs_btree.c
2227
xfs_btree_get_leaf_keys(cur, block, key);
fs/xfs/libxfs/xfs_btree.c
2229
xfs_btree_get_node_keys(cur, block, key);
fs/xfs/libxfs/xfs_btree.c
2256
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
2278
xfs_btree_get_keys(cur, block, lkey);
fs/xfs/libxfs/xfs_btree.c
2283
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
2286
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
229
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
2291
nlkey = xfs_btree_key_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
2292
nhkey = xfs_btree_high_key_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
2301
xfs_btree_get_node_keys(cur, block, lkey);
fs/xfs/libxfs/xfs_btree.c
2314
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
2316
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
2317
return __xfs_btree_updkeys(cur, level, block, bp, true);
fs/xfs/libxfs/xfs_btree.c
2328
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
2336
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
2338
return __xfs_btree_updkeys(cur, level, block, bp, false);
fs/xfs/libxfs/xfs_btree.c
2346
xfs_btree_get_keys(cur, block, &key);
fs/xfs/libxfs/xfs_btree.c
2351
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
2353
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
2358
kp = xfs_btree_key_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
2376
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
2383
block = xfs_btree_get_block(cur, 0, &bp);
fs/xfs/libxfs/xfs_btree.c
2386
error = xfs_btree_check_block(cur, block, 0, bp);
fs/xfs/libxfs/xfs_btree.c
239
if (!uuid_equal(&block->bb_u.s.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
2392
rp = xfs_btree_rec_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
241
if (block->bb_u.s.bb_blkno != cpu_to_be64(xfs_buf_daddr(bp)))
fs/xfs/libxfs/xfs_btree.c
245
if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops))
fs/xfs/libxfs/xfs_btree.c
247
if (be16_to_cpu(block->bb_level) != level)
fs/xfs/libxfs/xfs_btree.c
249
if (be16_to_cpu(block->bb_numrecs) >
fs/xfs/libxfs/xfs_btree.c
255
block->bb_u.s.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
258
block->bb_u.s.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
270
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
276
return __xfs_btree_check_memblock(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
278
return __xfs_btree_check_agblock(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
280
return __xfs_btree_check_fsblock(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
300
struct xfs_btree_block *block, /* generic btree block pointer */
fs/xfs/libxfs/xfs_btree.c
307
fa = __xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
3096
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
3106
int numrecs = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_btree.c
3109
rp = xfs_btree_rec_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
3155
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.c
3167
int numrecs = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_btree.c
3174
be16_add_cpu(&block->bb_level, 1);
fs/xfs/libxfs/xfs_btree.c
3182
xfs_btree_set_numrecs(block, 1);
fs/xfs/libxfs/xfs_btree.c
3183
kp = xfs_btree_key_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
3188
pp = xfs_btree_ptr_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
3228
struct xfs_btree_block *block; /* btree block */
fs/xfs/libxfs/xfs_btree.c
3241
block = xfs_btree_get_iroot(cur);
fs/xfs/libxfs/xfs_btree.c
3244
aptr = *xfs_btree_ptr_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
3267
memcpy(cblock, block, xfs_btree_block_len(cur));
fs/xfs/libxfs/xfs_btree.c
3277
error = xfs_btree_promote_node_iroot(cur, block, level, cbp,
fs/xfs/libxfs/xfs_btree.c
3282
xfs_btree_promote_leaf_iroot(cur, block, cbp, &nptr, cblock);
fs/xfs/libxfs/xfs_btree.c
3315
struct xfs_btree_block *block; /* one half of the old root block */
fs/xfs/libxfs/xfs_btree.c
3355
block = xfs_btree_get_block(cur, cur->bc_nlevels - 1, &bp);
fs/xfs/libxfs/xfs_btree.c
3358
error = xfs_btree_check_block(cur, block, cur->bc_nlevels - 1, bp);
fs/xfs/libxfs/xfs_btree.c
3363
xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB);
fs/xfs/libxfs/xfs_btree.c
3368
left = block;
fs/xfs/libxfs/xfs_btree.c
3378
right = block;
fs/xfs/libxfs/xfs_btree.c
3516
struct xfs_btree_block *block; /* btree block */
fs/xfs/libxfs/xfs_btree.c
3556
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
3558
numrecs = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_btree.c
3561
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
3569
xfs_btree_rec_addr(cur, ptr, block)));
fs/xfs/libxfs/xfs_btree.c
3572
xfs_btree_key_addr(cur, ptr, block)));
fs/xfs/libxfs/xfs_btree.c
3593
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
3594
numrecs = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_btree.c
3597
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
3613
kp = xfs_btree_key_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
3614
pp = xfs_btree_ptr_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
3633
xfs_btree_set_numrecs(block, numrecs);
fs/xfs/libxfs/xfs_btree.c
3639
xfs_btree_key_addr(cur, ptr + 1, block)));
fs/xfs/libxfs/xfs_btree.c
3646
rp = xfs_btree_rec_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
3652
xfs_btree_set_numrecs(block, ++numrecs);
fs/xfs/libxfs/xfs_btree.c
3657
xfs_btree_rec_addr(cur, ptr + 1, block)));
fs/xfs/libxfs/xfs_btree.c
3691
xfs_btree_get_keys(cur, block, lkey);
fs/xfs/libxfs/xfs_btree.c
3853
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
3866
block = cur->bc_ops->broot_realloc(cur, numrecs);
fs/xfs/libxfs/xfs_btree.c
3868
xfs_btree_set_numrecs(block, numrecs);
fs/xfs/libxfs/xfs_btree.c
3869
ASSERT(block->bb_numrecs == cblock->bb_numrecs);
fs/xfs/libxfs/xfs_btree.c
3872
kp = xfs_btree_key_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
3877
pp = xfs_btree_ptr_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
3888
be16_add_cpu(&block->bb_level, -1);
fs/xfs/libxfs/xfs_btree.c
3906
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
3935
block = xfs_btree_get_iroot(cur);
fs/xfs/libxfs/xfs_btree.c
3936
if (xfs_btree_get_numrecs(block) != 1)
fs/xfs/libxfs/xfs_btree.c
3953
xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_LEFTSIB);
fs/xfs/libxfs/xfs_btree.c
3955
xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB);
fs/xfs/libxfs/xfs_btree.c
4039
struct xfs_btree_block *block; /* btree block */
fs/xfs/libxfs/xfs_btree.c
4068
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
4069
numrecs = xfs_btree_get_numrecs(block);
fs/xfs/libxfs/xfs_btree.c
4072
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
409
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
4092
lkp = xfs_btree_key_addr(cur, ptr + 1, block);
fs/xfs/libxfs/xfs_btree.c
4093
lpp = xfs_btree_ptr_addr(cur, ptr + 1, block);
fs/xfs/libxfs/xfs_btree.c
4111
xfs_btree_rec_addr(cur, ptr + 1, block),
fs/xfs/libxfs/xfs_btree.c
4120
xfs_btree_set_numrecs(block, --numrecs);
fs/xfs/libxfs/xfs_btree.c
415
block->bb_u.l.bb_lsn = cpu_to_be64(bip->bli_item.li_lsn);
fs/xfs/libxfs/xfs_btree.c
4153
pp = xfs_btree_ptr_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
4192
xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB);
fs/xfs/libxfs/xfs_btree.c
4193
xfs_btree_get_sibling(cur, block, &lptr, XFS_BB_LEFTSIB);
fs/xfs/libxfs/xfs_btree.c
423
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
427
if (!xfs_log_check_lsn(mp, be64_to_cpu(block->bb_u.l.bb_lsn)))
fs/xfs/libxfs/xfs_btree.c
4277
ASSERT(xfs_btree_get_numrecs(block) >=
fs/xfs/libxfs/xfs_btree.c
4362
ASSERT(xfs_btree_get_numrecs(block) >=
fs/xfs/libxfs/xfs_btree.c
4389
lrecs + xfs_btree_get_numrecs(block) <=
fs/xfs/libxfs/xfs_btree.c
4396
right = block;
fs/xfs/libxfs/xfs_btree.c
4406
rrecs + xfs_btree_get_numrecs(block) <=
fs/xfs/libxfs/xfs_btree.c
4413
left = block;
fs/xfs/libxfs/xfs_btree.c
447
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
453
block->bb_u.s.bb_lsn = cpu_to_be64(bip->bli_item.li_lsn);
fs/xfs/libxfs/xfs_btree.c
461
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
4613
struct xfs_btree_block *block; /* btree block */
fs/xfs/libxfs/xfs_btree.c
4621
block = xfs_btree_get_block(cur, 0, &bp);
fs/xfs/libxfs/xfs_btree.c
4624
error = xfs_btree_check_block(cur, block, 0, bp);
fs/xfs/libxfs/xfs_btree.c
4632
if (ptr > xfs_btree_get_numrecs(block) || ptr <= 0) {
fs/xfs/libxfs/xfs_btree.c
4640
*recp = xfs_btree_rec_addr(cur, ptr, block);
fs/xfs/libxfs/xfs_btree.c
465
if (!xfs_log_check_lsn(mp, be64_to_cpu(block->bb_u.s.bb_lsn)))
fs/xfs/libxfs/xfs_btree.c
4653
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
4660
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
4668
xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB);
fs/xfs/libxfs/xfs_btree.c
4684
return xfs_btree_lookup_get_block(cur, level, &rptr, &block);
fs/xfs/libxfs/xfs_btree.c
4698
struct xfs_btree_block *block = NULL;
fs/xfs/libxfs/xfs_btree.c
4706
error = xfs_btree_lookup_get_block(cur, level, &lptr, &block);
fs/xfs/libxfs/xfs_btree.c
4714
ptr = xfs_btree_ptr_addr(cur, 1, block);
fs/xfs/libxfs/xfs_btree.c
4774
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
4778
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
4780
if (block->bb_u.l.bb_owner == cpu_to_be64(bbcoi->new_owner))
fs/xfs/libxfs/xfs_btree.c
4782
block->bb_u.l.bb_owner = cpu_to_be64(bbcoi->new_owner);
fs/xfs/libxfs/xfs_btree.c
4784
if (block->bb_u.s.bb_owner == cpu_to_be32(bbcoi->new_owner))
fs/xfs/libxfs/xfs_btree.c
4786
block->bb_u.s.bb_owner = cpu_to_be32(bbcoi->new_owner);
fs/xfs/libxfs/xfs_btree.c
4836
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
4840
if (!uuid_equal(&block->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
4842
if (block->bb_u.l.bb_blkno != cpu_to_be64(xfs_buf_daddr(bp)))
fs/xfs/libxfs/xfs_btree.c
4845
be64_to_cpu(block->bb_u.l.bb_owner) != owner)
fs/xfs/libxfs/xfs_btree.c
4857
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
4864
if (be16_to_cpu(block->bb_numrecs) > max_recs)
fs/xfs/libxfs/xfs_btree.c
4870
block->bb_u.l.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
4873
block->bb_u.l.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
4883
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
4891
if (be16_to_cpu(block->bb_numrecs) > max_recs)
fs/xfs/libxfs/xfs_btree.c
4897
block->bb_u.l.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
4901
block->bb_u.l.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
4918
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
4923
if (!uuid_equal(&block->bb_u.s.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
4925
if (block->bb_u.s.bb_blkno != cpu_to_be64(xfs_buf_daddr(bp)))
fs/xfs/libxfs/xfs_btree.c
4927
if (pag && be32_to_cpu(block->bb_u.s.bb_owner) != pag_agno(pag))
fs/xfs/libxfs/xfs_btree.c
4944
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_btree.c
4951
if (be16_to_cpu(block->bb_numrecs) > max_recs)
fs/xfs/libxfs/xfs_btree.c
4957
block->bb_u.s.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
4960
block->bb_u.s.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
5149
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
5158
error = xfs_btree_lookup_get_block(cur, level, &ptr, &block);
fs/xfs/libxfs/xfs_btree.c
5164
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
5171
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
5175
be16_to_cpu(block->bb_numrecs)) {
fs/xfs/libxfs/xfs_btree.c
5186
block);
fs/xfs/libxfs/xfs_btree.c
5212
lkp = xfs_btree_key_addr(cur, cur->bc_levels[level].ptr, block);
fs/xfs/libxfs/xfs_btree.c
5214
block);
fs/xfs/libxfs/xfs_btree.c
5215
pp = xfs_btree_ptr_addr(cur, cur->bc_levels[level].ptr, block);
fs/xfs/libxfs/xfs_btree.c
5231
&block);
fs/xfs/libxfs/xfs_btree.c
5237
error = xfs_btree_check_block(cur, block, level, bp);
fs/xfs/libxfs/xfs_btree.c
5500
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.c
5503
block = xfs_btree_get_block(cur, 0, &bp);
fs/xfs/libxfs/xfs_btree.c
5506
if (cur->bc_levels[0].ptr < xfs_btree_get_numrecs(block))
fs/xfs/libxfs/xfs_btree.c
5511
return block->bb_u.l.bb_rightsib != cpu_to_be64(NULLFSBLOCK);
fs/xfs/libxfs/xfs_btree.c
5513
return block->bb_u.s.bb_rightsib != cpu_to_be32(NULLAGBLOCK);
fs/xfs/libxfs/xfs_btree.c
765
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
768
((char *)block + xfs_btree_rec_offset(cur, n));
fs/xfs/libxfs/xfs_btree.c
778
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
781
((char *)block + xfs_btree_key_offset(cur, n));
fs/xfs/libxfs/xfs_btree.c
791
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
794
((char *)block + xfs_btree_high_key_offset(cur, n));
fs/xfs/libxfs/xfs_btree.c
804
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
806
int level = xfs_btree_get_level(block);
fs/xfs/libxfs/xfs_btree.c
808
ASSERT(block->bb_level != 0);
fs/xfs/libxfs/xfs_btree.c
811
((char *)block + xfs_btree_ptr_offset(cur, n, level));
fs/xfs/libxfs/xfs_btree.c
868
struct xfs_btree_block *block; /* generic btree block pointer */
fs/xfs/libxfs/xfs_btree.c
874
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
875
if (xfs_btree_check_block(cur, block, level, bp))
fs/xfs/libxfs/xfs_btree.c
880
if (!block->bb_numrecs)
fs/xfs/libxfs/xfs_btree.c
898
struct xfs_btree_block *block; /* generic btree block pointer */
fs/xfs/libxfs/xfs_btree.c
904
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.c
905
if (xfs_btree_check_block(cur, block, level, bp))
fs/xfs/libxfs/xfs_btree.c
910
if (!block->bb_numrecs)
fs/xfs/libxfs/xfs_btree.c
915
cur->bc_levels[level].ptr = be16_to_cpu(block->bb_numrecs);
fs/xfs/libxfs/xfs_btree.c
959
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
962
xfs_fsblock_t left = be64_to_cpu(block->bb_u.l.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
963
xfs_fsblock_t right = be64_to_cpu(block->bb_u.l.bb_rightsib);
fs/xfs/libxfs/xfs_btree.c
985
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.c
988
xfbno_t left = be64_to_cpu(block->bb_u.l.bb_leftsib);
fs/xfs/libxfs/xfs_btree.c
989
xfbno_t right = be64_to_cpu(block->bb_u.l.bb_rightsib);
fs/xfs/libxfs/xfs_btree.h
347
struct xfs_btree_block *block, int level, struct xfs_buf *bp);
fs/xfs/libxfs/xfs_btree.h
357
struct xfs_btree_block *block, /* generic btree block pointer */
fs/xfs/libxfs/xfs_btree.h
431
static inline int xfs_btree_get_numrecs(const struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.h
433
return be16_to_cpu(block->bb_numrecs);
fs/xfs/libxfs/xfs_btree.h
436
static inline void xfs_btree_set_numrecs(struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.h
439
block->bb_numrecs = cpu_to_be16(numrecs);
fs/xfs/libxfs/xfs_btree.h
442
static inline int xfs_btree_get_level(const struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree.h
444
return be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_btree.h
509
struct xfs_btree_block *block);
fs/xfs/libxfs/xfs_btree.h
511
struct xfs_btree_block *block);
fs/xfs/libxfs/xfs_btree.h
513
struct xfs_btree_block *block);
fs/xfs/libxfs/xfs_btree.h
515
struct xfs_btree_block *block);
fs/xfs/libxfs/xfs_btree.h
526
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree.h
529
struct xfs_btree_block *block, union xfs_btree_key *key);
fs/xfs/libxfs/xfs_btree.h
637
struct xfs_btree_block *block;
fs/xfs/libxfs/xfs_btree.h
640
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/libxfs/xfs_btree.h
643
return block->bb_u.l.bb_rightsib == cpu_to_be64(NULLFSBLOCK);
fs/xfs/libxfs/xfs_btree.h
644
return block->bb_u.s.bb_rightsib == cpu_to_be32(NULLAGBLOCK);
fs/xfs/libxfs/xfs_btree.h
650
const union xfs_btree_ptr *ptr, struct xfs_btree_block **block,
fs/xfs/libxfs/xfs_btree.h
654
struct xfs_btree_block **block, struct xfs_buf **bpp);
fs/xfs/libxfs/xfs_btree.h
656
struct xfs_btree_block *block, const union xfs_btree_ptr *ptr,
fs/xfs/libxfs/xfs_btree_staging.c
360
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_btree_staging.c
368
ret = get_records(cur, j, block, recs_this_block - j + 1, priv);
fs/xfs/libxfs/xfs_btree_staging.c
390
struct xfs_btree_block *block)
fs/xfs/libxfs/xfs_btree_staging.c
415
block_ptr = xfs_btree_ptr_addr(cur, j, block);
fs/xfs/libxfs/xfs_btree_staging.c
418
block_key = xfs_btree_key_addr(cur, j, block);
fs/xfs/libxfs/xfs_btree_staging.c
683
struct xfs_btree_block *block = NULL;
fs/xfs/libxfs/xfs_btree_staging.c
718
nr_this_block, &ptr, &bp, &block, priv);
fs/xfs/libxfs/xfs_btree_staging.c
726
block, priv);
fs/xfs/libxfs/xfs_btree_staging.c
748
block = NULL;
fs/xfs/libxfs/xfs_btree_staging.c
763
&ptr, &bp, &block, priv);
fs/xfs/libxfs/xfs_btree_staging.c
771
&child_ptr, block);
fs/xfs/libxfs/xfs_btree_staging.h
50
unsigned int idx, struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_dir2_priv.h
171
struct xfs_dir2_data_hdr *block, struct xfs_dir2_sf_hdr *sfhp);
fs/xfs/libxfs/xfs_ialloc_btree.c
295
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_ialloc_btree.c
299
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_ialloc_btree.c
319
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_ialloc_btree.h
30
#define XFS_INOBT_REC_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_ialloc_btree.h
32
((char *)(block) + \
fs/xfs/libxfs/xfs_ialloc_btree.h
36
#define XFS_INOBT_KEY_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_ialloc_btree.h
38
((char *)(block) + \
fs/xfs/libxfs/xfs_ialloc_btree.h
42
#define XFS_INOBT_PTR_ADDR(mp, block, index, maxrecs) \
fs/xfs/libxfs/xfs_ialloc_btree.h
44
((char *)(block) + \
fs/xfs/libxfs/xfs_log_format.h
58
static inline xfs_lsn_t xlog_assign_lsn(uint cycle, uint block)
fs/xfs/libxfs/xfs_log_format.h
60
return ((xfs_lsn_t)cycle << 32) | block;
fs/xfs/libxfs/xfs_refcount_btree.c
209
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_refcount_btree.c
214
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_refcount_btree.c
223
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_refcount_btree.h
29
#define XFS_REFCOUNT_REC_ADDR(block, index) \
fs/xfs/libxfs/xfs_refcount_btree.h
31
((char *)(block) + \
fs/xfs/libxfs/xfs_refcount_btree.h
35
#define XFS_REFCOUNT_KEY_ADDR(block, index) \
fs/xfs/libxfs/xfs_refcount_btree.h
37
((char *)(block) + \
fs/xfs/libxfs/xfs_refcount_btree.h
41
#define XFS_REFCOUNT_PTR_ADDR(block, index, maxrecs) \
fs/xfs/libxfs/xfs_refcount_btree.h
43
((char *)(block) + \
fs/xfs/libxfs/xfs_rmap_btree.c
304
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_rmap_btree.c
321
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_rmap_btree.c
330
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_rmap_btree.c
551
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_rmap_btree.c
556
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_rmap_btree.c
563
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_rmap_btree.h
23
#define XFS_RMAP_REC_ADDR(block, index) \
fs/xfs/libxfs/xfs_rmap_btree.h
25
((char *)(block) + XFS_RMAP_BLOCK_LEN + \
fs/xfs/libxfs/xfs_rmap_btree.h
28
#define XFS_RMAP_KEY_ADDR(block, index) \
fs/xfs/libxfs/xfs_rmap_btree.h
30
((char *)(block) + XFS_RMAP_BLOCK_LEN + \
fs/xfs/libxfs/xfs_rmap_btree.h
33
#define XFS_RMAP_HIGH_KEY_ADDR(block, index) \
fs/xfs/libxfs/xfs_rmap_btree.h
35
((char *)(block) + XFS_RMAP_BLOCK_LEN + \
fs/xfs/libxfs/xfs_rmap_btree.h
39
#define XFS_RMAP_PTR_ADDR(block, index, maxrecs) \
fs/xfs/libxfs/xfs_rmap_btree.h
41
((char *)(block) + XFS_RMAP_BLOCK_LEN + \
fs/xfs/libxfs/xfs_rtbitmap.c
152
xfs_fileoff_t block, /* block number in bitmap or summary */
fs/xfs/libxfs/xfs_rtbitmap.c
183
if (*cbpp && *coffp == block)
fs/xfs/libxfs/xfs_rtbitmap.c
195
error = xfs_bmapi_read(ip, block, 1, &map, &nmap, 0);
fs/xfs/libxfs/xfs_rtbitmap.c
227
*coffp = block;
fs/xfs/libxfs/xfs_rtbitmap.c
234
xfs_fileoff_t block)
fs/xfs/libxfs/xfs_rtbitmap.c
238
if (XFS_IS_CORRUPT(mp, block >= mp->m_sb.sb_rbmblocks)) {
fs/xfs/libxfs/xfs_rtbitmap.c
243
return xfs_rtbuf_get(args, block, XFS_RTGI_BITMAP);
fs/xfs/libxfs/xfs_rtbitmap.c
249
xfs_fileoff_t block)
fs/xfs/libxfs/xfs_rtbitmap.c
253
if (XFS_IS_CORRUPT(mp, block >= mp->m_rsumblocks)) {
fs/xfs/libxfs/xfs_rtbitmap.c
257
return xfs_rtbuf_get(args, block, XFS_RTGI_SUMMARY);
fs/xfs/libxfs/xfs_rtbitmap.c
272
xfs_fileoff_t block; /* bitmap block number */
fs/xfs/libxfs/xfs_rtbitmap.c
286
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
287
error = xfs_rtbitmap_read_buf(args, block);
fs/xfs/libxfs/xfs_rtbitmap.c
336
error = xfs_rtbitmap_read_buf(args, --block);
fs/xfs/libxfs/xfs_rtbitmap.c
374
error = xfs_rtbitmap_read_buf(args, --block);
fs/xfs/libxfs/xfs_rtbitmap.c
426
xfs_fileoff_t block; /* bitmap block number */
fs/xfs/libxfs/xfs_rtbitmap.c
442
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
443
error = xfs_rtbitmap_read_buf(args, block);
fs/xfs/libxfs/xfs_rtbitmap.c
491
error = xfs_rtbitmap_read_buf(args, ++block);
fs/xfs/libxfs/xfs_rtbitmap.c
529
error = xfs_rtbitmap_read_buf(args, ++block);
fs/xfs/libxfs/xfs_rtbitmap.c
667
xfs_fileoff_t block; /* bitmap block number */
fs/xfs/libxfs/xfs_rtbitmap.c
679
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
683
error = xfs_rtbitmap_read_buf(args, block);
fs/xfs/libxfs/xfs_rtbitmap.c
726
error = xfs_rtbitmap_read_buf(args, ++block);
fs/xfs/libxfs/xfs_rtbitmap.c
758
error = xfs_rtbitmap_read_buf(args, ++block);
fs/xfs/libxfs/xfs_rtbitmap.c
881
xfs_fileoff_t block; /* bitmap block number */
fs/xfs/libxfs/xfs_rtbitmap.c
893
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
897
error = xfs_rtbitmap_read_buf(args, block);
fs/xfs/libxfs/xfs_rtbitmap.c
945
error = xfs_rtbitmap_read_buf(args, ++block);
fs/xfs/libxfs/xfs_rtbitmap.c
984
error = xfs_rtbitmap_read_buf(args, ++block);
fs/xfs/libxfs/xfs_rtbitmap.h
386
int xfs_rtbitmap_read_buf(struct xfs_rtalloc_args *args, xfs_fileoff_t block);
fs/xfs/libxfs/xfs_rtbitmap.h
387
int xfs_rtsummary_read_buf(struct xfs_rtalloc_args *args, xfs_fileoff_t block);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
191
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
195
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_rtrefcount_btree.c
203
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_rtrefcount_btree.h
104
((char *)(block + 1) +
fs/xfs/libxfs/xfs_rtrefcount_btree.h
37
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
41
((char *)block + XFS_RTREFCOUNT_BLOCK_LEN +
fs/xfs/libxfs/xfs_rtrefcount_btree.h
47
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
51
((char *)block + XFS_RTREFCOUNT_BLOCK_LEN +
fs/xfs/libxfs/xfs_rtrefcount_btree.h
57
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
62
((char *)block + XFS_RTREFCOUNT_BLOCK_LEN +
fs/xfs/libxfs/xfs_rtrefcount_btree.h
79
struct xfs_rtrefcount_root *block,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
83
((char *)(block + 1) +
fs/xfs/libxfs/xfs_rtrefcount_btree.h
89
struct xfs_rtrefcount_root *block,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
93
((char *)(block + 1) +
fs/xfs/libxfs/xfs_rtrefcount_btree.h
99
struct xfs_rtrefcount_root *block,
fs/xfs/libxfs/xfs_rtrmap_btree.c
1023
struct xfs_btree_block *block = rtg_rmap(rtg)->i_df.if_broot;
fs/xfs/libxfs/xfs_rtrmap_btree.c
1027
if (block->bb_numrecs == 0)
fs/xfs/libxfs/xfs_rtrmap_btree.c
1030
xfs_btree_get_keys(cur, block, &key);
fs/xfs/libxfs/xfs_rtrmap_btree.c
246
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_rtrmap_btree.c
250
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_rtrmap_btree.c
258
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_rtrmap_btree.c
536
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/libxfs/xfs_rtrmap_btree.c
541
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/libxfs/xfs_rtrmap_btree.c
548
level = be16_to_cpu(block->bb_level);
fs/xfs/libxfs/xfs_rtrmap_btree.h
100
struct xfs_rtrmap_root *block,
fs/xfs/libxfs/xfs_rtrmap_btree.h
104
((char *)(block + 1) +
fs/xfs/libxfs/xfs_rtrmap_btree.h
110
struct xfs_rtrmap_root *block,
fs/xfs/libxfs/xfs_rtrmap_btree.h
115
((char *)(block + 1) +
fs/xfs/libxfs/xfs_rtrmap_btree.h
38
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_rtrmap_btree.h
42
((char *)block + XFS_RTRMAP_BLOCK_LEN +
fs/xfs/libxfs/xfs_rtrmap_btree.h
48
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_rtrmap_btree.h
52
((char *)block + XFS_RTRMAP_BLOCK_LEN +
fs/xfs/libxfs/xfs_rtrmap_btree.h
58
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_rtrmap_btree.h
62
((char *)block + XFS_RTRMAP_BLOCK_LEN +
fs/xfs/libxfs/xfs_rtrmap_btree.h
69
struct xfs_btree_block *block,
fs/xfs/libxfs/xfs_rtrmap_btree.h
74
((char *)block + XFS_RTRMAP_BLOCK_LEN +
fs/xfs/libxfs/xfs_rtrmap_btree.h
90
struct xfs_rtrmap_root *block,
fs/xfs/libxfs/xfs_rtrmap_btree.h
94
((char *)(block + 1) +
fs/xfs/scrub/alloc_repair.c
606
struct xfs_btree_block *block,
fs/xfs/scrub/alloc_repair.c
624
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/scrub/bmap.c
532
struct xfs_btree_block *block;
fs/xfs/scrub/bmap.c
544
block = xfs_btree_get_block(bs->cur, i, &bp);
fs/xfs/scrub/bmap.c
545
owner = be64_to_cpu(block->bb_u.l.bb_owner);
fs/xfs/scrub/bmap_repair.c
563
struct xfs_btree_block *block,
fs/xfs/scrub/bmap_repair.c
584
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/scrub/btree.c
147
struct xfs_btree_block *block;
fs/xfs/scrub/btree.c
151
block = xfs_btree_get_block(cur, 0, &bp);
fs/xfs/scrub/btree.c
152
rec = xfs_btree_rec_addr(cur, cur->bc_levels[0].ptr, block);
fs/xfs/scrub/btree.c
195
struct xfs_btree_block *block;
fs/xfs/scrub/btree.c
199
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/scrub/btree.c
200
key = xfs_btree_key_addr(cur, cur->bc_levels[level].ptr, block);
fs/xfs/scrub/btree.c
224
key = xfs_btree_high_key_addr(cur, cur->bc_levels[level].ptr, block);
fs/xfs/scrub/btree.c
322
struct xfs_btree_block *block)
fs/xfs/scrub/btree.c
330
xfs_btree_get_sibling(cur, block, &leftsib, XFS_BB_LEFTSIB);
fs/xfs/scrub/btree.c
331
xfs_btree_get_sibling(cur, block, &rightsib, XFS_BB_RIGHTSIB);
fs/xfs/scrub/btree.c
332
level = xfs_btree_get_level(block);
fs/xfs/scrub/btree.c
498
struct xfs_btree_block *block)
fs/xfs/scrub/btree.c
502
unsigned int numrecs = be16_to_cpu(block->bb_numrecs);
fs/xfs/scrub/btree.c
546
struct xfs_btree_block *block)
fs/xfs/scrub/btree.c
558
xfs_btree_get_keys(cur, block, &block_key);
fs/xfs/scrub/btree.c
640
struct xfs_btree_block *block)
fs/xfs/scrub/btree.c
654
xfs_btree_get_keys(cur, block, &block_keys);
fs/xfs/scrub/btree.c
693
struct xfs_btree_block *block;
fs/xfs/scrub/btree.c
731
error = xchk_btree_get_block(bs, level, &ptr, &block, &bp);
fs/xfs/scrub/btree.c
732
if (error || !block)
fs/xfs/scrub/btree.c
738
block = xfs_btree_get_block(cur, level, &bp);
fs/xfs/scrub/btree.c
743
be16_to_cpu(block->bb_numrecs)) {
fs/xfs/scrub/btree.c
744
xchk_btree_block_keys(bs, level, block);
fs/xfs/scrub/btree.c
756
block);
fs/xfs/scrub/btree.c
770
be16_to_cpu(block->bb_numrecs)) {
fs/xfs/scrub/btree.c
771
xchk_btree_block_keys(bs, level, block);
fs/xfs/scrub/btree.c
782
pp = xfs_btree_ptr_addr(cur, cur->bc_levels[level].ptr, block);
fs/xfs/scrub/btree.c
788
error = xchk_btree_get_block(bs, level, pp, &block, &bp);
fs/xfs/scrub/btree.c
789
if (error || !block)
fs/xfs/scrub/ialloc_repair.c
522
struct xfs_btree_block *block,
fs/xfs/scrub/ialloc_repair.c
540
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/scrub/ialloc_repair.c
552
struct xfs_btree_block *block,
fs/xfs/scrub/ialloc_repair.c
567
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/scrub/rcbag_btree.c
127
struct xfs_btree_block *block = XFS_BUF_TO_BLOCK(bp);
fs/xfs/scrub/rcbag_btree.c
132
if (!xfs_verify_magic(bp, block->bb_magic))
fs/xfs/scrub/rcbag_btree.c
139
level = be16_to_cpu(block->bb_level);
fs/xfs/scrub/rcbag_btree.h
38
#define RCBAG_REC_ADDR(block, index) \
fs/xfs/scrub/rcbag_btree.h
40
((char *)(block) + RCBAG_BLOCK_LEN + \
fs/xfs/scrub/rcbag_btree.h
43
#define RCBAG_KEY_ADDR(block, index) \
fs/xfs/scrub/rcbag_btree.h
45
((char *)(block) + RCBAG_BLOCK_LEN + \
fs/xfs/scrub/rcbag_btree.h
48
#define RCBAG_PTR_ADDR(block, index, maxrecs) \
fs/xfs/scrub/rcbag_btree.h
50
((char *)(block) + RCBAG_BLOCK_LEN + \
fs/xfs/scrub/refcount_repair.c
515
struct xfs_btree_block *block,
fs/xfs/scrub/refcount_repair.c
531
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/scrub/rmap_repair.c
1220
struct xfs_btree_block *block,
fs/xfs/scrub/rmap_repair.c
1244
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/scrub/rtrefcount_repair.c
557
struct xfs_btree_block *block,
fs/xfs/scrub/rtrefcount_repair.c
572
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/scrub/rtrmap_repair.c
645
struct xfs_btree_block *block,
fs/xfs/scrub/rtrmap_repair.c
669
block_rec = xfs_btree_rec_addr(cur, idx, block);
fs/xfs/xfs_aops.c
724
sector_t block)
fs/xfs/xfs_aops.c
741
return iomap_bmap(mapping, block, &xfs_read_iomap_ops);
fs/xfs/xfs_log_priv.h
546
xlog_crack_atomic_lsn(atomic64_t *lsn, uint *cycle, uint *block)
fs/xfs/xfs_log_priv.h
551
*block = BLOCK_LSN(val);
fs/xfs/xfs_log_priv.h
558
xlog_assign_atomic_lsn(atomic64_t *lsn, uint cycle, uint block)
fs/xfs/xfs_log_priv.h
560
atomic64_set(lsn, xlog_assign_lsn(cycle, block));
fs/xfs/xfs_log_recover.c
1486
int block,
fs/xfs/xfs_log_recover.c
1497
recp->h_lsn = cpu_to_be64(xlog_assign_lsn(cycle, block));
fs/xfs/xfs_sysfs.c
409
int block;
fs/xfs/xfs_sysfs.c
414
block = log->l_curr_block;
fs/xfs/xfs_sysfs.c
417
return sysfs_emit(buf, "%d:%d\n", cycle, block);
fs/xfs/xfs_sysfs.c
427
int block;
fs/xfs/xfs_sysfs.c
430
xlog_crack_atomic_lsn(&log->l_tail_lsn, &cycle, &block);
fs/xfs/xfs_sysfs.c
431
return sysfs_emit(buf, "%d:%d\n", cycle, block);
fs/xfs/xfs_trace.h
4343
__field(xfs_daddr_t, block)
fs/xfs/xfs_trace.h
4352
__entry->block = fsmap->fmr_physical;
fs/xfs/xfs_trace.h
4361
__entry->block,
include/crypto/hash.h
184
#define HASH_STATE_AND_BLOCK(state, block) ((state) + (block) + 1)
include/crypto/internal/blockhash.h
47
#define BLOCK_HASH_UPDATE(block, state, src, nbytes, bs, buf, buflen) \
include/crypto/internal/blockhash.h
48
BLOCK_HASH_UPDATE_BASE(block, state, src, nbytes, bs, 1, buf, buflen)
include/crypto/internal/blockhash.h
49
#define BLOCK_HASH_UPDATE_BLOCKS(block, state, src, nbytes, bs, buf, buflen) \
include/crypto/internal/blockhash.h
50
BLOCK_HASH_UPDATE_BASE(block, state, src, nbytes, bs, bs, buf, buflen)
include/crypto/md5.h
28
u32 block[MD5_BLOCK_WORDS];
include/crypto/sm3_base.h
49
} block = {};
include/crypto/sm3_base.h
61
memcpy(&block, src, len);
include/crypto/sm3_base.h
62
block.u8[len] = 0x80;
include/crypto/sm3_base.h
64
block.b64[bit_offset] = cpu_to_be64(sctx->count << 3);
include/crypto/sm3_base.h
65
block_fn(sctx, block.u8, (bit_offset + 1) * 8 / SM3_BLOCK_SIZE);
include/crypto/sm3_base.h
66
memzero_explicit(&block, sizeof(block));
include/drm/drm_buddy.h
101
return block->header & DRM_BUDDY_HEADER_ORDER;
include/drm/drm_buddy.h
105
drm_buddy_block_state(struct drm_buddy_block *block)
include/drm/drm_buddy.h
107
return block->header & DRM_BUDDY_HEADER_STATE;
include/drm/drm_buddy.h
111
drm_buddy_block_is_allocated(struct drm_buddy_block *block)
include/drm/drm_buddy.h
113
return drm_buddy_block_state(block) == DRM_BUDDY_ALLOCATED;
include/drm/drm_buddy.h
117
drm_buddy_block_is_clear(struct drm_buddy_block *block)
include/drm/drm_buddy.h
119
return block->header & DRM_BUDDY_HEADER_CLEAR;
include/drm/drm_buddy.h
123
drm_buddy_block_is_free(struct drm_buddy_block *block)
include/drm/drm_buddy.h
125
return drm_buddy_block_state(block) == DRM_BUDDY_FREE;
include/drm/drm_buddy.h
129
drm_buddy_block_is_split(struct drm_buddy_block *block)
include/drm/drm_buddy.h
131
return drm_buddy_block_state(block) == DRM_BUDDY_SPLIT;
include/drm/drm_buddy.h
136
struct drm_buddy_block *block)
include/drm/drm_buddy.h
138
return mm->chunk_size << drm_buddy_block_order(block);
include/drm/drm_buddy.h
146
drm_get_buddy(struct drm_buddy_block *block);
include/drm/drm_buddy.h
161
void drm_buddy_free_block(struct drm_buddy *mm, struct drm_buddy_block *block);
include/drm/drm_buddy.h
169
struct drm_buddy_block *block,
include/drm/drm_buddy.h
93
drm_buddy_block_offset(const struct drm_buddy_block *block)
include/drm/drm_buddy.h
95
return block->header & DRM_BUDDY_HEADER_OFFSET;
include/drm/drm_buddy.h
99
drm_buddy_block_order(struct drm_buddy_block *block)
include/drm/drm_edid.h
476
int (*read_block)(void *context, u8 *buf, unsigned int block, size_t len),
include/drm/intel/display_parent_interface.h
54
void (*block)(struct drm_device *drm);
include/linux/ata.h
1033
static inline bool lba_28_ok(u64 block, u32 n_block)
include/linux/ata.h
1036
return ((block + n_block) < ((1 << 28) - 1)) && (n_block <= ATA_MAX_SECTORS);
include/linux/ata.h
1039
static inline bool lba_48_ok(u64 block, u32 n_block)
include/linux/ata.h
1042
return ((block + n_block - 1) < ((u64)1 << 48)) && (n_block <= ATA_MAX_SECTORS_LBA48);
include/linux/blkdev.h
1277
static inline int sb_issue_discard(struct super_block *sb, sector_t block,
include/linux/blkdev.h
1281
block << (sb->s_blocksize_bits -
include/linux/blkdev.h
1287
static inline int sb_issue_zeroout(struct super_block *sb, sector_t block,
include/linux/blkdev.h
1291
block << (sb->s_blocksize_bits -
include/linux/buffer_head.h
214
void clean_bdev_aliases(struct block_device *bdev, sector_t block,
include/linux/buffer_head.h
224
struct buffer_head *__find_get_block(struct block_device *bdev, sector_t block,
include/linux/buffer_head.h
227
sector_t block, unsigned size);
include/linux/buffer_head.h
228
struct buffer_head *bdev_getblk(struct block_device *bdev, sector_t block,
include/linux/buffer_head.h
232
void __breadahead(struct block_device *, sector_t block, unsigned int size);
include/linux/buffer_head.h
234
sector_t block, unsigned size, gfp_t gfp);
include/linux/buffer_head.h
344
sb_bread(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
346
return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, __GFP_MOVABLE);
include/linux/buffer_head.h
350
sb_bread_unmovable(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
352
return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, 0);
include/linux/buffer_head.h
356
sb_breadahead(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
358
__breadahead(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
362
sector_t block, unsigned size)
include/linux/buffer_head.h
369
return bdev_getblk(bdev, block, size, gfp);
include/linux/buffer_head.h
373
sector_t block, unsigned size)
include/linux/buffer_head.h
380
return bdev_getblk(bdev, block, size, gfp);
include/linux/buffer_head.h
384
sector_t block)
include/linux/buffer_head.h
386
return __getblk(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
390
sector_t block, gfp_t gfp)
include/linux/buffer_head.h
392
return bdev_getblk(sb->s_bdev, block, sb->s_blocksize, gfp);
include/linux/buffer_head.h
396
sb_find_get_block(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
398
return __find_get_block(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
402
sb_find_get_block_nonatomic(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
404
return __find_get_block_nonatomic(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
408
map_bh(struct buffer_head *bh, struct super_block *sb, sector_t block)
include/linux/buffer_head.h
412
bh->b_blocknr = block;
include/linux/buffer_head.h
485
sector_t block, unsigned size)
include/linux/buffer_head.h
487
return __bread_gfp(bdev, block, size, __GFP_MOVABLE);
include/linux/dm-bufio.h
140
int dm_bufio_issue_discard(struct dm_bufio_client *c, sector_t block, sector_t count);
include/linux/dm-bufio.h
147
void dm_bufio_forget(struct dm_bufio_client *c, sector_t block);
include/linux/dm-bufio.h
154
void dm_bufio_forget_buffers(struct dm_bufio_client *c, sector_t block, sector_t n_blocks);
include/linux/dm-bufio.h
64
void *dm_bufio_read(struct dm_bufio_client *c, sector_t block,
include/linux/dm-bufio.h
67
void *dm_bufio_read_with_ioprio(struct dm_bufio_client *c, sector_t block,
include/linux/dm-bufio.h
74
void *dm_bufio_get(struct dm_bufio_client *c, sector_t block,
include/linux/dm-bufio.h
81
void *dm_bufio_new(struct dm_bufio_client *c, sector_t block,
include/linux/dm-bufio.h
90
sector_t block, unsigned int n_blocks);
include/linux/dm-bufio.h
93
sector_t block, unsigned int n_blocks,
include/linux/exportfs.h
155
u32 block;
include/linux/fs.h
2659
extern int bmap(struct inode *inode, sector_t *block);
include/linux/fs.h
2661
static inline int bmap(struct inode *inode, sector_t *block)
include/linux/hfs_common.h
170
__be16 block;
include/linux/iio/buffer-dma.h
144
struct iio_dma_buffer_block *block);
include/linux/iio/buffer-dma.h
148
void iio_dma_buffer_block_done(struct iio_dma_buffer_block *block);
include/linux/iio/buffer-dma.h
173
struct iio_dma_buffer_block *block);
include/linux/iio/buffer-dma.h
175
struct iio_dma_buffer_block *block,
include/linux/iio/buffer_impl.h
90
struct iio_dma_buffer_block *block);
include/linux/iio/buffer_impl.h
92
struct iio_dma_buffer_block *block,
include/linux/jbd2.h
1293
int (*j_bmap)(struct journal_s *journal, sector_t *block);
include/linux/jbd2.h
1423
unsigned long *block);
include/linux/jbd2.h
1424
int __jbd2_update_log_tail(journal_t *journal, tid_t tid, unsigned long block);
include/linux/jbd2.h
1425
void jbd2_update_log_tail(journal_t *journal, tid_t tid, unsigned long block);
include/linux/lockd/lockd.h
217
void nlmclnt_prepare_block(struct nlm_wait *block, struct nlm_host *host,
include/linux/lockd/lockd.h
219
void nlmclnt_queue_block(struct nlm_wait *block);
include/linux/lockd/lockd.h
220
__be32 nlmclnt_dequeue_block(struct nlm_wait *block);
include/linux/lockd/lockd.h
221
int nlmclnt_wait(struct nlm_wait *block, struct nlm_rqst *req, long timeout);
include/linux/lockd/xdr.h
68
u32 block;
include/linux/mfd/stmpe.h
156
enum stmpe_block block);
include/linux/mfd/tc3589x.h
154
unsigned int block;
include/linux/mtd/blktrans.h
49
unsigned long block, char *buffer);
include/linux/mtd/blktrans.h
51
unsigned long block, char *buffer);
include/linux/mtd/blktrans.h
53
unsigned long block, unsigned nr_blocks);
include/linux/mtd/inftl.h
51
int INFTL_formatblock(struct INFTLrecord *s, int block);
include/linux/mtd/nand.h
1104
int nanddev_bbt_markbad(struct nand_device *nand, unsigned int block);
include/linux/mtd/nftl.h
42
int NFTL_formatblock(struct NFTLrecord *s, int block);
include/linux/mtd/onenand.h
227
loff_t onenand_addr(struct onenand_chip *this, int block);
include/linux/nfs_xdr.h
581
unsigned char block : 1;
include/linux/percpu-rwsem.h
150
return atomic_read(&sem->block);
include/linux/percpu-rwsem.h
18
atomic_t block;
include/linux/percpu-rwsem.h
37
.block = ATOMIC_INIT(0), \
include/linux/platform_data/cros_ec_commands.h
1704
uint8_t block[EC_VBNV_BLOCK_SIZE];
include/linux/platform_data/cros_ec_commands.h
1708
uint8_t block[EC_VBNV_BLOCK_SIZE];
include/linux/suspend.h
478
extern bool pm_get_wakeup_count(unsigned int *count, bool block);
include/linux/thunderbolt.h
151
struct tb_property_dir *tb_property_parse_dir(const u32 *block,
include/linux/thunderbolt.h
153
ssize_t tb_property_format_dir(const struct tb_property_dir *dir, u32 *block,
include/media/cec.h
372
bool block);
include/media/cec.h
374
bool block);
include/media/cec.h
380
bool block);
include/media/cec.h
483
bool block)
include/media/cec.h
547
unsigned int block;
include/media/cec.h
564
for (block = 1; block < blocks; block++) {
include/media/cec.h
565
unsigned int offset = block * 128;
include/net/flow_offload.h
596
struct flow_block *block;
include/net/flow_offload.h
644
struct flow_block_cb *flow_block_cb_lookup(struct flow_block *block,
include/net/mac80211.h
6611
struct ieee80211_sta *pubsta, bool block);
include/net/netfilter/nf_flow_table.h
253
struct flow_block *block = &flow_table->flow_block;
include/net/netfilter/nf_flow_table.h
258
block_cb = flow_block_cb_lookup(block, cb, cb_priv);
include/net/netfilter/nf_flow_table.h
270
list_add_tail(&block_cb->list, &block->cb_list);
include/net/netfilter/nf_flow_table.h
286
struct flow_block *block = &flow_table->flow_block;
include/net/netfilter/nf_flow_table.h
290
block_cb = flow_block_cb_lookup(block, cb, cb_priv);
include/net/pkt_cls.h
111
static inline bool tcf_block_shared(struct tcf_block *block)
include/net/pkt_cls.h
116
static inline bool tcf_block_non_null_shared(struct tcf_block *block)
include/net/pkt_cls.h
137
static inline void tcf_block_put(struct tcf_block *block)
include/net/pkt_cls.h
142
void tcf_block_put_ext(struct tcf_block *block, struct Qdisc *q,
include/net/pkt_cls.h
147
static inline struct Qdisc *tcf_block_q(struct tcf_block *block)
include/net/pkt_cls.h
153
const struct tcf_block *block,
include/net/pkt_cls.h
182
struct Qdisc *q = tp->chain->block->q;
include/net/pkt_cls.h
206
struct Qdisc *q = tp->chain->block->q;
include/net/pkt_cls.h
38
struct tcf_block *block;
include/net/pkt_cls.h
47
struct tcf_chain *tcf_chain_get_by_act(struct tcf_block *block,
include/net/pkt_cls.h
50
struct tcf_chain *tcf_get_next_chain(struct tcf_block *block,
include/net/pkt_cls.h
54
void tcf_block_netif_keep_dst(struct tcf_block *block);
include/net/pkt_cls.h
595
int tc_setup_cb_call(struct tcf_block *block, enum tc_setup_type type,
include/net/pkt_cls.h
597
int tc_setup_cb_add(struct tcf_block *block, struct tcf_proto *tp,
include/net/pkt_cls.h
600
int tc_setup_cb_replace(struct tcf_block *block, struct tcf_proto *tp,
include/net/pkt_cls.h
605
int tc_setup_cb_destroy(struct tcf_block *block, struct tcf_proto *tp,
include/net/pkt_cls.h
608
int tc_setup_cb_reoffload(struct tcf_block *block, struct tcf_proto *tp,
include/net/pkt_cls.h
61
void tcf_block_put(struct tcf_block *block);
include/net/pkt_cls.h
62
void tcf_block_put_ext(struct tcf_block *block, struct Qdisc *q,
include/net/pkt_cls.h
67
static inline bool tcf_block_shared(struct tcf_block *block)
include/net/pkt_cls.h
69
return block->index;
include/net/pkt_cls.h
72
static inline bool tcf_block_non_null_shared(struct tcf_block *block)
include/net/pkt_cls.h
74
return block && block->index;
include/net/pkt_cls.h
80
static inline bool tcf_block_bypass_sw(struct tcf_block *block)
include/net/pkt_cls.h
82
return block && !atomic_read(&block->useswcnt);
include/net/pkt_cls.h
86
static inline struct Qdisc *tcf_block_q(struct tcf_block *block)
include/net/pkt_cls.h
88
WARN_ON(tcf_block_shared(block));
include/net/pkt_cls.h
89
return block->q;
include/net/pkt_cls.h
93
const struct tcf_block *block,
include/net/sch_generic.h
1430
struct tcf_block *block;
include/net/sch_generic.h
1458
struct tcf_block *block);
include/net/sch_generic.h
471
struct tcf_block *block;
include/soc/mscc/ocelot.h
859
struct ocelot_vcap_block block[3];
include/soc/mscc/ocelot_vcap.h
730
ocelot_vcap_block_find_filter_by_id(struct ocelot_vcap_block *block,
include/sound/emux_synth.h
169
void *block; /* sample block pointer (optional) */
include/sound/gus.h
518
int snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block);
include/sound/gus.h
536
struct snd_gf1_dma_block * block,
include/sound/pcm_oss.h
14
block:1,
include/sound/soundfont.h
37
struct snd_util_memblk *block; /* allocated data block */
include/sound/tas2781.h
251
struct tasdev_blk *block);
include/sound/util_mem.h
28
struct list_head block; /* block linked-list header */
include/trace/events/afs.h
1044
unsigned int block,
include/trace/events/afs.h
1050
TP_ARGS(dvnode, why, op, block, slot, f_vnode, f_unique, name),
include/trace/events/afs.h
1057
__field(unsigned int, block)
include/trace/events/afs.h
1071
__entry->block = block;
include/trace/events/afs.h
1083
__entry->block, __entry->slot,
include/trace/events/bcache.h
268
__field(unsigned, block )
include/trace/events/bcache.h
274
__entry->block = b->written;
include/trace/events/bcache.h
279
__entry->bucket, __entry->block, __entry->keys)
include/trace/events/block.h
3
#define TRACE_SYSTEM block
include/trace/events/btrfs.h
2270
__field( u64, block )
include/trace/events/btrfs.h
2280
__entry->block = eb->start;
include/trace/events/btrfs.h
2291
__entry->block, __entry->generation,
include/trace/events/btrfs.h
2314
__field( u64, block )
include/trace/events/btrfs.h
2321
__entry->block = eb->start;
include/trace/events/btrfs.h
2328
__entry->block, __entry->generation,
include/trace/events/ext4.h
1221
TP_PROTO(struct inode *inode, int is_metadata, __u64 block),
include/trace/events/ext4.h
1223
TP_ARGS(inode, is_metadata, block),
include/trace/events/ext4.h
1228
__field( __u64, block )
include/trace/events/ext4.h
1236
__entry->block = block;
include/trace/events/ext4.h
1244
__entry->mode, __entry->is_metadata, __entry->block)
include/trace/events/ext4.h
2605
__field(u64, block)
include/trace/events/ext4.h
2613
__entry->block = fsmap->fmr_physical;
include/trace/events/ext4.h
2621
__entry->block,
include/trace/events/ext4.h
760
unsigned long long block, unsigned int count),
include/trace/events/ext4.h
762
TP_ARGS(pa, block, count),
include/trace/events/ext4.h
767
__field( __u64, block )
include/trace/events/ext4.h
775
__entry->block = block;
include/trace/events/ext4.h
782
__entry->block, __entry->count)
include/trace/events/ext4.h
893
TP_PROTO(struct ext4_allocation_request *ar, unsigned long long block),
include/trace/events/ext4.h
895
TP_ARGS(ar, block),
include/trace/events/ext4.h
900
__field( __u64, block )
include/trace/events/ext4.h
914
__entry->block = block;
include/trace/events/ext4.h
929
__entry->len, __entry->block, __entry->logical,
include/trace/events/ext4.h
935
TP_PROTO(struct inode *inode, __u64 block, unsigned long count,
include/trace/events/ext4.h
938
TP_ARGS(inode, block, count, flags),
include/trace/events/ext4.h
943
__field( __u64, block )
include/trace/events/ext4.h
952
__entry->block = block;
include/trace/events/ext4.h
961
__entry->mode, __entry->block, __entry->count,
include/trace/events/nilfs2.h
169
unsigned long block),
include/trace/events/nilfs2.h
171
TP_ARGS(inode, ino, block),
include/trace/events/nilfs2.h
176
__field(unsigned long, block)
include/trace/events/nilfs2.h
182
__entry->block = block;
include/trace/events/nilfs2.h
188
__entry->block)
include/trace/events/smbus.h
167
__entry->len = data->block[0] + 1;
include/trace/events/smbus.h
169
memcpy(__entry->buf, data->block, __entry->len);
include/trace/events/smbus.h
58
__entry->len = data->block[0] + 1;
include/trace/events/smbus.h
60
memcpy(__entry->buf, data->block, __entry->len);
include/uapi/linux/fuse.h
957
uint64_t block;
include/uapi/linux/fuse.h
963
uint64_t block;
include/uapi/linux/i2c.h
145
__u8 block[I2C_SMBUS_BLOCK_MAX + 2]; /* block[0] is used for length */
include/uapi/linux/rio_mport_cdev.h
203
__u64 block; /* Pointer to array of <count> transfers */
include/uapi/linux/videodev2.h
2164
__u8 block;
include/uapi/sound/asequencer.h
262
unsigned char block; /**< Block number (optional) */
kernel/locking/percpu-rwsem.c
200
return per_cpu_sum(*sem->read_count) != 0 && !atomic_read(&sem->block);
kernel/locking/percpu-rwsem.c
24
atomic_set(&sem->block, 0);
kernel/locking/percpu-rwsem.c
276
atomic_set_release(&sem->block, 0);
kernel/locking/percpu-rwsem.c
73
if (likely(!atomic_read_acquire(&sem->block)))
kernel/locking/percpu-rwsem.c
86
if (atomic_read(&sem->block))
kernel/locking/percpu-rwsem.c
89
return atomic_xchg(&sem->block, 1) == 0;
kernel/power/snapshot.c
471
struct rtree_node *node, *block, **dst;
kernel/power/snapshot.c
497
block = alloc_rtree_node(gfp_mask, safe_needed, ca, &zone->leaves);
kernel/power/snapshot.c
498
if (!block)
kernel/power/snapshot.c
523
*dst = block;
lib/crypto/mips/md5.h
46
const u64 *block = (const u64 *)data;
lib/crypto/mips/md5.h
48
write_octeon_64bit_block_dword(block[0], 0);
lib/crypto/mips/md5.h
49
write_octeon_64bit_block_dword(block[1], 1);
lib/crypto/mips/md5.h
50
write_octeon_64bit_block_dword(block[2], 2);
lib/crypto/mips/md5.h
51
write_octeon_64bit_block_dword(block[3], 3);
lib/crypto/mips/md5.h
52
write_octeon_64bit_block_dword(block[4], 4);
lib/crypto/mips/md5.h
53
write_octeon_64bit_block_dword(block[5], 5);
lib/crypto/mips/md5.h
54
write_octeon_64bit_block_dword(block[6], 6);
lib/crypto/mips/md5.h
55
octeon_md5_start(block[7]);
lib/crypto/mips/sha1.h
65
const u64 *block = (const u64 *)data;
lib/crypto/mips/sha1.h
67
write_octeon_64bit_block_dword(block[0], 0);
lib/crypto/mips/sha1.h
68
write_octeon_64bit_block_dword(block[1], 1);
lib/crypto/mips/sha1.h
69
write_octeon_64bit_block_dword(block[2], 2);
lib/crypto/mips/sha1.h
70
write_octeon_64bit_block_dword(block[3], 3);
lib/crypto/mips/sha1.h
71
write_octeon_64bit_block_dword(block[4], 4);
lib/crypto/mips/sha1.h
72
write_octeon_64bit_block_dword(block[5], 5);
lib/crypto/mips/sha1.h
73
write_octeon_64bit_block_dword(block[6], 6);
lib/crypto/mips/sha1.h
74
octeon_sha1_start(block[7]);
lib/crypto/mips/sha256.h
39
const u64 *block = (const u64 *)data;
lib/crypto/mips/sha256.h
41
write_octeon_64bit_block_dword(block[0], 0);
lib/crypto/mips/sha256.h
42
write_octeon_64bit_block_dword(block[1], 1);
lib/crypto/mips/sha256.h
43
write_octeon_64bit_block_dword(block[2], 2);
lib/crypto/mips/sha256.h
44
write_octeon_64bit_block_dword(block[3], 3);
lib/crypto/mips/sha256.h
45
write_octeon_64bit_block_dword(block[4], 4);
lib/crypto/mips/sha256.h
46
write_octeon_64bit_block_dword(block[5], 5);
lib/crypto/mips/sha256.h
47
write_octeon_64bit_block_dword(block[6], 6);
lib/crypto/mips/sha256.h
48
octeon_sha256_start(block[7]);
lib/crypto/mips/sha512.h
43
const u64 *block = (const u64 *)data;
lib/crypto/mips/sha512.h
45
write_octeon_64bit_block_sha512(block[0], 0);
lib/crypto/mips/sha512.h
46
write_octeon_64bit_block_sha512(block[1], 1);
lib/crypto/mips/sha512.h
47
write_octeon_64bit_block_sha512(block[2], 2);
lib/crypto/mips/sha512.h
48
write_octeon_64bit_block_sha512(block[3], 3);
lib/crypto/mips/sha512.h
49
write_octeon_64bit_block_sha512(block[4], 4);
lib/crypto/mips/sha512.h
50
write_octeon_64bit_block_sha512(block[5], 5);
lib/crypto/mips/sha512.h
51
write_octeon_64bit_block_sha512(block[6], 6);
lib/crypto/mips/sha512.h
52
write_octeon_64bit_block_sha512(block[7], 7);
lib/crypto/mips/sha512.h
53
write_octeon_64bit_block_sha512(block[8], 8);
lib/crypto/mips/sha512.h
54
write_octeon_64bit_block_sha512(block[9], 9);
lib/crypto/mips/sha512.h
55
write_octeon_64bit_block_sha512(block[10], 10);
lib/crypto/mips/sha512.h
56
write_octeon_64bit_block_sha512(block[11], 11);
lib/crypto/mips/sha512.h
57
write_octeon_64bit_block_sha512(block[12], 12);
lib/crypto/mips/sha512.h
58
write_octeon_64bit_block_sha512(block[13], 13);
lib/crypto/mips/sha512.h
59
write_octeon_64bit_block_sha512(block[14], 14);
lib/crypto/mips/sha512.h
60
octeon_sha512_start(block[15]);
lib/crypto/mldsa.c
138
u8 block[SHAKE128_BLOCK_SIZE + 1];
lib/crypto/mldsa.c
416
u8 block[SHAKE128_BLOCK_SIZE + 1])
lib/crypto/mldsa.c
422
shake_squeeze(shake, block, SHAKE128_BLOCK_SIZE);
lib/crypto/mldsa.c
423
block[SHAKE128_BLOCK_SIZE] = 0; /* for KMSAN */
lib/crypto/mldsa.c
426
u32 x = get_unaligned_le32(&block[j]) & 0x7fffff;
lib/crypto/mldsa.c
600
ws->block);
lib/test_bpf.c
2160
int block = 64; /* Increase for more tests per MSB position */
lib/test_bpf.c
2161
int len = 3 + 8 * 63 * block * 2;
lib/test_bpf.c
2173
for (adj = -block / 2; adj < block / 2; adj++) {
lib/tests/printf_kunit.c
30
#define NOWARN(option, comment, block) \
lib/tests/printf_kunit.c
33
block \
lib/xz/xz_dec_stream.c
231
s->block.compressed += b->in_pos - s->in_start;
lib/xz/xz_dec_stream.c
232
s->block.uncompressed += b->out_pos - s->out_start;
lib/xz/xz_dec_stream.c
238
if (s->block.compressed > s->block_header.compressed
lib/xz/xz_dec_stream.c
239
|| s->block.uncompressed
lib/xz/xz_dec_stream.c
250
!= s->block.compressed)
lib/xz/xz_dec_stream.c
255
!= s->block.uncompressed)
lib/xz/xz_dec_stream.c
258
s->block.hash.unpadded += s->block_header.size
lib/xz/xz_dec_stream.c
259
+ s->block.compressed;
lib/xz/xz_dec_stream.c
262
s->block.hash.unpadded += check_sizes[s->check_type];
lib/xz/xz_dec_stream.c
265
s->block.hash.unpadded += 4;
lib/xz/xz_dec_stream.c
268
s->block.hash.uncompressed += s->block.uncompressed;
lib/xz/xz_dec_stream.c
269
s->block.hash.crc32 = xz_crc32(
lib/xz/xz_dec_stream.c
270
(const uint8_t *)&s->block.hash,
lib/xz/xz_dec_stream.c
271
sizeof(s->block.hash), s->block.hash.crc32);
lib/xz/xz_dec_stream.c
273
++s->block.count;
lib/xz/xz_dec_stream.c
315
if (s->index.count != s->block.count)
lib/xz/xz_dec_stream.c
542
s->block.compressed = 0;
lib/xz/xz_dec_stream.c
543
s->block.uncompressed = 0;
lib/xz/xz_dec_stream.c
641
while (s->block.compressed & 3) {
lib/xz/xz_dec_stream.c
648
++s->block.compressed;
lib/xz/xz_dec_stream.c
695
if (!memeq(&s->block.hash, &s->index.hash,
lib/xz/xz_dec_stream.c
696
sizeof(s->block.hash)))
lib/xz/xz_dec_stream.c
821
memzero(&s->block, sizeof(s->block));
lib/xz/xz_dec_stream.c
93
} block;
lib/zstd/compress/zstd_compress.c
7295
BlockSummary const block = ZSTD_get1BlockSummary(inSeqs, nbSequences);
lib/zstd/compress/zstd_compress.c
7296
U32 const lastBlock = (block.nbSequences == nbSequences);
lib/zstd/compress/zstd_compress.c
7297
FORWARD_IF_ERROR(block.nbSequences, "Error while trying to determine nb of sequences for a block");
lib/zstd/compress/zstd_compress.c
7298
assert(block.nbSequences <= nbSequences);
lib/zstd/compress/zstd_compress.c
7299
RETURN_ERROR_IF(block.litSize > litSize, externalSequences_invalid, "discrepancy: Sequences require more literals than present in buffer");
lib/zstd/compress/zstd_compress.c
7303
inSeqs, block.nbSequences,
lib/zstd/compress/zstd_compress.c
7306
inSeqs += block.nbSequences;
lib/zstd/compress/zstd_compress.c
7307
nbSequences -= block.nbSequences;
lib/zstd/compress/zstd_compress.c
7308
remaining -= block.blockSize;
lib/zstd/compress/zstd_compress.c
7319
literals, block.litSize,
lib/zstd/compress/zstd_compress.c
7329
litSize -= block.litSize;
lib/zstd/compress/zstd_compress.c
7330
literals = (const char*)literals + block.litSize;
mm/dmapool.c
106
pool->name, block);
mm/dmapool.c
118
memset(block, POOL_POISON_ALLOCATED, pool->size);
mm/dmapool.c
136
struct dma_block *block = pool->next_block;
mm/dmapool.c
146
while (block) {
mm/dmapool.c
147
if (block != vaddr) {
mm/dmapool.c
148
block = block->next_block;
mm/dmapool.c
165
static void pool_check_block(struct dma_pool *pool, struct dma_block *block,
mm/dmapool.c
184
struct dma_block *block = pool->next_block;
mm/dmapool.c
186
if (block) {
mm/dmapool.c
187
pool->next_block = block->next_block;
mm/dmapool.c
190
return block;
mm/dmapool.c
193
static void pool_block_push(struct dma_pool *pool, struct dma_block *block,
mm/dmapool.c
196
block->dma = dma;
mm/dmapool.c
197
block->next_block = pool->next_block;
mm/dmapool.c
198
pool->next_block = block;
mm/dmapool.c
306
struct dma_block *block, *first = NULL, *last = NULL;
mm/dmapool.c
316
block = page->vaddr + offset;
mm/dmapool.c
317
block->dma = page->dma + offset;
mm/dmapool.c
318
block->next_block = NULL;
mm/dmapool.c
321
last->next_block = block;
mm/dmapool.c
323
first = block;
mm/dmapool.c
324
last = block;
mm/dmapool.c
410
struct dma_block *block;
mm/dmapool.c
417
block = pool_block_pop(pool);
mm/dmapool.c
418
if (!block) {
mm/dmapool.c
431
block = pool_block_pop(pool);
mm/dmapool.c
435
*handle = block->dma;
mm/dmapool.c
436
pool_check_block(pool, block, mem_flags);
mm/dmapool.c
438
memset(block, 0, pool->size);
mm/dmapool.c
440
return block;
mm/dmapool.c
455
struct dma_block *block = vaddr;
mm/dmapool.c
460
pool_block_push(pool, block, dma);
mm/dmapool.c
96
static void pool_check_block(struct dma_pool *pool, struct dma_block *block,
mm/dmapool.c
99
u8 *data = (void *)block;
mm/page_io.c
126
sector_t block;
mm/page_io.c
128
block = probe_block + block_in_page;
mm/page_io.c
129
ret = bmap(inode, &block);
mm/page_io.c
130
if (ret || !block)
mm/page_io.c
133
if (block != first_block + block_in_page) {
mm/percpu.c
1033
for (block = s_block + 1; block < e_block; block++) {
mm/percpu.c
1034
block->first_free = 0;
mm/percpu.c
1035
block->scan_hint = 0;
mm/percpu.c
1036
block->contig_hint_start = 0;
mm/percpu.c
1037
block->contig_hint = PCPU_BITMAP_BLOCK_BITS;
mm/percpu.c
1038
block->left_free = PCPU_BITMAP_BLOCK_BITS;
mm/percpu.c
1039
block->right_free = PCPU_BITMAP_BLOCK_BITS;
mm/percpu.c
1315
static void pcpu_init_md_block(struct pcpu_block_md *block, int nr_bits)
mm/percpu.c
1317
block->scan_hint = 0;
mm/percpu.c
1318
block->contig_hint = nr_bits;
mm/percpu.c
1319
block->left_free = nr_bits;
mm/percpu.c
1320
block->right_free = nr_bits;
mm/percpu.c
1321
block->first_free = 0;
mm/percpu.c
1322
block->nr_bits = nr_bits;
mm/percpu.c
2101
struct pcpu_block_md *block;
mm/percpu.c
2143
block = chunk->md_blocks + i;
mm/percpu.c
2144
if (block->contig_hint == PCPU_BITMAP_BLOCK_BITS &&
mm/percpu.c
317
static bool pcpu_check_block_hint(struct pcpu_block_md *block, int bits,
mm/percpu.c
320
int bit_off = ALIGN(block->contig_hint_start, align) -
mm/percpu.c
321
block->contig_hint_start;
mm/percpu.c
323
return bit_off + bits <= block->contig_hint;
mm/percpu.c
337
static int pcpu_next_hint(struct pcpu_block_md *block, int alloc_bits)
mm/percpu.c
346
if (block->scan_hint &&
mm/percpu.c
347
block->contig_hint_start > block->scan_hint_start &&
mm/percpu.c
348
alloc_bits > block->scan_hint)
mm/percpu.c
349
return block->scan_hint_start + block->scan_hint;
mm/percpu.c
351
return block->first_free;
mm/percpu.c
370
struct pcpu_block_md *block;
mm/percpu.c
373
for (block = chunk->md_blocks + i; i < pcpu_chunk_nr_blocks(chunk);
mm/percpu.c
374
block++, i++) {
mm/percpu.c
377
*bits += block->left_free;
mm/percpu.c
378
if (block->left_free == PCPU_BITMAP_BLOCK_BITS)
mm/percpu.c
391
*bits = block->contig_hint;
mm/percpu.c
392
if (*bits && block->contig_hint_start >= block_off &&
mm/percpu.c
393
*bits + block->contig_hint_start < PCPU_BITMAP_BLOCK_BITS) {
mm/percpu.c
395
block->contig_hint_start);
mm/percpu.c
401
*bits = block->right_free;
mm/percpu.c
402
*bit_off = (i + 1) * PCPU_BITMAP_BLOCK_BITS - block->right_free;
mm/percpu.c
425
struct pcpu_block_md *block;
mm/percpu.c
428
for (block = chunk->md_blocks + i; i < pcpu_chunk_nr_blocks(chunk);
mm/percpu.c
429
block++, i++) {
mm/percpu.c
432
*bits += block->left_free;
mm/percpu.c
435
if (block->left_free == PCPU_BITMAP_BLOCK_BITS)
mm/percpu.c
440
*bits = ALIGN(block->contig_hint_start, align) -
mm/percpu.c
441
block->contig_hint_start;
mm/percpu.c
446
if (block->contig_hint &&
mm/percpu.c
447
block->contig_hint_start >= block_off &&
mm/percpu.c
448
block->contig_hint >= *bits + alloc_bits) {
mm/percpu.c
449
int start = pcpu_next_hint(block, alloc_bits);
mm/percpu.c
451
*bits += alloc_bits + block->contig_hint_start -
mm/percpu.c
459
*bit_off = ALIGN(PCPU_BITMAP_BLOCK_BITS - block->right_free,
mm/percpu.c
630
static void pcpu_block_update(struct pcpu_block_md *block, int start, int end)
mm/percpu.c
634
block->first_free = min(block->first_free, start);
mm/percpu.c
636
block->left_free = contig;
mm/percpu.c
638
if (end == block->nr_bits)
mm/percpu.c
639
block->right_free = contig;
mm/percpu.c
641
if (contig > block->contig_hint) {
mm/percpu.c
643
if (start > block->contig_hint_start) {
mm/percpu.c
644
if (block->contig_hint > block->scan_hint) {
mm/percpu.c
645
block->scan_hint_start =
mm/percpu.c
646
block->contig_hint_start;
mm/percpu.c
647
block->scan_hint = block->contig_hint;
mm/percpu.c
648
} else if (start < block->scan_hint_start) {
mm/percpu.c
654
block->scan_hint = 0;
mm/percpu.c
657
block->scan_hint = 0;
mm/percpu.c
659
block->contig_hint_start = start;
mm/percpu.c
660
block->contig_hint = contig;
mm/percpu.c
661
} else if (contig == block->contig_hint) {
mm/percpu.c
662
if (block->contig_hint_start &&
mm/percpu.c
664
__ffs(start) > __ffs(block->contig_hint_start))) {
mm/percpu.c
666
block->contig_hint_start = start;
mm/percpu.c
667
if (start < block->scan_hint_start &&
mm/percpu.c
668
block->contig_hint > block->scan_hint)
mm/percpu.c
669
block->scan_hint = 0;
mm/percpu.c
670
} else if (start > block->scan_hint_start ||
mm/percpu.c
671
block->contig_hint > block->scan_hint) {
mm/percpu.c
677
block->scan_hint_start = start;
mm/percpu.c
678
block->scan_hint = contig;
mm/percpu.c
686
if ((start < block->contig_hint_start &&
mm/percpu.c
687
(contig > block->scan_hint ||
mm/percpu.c
688
(contig == block->scan_hint &&
mm/percpu.c
689
start > block->scan_hint_start)))) {
mm/percpu.c
690
block->scan_hint_start = start;
mm/percpu.c
691
block->scan_hint = contig;
mm/percpu.c
718
struct pcpu_block_md *block;
mm/percpu.c
724
block = chunk->md_blocks + s_index;
mm/percpu.c
730
pcpu_block_update(block, s_off, e_off);
mm/percpu.c
776
struct pcpu_block_md *block = chunk->md_blocks + index;
mm/percpu.c
781
if (block->scan_hint) {
mm/percpu.c
782
start = block->scan_hint_start + block->scan_hint;
mm/percpu.c
783
block->contig_hint_start = block->scan_hint_start;
mm/percpu.c
784
block->contig_hint = block->scan_hint;
mm/percpu.c
785
block->scan_hint = 0;
mm/percpu.c
787
start = block->first_free;
mm/percpu.c
788
block->contig_hint = 0;
mm/percpu.c
791
block->right_free = 0;
mm/percpu.c
795
pcpu_block_update(block, start, end);
mm/percpu.c
813
struct pcpu_block_md *s_block, *e_block, *block;
mm/percpu.c
908
for (block = s_block + 1; block < e_block; block++) {
mm/percpu.c
909
block->scan_hint = 0;
mm/percpu.c
910
block->contig_hint = 0;
mm/percpu.c
911
block->left_free = 0;
mm/percpu.c
912
block->right_free = 0;
mm/percpu.c
967
struct pcpu_block_md *s_block, *e_block, *block;
net/can/j1939/j1939-priv.h
285
unsigned int block;
net/can/j1939/transport.c
1457
else if (dat[1] > session->pkt.block /* 0xff for etp */)
net/can/j1939/transport.c
1660
session->pkt.block = 0xff;
net/can/j1939/transport.c
1667
session->pkt.block = min(dat[3], dat[4]);
net/can/j1939/transport.c
2058
session->pkt.block = skcb->addr.type == J1939_ETP ? 255 :
net/can/j1939/transport.c
928
len = min3(len, session->pkt.block, j1939_tp_block ?: 255);
net/core/dev.c
4396
if (tcf_block_bypass_sw(miniq->block))
net/core/dev.c
4404
ret = tcf_classify(skb, miniq->block, miniq->filter_list, &res, false);
net/core/flow_offload.c
287
struct flow_block_cb *flow_block_cb_lookup(struct flow_block *block,
net/core/flow_offload.c
292
list_for_each_entry(block_cb, &block->cb_list, list) {
net/core/flow_offload.c
362
block_cb = flow_block_cb_lookup(f->block, cb, cb_ident);
net/dsa/user.c
1709
block_cb = flow_block_cb_lookup(f->block, cb, dev);
net/hsr/hsr_framereg.c
310
static bool hsr_seq_block_is_old(struct hsr_seq_block *block)
net/hsr/hsr_framereg.c
314
return time_is_before_jiffies(block->time + expiry);
net/hsr/hsr_framereg.c
318
struct hsr_seq_block *block)
net/hsr/hsr_framereg.c
320
if (block->time)
net/hsr/hsr_framereg.c
321
xa_erase(&node->seq_blocks, block->block_idx);
net/hsr/hsr_framereg.c
322
block->time = 0;
net/hsr/hsr_framereg.c
335
struct hsr_seq_block *block, *res;
net/hsr/hsr_framereg.c
338
block = xa_load(&node->seq_blocks, block_idx);
net/hsr/hsr_framereg.c
340
if (block && hsr_seq_block_is_old(block)) {
net/hsr/hsr_framereg.c
341
hsr_forget_seq_block(node, block);
net/hsr/hsr_framereg.c
342
block = NULL;
net/hsr/hsr_framereg.c
345
if (!block) {
net/hsr/hsr_framereg.c
347
block = node->block_buf + node->next_block * block_sz;
net/hsr/hsr_framereg.c
348
hsr_forget_seq_block(node, block);
net/hsr/hsr_framereg.c
350
memset(block, 0, block_sz);
net/hsr/hsr_framereg.c
351
block->time = jiffies;
net/hsr/hsr_framereg.c
352
block->block_idx = block_idx;
net/hsr/hsr_framereg.c
354
res = xa_store(&node->seq_blocks, block_idx, block, GFP_ATOMIC);
net/hsr/hsr_framereg.c
356
block->time = 0;
net/hsr/hsr_framereg.c
364
return block;
net/hsr/hsr_framereg.c
591
struct hsr_seq_block *block;
net/hsr/hsr_framereg.c
603
block = hsr_get_seq_block(node, block_idx);
net/hsr/hsr_framereg.c
604
if (!block)
net/hsr/hsr_framereg.c
608
if (__test_and_set_bit(seq_bit, block->seq_nrs[port_type]))
net/hsr/hsr_framereg.c
818
struct hsr_seq_block *block;
net/hsr/hsr_framereg.c
828
block = node->block_buf + block_off * block_sz;
net/hsr/hsr_framereg.c
830
if (!bitmap_empty(block->seq_nrs[HSR_PT_SLAVE_B - 1],
net/hsr/hsr_framereg.c
832
seq_bit = find_last_bit(block->seq_nrs[HSR_PT_SLAVE_B - 1],
net/hsr/hsr_framereg.c
834
*if1_seq = (block->block_idx << HSR_SEQ_BLOCK_SHIFT) | seq_bit;
net/hsr/hsr_framereg.c
836
if (!bitmap_empty(block->seq_nrs[HSR_PT_SLAVE_A - 1],
net/hsr/hsr_framereg.c
838
seq_bit = find_last_bit(block->seq_nrs[HSR_PT_SLAVE_A - 1],
net/hsr/hsr_framereg.c
840
*if2_seq = (block->block_idx << HSR_SEQ_BLOCK_SHIFT) | seq_bit;
net/hsr/prp_dup_discard_test.c
112
block = hsr_get_seq_block(&data->node, block_idx);
net/hsr/prp_dup_discard_test.c
113
block->time = jiffies - msecs_to_jiffies(HSR_ENTRY_FORGET_TIME) - 1;
net/hsr/prp_dup_discard_test.c
45
struct hsr_seq_block *block;
net/hsr/prp_dup_discard_test.c
48
block = xa_load(&data->node.seq_blocks, block_idx);
net/hsr/prp_dup_discard_test.c
49
KUNIT_EXPECT_NOT_NULL(test, block);
net/hsr/prp_dup_discard_test.c
52
KUNIT_EXPECT_TRUE(test, test_bit(seq_bit, block->seq_nrs[0]));
net/hsr/prp_dup_discard_test.c
59
struct hsr_seq_block *block;
net/hsr/prp_dup_discard_test.c
62
block = hsr_get_seq_block(&data->node, block_idx);
net/hsr/prp_dup_discard_test.c
63
KUNIT_EXPECT_NOT_NULL(test, block);
net/hsr/prp_dup_discard_test.c
66
KUNIT_EXPECT_FALSE(test, test_bit(seq_bit, block->seq_nrs[0]));
net/hsr/prp_dup_discard_test.c
98
struct hsr_seq_block *block;
net/ipv6/addrconf.c
1359
static int ipv6_create_tempaddr(struct inet6_ifaddr *ifp, bool block)
net/ipv6/addrconf.c
1481
ift = ipv6_add_addr(idev, &cfg, block, NULL);
net/mac80211/michael.c
57
size_t block, blocks, left;
net/mac80211/michael.c
66
for (block = 0; block < blocks; block++)
net/mac80211/michael.c
67
michael_block(&mctx, get_unaligned_le32(&data[block * 4]));
net/mac80211/sta_info.c
2290
struct ieee80211_sta *pubsta, bool block)
net/mac80211/sta_info.c
2294
trace_api_sta_block_awake(sta->local, pubsta, block);
net/mac80211/sta_info.c
2296
if (block) {
net/mac80211/trace.h
2945
struct ieee80211_sta *sta, bool block),
net/mac80211/trace.h
2947
TP_ARGS(local, sta, block),
net/mac80211/trace.h
2952
__field(bool, block)
net/mac80211/trace.h
2958
__entry->block = block;
net/mac80211/trace.h
2963
LOCAL_PR_ARG, STA_PR_ARG, __entry->block
net/netfilter/nf_flow_table_offload.c
1198
bo->block = &flowtable->flow_block;
net/netfilter/nf_tables_offload.c
379
bo->block = &basechain->flow_block;
net/sched/act_api.c
94
*newchain = tcf_chain_get_by_act(tp->chain->block, chain_index);
net/sched/act_ct.c
378
struct flow_block *block;
net/sched/act_ct.c
384
block = &ct_ft->nf_ft.flow_block;
net/sched/act_ct.c
386
WARN_ON(!list_empty(&block->cb_list));
net/sched/act_mirred.c
338
struct tcf_block *block, int m_eaction,
net/sched/act_mirred.c
349
xa_for_each(&block->ports, index, dev) {
net/sched/act_mirred.c
372
struct tcf_block *block, int m_eaction,
net/sched/act_mirred.c
378
xa_for_each(&block->ports, index, dev) {
net/sched/act_mirred.c
395
struct tcf_block *block;
net/sched/act_mirred.c
405
block = tcf_block_lookup(dev_net(skb->dev), blockid);
net/sched/act_mirred.c
406
if (!block || xa_empty(&block->ports)) {
net/sched/act_mirred.c
412
return tcf_blockcast_redir(skb, m, block, m_eaction,
net/sched/act_mirred.c
416
return tcf_blockcast_mirror(skb, m, block, m_eaction, exception_ifindex,
net/sched/cls_api.c
1004
static void tcf_block_remove(struct tcf_block *block, struct net *net)
net/sched/cls_api.c
1009
idr_remove(&tn->idr, block->index);
net/sched/cls_api.c
1017
struct tcf_block *block;
net/sched/cls_api.c
1019
block = kzalloc_obj(*block);
net/sched/cls_api.c
1020
if (!block) {
net/sched/cls_api.c
1024
mutex_init(&block->lock);
net/sched/cls_api.c
1025
mutex_init(&block->proto_destroy_lock);
net/sched/cls_api.c
1026
init_rwsem(&block->cb_lock);
net/sched/cls_api.c
1027
flow_block_init(&block->flow_block);
net/sched/cls_api.c
1028
INIT_LIST_HEAD(&block->chain_list);
net/sched/cls_api.c
1029
INIT_LIST_HEAD(&block->owner_list);
net/sched/cls_api.c
1030
INIT_LIST_HEAD(&block->chain0.filter_chain_list);
net/sched/cls_api.c
1032
refcount_set(&block->refcnt, 1);
net/sched/cls_api.c
1033
block->net = net;
net/sched/cls_api.c
1034
block->index = block_index;
net/sched/cls_api.c
1035
xa_init(&block->ports);
net/sched/cls_api.c
1038
if (!tcf_block_shared(block))
net/sched/cls_api.c
1039
block->q = q;
net/sched/cls_api.c
1040
return block;
net/sched/cls_api.c
1053
struct tcf_block *block;
net/sched/cls_api.c
1056
block = tcf_block_lookup(net, block_index);
net/sched/cls_api.c
1057
if (block && !refcount_inc_not_zero(&block->refcnt))
net/sched/cls_api.c
1058
block = NULL;
net/sched/cls_api.c
1061
return block;
net/sched/cls_api.c
1065
__tcf_get_next_chain(struct tcf_block *block, struct tcf_chain *chain)
net/sched/cls_api.c
1067
mutex_lock(&block->lock);
net/sched/cls_api.c
1069
chain = list_is_last(&chain->list, &block->chain_list) ?
net/sched/cls_api.c
1072
chain = list_first_entry_or_null(&block->chain_list,
net/sched/cls_api.c
1077
chain = list_is_last(&chain->list, &block->chain_list) ?
net/sched/cls_api.c
1082
mutex_unlock(&block->lock);
net/sched/cls_api.c
1097
tcf_get_next_chain(struct tcf_block *block, struct tcf_chain *chain)
net/sched/cls_api.c
1099
struct tcf_chain *chain_next = __tcf_get_next_chain(block, chain);
net/sched/cls_api.c
1161
static void tcf_block_flush_all_chains(struct tcf_block *block, bool rtnl_held)
net/sched/cls_api.c
1168
for (chain = tcf_get_next_chain(block, NULL);
net/sched/cls_api.c
1170
chain = tcf_get_next_chain(block, chain)) {
net/sched/cls_api.c
1280
struct tcf_block *block;
net/sched/cls_api.c
1283
block = tcf_block_refcnt_get(net, block_index);
net/sched/cls_api.c
1284
if (!block) {
net/sched/cls_api.c
1291
block = cops->tcf_block(q, cl, extack);
net/sched/cls_api.c
1292
if (!block)
net/sched/cls_api.c
1295
if (tcf_block_shared(block)) {
net/sched/cls_api.c
1306
refcount_inc(&block->refcnt);
net/sched/cls_api.c
1309
return block;
net/sched/cls_api.c
1312
static void __tcf_block_put(struct tcf_block *block, struct Qdisc *q,
net/sched/cls_api.c
1315
if (refcount_dec_and_mutex_lock(&block->refcnt, &block->lock)) {
net/sched/cls_api.c
1322
bool free_block = list_empty(&block->chain_list);
net/sched/cls_api.c
1324
mutex_unlock(&block->lock);
net/sched/cls_api.c
1325
if (tcf_block_shared(block))
net/sched/cls_api.c
1326
tcf_block_remove(block, block->net);
net/sched/cls_api.c
1329
tcf_block_offload_unbind(block, q, ei);
net/sched/cls_api.c
1332
tcf_block_destroy(block);
net/sched/cls_api.c
1334
tcf_block_flush_all_chains(block, rtnl_held);
net/sched/cls_api.c
1336
tcf_block_offload_unbind(block, q, ei);
net/sched/cls_api.c
1340
static void tcf_block_refcnt_put(struct tcf_block *block, bool rtnl_held)
net/sched/cls_api.c
1342
__tcf_block_put(block, NULL, NULL, rtnl_held);
net/sched/cls_api.c
1354
struct tcf_block *block;
net/sched/cls_api.c
1367
block = __tcf_block_find(net, *q, *cl, ifindex, block_index, extack);
net/sched/cls_api.c
1368
if (IS_ERR(block)) {
net/sched/cls_api.c
1369
err = PTR_ERR(block);
net/sched/cls_api.c
1373
return block;
net/sched/cls_api.c
1383
static void tcf_block_release(struct Qdisc *q, struct tcf_block *block,
net/sched/cls_api.c
1386
if (!IS_ERR_OR_NULL(block))
net/sched/cls_api.c
1387
tcf_block_refcnt_put(block, rtnl_held);
net/sched/cls_api.c
1404
tcf_block_owner_netif_keep_dst(struct tcf_block *block,
net/sched/cls_api.c
1408
if (block->keep_dst &&
net/sched/cls_api.c
1414
void tcf_block_netif_keep_dst(struct tcf_block *block)
net/sched/cls_api.c
1418
block->keep_dst = true;
net/sched/cls_api.c
1419
list_for_each_entry(item, &block->owner_list, list)
net/sched/cls_api.c
1420
tcf_block_owner_netif_keep_dst(block, item->q,
net/sched/cls_api.c
1425
static int tcf_block_owner_add(struct tcf_block *block,
net/sched/cls_api.c
1436
list_add(&item->list, &block->owner_list);
net/sched/cls_api.c
1440
static void tcf_block_owner_del(struct tcf_block *block,
net/sched/cls_api.c
1446
list_for_each_entry(item, &block->owner_list, list) {
net/sched/cls_api.c
1456
static bool tcf_block_tracks_dev(struct tcf_block *block,
net/sched/cls_api.c
1459
return tcf_block_shared(block) &&
net/sched/cls_api.c
1470
struct tcf_block *block = NULL;
net/sched/cls_api.c
1475
block = tcf_block_refcnt_get(net, ei->block_index);
net/sched/cls_api.c
1477
if (!block) {
net/sched/cls_api.c
1478
block = tcf_block_create(net, q, ei->block_index, extack);
net/sched/cls_api.c
1479
if (IS_ERR(block))
net/sched/cls_api.c
1480
return PTR_ERR(block);
net/sched/cls_api.c
1481
if (tcf_block_shared(block)) {
net/sched/cls_api.c
1482
err = tcf_block_insert(block, net, extack);
net/sched/cls_api.c
1488
err = tcf_block_owner_add(block, q, ei->binder_type);
net/sched/cls_api.c
1492
tcf_block_owner_netif_keep_dst(block, q, ei->binder_type);
net/sched/cls_api.c
1494
err = tcf_chain0_head_change_cb_add(block, ei, extack);
net/sched/cls_api.c
1498
err = tcf_block_offload_bind(block, q, ei, extack);
net/sched/cls_api.c
1502
if (tcf_block_tracks_dev(block, ei)) {
net/sched/cls_api.c
1503
err = xa_insert(&block->ports, dev->ifindex, dev, GFP_KERNEL);
net/sched/cls_api.c
1510
*p_block = block;
net/sched/cls_api.c
1514
tcf_block_offload_unbind(block, q, ei);
net/sched/cls_api.c
1516
tcf_chain0_head_change_cb_del(block, ei);
net/sched/cls_api.c
1518
tcf_block_owner_del(block, q, ei->binder_type);
net/sched/cls_api.c
1521
tcf_block_refcnt_put(block, true);
net/sched/cls_api.c
1550
void tcf_block_put_ext(struct tcf_block *block, struct Qdisc *q,
net/sched/cls_api.c
1555
if (!block)
net/sched/cls_api.c
1557
if (tcf_block_tracks_dev(block, ei))
net/sched/cls_api.c
1558
xa_erase(&block->ports, dev->ifindex);
net/sched/cls_api.c
1559
tcf_chain0_head_change_cb_del(block, ei);
net/sched/cls_api.c
1560
tcf_block_owner_del(block, q, ei->binder_type);
net/sched/cls_api.c
1562
__tcf_block_put(block, q, ei, true);
net/sched/cls_api.c
1566
void tcf_block_put(struct tcf_block *block)
net/sched/cls_api.c
1570
if (!block)
net/sched/cls_api.c
1572
tcf_block_put_ext(block, block->q, &ei);
net/sched/cls_api.c
1578
tcf_block_playback_offloads(struct tcf_block *block, flow_setup_cb_t *cb,
net/sched/cls_api.c
1586
lockdep_assert_held(&block->cb_lock);
net/sched/cls_api.c
1588
for (chain = __tcf_get_next_chain(block, NULL);
net/sched/cls_api.c
1591
chain = __tcf_get_next_chain(block, chain),
net/sched/cls_api.c
1621
tcf_block_playback_offloads(block, cb, cb_priv, false, offload_in_use,
net/sched/cls_api.c
1626
static int tcf_block_bind(struct tcf_block *block,
net/sched/cls_api.c
1632
lockdep_assert_held(&block->cb_lock);
net/sched/cls_api.c
1635
err = tcf_block_playback_offloads(block, block_cb->cb,
net/sched/cls_api.c
1637
tcf_block_offload_in_use(block),
net/sched/cls_api.c
1642
block->lockeddevcnt++;
net/sched/cls_api.c
1646
list_splice(&bo->cb_list, &block->flow_block.cb_list);
net/sched/cls_api.c
1655
tcf_block_playback_offloads(block, block_cb->cb,
net/sched/cls_api.c
1657
tcf_block_offload_in_use(block),
net/sched/cls_api.c
1660
block->lockeddevcnt--;
net/sched/cls_api.c
1668
static void tcf_block_unbind(struct tcf_block *block,
net/sched/cls_api.c
1673
lockdep_assert_held(&block->cb_lock);
net/sched/cls_api.c
1676
tcf_block_playback_offloads(block, block_cb->cb,
net/sched/cls_api.c
1678
tcf_block_offload_in_use(block),
net/sched/cls_api.c
1683
block->lockeddevcnt--;
net/sched/cls_api.c
1687
static int tcf_block_setup(struct tcf_block *block,
net/sched/cls_api.c
1694
err = tcf_block_bind(block, bo);
net/sched/cls_api.c
1698
tcf_block_unbind(block, bo);
net/sched/cls_api.c
1792
tp->chain->block->index,
net/sched/cls_api.c
1806
const struct tcf_block *block,
net/sched/cls_api.c
181
struct tcf_block *block = chain->block;
net/sched/cls_api.c
1823
if (block) {
net/sched/cls_api.c
183
mutex_lock(&block->proto_destroy_lock);
net/sched/cls_api.c
184
hash_add_rcu(block->proto_destroy_ht, &tp->destroy_ht_node,
net/sched/cls_api.c
1844
fchain = tcf_chain_lookup_rcu(block, chain);
net/sched/cls_api.c
186
mutex_unlock(&block->proto_destroy_lock);
net/sched/cls_api.c
205
hash_for_each_possible_rcu(chain->block->proto_destroy_ht, iter,
net/sched/cls_api.c
2051
struct tcf_proto *tp, struct tcf_block *block,
net/sched/cls_api.c
2074
tcm->tcm_block_index = block->index;
net/sched/cls_api.c
2117
struct tcf_block *block,
net/sched/cls_api.c
2132
ret = tcf_fill_node(net, skb, tp, block, q, parent, fh, portid,
net/sched/cls_api.c
2148
struct tcf_block *block, struct Qdisc *q,
net/sched/cls_api.c
2159
skb = tfilter_notify_prep(net, oskb, n, tp, block, q, parent, fh, event,
net/sched/cls_api.c
2174
struct tcf_block *block, struct Qdisc *q,
net/sched/cls_api.c
2185
skb = tfilter_notify_prep(net, oskb, n, tp, block, q, parent, fh,
net/sched/cls_api.c
220
struct tcf_block *block = chain->block;
net/sched/cls_api.c
2207
struct tcf_block *block, struct Qdisc *q,
net/sched/cls_api.c
2216
tfilter_notify(net, oskb, n, tp, block, q, parent, NULL,
net/sched/cls_api.c
222
mutex_lock(&block->proto_destroy_lock);
net/sched/cls_api.c
2231
static bool is_ingress_or_clsact(struct tcf_block *block, struct Qdisc *q)
net/sched/cls_api.c
2233
return tcf_block_shared(block) || (q && !!(q->flags & TCQ_F_INGRESS));
net/sched/cls_api.c
225
mutex_unlock(&block->proto_destroy_lock);
net/sched/cls_api.c
2251
struct tcf_block *block;
net/sched/cls_api.c
2275
block = NULL;
net/sched/cls_api.c
2320
block = __tcf_block_find(net, q, cl, t->tcm_ifindex, t->tcm_block_index,
net/sched/cls_api.c
2322
if (IS_ERR(block)) {
net/sched/cls_api.c
2323
err = PTR_ERR(block);
net/sched/cls_api.c
2326
block->classid = parent;
net/sched/cls_api.c
2334
chain = tcf_chain_get(block, chain_index, true);
net/sched/cls_api.c
2428
if (is_ingress_or_clsact(block, q))
net/sched/cls_api.c
2433
tfilter_notify(net, skb, n, tp, block, q, parent, fh,
net/sched/cls_api.c
2452
tcf_block_release(q, block, rtnl_held);
net/sched/cls_api.c
2486
struct tcf_block *block = NULL;
net/sched/cls_api.c
2534
block = __tcf_block_find(net, q, cl, t->tcm_ifindex, t->tcm_block_index,
net/sched/cls_api.c
2536
if (IS_ERR(block)) {
net/sched/cls_api.c
2537
err = PTR_ERR(block);
net/sched/cls_api.c
2547
chain = tcf_chain_get(block, chain_index, false);
net/sched/cls_api.c
2562
tfilter_notify_chain(net, skb, block, q, parent, n,
net/sched/cls_api.c
2589
tfilter_notify(net, skb, n, tp, block, q, parent, fh,
net/sched/cls_api.c
2604
err = tfilter_del_notify(net, skb, n, tp, block, q, parent, fh,
net/sched/cls_api.c
2619
tcf_block_release(q, block, rtnl_held);
net/sched/cls_api.c
2645
struct tcf_block *block = NULL;
net/sched/cls_api.c
2692
block = __tcf_block_find(net, q, cl, t->tcm_ifindex, t->tcm_block_index,
net/sched/cls_api.c
2694
if (IS_ERR(block)) {
net/sched/cls_api.c
2695
err = PTR_ERR(block);
net/sched/cls_api.c
2705
chain = tcf_chain_get(block, chain_index, false);
net/sched/cls_api.c
2735
err = tfilter_notify(net, skb, n, tp, block, q, parent,
net/sched/cls_api.c
2748
tcf_block_release(q, block, rtnl_held);
net/sched/cls_api.c
2760
struct tcf_block *block;
net/sched/cls_api.c
2771
return tcf_fill_node(net, a->skb, tp, a->block, a->q, a->parent,
net/sched/cls_api.c
2782
struct tcf_block *block = chain->block;
net/sched/cls_api.c
2805
if (tcf_fill_node(net, skb, tp, block, q, parent, NULL,
net/sched/cls_api.c
2817
arg.block = block;
net/sched/cls_api.c
2850
struct tcf_block *block;
net/sched/cls_api.c
2874
block = tcf_block_refcnt_get(net, tcm->tcm_block_index);
net/sched/cls_api.c
2875
if (!block)
net/sched/cls_api.c
2910
block = cops->tcf_block(q, cl, NULL);
net/sched/cls_api.c
2911
if (!block)
net/sched/cls_api.c
2913
parent = block->classid;
net/sched/cls_api.c
2914
if (tcf_block_shared(block))
net/sched/cls_api.c
2921
for (chain = __tcf_get_next_chain(block, NULL);
net/sched/cls_api.c
2924
chain = __tcf_get_next_chain(block, chain),
net/sched/cls_api.c
2938
tcf_block_refcnt_put(block, true);
net/sched/cls_api.c
2951
struct tcf_block *block,
net/sched/cls_api.c
2973
if (block->q) {
net/sched/cls_api.c
2974
tcm->tcm_ifindex = qdisc_dev(block->q)->ifindex;
net/sched/cls_api.c
2975
tcm->tcm_parent = block->q->handle;
net/sched/cls_api.c
2978
tcm->tcm_block_index = block->index;
net/sched/cls_api.c
3010
struct tcf_block *block = chain->block;
net/sched/cls_api.c
3011
struct net *net = block->net;
net/sched/cls_api.c
3023
chain->index, net, skb, block, portid,
net/sched/cls_api.c
3040
struct tcf_block *block, struct sk_buff *oskb,
net/sched/cls_api.c
3044
struct net *net = block->net;
net/sched/cls_api.c
3055
block, portid, seq, flags, RTM_DELCHAIN, NULL) <= 0) {
net/sched/cls_api.c
3123
struct tcf_block *block;
net/sched/cls_api.c
3138
block = tcf_block_find(net, &q, &parent, &cl,
net/sched/cls_api.c
3140
if (IS_ERR(block))
net/sched/cls_api.c
3141
return PTR_ERR(block);
net/sched/cls_api.c
3150
mutex_lock(&block->lock);
net/sched/cls_api.c
3151
chain = tcf_chain_lookup(block, chain_index);
net/sched/cls_api.c
3170
chain = tcf_chain_create(block, chain_index);
net/sched/cls_api.c
3195
mutex_unlock(&block->lock);
net/sched/cls_api.c
3209
tfilter_notify_chain(net, skb, block, q, parent, n,
net/sched/cls_api.c
3233
tcf_block_release(q, block, true);
net/sched/cls_api.c
3240
mutex_unlock(&block->lock);
net/sched/cls_api.c
3250
struct tcf_block *block;
net/sched/cls_api.c
3266
block = tcf_block_refcnt_get(net, tcm->tcm_block_index);
net/sched/cls_api.c
3267
if (!block)
net/sched/cls_api.c
3295
block = cops->tcf_block(q, cl, NULL);
net/sched/cls_api.c
3296
if (!block)
net/sched/cls_api.c
3298
if (tcf_block_shared(block))
net/sched/cls_api.c
3305
mutex_lock(&block->lock);
net/sched/cls_api.c
3306
list_for_each_entry(chain, &block->chain_list, list) {
net/sched/cls_api.c
3317
chain->index, net, skb, block,
net/sched/cls_api.c
3325
mutex_unlock(&block->lock);
net/sched/cls_api.c
3328
tcf_block_refcnt_put(block, true);
net/sched/cls_api.c
3554
static void tcf_block_offload_inc(struct tcf_block *block, u32 *flags)
net/sched/cls_api.c
3559
atomic_inc(&block->offloadcnt);
net/sched/cls_api.c
3562
static void tcf_block_offload_dec(struct tcf_block *block, u32 *flags)
net/sched/cls_api.c
3567
atomic_dec(&block->offloadcnt);
net/sched/cls_api.c
3570
static void tc_cls_offload_cnt_update(struct tcf_block *block,
net/sched/cls_api.c
3574
lockdep_assert_held(&block->cb_lock);
net/sched/cls_api.c
3579
tcf_block_offload_inc(block, flags);
net/sched/cls_api.c
3584
tcf_block_offload_dec(block, flags);
net/sched/cls_api.c
3590
tc_cls_offload_cnt_reset(struct tcf_block *block, struct tcf_proto *tp,
net/sched/cls_api.c
3593
lockdep_assert_held(&block->cb_lock);
net/sched/cls_api.c
3596
tcf_block_offload_dec(block, flags);
net/sched/cls_api.c
3602
__tc_setup_cb_call(struct tcf_block *block, enum tc_setup_type type,
net/sched/cls_api.c
3609
list_for_each_entry(block_cb, &block->flow_block.cb_list, list) {
net/sched/cls_api.c
3621
int tc_setup_cb_call(struct tcf_block *block, enum tc_setup_type type,
net/sched/cls_api.c
3624
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/cls_api.c
3630
down_read(&block->cb_lock);
net/sched/cls_api.c
3635
if (!rtnl_held && !take_rtnl && block->lockeddevcnt) {
net/sched/cls_api.c
3636
up_read(&block->cb_lock);
net/sched/cls_api.c
3641
ok_count = __tc_setup_cb_call(block, type, type_data, err_stop);
net/sched/cls_api.c
3643
up_read(&block->cb_lock);
net/sched/cls_api.c
3656
int tc_setup_cb_add(struct tcf_block *block, struct tcf_proto *tp,
net/sched/cls_api.c
3660
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/cls_api.c
3666
down_read(&block->cb_lock);
net/sched/cls_api.c
3671
if (!rtnl_held && !take_rtnl && block->lockeddevcnt) {
net/sched/cls_api.c
3672
up_read(&block->cb_lock);
net/sched/cls_api.c
3678
if (block->nooffloaddevcnt && err_stop) {
net/sched/cls_api.c
3683
ok_count = __tc_setup_cb_call(block, type, type_data, err_stop);
net/sched/cls_api.c
3690
tc_cls_offload_cnt_update(block, tp, in_hw_count, flags,
net/sched/cls_api.c
3693
up_read(&block->cb_lock);
net/sched/cls_api.c
3706
int tc_setup_cb_replace(struct tcf_block *block, struct tcf_proto *tp,
net/sched/cls_api.c
3712
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/cls_api.c
3718
down_read(&block->cb_lock);
net/sched/cls_api.c
3723
if (!rtnl_held && !take_rtnl && block->lockeddevcnt) {
net/sched/cls_api.c
3724
up_read(&block->cb_lock);
net/sched/cls_api.c
3730
if (block->nooffloaddevcnt && err_stop) {
net/sched/cls_api.c
3735
tc_cls_offload_cnt_reset(block, tp, old_in_hw_count, old_flags);
net/sched/cls_api.c
3739
ok_count = __tc_setup_cb_call(block, type, type_data, err_stop);
net/sched/cls_api.c
3746
tc_cls_offload_cnt_update(block, tp, new_in_hw_count,
net/sched/cls_api.c
3749
up_read(&block->cb_lock);
net/sched/cls_api.c
3760
int tc_setup_cb_destroy(struct tcf_block *block, struct tcf_proto *tp,
net/sched/cls_api.c
3764
bool take_rtnl = READ_ONCE(block->lockeddevcnt) && !rtnl_held;
net/sched/cls_api.c
3770
down_read(&block->cb_lock);
net/sched/cls_api.c
3775
if (!rtnl_held && !take_rtnl && block->lockeddevcnt) {
net/sched/cls_api.c
3776
up_read(&block->cb_lock);
net/sched/cls_api.c
3781
ok_count = __tc_setup_cb_call(block, type, type_data, err_stop);
net/sched/cls_api.c
3783
tc_cls_offload_cnt_reset(block, tp, in_hw_count, flags);
net/sched/cls_api.c
3787
up_read(&block->cb_lock);
net/sched/cls_api.c
3794
int tc_setup_cb_reoffload(struct tcf_block *block, struct tcf_proto *tp,
net/sched/cls_api.c
3805
tc_cls_offload_cnt_update(block, tp, in_hw_count, flags, 1,
net/sched/cls_api.c
3991
return tcf_block_get_ext(&qe->block, sch, &qe->info, extack);
net/sched/cls_api.c
3998
tcf_block_put_ext(qe->block, sch, &qe->info);
net/sched/cls_api.c
417
struct tcf_block *block = tp->chain->block;
net/sched/cls_api.c
422
if (!atomic_dec_return(&block->useswcnt))
net/sched/cls_api.c
436
if (counted && atomic_inc_return(&block->useswcnt) == 1)
net/sched/cls_api.c
489
#define ASSERT_BLOCK_LOCKED(block) \
net/sched/cls_api.c
490
lockdep_assert_held(&(block)->lock)
net/sched/cls_api.c
498
static struct tcf_chain *tcf_chain_create(struct tcf_block *block,
net/sched/cls_api.c
503
ASSERT_BLOCK_LOCKED(block);
net/sched/cls_api.c
508
list_add_tail_rcu(&chain->list, &block->chain_list);
net/sched/cls_api.c
510
chain->block = block;
net/sched/cls_api.c
514
block->chain0.chain = chain;
net/sched/cls_api.c
529
struct tcf_block *block = chain->block;
net/sched/cls_api.c
534
mutex_lock(&block->lock);
net/sched/cls_api.c
535
list_for_each_entry(item, &block->chain0.filter_chain_list, list)
net/sched/cls_api.c
537
mutex_unlock(&block->lock);
net/sched/cls_api.c
544
struct tcf_block *block = chain->block;
net/sched/cls_api.c
546
ASSERT_BLOCK_LOCKED(block);
net/sched/cls_api.c
550
block->chain0.chain = NULL;
net/sched/cls_api.c
552
if (list_empty(&block->chain_list) &&
net/sched/cls_api.c
553
refcount_read(&block->refcnt) == 0)
net/sched/cls_api.c
559
static void tcf_block_destroy(struct tcf_block *block)
net/sched/cls_api.c
561
mutex_destroy(&block->lock);
net/sched/cls_api.c
562
mutex_destroy(&block->proto_destroy_lock);
net/sched/cls_api.c
563
xa_destroy(&block->ports);
net/sched/cls_api.c
564
kfree_rcu(block, rcu);
net/sched/cls_api.c
569
struct tcf_block *block = chain->block;
net/sched/cls_api.c
574
tcf_block_destroy(block);
net/sched/cls_api.c
579
ASSERT_BLOCK_LOCKED(chain->block);
net/sched/cls_api.c
586
ASSERT_BLOCK_LOCKED(chain->block);
net/sched/cls_api.c
594
static struct tcf_chain *tcf_chain_lookup(struct tcf_block *block,
net/sched/cls_api.c
599
ASSERT_BLOCK_LOCKED(block);
net/sched/cls_api.c
601
list_for_each_entry(chain, &block->chain_list, list) {
net/sched/cls_api.c
609
static struct tcf_chain *tcf_chain_lookup_rcu(const struct tcf_block *block,
net/sched/cls_api.c
614
list_for_each_entry_rcu(chain, &block->chain_list, list) {
net/sched/cls_api.c
626
static struct tcf_chain *__tcf_chain_get(struct tcf_block *block,
net/sched/cls_api.c
633
mutex_lock(&block->lock);
net/sched/cls_api.c
634
chain = tcf_chain_lookup(block, chain_index);
net/sched/cls_api.c
640
chain = tcf_chain_create(block, chain_index);
net/sched/cls_api.c
648
mutex_unlock(&block->lock);
net/sched/cls_api.c
662
mutex_unlock(&block->lock);
net/sched/cls_api.c
666
static struct tcf_chain *tcf_chain_get(struct tcf_block *block, u32 chain_index,
net/sched/cls_api.c
669
return __tcf_chain_get(block, chain_index, create, false);
net/sched/cls_api.c
672
struct tcf_chain *tcf_chain_get_by_act(struct tcf_block *block, u32 chain_index)
net/sched/cls_api.c
674
return __tcf_chain_get(block, chain_index, true, true);
net/sched/cls_api.c
682
struct tcf_block *block, struct sk_buff *oskb,
net/sched/cls_api.c
688
struct tcf_block *block = chain->block;
net/sched/cls_api.c
694
mutex_lock(&block->lock);
net/sched/cls_api.c
697
mutex_unlock(&block->lock);
net/sched/cls_api.c
718
chain->index, block, NULL, 0, 0);
net/sched/cls_api.c
725
mutex_unlock(&block->lock);
net/sched/cls_api.c
773
static int tcf_block_setup(struct tcf_block *block,
net/sched/cls_api.c
786
bo->block = flow_block;
net/sched/cls_api.c
794
static void tcf_block_unbind(struct tcf_block *block,
net/sched/cls_api.c
799
struct tcf_block *block = block_cb->indr.data;
net/sched/cls_api.c
807
&block->flow_block, tcf_block_shared(block),
net/sched/cls_api.c
810
down_write(&block->cb_lock);
net/sched/cls_api.c
813
tcf_block_unbind(block, &bo);
net/sched/cls_api.c
814
up_write(&block->cb_lock);
net/sched/cls_api.c
818
static bool tcf_block_offload_in_use(struct tcf_block *block)
net/sched/cls_api.c
820
return atomic_read(&block->offloadcnt);
net/sched/cls_api.c
823
static int tcf_block_offload_cmd(struct tcf_block *block,
net/sched/cls_api.c
832
&block->flow_block, tcf_block_shared(block),
net/sched/cls_api.c
845
return tcf_block_setup(block, &bo);
net/sched/cls_api.c
848
flow_indr_dev_setup_offload(dev, sch, TC_SETUP_BLOCK, block, &bo,
net/sched/cls_api.c
850
tcf_block_setup(block, &bo);
net/sched/cls_api.c
855
static int tcf_block_offload_bind(struct tcf_block *block, struct Qdisc *q,
net/sched/cls_api.c
862
down_write(&block->cb_lock);
net/sched/cls_api.c
869
tcf_block_offload_in_use(block)) {
net/sched/cls_api.c
875
err = tcf_block_offload_cmd(block, dev, q, ei, FLOW_BLOCK_BIND, extack);
net/sched/cls_api.c
881
up_write(&block->cb_lock);
net/sched/cls_api.c
885
if (tcf_block_offload_in_use(block))
net/sched/cls_api.c
889
block->nooffloaddevcnt++;
net/sched/cls_api.c
891
up_write(&block->cb_lock);
net/sched/cls_api.c
895
static void tcf_block_offload_unbind(struct tcf_block *block, struct Qdisc *q,
net/sched/cls_api.c
901
down_write(&block->cb_lock);
net/sched/cls_api.c
902
err = tcf_block_offload_cmd(block, dev, q, ei, FLOW_BLOCK_UNBIND, NULL);
net/sched/cls_api.c
905
up_write(&block->cb_lock);
net/sched/cls_api.c
909
WARN_ON(block->nooffloaddevcnt-- == 0);
net/sched/cls_api.c
910
up_write(&block->cb_lock);
net/sched/cls_api.c
914
tcf_chain0_head_change_cb_add(struct tcf_block *block,
net/sched/cls_api.c
929
mutex_lock(&block->lock);
net/sched/cls_api.c
930
chain0 = block->chain0.chain;
net/sched/cls_api.c
934
list_add(&item->list, &block->chain0.filter_chain_list);
net/sched/cls_api.c
935
mutex_unlock(&block->lock);
net/sched/cls_api.c
946
mutex_lock(&block->lock);
net/sched/cls_api.c
947
list_add(&item->list, &block->chain0.filter_chain_list);
net/sched/cls_api.c
948
mutex_unlock(&block->lock);
net/sched/cls_api.c
958
tcf_chain0_head_change_cb_del(struct tcf_block *block,
net/sched/cls_api.c
963
mutex_lock(&block->lock);
net/sched/cls_api.c
964
list_for_each_entry(item, &block->chain0.filter_chain_list, list) {
net/sched/cls_api.c
968
if (block->chain0.chain)
net/sched/cls_api.c
971
mutex_unlock(&block->lock);
net/sched/cls_api.c
977
mutex_unlock(&block->lock);
net/sched/cls_api.c
988
static int tcf_block_insert(struct tcf_block *block, struct net *net,
net/sched/cls_api.c
996
err = idr_alloc_u32(&tn->idr, block, &block->index, block->index,
net/sched/cls_bpf.c
147
struct tcf_block *block = tp->chain->block;
net/sched/cls_bpf.c
165
err = tc_setup_cb_replace(block, tp, TC_SETUP_CLSBPF, &cls_bpf,
net/sched/cls_bpf.c
171
err = tc_setup_cb_add(block, tp, TC_SETUP_CLSBPF, &cls_bpf,
net/sched/cls_bpf.c
175
err = tc_setup_cb_destroy(block, tp, TC_SETUP_CLSBPF, &cls_bpf,
net/sched/cls_bpf.c
228
struct tcf_block *block = tp->chain->block;
net/sched/cls_bpf.c
238
tc_setup_cb_call(block, TC_SETUP_CLSBPF, &cls_bpf, false, true);
net/sched/cls_bpf.c
402
tcf_block_netif_keep_dst(tp->chain->block);
net/sched/cls_bpf.c
654
struct tcf_block *block = tp->chain->block;
net/sched/cls_bpf.c
672
err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSBPF,
net/sched/cls_flow.c
506
struct tcf_block *block = tp->chain->block;
net/sched/cls_flow.c
509
if (tcf_block_shared(block)) {
net/sched/cls_flow.c
515
q = tcf_block_q(block);
net/sched/cls_flow.c
529
tcf_block_netif_keep_dst(tp->chain->block);
net/sched/cls_flower.c
2684
struct tcf_block *block = tp->chain->block;
net/sched/cls_flower.c
2725
err = tc_setup_cb_reoffload(block, tp, add, cb,
net/sched/cls_flower.c
2771
struct tcf_block *block = chain->block;
net/sched/cls_flower.c
2787
tc_setup_cb_call(block, TC_SETUP_CLSFLOWER, &cls_flower, false, true);
net/sched/cls_flower.c
2797
struct tcf_block *block = chain->block;
net/sched/cls_flower.c
2803
tc_setup_cb_call(block, TC_SETUP_CLSFLOWER, &cls_flower, false, true);
net/sched/cls_flower.c
451
struct tcf_block *block = tp->chain->block;
net/sched/cls_flower.c
458
tc_setup_cb_destroy(block, tp, TC_SETUP_CLSFLOWER, &cls_flower, false,
net/sched/cls_flower.c
467
struct tcf_block *block = tp->chain->block;
net/sched/cls_flower.c
492
err = tc_setup_cb_add(block, tp, TC_SETUP_CLSFLOWER, &cls_flower,
net/sched/cls_flower.c
511
struct tcf_block *block = tp->chain->block;
net/sched/cls_flower.c
519
tc_setup_cb_call(block, TC_SETUP_CLSFLOWER, &cls_flower, false,
net/sched/cls_fw.c
254
if (tcf_block_shared(tp->chain->block)) {
net/sched/cls_fw.c
77
struct Qdisc *q = tcf_block_q(tp->chain->block);
net/sched/cls_matchall.c
111
err = tc_setup_cb_add(block, tp, TC_SETUP_CLSMATCHALL, &cls_mall,
net/sched/cls_matchall.c
280
struct tcf_block *block = tp->chain->block;
net/sched/cls_matchall.c
303
err = tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSMATCHALL,
net/sched/cls_matchall.c
317
struct tcf_block *block = tp->chain->block;
net/sched/cls_matchall.c
323
tc_setup_cb_call(block, TC_SETUP_CLSMATCHALL, &cls_mall, false, true);
net/sched/cls_matchall.c
74
struct tcf_block *block = tp->chain->block;
net/sched/cls_matchall.c
80
tc_setup_cb_destroy(block, tp, TC_SETUP_CLSMATCHALL, &cls_mall, false,
net/sched/cls_matchall.c
90
struct tcf_block *block = tp->chain->block;
net/sched/cls_route.c
541
tcf_block_netif_keep_dst(tp->chain->block);
net/sched/cls_u32.c
1258
struct tcf_block *block = tp->chain->block;
net/sched/cls_u32.c
1282
return tc_setup_cb_reoffload(block, tp, add, cb, TC_SETUP_CLSU32,
net/sched/cls_u32.c
332
struct tcf_block *block = tp->chain->block;
net/sched/cls_u32.c
340
if (tcf_block_shared(block))
net/sched/cls_u32.c
341
return block;
net/sched/cls_u32.c
343
return block->q;
net/sched/cls_u32.c
491
struct tcf_block *block = tp->chain->block;
net/sched/cls_u32.c
500
tc_setup_cb_call(block, TC_SETUP_CLSU32, &cls_u32, false, true);
net/sched/cls_u32.c
506
struct tcf_block *block = tp->chain->block;
net/sched/cls_u32.c
518
err = tc_setup_cb_call(block, TC_SETUP_CLSU32, &cls_u32, skip_sw, true);
net/sched/cls_u32.c
535
struct tcf_block *block = tp->chain->block;
net/sched/cls_u32.c
542
tc_setup_cb_destroy(block, tp, TC_SETUP_CLSU32, &cls_u32, false,
net/sched/cls_u32.c
550
struct tcf_block *block = tp->chain->block;
net/sched/cls_u32.c
572
err = tc_setup_cb_add(block, tp, TC_SETUP_CLSU32, &cls_u32, skip_sw,
net/sched/ematch.c
177
struct net *net = tp->chain->block->net;
net/sched/sch_api.c
2089
struct Qdisc *q = tcf_block_q(tp->chain->block);
net/sched/sch_api.c
2110
struct tcf_block *block;
net/sched/sch_api.c
2113
block = cops->tcf_block(q, cl, NULL);
net/sched/sch_api.c
2114
if (!block)
net/sched/sch_api.c
2116
for (chain = tcf_get_next_chain(block, NULL);
net/sched/sch_api.c
2118
chain = tcf_get_next_chain(block, chain)) {
net/sched/sch_cake.c
221
struct tcf_block *block;
net/sched/sch_cake.c
2797
tcf_block_put(q->block);
net/sched/sch_cake.c
2841
err = tcf_block_get(&qd->block, &qd->filter_list, sch, extack);
net/sched/sch_cake.c
3105
return q->block;
net/sched/sch_drr.c
189
return q->block;
net/sched/sch_drr.c
34
struct tcf_block *block;
net/sched/sch_drr.c
419
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_drr.c
451
tcf_block_put(q->block);
net/sched/sch_ets.c
371
return q->block;
net/sched/sch_ets.c
51
struct tcf_block *block;
net/sched/sch_ets.c
710
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_ets.c
740
tcf_block_put(q->block);
net/sched/sch_fq_codel.c
464
tcf_block_put(q->block);
net/sched/sch_fq_codel.c
494
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_fq_codel.c
52
struct tcf_block *block;
net/sched/sch_fq_codel.c
634
return q->block;
net/sched/sch_fq_pie.c
447
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_fq_pie.c
557
tcf_block_put(q->block);
net/sched/sch_fq_pie.c
57
struct tcf_block *block;
net/sched/sch_generic.c
1572
struct tcf_block *block)
net/sched/sch_generic.c
1574
miniqp->miniq1.block = block;
net/sched/sch_generic.c
1575
miniqp->miniq2.block = block;
net/sched/sch_hfsc.c
1034
err = tcf_block_get(&cl->block, &cl->filter_list, sch, extack);
net/sched/sch_hfsc.c
1044
tcf_block_put(cl->block);
net/sched/sch_hfsc.c
1095
tcf_block_put(cl->block);
net/sched/sch_hfsc.c
118
struct tcf_block *block;
net/sched/sch_hfsc.c
1266
return cl->block;
net/sched/sch_hfsc.c
1413
err = tcf_block_get(&q->root.block, &q->root.filter_list, sch, extack);
net/sched/sch_hfsc.c
1507
tcf_block_put(cl->block);
net/sched/sch_hfsc.c
1508
cl->block = NULL;
net/sched/sch_htb.c
104
struct tcf_block *block;
net/sched/sch_htb.c
1068
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_htb.c
156
struct tcf_block *block;
net/sched/sch_htb.c
1610
tcf_block_put(cl->block);
net/sched/sch_htb.c
1631
tcf_block_put(q->block);
net/sched/sch_htb.c
1635
tcf_block_put(cl->block);
net/sched/sch_htb.c
1636
cl->block = NULL;
net/sched/sch_htb.c
1854
err = tcf_block_get(&cl->block, &cl->filter_list, sch, extack);
net/sched/sch_htb.c
2065
tcf_block_put(cl->block);
net/sched/sch_htb.c
2077
return cl ? cl->block : q->block;
net/sched/sch_ingress.c
103
err = tcf_block_get_ext(&q->block, sch, &q->block_info, extack);
net/sched/sch_ingress.c
107
mini_qdisc_pair_block_init(&q->miniqp, q->block);
net/sched/sch_ingress.c
121
tcf_block_put_ext(q->block, sch, &q->block_info);
net/sched/sch_ingress.c
19
struct tcf_block *block;
net/sched/sch_ingress.c
53
return q->block;
net/sched/sch_multiq.c
164
tcf_block_put(q->block);
net/sched/sch_multiq.c
24
struct tcf_block *block;
net/sched/sch_multiq.c
246
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_multiq.c
368
return q->block;
net/sched/sch_prio.c
170
tcf_block_put(q->block);
net/sched/sch_prio.c
238
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_prio.c
24
struct tcf_block *block;
net/sched/sch_prio.c
391
return q->block;
net/sched/sch_qfq.c
1447
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_qfq.c
1501
tcf_block_put(q->block);
net/sched/sch_qfq.c
180
struct tcf_block *block;
net/sched/sch_qfq.c
582
return q->block;
net/sched/sch_sfb.c
472
tcf_block_put(q->block);
net/sched/sch_sfb.c
55
struct tcf_block *block;
net/sched/sch_sfb.c
561
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_sfb.c
676
return q->block;
net/sched/sch_sfq.c
118
struct tcf_block *block;
net/sched/sch_sfq.c
766
tcf_block_put(q->block);
net/sched/sch_sfq.c
784
err = tcf_block_get(&q->block, &q->filter_list, sch, extack);
net/sched/sch_sfq.c
894
return q->block;
scripts/mod/sumversion.c
158
le32_to_cpu_array(ctx->block, ARRAY_SIZE(ctx->block));
scripts/mod/sumversion.c
159
md4_transform(ctx->hash, ctx->block);
scripts/mod/sumversion.c
174
const uint32_t avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f);
scripts/mod/sumversion.c
179
memcpy((char *)mctx->block + (sizeof(mctx->block) - avail),
scripts/mod/sumversion.c
184
memcpy((char *)mctx->block + (sizeof(mctx->block) - avail),
scripts/mod/sumversion.c
191
while (len >= sizeof(mctx->block)) {
scripts/mod/sumversion.c
192
memcpy(mctx->block, data, sizeof(mctx->block));
scripts/mod/sumversion.c
194
data += sizeof(mctx->block);
scripts/mod/sumversion.c
195
len -= sizeof(mctx->block);
scripts/mod/sumversion.c
198
memcpy(mctx->block, data, len);
scripts/mod/sumversion.c
204
char *p = (char *)mctx->block + offset;
scripts/mod/sumversion.c
211
p = (char *)mctx->block;
scripts/mod/sumversion.c
216
mctx->block[14] = mctx->byte_count << 3;
scripts/mod/sumversion.c
217
mctx->block[15] = mctx->byte_count >> 29;
scripts/mod/sumversion.c
218
le32_to_cpu_array(mctx->block, (sizeof(mctx->block) -
scripts/mod/sumversion.c
220
md4_transform(mctx->hash, mctx->block);
scripts/mod/sumversion.c
44
uint32_t block[MD4_BLOCK_WORDS];
sound/aoa/codecs/tas.c
155
u8 block[6];
sound/aoa/codecs/tas.c
175
block[0] = tmp>>20;
sound/aoa/codecs/tas.c
176
block[1] = tmp>>12;
sound/aoa/codecs/tas.c
177
block[2] = tmp>>4;
sound/aoa/codecs/tas.c
179
block[3] = tmp>>20;
sound/aoa/codecs/tas.c
180
block[4] = tmp>>12;
sound/aoa/codecs/tas.c
181
block[5] = tmp>>4;
sound/aoa/codecs/tas.c
182
tas_write_reg(tas, TAS_REG_VOL, 6, block);
sound/aoa/codecs/tas.c
187
u8 block[9];
sound/aoa/codecs/tas.c
195
block[3*i+0] = tmp>>16;
sound/aoa/codecs/tas.c
196
block[3*i+1] = tmp>>8;
sound/aoa/codecs/tas.c
197
block[3*i+2] = tmp;
sound/aoa/codecs/tas.c
199
tas_write_reg(tas, TAS_REG_LMIX, 9, block);
sound/aoa/codecs/tas.c
205
block[3*i+0] = tmp>>16;
sound/aoa/codecs/tas.c
206
block[3*i+1] = tmp>>8;
sound/aoa/codecs/tas.c
207
block[3*i+2] = tmp;
sound/aoa/codecs/tas.c
209
tas_write_reg(tas, TAS_REG_RMIX, 9, block);
sound/core/oss/pcm_oss.c
2368
else if (setup->block)
sound/core/oss/pcm_oss.c
2963
setup->block ? " block" : "",
sound/core/oss/pcm_oss.c
3023
template.block = 1;
sound/core/seq/seq_system.h
20
static inline void snd_seq_system_ump_notify(int client, int block, int type,
sound/core/seq/seq_system.h
26
snd_seq_system_broadcast(client, block, type, atomic);
sound/drivers/opl3/opl3_midi.c
112
((block << 2) & OPL3_BLOCKNUM_MASK);
sound/drivers/opl3/opl3_midi.c
89
int block = ((note / 12) & 0x07) - 1;
sound/isa/gus/gus_dma.c
101
block = NULL;
sound/isa/gus/gus_dma.c
103
if (block) {
sound/isa/gus/gus_dma.c
104
gus->gf1.dma_ack = block->ack;
sound/isa/gus/gus_dma.c
105
gus->gf1.dma_private_data = block->private_data;
sound/isa/gus/gus_dma.c
107
return block;
sound/isa/gus/gus_dma.c
113
struct snd_gf1_dma_block *block;
sound/isa/gus/gus_dma.c
125
block = snd_gf1_dma_next_block(gus);
sound/isa/gus/gus_dma.c
127
if (!block)
sound/isa/gus/gus_dma.c
129
snd_gf1_dma_program(gus, block->addr, block->buf_addr, block->count, (unsigned short) block->cmd);
sound/isa/gus/gus_dma.c
130
kfree(block);
sound/isa/gus/gus_dma.c
134
block->addr, block->buf_addr, block->count, block->cmd);
sound/isa/gus/gus_dma.c
154
struct snd_gf1_dma_block *block;
sound/isa/gus/gus_dma.c
162
while ((block = gus->gf1.dma_data_pcm)) {
sound/isa/gus/gus_dma.c
163
gus->gf1.dma_data_pcm = block->next;
sound/isa/gus/gus_dma.c
164
kfree(block);
sound/isa/gus/gus_dma.c
166
while ((block = gus->gf1.dma_data_synth)) {
sound/isa/gus/gus_dma.c
167
gus->gf1.dma_data_synth = block->next;
sound/isa/gus/gus_dma.c
168
kfree(block);
sound/isa/gus/gus_dma.c
181
struct snd_gf1_dma_block *block;
sound/isa/gus/gus_dma.c
184
block = kmalloc(sizeof(*block), atomic ? GFP_ATOMIC : GFP_KERNEL);
sound/isa/gus/gus_dma.c
185
if (!block)
sound/isa/gus/gus_dma.c
188
*block = *__block;
sound/isa/gus/gus_dma.c
189
block->next = NULL;
sound/isa/gus/gus_dma.c
193
block->addr, (long) block->buffer, block->count,
sound/isa/gus/gus_dma.c
194
block->cmd);
sound/isa/gus/gus_dma.c
206
gus->gf1.dma_data_synth_last->next = block;
sound/isa/gus/gus_dma.c
207
gus->gf1.dma_data_synth_last = block;
sound/isa/gus/gus_dma.c
210
gus->gf1.dma_data_synth_last = block;
sound/isa/gus/gus_dma.c
214
gus->gf1.dma_data_pcm_last->next = block;
sound/isa/gus/gus_dma.c
215
gus->gf1.dma_data_pcm_last = block;
sound/isa/gus/gus_dma.c
218
gus->gf1.dma_data_pcm_last = block;
sound/isa/gus/gus_dma.c
81
struct snd_gf1_dma_block *block;
sound/isa/gus/gus_dma.c
85
block = gus->gf1.dma_data_pcm;
sound/isa/gus/gus_dma.c
86
if (gus->gf1.dma_data_pcm_last == block) {
sound/isa/gus/gus_dma.c
90
gus->gf1.dma_data_pcm = block->next;
sound/isa/gus/gus_dma.c
93
block = gus->gf1.dma_data_synth;
sound/isa/gus/gus_dma.c
94
if (gus->gf1.dma_data_synth_last == block) {
sound/isa/gus/gus_dma.c
98
gus->gf1.dma_data_synth = block->next;
sound/isa/gus/gus_mem.c
105
struct snd_gf1_mem_block *block;
sound/isa/gus/gus_mem.c
110
for (block = alloc->first; block; block = block->next)
sound/isa/gus/gus_mem.c
111
if (!memcmp(share_id, block->share_id,
sound/isa/gus/gus_mem.c
112
sizeof(block->share_id)))
sound/isa/gus/gus_mem.c
113
return block;
sound/isa/gus/gus_mem.c
118
struct snd_gf1_mem_block * block,
sound/isa/gus/gus_mem.c
129
block->flags = w_16 ? SNDRV_GF1_MEM_BLOCK_16BIT : 0;
sound/isa/gus/gus_mem.c
130
block->owner = SNDRV_GF1_MEM_OWNER_DRIVER;
sound/isa/gus/gus_mem.c
131
block->share = 0;
sound/isa/gus/gus_mem.c
132
block->share_id[0] = block->share_id[1] =
sound/isa/gus/gus_mem.c
133
block->share_id[2] = block->share_id[3] = 0;
sound/isa/gus/gus_mem.c
134
block->name = NULL;
sound/isa/gus/gus_mem.c
135
block->prev = block->next = NULL;
sound/isa/gus/gus_mem.c
153
block->ptr = ptr1;
sound/isa/gus/gus_mem.c
154
block->size = size;
sound/isa/gus/gus_mem.c
161
block->ptr = info[idx].address;
sound/isa/gus/gus_mem.c
162
block->size = size;
sound/isa/gus/gus_mem.c
173
struct snd_gf1_mem_block block, *nblock;
sound/isa/gus/gus_mem.c
189
if (snd_gf1_mem_find(alloc, &block, size, w_16, align) < 0)
sound/isa/gus/gus_mem.c
19
snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block,
sound/isa/gus/gus_mem.c
192
memcpy(&block.share_id, share_id, sizeof(block.share_id));
sound/isa/gus/gus_mem.c
193
block.owner = owner;
sound/isa/gus/gus_mem.c
194
nblock = snd_gf1_mem_xalloc(alloc, &block, name);
sound/isa/gus/gus_mem.c
200
struct snd_gf1_mem_block *block;
sound/isa/gus/gus_mem.c
203
block = snd_gf1_mem_look(alloc, address);
sound/isa/gus/gus_mem.c
204
if (block)
sound/isa/gus/gus_mem.c
205
return snd_gf1_mem_xfree(alloc, block);
sound/isa/gus/gus_mem.c
212
struct snd_gf1_mem_block block;
sound/isa/gus/gus_mem.c
220
memset(&block, 0, sizeof(block));
sound/isa/gus/gus_mem.c
221
block.owner = SNDRV_GF1_MEM_OWNER_DRIVER;
sound/isa/gus/gus_mem.c
223
block.ptr = 0;
sound/isa/gus/gus_mem.c
224
block.size = 1024;
sound/isa/gus/gus_mem.c
225
if (!snd_gf1_mem_xalloc(alloc, &block, "InterWave LFOs"))
sound/isa/gus/gus_mem.c
228
block.ptr = gus->gf1.default_voice_address;
sound/isa/gus/gus_mem.c
229
block.size = 4;
sound/isa/gus/gus_mem.c
230
if (!snd_gf1_mem_xalloc(alloc, &block, "Voice default (NULL's)"))
sound/isa/gus/gus_mem.c
241
struct snd_gf1_mem_block *block, *nblock;
sound/isa/gus/gus_mem.c
244
block = alloc->first;
sound/isa/gus/gus_mem.c
245
while (block) {
sound/isa/gus/gus_mem.c
246
nblock = block->next;
sound/isa/gus/gus_mem.c
247
snd_gf1_mem_xfree(alloc, block);
sound/isa/gus/gus_mem.c
248
block = nblock;
sound/isa/gus/gus_mem.c
259
struct snd_gf1_mem_block *block;
sound/isa/gus/gus_mem.c
27
*nblock = *block;
sound/isa/gus/gus_mem.c
277
for (block = alloc->first, i = 0; block; block = block->next, i++) {
sound/isa/gus/gus_mem.c
278
used += block->size;
sound/isa/gus/gus_mem.c
279
snd_iprintf(buffer, "Block %i onboard 0x%x size %i (0x%x):\n", i, block->ptr, block->size, block->size);
sound/isa/gus/gus_mem.c
280
if (block->share ||
sound/isa/gus/gus_mem.c
281
block->share_id[0] || block->share_id[1] ||
sound/isa/gus/gus_mem.c
282
block->share_id[2] || block->share_id[3])
sound/isa/gus/gus_mem.c
284
block->share,
sound/isa/gus/gus_mem.c
285
block->share_id[0], block->share_id[1],
sound/isa/gus/gus_mem.c
286
block->share_id[2], block->share_id[3]);
sound/isa/gus/gus_mem.c
288
block->flags & SNDRV_GF1_MEM_BLOCK_16BIT ? " 16-bit" : "");
sound/isa/gus/gus_mem.c
290
switch (block->owner) {
sound/isa/gus/gus_mem.c
292
snd_iprintf(buffer, "driver - %s\n", block->name);
sound/isa/gus/gus_mem.c
60
int snd_gf1_mem_xfree(struct snd_gf1_mem * alloc, struct snd_gf1_mem_block * block)
sound/isa/gus/gus_mem.c
62
if (block->share) { /* ok.. shared block */
sound/isa/gus/gus_mem.c
63
block->share--;
sound/isa/gus/gus_mem.c
66
if (alloc->first == block) {
sound/isa/gus/gus_mem.c
67
alloc->first = block->next;
sound/isa/gus/gus_mem.c
68
if (block->next)
sound/isa/gus/gus_mem.c
69
block->next->prev = NULL;
sound/isa/gus/gus_mem.c
71
block->prev->next = block->next;
sound/isa/gus/gus_mem.c
72
if (block->next)
sound/isa/gus/gus_mem.c
73
block->next->prev = block->prev;
sound/isa/gus/gus_mem.c
75
if (alloc->last == block) {
sound/isa/gus/gus_mem.c
76
alloc->last = block->prev;
sound/isa/gus/gus_mem.c
77
if (block->prev)
sound/isa/gus/gus_mem.c
78
block->prev->next = NULL;
sound/isa/gus/gus_mem.c
80
block->next->prev = block->prev;
sound/isa/gus/gus_mem.c
81
if (block->prev)
sound/isa/gus/gus_mem.c
82
block->prev->next = block->next;
sound/isa/gus/gus_mem.c
84
kfree(block->name);
sound/isa/gus/gus_mem.c
85
kfree(block);
sound/isa/gus/gus_mem.c
92
struct snd_gf1_mem_block *block;
sound/isa/gus/gus_mem.c
94
for (block = alloc->first; block; block = block->next) {
sound/isa/gus/gus_mem.c
95
if (block->ptr == address) {
sound/isa/gus/gus_mem.c
96
return block;
sound/isa/gus/gus_pcm.c
391
struct snd_gf1_mem_block *block;
sound/isa/gus/gus_pcm.c
396
block = snd_gf1_mem_alloc(&gus->gf1.mem_alloc,
sound/isa/gus/gus_pcm.c
401
if (!block)
sound/isa/gus/gus_pcm.c
403
pcmp->memory = block->ptr;
sound/isa/gus/gus_pcm.c
64
struct snd_gf1_dma_block block;
sound/isa/gus/gus_pcm.c
70
memset(&block, 0, sizeof(block));
sound/isa/gus/gus_pcm.c
71
block.cmd = SNDRV_GF1_DMA_IRQ;
sound/isa/gus/gus_pcm.c
73
block.cmd |= SNDRV_GF1_DMA_UNSIGNED;
sound/isa/gus/gus_pcm.c
75
block.cmd |= SNDRV_GF1_DMA_16BIT;
sound/isa/gus/gus_pcm.c
76
block.addr = addr & ~31;
sound/isa/gus/gus_pcm.c
77
block.buffer = runtime->dma_area + offset;
sound/isa/gus/gus_pcm.c
78
block.buf_addr = runtime->dma_addr + offset;
sound/isa/gus/gus_pcm.c
79
block.count = count;
sound/isa/gus/gus_pcm.c
80
block.private_data = pcmp;
sound/isa/gus/gus_pcm.c
81
block.ack = snd_gf1_pcm_block_change_ack;
sound/isa/gus/gus_pcm.c
82
if (!snd_gf1_dma_transfer_block(pcmp->gus, &block, 0, 0))
sound/isa/sb/emu8000_patch.c
158
sp->block = snd_util_mem_alloc(hdr, truesize * 2);
sound/isa/sb/emu8000_patch.c
159
if (sp->block == NULL) {
sound/isa/sb/emu8000_patch.c
173
dram_offset = EMU8000_DRAM_OFFSET + (sp->block->offset >> 1);
sound/isa/sb/emu8000_patch.c
192
if (! sp->block->offset) {
sound/isa/sb/emu8000_patch.c
263
if (sp->block) {
sound/isa/sb/emu8000_patch.c
264
snd_util_mem_free(hdr, sp->block);
sound/isa/sb/emu8000_patch.c
265
sp->block = NULL;
sound/isa/sb/emu8000_pcm.c
44
struct snd_util_memblk *block;
sound/isa/sb/emu8000_pcm.c
524
if (rec->block) {
sound/isa/sb/emu8000_pcm.c
526
snd_util_mem_free(rec->emu->memhdr, rec->block);
sound/isa/sb/emu8000_pcm.c
527
rec->block = NULL;
sound/isa/sb/emu8000_pcm.c
531
rec->block = snd_util_mem_alloc(rec->emu->memhdr, rec->allocated_bytes);
sound/isa/sb/emu8000_pcm.c
532
if (! rec->block)
sound/isa/sb/emu8000_pcm.c
534
rec->offset = EMU8000_DRAM_OFFSET + (rec->block->offset >> 1); /* in word */
sound/isa/sb/emu8000_pcm.c
548
if (rec->block) {
sound/isa/sb/emu8000_pcm.c
554
snd_util_mem_free(rec->emu->memhdr, rec->block);
sound/isa/sb/emu8000_pcm.c
555
rec->block = NULL;
sound/pci/bt87x.c
239
u32 block = i * 16 / periods;
sound/pci/bt87x.c
241
cmd |= block << RISC_SET_STATUS_SHIFT;
sound/pci/bt87x.c
242
cmd |= (~block & 0xf) << RISC_RESET_STATUS_SHIFT;
sound/pci/ctxfi/ctvmem.c
122
struct ct_vm_block *block;
sound/pci/ctxfi/ctvmem.c
128
block = get_vm_block(vm, size, atc);
sound/pci/ctxfi/ctvmem.c
129
if (block == NULL) {
sound/pci/ctxfi/ctvmem.c
136
pte_start = (block->addr >> CT_PAGE_SHIFT);
sound/pci/ctxfi/ctvmem.c
137
pages = block->size >> CT_PAGE_SHIFT;
sound/pci/ctxfi/ctvmem.c
144
block->size = size;
sound/pci/ctxfi/ctvmem.c
145
return block;
sound/pci/ctxfi/ctvmem.c
148
static void ct_vm_unmap(struct ct_vm *vm, struct ct_vm_block *block)
sound/pci/ctxfi/ctvmem.c
151
put_vm_block(vm, block);
sound/pci/ctxfi/ctvmem.c
168
struct ct_vm_block *block;
sound/pci/ctxfi/ctvmem.c
198
block = kzalloc_obj(*block);
sound/pci/ctxfi/ctvmem.c
199
if (NULL != block) {
sound/pci/ctxfi/ctvmem.c
200
block->addr = 0;
sound/pci/ctxfi/ctvmem.c
201
block->size = vm->size;
sound/pci/ctxfi/ctvmem.c
202
list_add(&block->list, &vm->unused);
sound/pci/ctxfi/ctvmem.c
32
struct ct_vm_block *block, *entry;
sound/pci/ctxfi/ctvmem.c
58
block = kzalloc_obj(*block);
sound/pci/ctxfi/ctvmem.c
59
if (!block)
sound/pci/ctxfi/ctvmem.c
62
block->addr = entry->addr;
sound/pci/ctxfi/ctvmem.c
63
block->size = size;
sound/pci/ctxfi/ctvmem.c
64
list_add(&block->list, &vm->used);
sound/pci/ctxfi/ctvmem.c
69
return block;
sound/pci/ctxfi/ctvmem.c
72
static void put_vm_block(struct ct_vm *vm, struct ct_vm_block *block)
sound/pci/ctxfi/ctvmem.c
77
block->size = CT_PAGE_ALIGN(block->size);
sound/pci/ctxfi/ctvmem.c
80
list_del(&block->list);
sound/pci/ctxfi/ctvmem.c
81
vm->size += block->size;
sound/pci/ctxfi/ctvmem.c
85
if (entry->addr >= (block->addr + block->size))
sound/pci/ctxfi/ctvmem.c
89
list_add_tail(&block->list, &vm->unused);
sound/pci/ctxfi/ctvmem.c
90
entry = block;
sound/pci/ctxfi/ctvmem.c
92
if ((block->addr + block->size) == entry->addr) {
sound/pci/ctxfi/ctvmem.c
93
entry->addr = block->addr;
sound/pci/ctxfi/ctvmem.c
94
entry->size += block->size;
sound/pci/ctxfi/ctvmem.c
95
kfree(block);
sound/pci/ctxfi/ctvmem.c
97
__list_add(&block->list, pos->prev, pos);
sound/pci/ctxfi/ctvmem.c
98
entry = block;
sound/pci/ctxfi/ctvmem.h
53
void (*unmap)(struct ct_vm *, struct ct_vm_block *block);
sound/pci/emu10k1/emu10k1_callback.c
148
if (vp->block) {
sound/pci/emu10k1/emu10k1_callback.c
150
emem = (struct snd_emu10k1_memblk *)vp->block;
sound/pci/emu10k1/emu10k1_callback.c
327
emem = (struct snd_emu10k1_memblk *)vp->block;
sound/pci/emu10k1/emu10k1_callback.c
460
emem = (struct snd_emu10k1_memblk *)vp->block;
sound/pci/emu10k1/emu10k1_patch.c
100
sp->block = snd_emu10k1_synth_alloc(emu, blocksize);
sound/pci/emu10k1/emu10k1_patch.c
101
if (sp->block == NULL) {
sound/pci/emu10k1/emu10k1_patch.c
113
snd_emu10k1_synth_memset(emu, sp->block, offset, size, fill);
sound/pci/emu10k1/emu10k1_patch.c
119
if (snd_emu10k1_synth_copy_from_user(emu, sp->block, offset, data, size, xor))
sound/pci/emu10k1/emu10k1_patch.c
126
if (snd_emu10k1_synth_copy_from_user(emu, sp->block, offset, data, size, xor))
sound/pci/emu10k1/emu10k1_patch.c
135
if (snd_emu10k1_synth_copy_from_user(emu, sp->block, offset, data, size, xor))
sound/pci/emu10k1/emu10k1_patch.c
141
snd_emu10k1_synth_memset(emu, sp->block, offset, blocksize - offset, fill);
sound/pci/emu10k1/emu10k1_patch.c
146
snd_emu10k1_synth_free(emu, sp->block);
sound/pci/emu10k1/emu10k1_patch.c
147
sp->block = NULL;
sound/pci/emu10k1/emu10k1_patch.c
164
if (sp->block) {
sound/pci/emu10k1/emu10k1_patch.c
165
snd_emu10k1_synth_free(emu, sp->block);
sound/pci/emu10k1/emu10k1_patch.c
166
sp->block = NULL;
sound/pci/emu10k1/memory.c
214
list_for_each(p, &emu->memhdr->block) {
sound/pci/emu10k1/memory.c
447
if (p != &hdr->block) {
sound/pci/emu10k1/memory.c
454
if (p != &hdr->block) {
sound/pci/ice1712/prodigy_hifi.c
173
unsigned int block;
sound/pci/ice1712/prodigy_hifi.c
181
block = (reg << 9) | (data & 0x1ff);
sound/pci/ice1712/prodigy_hifi.c
182
wm8766_spi_send_word(ice, block); /* REGISTER ADDRESS */
sound/pci/ice1712/prodigy_hifi.c
213
unsigned int block;
sound/pci/ice1712/prodigy_hifi.c
219
block = ((AK4396_ADDR & 0x03) << 14) | (1 << 13) |
sound/pci/ice1712/prodigy_hifi.c
221
ak4396_send_word(ice, block); /* REGISTER ADDRESS */
sound/pci/trident/trident_memory.c
116
list_for_each(p, &hdr->block) {
sound/ppc/tumbler.c
200
unsigned char block[6];
sound/ppc/tumbler.c
223
block[0] = (left_vol >> 16) & 0xff;
sound/ppc/tumbler.c
224
block[1] = (left_vol >> 8) & 0xff;
sound/ppc/tumbler.c
225
block[2] = (left_vol >> 0) & 0xff;
sound/ppc/tumbler.c
227
block[3] = (right_vol >> 16) & 0xff;
sound/ppc/tumbler.c
228
block[4] = (right_vol >> 8) & 0xff;
sound/ppc/tumbler.c
229
block[5] = (right_vol >> 0) & 0xff;
sound/ppc/tumbler.c
232
block) < 0) {
sound/ppc/tumbler.c
491
unsigned char block[4];
sound/ppc/tumbler.c
503
block[i] = (vol >> ((info->bytes - i - 1) * 8)) & 0xff;
sound/ppc/tumbler.c
505
info->bytes, block) < 0) {
sound/ppc/tumbler.c
630
unsigned char block[9];
sound/ppc/tumbler.c
642
block[i * 3 + j] = (vol >> ((2 - j) * 8)) & 0xff;
sound/ppc/tumbler.c
645
9, block) < 0) {
sound/soc/codecs/tas2781-fmwlib.c
1076
struct tasdevice_priv *tasdevice, struct tasdev_blk *block)
sound/soc/codecs/tas2781-fmwlib.c
1078
const unsigned int blk_size = block->blk_size;
sound/soc/codecs/tas2781-fmwlib.c
1080
unsigned char *data = block->data;
sound/soc/codecs/tas2781-fmwlib.c
1082
for (i = 0, length = 0; i < block->nr_subblocks; i++) {
sound/soc/codecs/tas2781-fmwlib.c
1084
block->dev_idx, blk_size - length);
sound/soc/codecs/tas2781-fmwlib.c
1162
struct tasdev_blk *block, const struct firmware *fmw, int offset)
sound/soc/codecs/tas2781-fmwlib.c
1172
block->type = get_unaligned_be32(&data[offset]);
sound/soc/codecs/tas2781-fmwlib.c
1181
block->is_pchksum_present = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
1184
block->pchksum = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
1187
block->is_ychksum_present = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
1190
block->ychksum = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
1193
block->is_pchksum_present = 0;
sound/soc/codecs/tas2781-fmwlib.c
1194
block->is_ychksum_present = 0;
sound/soc/codecs/tas2781-fmwlib.c
1197
block->nr_cmds = get_unaligned_be32(&data[offset]);
sound/soc/codecs/tas2781-fmwlib.c
1200
n = block->nr_cmds * 4;
sound/soc/codecs/tas2781-fmwlib.c
1209
block->data = kmemdup(&data[offset], n, GFP_KERNEL);
sound/soc/codecs/tas2781-fmwlib.c
1210
if (!block->data) {
sound/soc/codecs/tas2781-fmwlib.c
1632
struct tasdev_blk *block, int chn, unsigned char book,
sound/soc/codecs/tas2781-fmwlib.c
1653
block->nr_retry--;
sound/soc/codecs/tas2781-fmwlib.c
1654
if (block->nr_retry > 0)
sound/soc/codecs/tas2781-fmwlib.c
1657
set_err_prg_cfg(block->type, &tas_priv->tasdevice[chn]);
sound/soc/codecs/tas2781-fmwlib.c
1664
struct tasdev_blk *block, int chn, unsigned char book,
sound/soc/codecs/tas2781-fmwlib.c
1676
if (block->is_ychksum_present)
sound/soc/codecs/tas2781-fmwlib.c
1677
ret = tasdev_bytes_chksum(tas_priv, block, chn,
sound/soc/codecs/tas2781-fmwlib.c
1684
if (block->is_ychksum_present)
sound/soc/codecs/tas2781-fmwlib.c
1685
ret = tasdev_bytes_chksum(tas_priv, block, chn, book,
sound/soc/codecs/tas2781-fmwlib.c
1689
if (!block->is_ychksum_present || ret >= 0) {
sound/soc/codecs/tas2781-fmwlib.c
1700
struct tasdev_blk *block, int chn)
sound/soc/codecs/tas2781-fmwlib.c
1709
set_err_prg_cfg(block->type, &tas_priv->tasdevice[chn]);
sound/soc/codecs/tas2781-fmwlib.c
1713
if ((nr_value & 0xff) != block->pchksum) {
sound/soc/codecs/tas2781-fmwlib.c
1717
block->pchksum, (nr_value & 0xff));
sound/soc/codecs/tas2781-fmwlib.c
1720
block->nr_retry--;
sound/soc/codecs/tas2781-fmwlib.c
1722
if (block->nr_retry <= 0)
sound/soc/codecs/tas2781-fmwlib.c
1723
set_err_prg_cfg(block->type,
sound/soc/codecs/tas2781-fmwlib.c
1733
struct tasdev_blk *block, int chn)
sound/soc/codecs/tas2781-fmwlib.c
1746
while (block->nr_retry > 0) {
sound/soc/codecs/tas2781-fmwlib.c
1747
if (block->is_pchksum_present) {
sound/soc/codecs/tas2781-fmwlib.c
1754
if (block->is_ychksum_present)
sound/soc/codecs/tas2781-fmwlib.c
1759
while (nr_cmds < block->nr_cmds) {
sound/soc/codecs/tas2781-fmwlib.c
1760
data = block->data + nr_cmds * 4;
sound/soc/codecs/tas2781-fmwlib.c
1775
if (block->is_ychksum_present) {
sound/soc/codecs/tas2781-fmwlib.c
1777
block, chn, book, page, offset,
sound/soc/codecs/tas2781-fmwlib.c
1799
block, chn, book, page, offset, data,
sound/soc/codecs/tas2781-fmwlib.c
1806
if (block->nr_retry > 0)
sound/soc/codecs/tas2781-fmwlib.c
1811
if (block->is_pchksum_present) {
sound/soc/codecs/tas2781-fmwlib.c
1812
ret = tasdev_block_chksum(tas_priv, block, chn);
sound/soc/codecs/tas2781-fmwlib.c
1814
if (block->nr_retry > 0)
sound/soc/codecs/tas2781-fmwlib.c
1820
if (block->is_ychksum_present) {
sound/soc/codecs/tas2781-fmwlib.c
1824
block->ychksum, crc_chksum);
sound/soc/codecs/tas2781-fmwlib.c
1839
struct tasdev_blk *block)
sound/soc/codecs/tas2781-fmwlib.c
1846
switch (block->type) {
sound/soc/codecs/tas2781-fmwlib.c
1877
block->type);
sound/soc/codecs/tas2781-fmwlib.c
1882
block->nr_retry = 6;
sound/soc/codecs/tas2781-fmwlib.c
1885
ret = tasdev_load_blk(tas_priv, block, chn);
sound/soc/codecs/tas2781-fmwlib.c
1888
chn, block->type);
sound/soc/codecs/tas2781-fmwlib.c
2373
struct tasdev_blk *block;
sound/soc/codecs/tas2781-fmwlib.c
2392
block = &(im->dev_blks[blks]);
sound/soc/codecs/tas2781-fmwlib.c
2393
if (!block)
sound/soc/codecs/tas2781-fmwlib.c
2395
kfree(block->data);
sound/soc/codecs/tas2781-fmwlib.c
2453
struct tasdev_blk *block;
sound/soc/codecs/tas2781-fmwlib.c
2458
block = &(dev_data->dev_blks[i]);
sound/soc/codecs/tas2781-fmwlib.c
2459
ret = tas_priv->tasdevice_load_block(tas_priv, block);
sound/soc/codecs/tas2781-fmwlib.c
417
struct tasdev_blk *block)
sound/soc/codecs/tas2781-fmwlib.c
437
if (block->type == p[i].blktyp) {
sound/soc/codecs/tas2781-fmwlib.c
447
struct tasdev_blk *block, const struct firmware *fmw, int offset)
sound/soc/codecs/tas2781-fmwlib.c
460
block->type = get_unaligned_be32(&data[offset]);
sound/soc/codecs/tas2781-fmwlib.c
463
block->is_pchksum_present = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
466
block->pchksum = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
469
block->is_ychksum_present = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
472
block->ychksum = data[offset];
sound/soc/codecs/tas2781-fmwlib.c
475
block->blk_size = get_unaligned_be32(&data[offset]);
sound/soc/codecs/tas2781-fmwlib.c
478
block->nr_subblocks = get_unaligned_be32(&data[offset]);
sound/soc/codecs/tas2781-fmwlib.c
487
block->dev_idx = map_dev_idx(tas_fmw, block);
sound/soc/codecs/tas2781-fmwlib.c
489
if (offset + block->blk_size > fmw->size) {
sound/soc/codecs/tas2781-fmwlib.c
495
block->data = kmemdup(&data[offset], block->blk_size, GFP_KERNEL);
sound/soc/codecs/tas2781-fmwlib.c
496
if (!block->data) {
sound/soc/codecs/tas2781-fmwlib.c
500
offset += block->blk_size;
sound/soc/codecs/wm8994.c
395
static void wm8994_set_retune_mobile(struct snd_soc_component *component, int block)
sound/soc/codecs/wm8994.c
400
int base = wm8994_retune_mobile_base[block];
sound/soc/codecs/wm8994.c
406
switch (block) {
sound/soc/codecs/wm8994.c
420
cfg = wm8994->retune_mobile_cfg[block];
sound/soc/codecs/wm8994.c
435
block,
sound/soc/codecs/wm8994.c
472
int block = wm8994_get_retune_mobile_block(kcontrol->id.name);
sound/soc/codecs/wm8994.c
475
if (block < 0)
sound/soc/codecs/wm8994.c
476
return block;
sound/soc/codecs/wm8994.c
481
wm8994->retune_mobile_cfg[block] = value;
sound/soc/codecs/wm8994.c
483
wm8994_set_retune_mobile(component, block);
sound/soc/codecs/wm8994.c
493
int block = wm8994_get_retune_mobile_block(kcontrol->id.name);
sound/soc/codecs/wm8994.c
495
if (block < 0)
sound/soc/codecs/wm8994.c
496
return block;
sound/soc/codecs/wm8994.c
498
ucontrol->value.enumerated.item[0] = wm8994->retune_mobile_cfg[block];
sound/soc/codecs/wm8996.c
334
static void wm8996_set_retune_mobile(struct snd_soc_component *component, int block)
sound/soc/codecs/wm8996.c
343
switch (block) {
sound/soc/codecs/wm8996.c
366
cfg = wm8996->retune_mobile_cfg[block];
sound/soc/codecs/wm8996.c
381
block,
sound/soc/codecs/wm8996.c
415
int block = wm8996_get_retune_mobile_block(kcontrol->id.name);
sound/soc/codecs/wm8996.c
418
if (block < 0)
sound/soc/codecs/wm8996.c
419
return block;
sound/soc/codecs/wm8996.c
424
wm8996->retune_mobile_cfg[block] = value;
sound/soc/codecs/wm8996.c
426
wm8996_set_retune_mobile(component, block);
sound/soc/codecs/wm8996.c
436
int block = wm8996_get_retune_mobile_block(kcontrol->id.name);
sound/soc/codecs/wm8996.c
438
if (block < 0)
sound/soc/codecs/wm8996.c
439
return block;
sound/soc/codecs/wm8996.c
440
ucontrol->value.enumerated.item[0] = wm8996->retune_mobile_cfg[block];
sound/soc/intel/atom/sst-atom-controls.c
24
u8 ipc_msg, u8 block,
sound/soc/intel/atom/sst-atom-controls.c
32
byte_data->block = block;
sound/soc/intel/atom/sst-atom-controls.c
48
u8 ipc_msg, u8 block, u8 task_id, u8 pipe_id,
sound/soc/intel/atom/sst-atom-controls.c
56
block, task_id, pipe_id, len, cmd_data);
sound/soc/intel/atom/sst-atom-controls.c
73
u8 ipc_msg, u8 block, u8 task_id, u8 pipe_id,
sound/soc/intel/atom/sst-atom-controls.c
79
ret = sst_fill_and_send_cmd_unlocked(drv, ipc_msg, block,
sound/soc/intel/atom/sst-mfld-dsp.h
512
u8 block;
sound/soc/intel/atom/sst/sst.c
514
struct sst_block *block;
sound/soc/intel/atom/sst/sst.c
538
block = sst_create_block(ctx, 0, FW_DWNL_ID);
sound/soc/intel/atom/sst/sst.c
539
if (block == NULL)
sound/soc/intel/atom/sst/sst.c
545
ret = sst_wait_timeout(ctx, block);
sound/soc/intel/atom/sst/sst.c
570
sst_free_block(ctx, block);
sound/soc/intel/atom/sst/sst.h
462
struct sst_block *block);
sound/soc/intel/atom/sst/sst.h
476
struct intel_sst_drv *sst_drv_ctx, struct sst_block **block,
sound/soc/intel/atom/sst/sst_ipc.c
66
struct sst_block *block;
sound/soc/intel/atom/sst/sst_ipc.c
71
list_for_each_entry(block, &ctx->block_list, node) {
sound/soc/intel/atom/sst/sst_ipc.c
72
dev_dbg(ctx->dev, "Block ipc %d, drv_id %d\n", block->msg_id,
sound/soc/intel/atom/sst/sst_ipc.c
73
block->drv_id);
sound/soc/intel/atom/sst/sst_ipc.c
74
if (block->msg_id == ipc && block->drv_id == drv_id) {
sound/soc/intel/atom/sst/sst_ipc.c
76
block->ret_code = result;
sound/soc/intel/atom/sst/sst_ipc.c
77
block->data = data;
sound/soc/intel/atom/sst/sst_ipc.c
78
block->size = size;
sound/soc/intel/atom/sst/sst_ipc.c
79
block->condition = true;
sound/soc/intel/atom/sst/sst_ipc.c
94
struct sst_block *block, *__block;
sound/soc/intel/atom/sst/sst_ipc.c
98
list_for_each_entry_safe(block, __block, &ctx->block_list, node) {
sound/soc/intel/atom/sst/sst_ipc.c
99
if (block == freed) {
sound/soc/intel/atom/sst/sst_loader.c
175
struct fw_block_info *block;
sound/soc/intel/atom/sst/sst_loader.c
185
block = (void *)module + sizeof(*module);
sound/soc/intel/atom/sst/sst_loader.c
188
if (block->size <= 0) {
sound/soc/intel/atom/sst/sst_loader.c
192
switch (block->type) {
sound/soc/intel/atom/sst/sst_loader.c
203
block = (void *)block + sizeof(*block) + block->size;
sound/soc/intel/atom/sst/sst_loader.c
207
block->type, count);
sound/soc/intel/atom/sst/sst_loader.c
212
ram_iomem + block->ram_offset,
sound/soc/intel/atom/sst/sst_loader.c
213
(void *)block + sizeof(*block), block->size, 1);
sound/soc/intel/atom/sst/sst_loader.c
217
block = (void *)block + sizeof(*block) + block->size;
sound/soc/intel/atom/sst/sst_loader.c
395
struct sst_block *block;
sound/soc/intel/atom/sst/sst_loader.c
409
block = sst_create_block(sst_drv_ctx, 0, FW_DWNL_ID);
sound/soc/intel/atom/sst/sst_loader.c
410
if (block == NULL)
sound/soc/intel/atom/sst/sst_loader.c
433
ret_val = sst_wait_timeout(sst_drv_ctx, block);
sound/soc/intel/atom/sst/sst_loader.c
445
sst_free_block(sst_drv_ctx, block);
sound/soc/intel/atom/sst/sst_pvt.c
100
block->ret_code);
sound/soc/intel/atom/sst/sst_pvt.c
101
retval = -block->ret_code;
sound/soc/intel/atom/sst/sst_pvt.c
103
block->on = false;
sound/soc/intel/atom/sst/sst_pvt.c
106
block->condition, block->msg_id, sst_drv_ctx->sst_state);
sound/soc/intel/atom/sst/sst_pvt.c
154
struct intel_sst_drv *sst_drv_ctx, struct sst_block **block,
sound/soc/intel/atom/sst/sst_pvt.c
162
*block = sst_create_block(sst_drv_ctx, msg_id, drv_id);
sound/soc/intel/atom/sst/sst_pvt.c
163
if (*block == NULL) {
sound/soc/intel/atom/sst/sst_pvt.c
192
struct sst_block *block = NULL;
sound/soc/intel/atom/sst/sst_pvt.c
203
&msg, large, sst, &block, ipc_msg, pvt_id);
sound/soc/intel/atom/sst/sst_pvt.c
239
ret = sst_wait_timeout(sst, block);
sound/soc/intel/atom/sst/sst_pvt.c
243
if (data && block->data) {
sound/soc/intel/atom/sst/sst_pvt.c
244
*data = kmemdup(block->data, block->size, GFP_KERNEL);
sound/soc/intel/atom/sst/sst_pvt.c
253
sst_free_block(sst, block);
sound/soc/intel/atom/sst/sst_pvt.c
81
int sst_wait_timeout(struct intel_sst_drv *sst_drv_ctx, struct sst_block *block)
sound/soc/intel/atom/sst/sst_pvt.c
92
block->condition, block->msg_id, block->drv_id);
sound/soc/intel/atom/sst/sst_pvt.c
94
block->condition,
sound/soc/intel/atom/sst/sst_pvt.c
98
block->condition);
sound/soc/intel/atom/sst/sst_stream.c
175
struct sst_block *block = NULL;
sound/soc/intel/atom/sst/sst_stream.c
176
u8 bytes_block = bytes->block;
sound/soc/intel/atom/sst/sst_stream.c
195
block = sst_create_block(sst_drv_ctx, bytes->ipc_msg, pvt_id);
sound/soc/intel/atom/sst/sst_stream.c
196
if (block == NULL) {
sound/soc/intel/atom/sst/sst_stream.c
208
ret = sst_wait_timeout(sst_drv_ctx, block);
sound/soc/intel/atom/sst/sst_stream.c
211
sst_free_block(sst_drv_ctx, block);
sound/soc/intel/atom/sst/sst_stream.c
221
unsigned char *r = block->data;
sound/soc/intel/atom/sst/sst_stream.c
229
sst_free_block(sst_drv_ctx, block);
sound/soc/sof/intel/telemetry.c
22
struct xtensa_arch_block *block;
sound/soc/sof/intel/telemetry.c
42
block = kmalloc_obj(*block);
sound/soc/sof/intel/telemetry.c
43
if (!block)
sound/soc/sof/intel/telemetry.c
46
sof_mailbox_read(sdev, slot_offset + sizeof(*telemetry_data), block, sizeof(*block));
sound/soc/sof/intel/telemetry.c
47
if (block->soc != XTENSA_SOC_INTEL_ADSP) {
sound/soc/sof/intel/telemetry.c
48
dev_err(sdev->dev, "%s invalid SOC %d\n", invalid_slot_msg, block->soc);
sound/soc/sof/intel/telemetry.c
62
switch (block->toolchain) {
sound/soc/sof/intel/telemetry.c
78
xoops->exccause = block->exccause;
sound/soc/sof/intel/telemetry.c
79
xoops->excvaddr = block->excvaddr;
sound/soc/sof/intel/telemetry.c
80
xoops->epc1 = block->pc;
sound/soc/sof/intel/telemetry.c
81
xoops->ps = block->ps;
sound/soc/sof/intel/telemetry.c
82
xoops->sar = block->sar;
sound/soc/sof/intel/telemetry.c
85
memcpy((void *)xoops->ar, block->ar, XTENSA_CORE_AR_REGS_COUNT * sizeof(u32));
sound/soc/sof/intel/telemetry.c
92
kfree(block);
sound/soc/sof/ipc3-loader.c
238
struct snd_sof_blk_hdr *block;
sound/soc/sof/ipc3-loader.c
246
block = (struct snd_sof_blk_hdr *)((u8 *)module + sizeof(*module));
sound/soc/sof/ipc3-loader.c
252
if (remaining < sizeof(*block)) {
sound/soc/sof/ipc3-loader.c
258
remaining -= sizeof(*block);
sound/soc/sof/ipc3-loader.c
260
if (block->size == 0) {
sound/soc/sof/ipc3-loader.c
264
block->type, block->offset);
sound/soc/sof/ipc3-loader.c
268
switch (block->type) {
sound/soc/sof/ipc3-loader.c
275
offset = block->offset;
sound/soc/sof/ipc3-loader.c
279
__func__, block->type, count);
sound/soc/sof/ipc3-loader.c
284
count, block->type, block->size, offset);
sound/soc/sof/ipc3-loader.c
287
if (block->size % sizeof(u32)) {
sound/soc/sof/ipc3-loader.c
289
__func__, block->size);
sound/soc/sof/ipc3-loader.c
292
ret = snd_sof_dsp_block_write(sdev, block->type, offset,
sound/soc/sof/ipc3-loader.c
293
block + 1, block->size);
sound/soc/sof/ipc3-loader.c
296
__func__, block->type);
sound/soc/sof/ipc3-loader.c
300
if (remaining < block->size) {
sound/soc/sof/ipc3-loader.c
306
remaining -= block->size;
sound/soc/sof/ipc3-loader.c
308
block = (struct snd_sof_blk_hdr *)((u8 *)block + sizeof(*block)
sound/soc/sof/ipc3-loader.c
309
+ block->size);
sound/synth/emux/emux_synth.c
102
vp->block = vp->zone->sample->block;
sound/synth/emux/emux_synth.c
104
vp->block = NULL;
sound/synth/emux/emux_synth.c
480
vp->block = NULL;
sound/synth/util_mem.c
107
if (prev == &hdr->block)
sound/synth/util_mem.c
34
INIT_LIST_HEAD(&hdr->block);
sound/synth/util_mem.c
49
while ((p = hdr->block.next) != &hdr->block) {
sound/synth/util_mem.c
78
list_for_each(p, &hdr->block) {
tools/lib/subcmd/run-command.c
170
static int wait_or_whine(struct child_process *cmd, bool block)
tools/lib/subcmd/run-command.c
177
pid_t waiting = waitpid(cmd->pid, &status, block ? 0 : WNOHANG);
tools/lib/subcmd/run-command.c
179
if (!block && waiting == 0)
tools/perf/util/annotate-data.c
1568
Dwarf_Block block;
tools/perf/util/annotate-data.c
1572
dwarf_formblock(&attr, &block) == 0 && block.length == 1) {
tools/perf/util/annotate-data.c
1573
switch (*block.data) {
tools/perf/util/annotate-data.c
1575
fbreg = dloc->fbreg = *block.data - DW_OP_reg0;
tools/testing/selftests/bpf/prog_tests/tc_links.c
1798
static int qdisc_replace(int ifindex, const char *kind, bool block)
tools/testing/selftests/bpf/prog_tests/tc_links.c
1817
if (block)
tools/testing/selftests/bpf/progs/refcounted_kptr.c
466
bpf_spin_lock(&block);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
470
bpf_spin_unlock(&block);
tools/testing/selftests/bpf/progs/refcounted_kptr.c
48
private(C) struct bpf_spin_lock block;
tools/testing/selftests/bpf/progs/stream.c
275
bpf_spin_lock(&block);
tools/testing/selftests/bpf/progs/stream.c
284
bpf_spin_unlock(&block);
tools/testing/selftests/bpf/progs/stream.c
47
private(STREAM) struct bpf_spin_lock block;
tools/testing/selftests/drivers/net/hw/toeplitz.c
234
struct tpacket_block_desc *block;
tools/testing/selftests/drivers/net/hw/toeplitz.c
238
block = (void *)(ring->mmap + ring->idx * ring_block_sz);
tools/testing/selftests/drivers/net/hw/toeplitz.c
239
if (!(block->hdr.bh1.block_status & TP_STATUS_USER))
tools/testing/selftests/drivers/net/hw/toeplitz.c
242
frame = (char *)block;
tools/testing/selftests/drivers/net/hw/toeplitz.c
243
frame += block->hdr.bh1.offset_to_first_pkt;
tools/testing/selftests/drivers/net/hw/toeplitz.c
245
for (i = 0; i < block->hdr.bh1.num_pkts; i++) {
tools/testing/selftests/drivers/net/hw/toeplitz.c
250
block->hdr.bh1.block_status = TP_STATUS_KERNEL;