arch/alpha/kernel/smc37c669.c
1976
wb( &SMC37c669->index_port, SMC37c669_CONFIG_ON_KEY );
arch/alpha/kernel/smc37c669.c
1977
wb( &SMC37c669->index_port, SMC37c669_CONFIG_ON_KEY );
arch/alpha/kernel/smc37c669.c
1981
wb( &SMC37c669->index_port, SMC37c669_CONFIG_OFF_KEY );
arch/alpha/kernel/smc37c669.c
2011
wb(&SMC37c669->index_port, index);
arch/alpha/kernel/smc37c669.c
2045
wb( &SMC37c669->index_port, index );
arch/alpha/kernel/smc37c669.c
2046
wb( &SMC37c669->data_port, data );
arch/mips/include/asm/sn/sn0/hubio.h
892
wb: 1, /* 15: writeback pending. */
arch/s390/include/asm/pci_clp.h
95
u64 wb;
arch/s390/pci/pci_clp.c
190
zdev->bars[i].mio_wb = (void __iomem *) response->mio.addr[i].wb;
arch/xtensa/kernel/ptrace.c
303
unsigned long wb = regs->windowbase;
arch/xtensa/kernel/ptrace.c
305
tmp = ((ws >> wb) | (ws << (WSBITS - wb))) &
arch/xtensa/kernel/signal.c
120
regs->windowstart = 1 << wb;
arch/xtensa/kernel/signal.c
60
const unsigned long wb = regs->windowbase;
arch/xtensa/kernel/signal.c
73
wm = (ws >> wb) | (ws << (XCHAL_NUM_AREGS / 4 - wb));
certs/extract-cert.c
166
if (wb && !x509) {
certs/extract-cert.c
179
BIO_free(wb);
certs/extract-cert.c
47
static BIO *wb;
certs/extract-cert.c
55
if (!wb) {
certs/extract-cert.c
56
wb = BIO_new_file(cert_dst, "wb");
certs/extract-cert.c
57
ERR(!wb, "%s", cert_dst);
certs/extract-cert.c
60
ERR(!i2d_X509_bio(wb, x509), "%s", cert_dst);
drivers/char/ps3flash.c
258
int wb;
drivers/char/ps3flash.c
265
wb = ps3flash_writeback(ps3flash_dev);
drivers/char/ps3flash.c
266
if (wb)
drivers/char/ps3flash.c
267
return wb;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
202
u32 wb = 0;
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
418
wb = le32_to_cpup((__le32 *)(op->buf + nwait * 4));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
419
wb &= GENMASK((nbw * 8) - 1, 0);
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
426
wb |= ((1 << 7) << (nbw * 8));
drivers/crypto/allwinner/sun4i-ss/sun4i-ss-hash.c
427
((__le32 *)bf)[j++] = cpu_to_le32(wb);
drivers/gpu/drm/amd/amdgpu/amdgpu.h
494
uint32_t *wb;
drivers/gpu/drm/amd/amdgpu/amdgpu.h
525
int amdgpu_device_wb_get(struct amdgpu_device *adev, u32 *wb);
drivers/gpu/drm/amd/amdgpu/amdgpu.h
526
void amdgpu_device_wb_free(struct amdgpu_device *adev, u32 wb);
drivers/gpu/drm/amd/amdgpu/amdgpu.h
961
struct amdgpu_wb wb;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1714
if (adev->wb.wb_obj) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1715
amdgpu_bo_free_kernel(&adev->wb.wb_obj,
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1716
&adev->wb.gpu_addr,
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1717
(void **)&adev->wb.wb);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1718
adev->wb.wb_obj = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1735
if (adev->wb.wb_obj == NULL) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1739
&adev->wb.wb_obj, &adev->wb.gpu_addr,
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1740
(void **)&adev->wb.wb);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1746
adev->wb.num_wb = AMDGPU_MAX_WB;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1747
memset(&adev->wb.used, 0, sizeof(adev->wb.used));
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1750
memset((char *)adev->wb.wb, 0, AMDGPU_MAX_WB * sizeof(uint32_t) * 8);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1765
int amdgpu_device_wb_get(struct amdgpu_device *adev, u32 *wb)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1769
spin_lock_irqsave(&adev->wb.lock, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1770
offset = find_first_zero_bit(adev->wb.used, adev->wb.num_wb);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1771
if (offset < adev->wb.num_wb) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1772
__set_bit(offset, adev->wb.used);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1773
spin_unlock_irqrestore(&adev->wb.lock, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1774
*wb = offset << 3; /* convert to dw offset */
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1777
spin_unlock_irqrestore(&adev->wb.lock, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1790
void amdgpu_device_wb_free(struct amdgpu_device *adev, u32 wb)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1794
wb >>= 3;
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1795
spin_lock_irqsave(&adev->wb.lock, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1796
if (wb < adev->wb.num_wb)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1797
__clear_bit(wb, adev->wb.used);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
1798
spin_unlock_irqrestore(&adev->wb.lock, flags);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
4493
spin_lock_init(&adev->wb.lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
1123
value = adev->wb.wb[reg_val_offs];
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
100
ih->rptr_addr = adev->wb.gpu_addr + rptr_offs * 4;
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
101
ih->rptr_cpu = &adev->wb.wb[rptr_offs];
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
98
ih->wptr_addr = adev->wb.gpu_addr + wptr_offs * 4;
drivers/gpu/drm/amd/amdgpu/amdgpu_ih.c
99
ih->wptr_cpu = &adev->wb.wb[wptr_offs];
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
193
adev->wb.gpu_addr + (adev->mes.sch_ctx_offs[i] * 4);
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
195
(uint64_t *)&adev->wb.wb[adev->mes.sch_ctx_offs[i]];
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
205
adev->mes.query_status_fence_gpu_addr[i] = adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
208
(uint64_t *)&adev->wb.wb[adev->mes.query_status_fence_offs[i]];
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
482
read_val_gpu_addr = adev->wb.gpu_addr + (addr_offset * 4);
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
483
read_val_ptr = (uint32_t *)&adev->wb.wb[addr_offset];
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c
227
(ring->adev->wb.gpu_addr + offset * 4)
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c
230
(&ring->adev->wb.wb[offset])
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h
441
#define amdgpu_ring_emit_gds_switch(r, v, db, ds, wb, ws, ab, as) (r)->funcs->emit_gds_switch((r), (v), (db), (ds), (wb), (ws), (ab), (as))
drivers/gpu/drm/amd/amdgpu/amdgpu_umsch_mm.c
281
adev->umsch_mm.sch_ctx_gpu_addr = adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
788
adev->wb.wb[index] = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
789
wb_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
804
if (le32_to_cpu(adev->wb.wb[index]) == test_pattern)
drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
832
adev->wb.wb[index] = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
833
wb_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/amdgpu_vpe.c
859
ret = (le32_to_cpu(adev->wb.wb[index]) == test_pattern) ? 0 : -EINVAL;
drivers/gpu/drm/amd/amdgpu/cik_sdma.c
613
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/cik_sdma.c
615
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/cik_sdma.c
629
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/cik_sdma.c
666
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/cik_sdma.c
668
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/cik_sdma.c
693
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4087
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4088
adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4089
cpu_ptr = &adev->wb.wb[index];
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
8988
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
8990
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
625
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
626
adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6268
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
627
cpu_ptr = &adev->wb.wb[index];
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6270
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
4642
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
4644
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
514
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
515
adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
516
cpu_ptr = &adev->wb.wb[index];
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
328
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
329
adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
330
cpu_ptr = &adev->wb.wb[index];
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
3457
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
3459
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
6343
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
6345
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
883
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
884
adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
910
tmp = adev->wb.wb[index];
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
1239
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
1240
adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
1266
tmp = adev->wb.wb[index];
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4223
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4225
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4257
value = (uint64_t)adev->wb.wb[reg_val_offs] |
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4258
(uint64_t)adev->wb.wb[reg_val_offs + 1 ] << 32ULL;
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
5888
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
5890
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
1918
wb_gpu_addr = adev->wb.gpu_addr + (ring->rptr_offs * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
1924
wb_gpu_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
204
uint64_t wptr_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2190
atomic64_set((atomic64_t *)&adev->wb.wb[ring->wptr_offs], 0);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2923
return ring->adev->wb.wb[ring->rptr_offs]; /* gfx9 hardware is 32bit rptr */
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2932
wptr = atomic64_read((atomic64_t *)&ring->adev->wb.wb[ring->wptr_offs]);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2944
atomic64_set((atomic64_t *)&adev->wb.wb[ring->wptr_offs], ring->wptr);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2991
amdgpu_ring_write(ring, lower_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2993
amdgpu_ring_write(ring, upper_32_bits(adev->wb.gpu_addr +
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
466
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
467
adev->wb.wb[index] = cpu_to_le32(0xCAFEDEAD);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
493
tmp = adev->wb.wb[index];
drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c
696
return adev->wb.wb[ring->wptr_offs];
drivers/gpu/drm/amd/amdgpu/jpeg_v4_0_3.c
721
adev->wb.wb[ring->wptr_offs] = lower_32_bits(ring->wptr);
drivers/gpu/drm/amd/amdgpu/jpeg_v5_0_1.c
629
return adev->wb.wb[ring->wptr_offs];
drivers/gpu/drm/amd/amdgpu/jpeg_v5_0_1.c
647
adev->wb.wb[ring->wptr_offs] = lower_32_bits(ring->wptr);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
471
fence_gpu_addr = adev->wb.gpu_addr + (fence_offset * 4);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
472
fence_ptr = (u64 *)&adev->wb.wb[fence_offset];
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
199
status_gpu_addr = adev->wb.gpu_addr + (status_offset * 4);
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
200
status_ptr = (u64 *)&adev->wb.wb[status_offset];
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
182
status_gpu_addr = adev->wb.gpu_addr + (status_offset * 4);
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
183
status_ptr = (u64 *)&adev->wb.wb[status_offset];
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
179
status_gpu_addr = adev->wb.gpu_addr + (status_offset * 4);
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
180
status_ptr = (u64 *)&adev->wb.wb[status_offset];
drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
544
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
546
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
561
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
598
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
600
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v2_4.c
629
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
387
u32 *wb = (u32 *)ring->wptr_cpu_addr;
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
389
WRITE_ONCE(*wb, ring->wptr << 2);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
392
u32 *wb = (u32 *)ring->wptr_cpu_addr;
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
394
WRITE_ONCE(*wb, ring->wptr << 2);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
818
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
820
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
835
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
872
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
874
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v3_0.c
903
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
1476
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
1478
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
1493
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
1530
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
1532
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
1561
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
704
u64 *wb = (u64 *)ring->wptr_cpu_addr;
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
714
WRITE_ONCE(*wb, (ring->wptr << 2));
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
769
u64 *wb = (u64 *)ring->wptr_cpu_addr;
drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c
772
WRITE_ONCE(*wb, (ring->wptr << 2));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
1072
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
1074
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
1089
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
1126
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
1128
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
1157
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
225
rptr = READ_ONCE(*((u64 *)&ring->adev->wb.wb[ring->rptr_offs]));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
245
wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
271
u64 *wb = (u64 *)&adev->wb.wb[ring->wptr_offs];
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
281
WRITE_ONCE(*wb, (ring->wptr << 2));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
314
wptr = READ_ONCE(*((u64 *)&adev->wb.wb[ring->wptr_offs]));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
336
u64 *wb = (u64 *)&adev->wb.wb[ring->wptr_offs];
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
339
WRITE_ONCE(*wb, (ring->wptr << 2));
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
704
upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
706
lower_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
758
wptr_gpu_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
831
upper_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
833
lower_32_bits(adev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC);
drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c
865
wptr_gpu_addr = adev->wb.gpu_addr + (ring->wptr_offs * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1029
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1030
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1048
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1093
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1094
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c
1128
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
1027
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
929
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
930
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
948
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
993
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v5_2.c
994
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1000
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1001
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1034
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
936
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
937
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
955
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1016
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1017
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1050
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
952
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
953
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
971
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1006
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1007
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1040
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
942
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
943
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
961
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/si_dma.c
221
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/si_dma.c
223
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/si_dma.c
236
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/amdgpu/si_dma.c
273
gpu_addr = adev->wb.gpu_addr + (index * 4);
drivers/gpu/drm/amd/amdgpu/si_dma.c
275
adev->wb.wb[index] = cpu_to_le32(tmp);
drivers/gpu/drm/amd/amdgpu/si_dma.c
298
tmp = le32_to_cpu(adev->wb.wb[index]);
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1006
pipes[pipe_cnt].dout.wb.wb_src_height = wb_info->dwb_params.cnv_params.crop_height;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1007
pipes[pipe_cnt].dout.wb.wb_src_width = wb_info->dwb_params.cnv_params.crop_width;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1008
pipes[pipe_cnt].dout.wb.wb_dst_width = wb_info->dwb_params.dest_width;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1009
pipes[pipe_cnt].dout.wb.wb_dst_height = wb_info->dwb_params.dest_height;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1010
pipes[pipe_cnt].dout.wb.wb_htaps_luma = 1;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1011
pipes[pipe_cnt].dout.wb.wb_vtaps_luma = 1;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1012
pipes[pipe_cnt].dout.wb.wb_htaps_chroma = wb_info->dwb_params.scaler_taps.h_taps_c;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1013
pipes[pipe_cnt].dout.wb.wb_vtaps_chroma = wb_info->dwb_params.scaler_taps.v_taps_c;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1014
pipes[pipe_cnt].dout.wb.wb_hratio = 1.0;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1015
pipes[pipe_cnt].dout.wb.wb_vratio = 1.0;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1018
pipes[pipe_cnt].dout.wb.wb_pixel_format = dm_420_8;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1020
pipes[pipe_cnt].dout.wb.wb_pixel_format = dm_420_10;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
1022
pipes[pipe_cnt].dout.wb.wb_pixel_format = dm_444_32;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2551
pipes[pipe_cnt].dout.wb = dout_wb;
drivers/gpu/drm/amd/display/dc/dml/dcn30/dcn30_fpu.c
262
pipes[pipe_cnt].dout.wb = dout_wb;
drivers/gpu/drm/amd/display/dc/dml/display_mode_structs.h
479
struct writeback_st wb;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
662
dout->wb.wb_src_height;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
664
dout->wb.wb_src_width;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
666
dout->wb.wb_dst_width;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
668
dout->wb.wb_dst_height;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
670
dout->wb.wb_hratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
672
dout->wb.wb_vratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
674
(enum source_format_class) (dout->wb.wb_pixel_format);
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
676
dout->wb.wb_htaps_luma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
678
dout->wb.wb_vtaps_luma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
680
dout->wb.wb_htaps_luma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
682
dout->wb.wb_vtaps_luma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
684
dout->wb.wb_htaps_chroma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
686
dout->wb.wb_vtaps_chroma;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
688
dout->wb.wb_hratio;
drivers/gpu/drm/amd/display/dc/dml/display_mode_vba.c
690
dout->wb.wb_vratio;
drivers/gpu/drm/i915/i915_active.c
638
struct wait_barrier *wb = container_of(wq, typeof(*wb), base);
drivers/gpu/drm/i915/i915_active.c
640
if (i915_active_is_idle(wb->ref)) {
drivers/gpu/drm/i915/i915_active.c
651
struct wait_barrier *wb;
drivers/gpu/drm/i915/i915_active.c
653
wb = kmalloc_obj(*wb);
drivers/gpu/drm/i915/i915_active.c
654
if (unlikely(!wb))
drivers/gpu/drm/i915/i915_active.c
659
kfree(wb);
drivers/gpu/drm/i915/i915_active.c
663
wb->base.flags = 0;
drivers/gpu/drm/i915/i915_active.c
664
wb->base.func = barrier_wake;
drivers/gpu/drm/i915/i915_active.c
665
wb->base.private = fence;
drivers/gpu/drm/i915/i915_active.c
666
wb->ref = ref;
drivers/gpu/drm/i915/i915_active.c
668
add_wait_queue(__var_waitqueue(ref), &wb->base);
drivers/gpu/drm/lima/lima_pp.c
158
static void lima_pp_write_frame(struct lima_ip *ip, u32 *frame, u32 *wb)
drivers/gpu/drm/lima/lima_pp.c
167
writel(wb[n++], ip->iomem + LIMA_PP_WB(i) + j * 4);
drivers/gpu/drm/lima/lima_pp.c
377
lima_pp_write_frame(ip, frame->frame, frame->wb);
drivers/gpu/drm/lima/lima_pp.c
404
lima_pp_write_frame(ip, frame->frame, frame->wb);
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_10_0_sm8650.h
443
.wb = sm8650_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_12_0_sm8750.h
484
.wb = sm8750_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_12_2_glymur.h
531
.wb = glymur_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_13_0_kaanapali.h
482
.wb = kaanapali_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_5_0_sm8150.h
381
.wb = sm8150_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_5_1_sc8180x.h
405
.wb = sc8180x_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_5_2_sm7150.h
311
.wb = sm7150_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_5_3_sm6150.h
246
.wb = sm6150_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_5_4_sm6125.h
217
.wb = sm6125_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_6_0_sm8250.h
384
.wb = sm8250_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_6_2_sc7180.h
213
.wb = sc7180_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_6_4_sm6350.h
229
.wb = sm6350_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_7_0_sm8350.h
392
.wb = sm8350_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_7_2_sc7280.h
263
.wb = sc7280_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_8_1_sm8450.h
405
.wb = sm8450_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_8_4_sa8775p.h
445
.wb = sa8775p_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_9_0_sm8550.h
400
.wb = sm8550_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_9_1_sar2130p.h
400
.wb = sar2130p_wb,
drivers/gpu/drm/msm/disp/dpu1/catalog/dpu_9_2_x1e80100.h
441
.wb = x1e80100_wb,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c
2335
intf_cfg.wb = phys_enc->hw_wb->idx;
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_wb.c
241
intf_cfg.wb = hw_wb->idx;
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_wb.c
262
intf_cfg.wb = hw_wb->idx;
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_wb.c
38
static bool _dpu_encoder_phys_wb_clk_force_ctrl(struct dpu_hw_wb *wb,
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_wb.c
42
if (wb->ops.setup_clk_force_ctrl) {
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_wb.c
43
*forced_on = wb->ops.setup_clk_force_ctrl(wb, enable);
drivers/gpu/drm/msm/disp/dpu1/dpu_encoder_phys_wb.c
48
*forced_on = mdp->ops.setup_clk_force_ctrl(mdp, wb->caps->clk_ctrl, enable);
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_catalog.h
750
const struct dpu_wb_cfg *wb;
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
303
enum dpu_wb wb)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
305
switch (wb) {
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
324
enum dpu_wb wb)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
326
ctx->pending_wb_flush_mask |= BIT(wb - WB_0);
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
600
if (cfg->wb)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
601
wb_active |= BIT(cfg->wb - WB_0);
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
639
if (cfg->wb)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
640
intf_cfg |= (cfg->wb & 0x3) + 2;
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
718
if (cfg->wb) {
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.c
720
wb_active &= ~BIT(cfg->wb - WB_0);
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_ctl.h
52
enum dpu_wb wb;
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_wb.c
126
static void dpu_hw_wb_roi(struct dpu_hw_wb *ctx, struct dpu_hw_wb_cfg *wb)
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_wb.c
131
image_size = (wb->dest.height << 16) | wb->dest.width;
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_wb.c
133
out_size = (drm_rect_height(&wb->roi) << 16) | drm_rect_width(&wb->roi);
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_wb.h
37
struct dpu_hw_wb_cfg *wb);
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_wb.h
40
struct dpu_hw_wb_cfg *wb,
drivers/gpu/drm/msm/disp/dpu1/dpu_hw_wb.h
44
struct dpu_hw_wb_cfg *wb);
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
1021
msm_disp_snapshot_add_block(disp_state, cat->wb[i].len,
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
1022
dpu_kms->mmio + cat->wb[i].base, "%s",
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
1023
cat->wb[i].name);
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
785
if (dpu_kms->catalog->wb[i].id == WB_2) {
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
787
dpu_kms->catalog->wb[i].format_list,
drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c
788
dpu_kms->catalog->wb[i].num_formats);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
117
const struct dpu_wb_cfg *wb = &cat->wb[i];
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
119
hw = dpu_hw_wb_init(dev, wb, mmio, cat->mdss_ver);
drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c
125
rm->hw_wb[wb->id - WB_0] = hw;
drivers/gpu/drm/msm/disp/mdp5/mdp5_cfg.h
111
struct mdp5_wb_block wb;
drivers/gpu/drm/radeon/cik.c
3739
} else if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/cik.c
4064
WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
drivers/gpu/drm/radeon/cik.c
4082
WREG32(CP_RB0_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC);
drivers/gpu/drm/radeon/cik.c
4083
WREG32(CP_RB0_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/cik.c
4088
if (!rdev->wb.enabled)
drivers/gpu/drm/radeon/cik.c
4118
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/cik.c
4119
rptr = rdev->wb.wb[ring->rptr_offs/4];
drivers/gpu/drm/radeon/cik.c
4144
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/cik.c
4145
rptr = rdev->wb.wb[ring->rptr_offs/4];
drivers/gpu/drm/radeon/cik.c
4162
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/cik.c
4164
wptr = rdev->wb.wb[ring->wptr_offs/4];
drivers/gpu/drm/radeon/cik.c
4180
rdev->wb.wb[ring->wptr_offs/4] = ring->wptr;
drivers/gpu/drm/radeon/cik.c
4677
wb_gpu_addr = rdev->wb.gpu_addr + CIK_WB_CP1_WPTR_OFFSET;
drivers/gpu/drm/radeon/cik.c
4679
wb_gpu_addr = rdev->wb.gpu_addr + CIK_WB_CP2_WPTR_OFFSET;
drivers/gpu/drm/radeon/cik.c
4688
wb_gpu_addr = rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET;
drivers/gpu/drm/radeon/cik.c
4690
wb_gpu_addr = rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET;
drivers/gpu/drm/radeon/cik.c
6975
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/cik.c
6979
WREG32(IH_RB_WPTR_ADDR_LO, (rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFFFFFFFC);
drivers/gpu/drm/radeon/cik.c
6980
WREG32(IH_RB_WPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/cik.c
7485
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/cik.c
7486
wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]);
drivers/gpu/drm/radeon/cik_sdma.c
138
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/cik_sdma.c
400
upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFF);
drivers/gpu/drm/radeon/cik_sdma.c
402
((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC));
drivers/gpu/drm/radeon/cik_sdma.c
404
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/cik_sdma.c
658
gpu_addr = rdev->wb.gpu_addr + index;
drivers/gpu/drm/radeon/cik_sdma.c
661
rdev->wb.wb[index/4] = cpu_to_le32(tmp);
drivers/gpu/drm/radeon/cik_sdma.c
67
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/cik_sdma.c
676
tmp = le32_to_cpu(rdev->wb.wb[index/4]);
drivers/gpu/drm/radeon/cik_sdma.c
68
rptr = rdev->wb.wb[ring->rptr_offs/4];
drivers/gpu/drm/radeon/cik_sdma.c
715
gpu_addr = rdev->wb.gpu_addr + index;
drivers/gpu/drm/radeon/cik_sdma.c
718
rdev->wb.wb[index/4] = cpu_to_le32(tmp);
drivers/gpu/drm/radeon/cik_sdma.c
750
tmp = le32_to_cpu(rdev->wb.wb[index/4]);
drivers/gpu/drm/radeon/evergreen.c
2948
} else if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/evergreen.c
3107
((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC));
drivers/gpu/drm/radeon/evergreen.c
3108
WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/evergreen.c
3109
WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
drivers/gpu/drm/radeon/evergreen.c
3111
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/evergreen.c
4686
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/evergreen.c
4687
wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]);
drivers/gpu/drm/radeon/evergreen_dma.c
71
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/ni.c
1453
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/ni.c
1454
rptr = rdev->wb.wb[ring->rptr_offs/4];
drivers/gpu/drm/radeon/ni.c
1659
WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
drivers/gpu/drm/radeon/ni.c
1676
addr = rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET;
drivers/gpu/drm/radeon/ni_dma.c
127
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/ni_dma.c
222
upper_32_bits(rdev->wb.gpu_addr + wb_offset) & 0xFF);
drivers/gpu/drm/radeon/ni_dma.c
224
((rdev->wb.gpu_addr + wb_offset) & 0xFFFFFFFC));
drivers/gpu/drm/radeon/ni_dma.c
226
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/ni_dma.c
57
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/ni_dma.c
58
rptr = rdev->wb.wb[ring->rptr_offs/4];
drivers/gpu/drm/radeon/r100.c
1098
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r100.c
1099
rptr = le32_to_cpu(rdev->wb.wb[ring->rptr_offs/4]);
drivers/gpu/drm/radeon/r100.c
1217
S_00070C_RB_RPTR_ADDR((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) >> 2));
drivers/gpu/drm/radeon/r100.c
1218
WREG32(R_000774_SCRATCH_ADDR, rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET);
drivers/gpu/drm/radeon/r100.c
1220
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r600.c
2623
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r600.c
2624
rptr = rdev->wb.wb[ring->rptr_offs/4];
drivers/gpu/drm/radeon/r600.c
2749
((rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC));
drivers/gpu/drm/radeon/r600.c
2750
WREG32(CP_RB_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/r600.c
2751
WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
drivers/gpu/drm/radeon/r600.c
2753
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r600.c
2877
if (rdev->wb.use_event) {
drivers/gpu/drm/radeon/r600.c
3377
} else if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/r600.c
3714
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r600.c
3718
WREG32(IH_RB_WPTR_ADDR_LO, (rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFFFFFFFC);
drivers/gpu/drm/radeon/r600.c
3719
WREG32(IH_RB_WPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/r600.c
4041
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r600.c
4042
wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]);
drivers/gpu/drm/radeon/r600_dma.c
143
upper_32_bits(rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/r600_dma.c
145
((rdev->wb.gpu_addr + R600_WB_DMA_RPTR_OFFSET) & 0xFFFFFFFC));
drivers/gpu/drm/radeon/r600_dma.c
147
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r600_dma.c
243
gpu_addr = rdev->wb.gpu_addr + index;
drivers/gpu/drm/radeon/r600_dma.c
246
rdev->wb.wb[index/4] = cpu_to_le32(tmp);
drivers/gpu/drm/radeon/r600_dma.c
260
tmp = le32_to_cpu(rdev->wb.wb[index/4]);
drivers/gpu/drm/radeon/r600_dma.c
350
gpu_addr = rdev->wb.gpu_addr + index;
drivers/gpu/drm/radeon/r600_dma.c
381
tmp = le32_to_cpu(rdev->wb.wb[index/4]);
drivers/gpu/drm/radeon/r600_dma.c
408
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/r600_dma.c
55
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/r600_dma.c
56
rptr = rdev->wb.wb[ring->rptr_offs/4];
drivers/gpu/drm/radeon/radeon.h
1125
volatile uint32_t *wb;
drivers/gpu/drm/radeon/radeon.h
2379
struct radeon_wb wb;
drivers/gpu/drm/radeon/radeon_device.c
424
rdev->wb.enabled = false;
drivers/gpu/drm/radeon/radeon_device.c
438
if (rdev->wb.wb_obj) {
drivers/gpu/drm/radeon/radeon_device.c
439
if (!radeon_bo_reserve(rdev->wb.wb_obj, false)) {
drivers/gpu/drm/radeon/radeon_device.c
440
radeon_bo_kunmap(rdev->wb.wb_obj);
drivers/gpu/drm/radeon/radeon_device.c
441
radeon_bo_unpin(rdev->wb.wb_obj);
drivers/gpu/drm/radeon/radeon_device.c
442
radeon_bo_unreserve(rdev->wb.wb_obj);
drivers/gpu/drm/radeon/radeon_device.c
444
radeon_bo_unref(&rdev->wb.wb_obj);
drivers/gpu/drm/radeon/radeon_device.c
445
rdev->wb.wb = NULL;
drivers/gpu/drm/radeon/radeon_device.c
446
rdev->wb.wb_obj = NULL;
drivers/gpu/drm/radeon/radeon_device.c
463
if (rdev->wb.wb_obj == NULL) {
drivers/gpu/drm/radeon/radeon_device.c
466
&rdev->wb.wb_obj);
drivers/gpu/drm/radeon/radeon_device.c
471
r = radeon_bo_reserve(rdev->wb.wb_obj, false);
drivers/gpu/drm/radeon/radeon_device.c
476
r = radeon_bo_pin(rdev->wb.wb_obj, RADEON_GEM_DOMAIN_GTT,
drivers/gpu/drm/radeon/radeon_device.c
477
&rdev->wb.gpu_addr);
drivers/gpu/drm/radeon/radeon_device.c
479
radeon_bo_unreserve(rdev->wb.wb_obj);
drivers/gpu/drm/radeon/radeon_device.c
484
r = radeon_bo_kmap(rdev->wb.wb_obj, (void **)&rdev->wb.wb);
drivers/gpu/drm/radeon/radeon_device.c
485
radeon_bo_unreserve(rdev->wb.wb_obj);
drivers/gpu/drm/radeon/radeon_device.c
494
memset((char *)rdev->wb.wb, 0, RADEON_GPU_PAGE_SIZE);
drivers/gpu/drm/radeon/radeon_device.c
496
rdev->wb.use_event = false;
drivers/gpu/drm/radeon/radeon_device.c
499
rdev->wb.enabled = false;
drivers/gpu/drm/radeon/radeon_device.c
503
rdev->wb.enabled = false;
drivers/gpu/drm/radeon/radeon_device.c
506
rdev->wb.enabled = false;
drivers/gpu/drm/radeon/radeon_device.c
508
rdev->wb.enabled = true;
drivers/gpu/drm/radeon/radeon_device.c
511
rdev->wb.use_event = true;
drivers/gpu/drm/radeon/radeon_device.c
517
rdev->wb.enabled = true;
drivers/gpu/drm/radeon/radeon_device.c
518
rdev->wb.use_event = true;
drivers/gpu/drm/radeon/radeon_device.c
521
dev_info(rdev->dev, "WB %sabled\n", rdev->wb.enabled ? "en" : "dis");
drivers/gpu/drm/radeon/radeon_fence.c
71
if (likely(rdev->wb.enabled || !drv->scratch_reg)) {
drivers/gpu/drm/radeon/radeon_fence.c
765
if (rdev->wb.use_event || !radeon_ring_supports_scratch_reg(rdev, &rdev->ring[ring])) {
drivers/gpu/drm/radeon/radeon_fence.c
769
rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4];
drivers/gpu/drm/radeon/radeon_fence.c
770
rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr +
drivers/gpu/drm/radeon/radeon_fence.c
789
rdev->fence_drv[ring].cpu_addr = &rdev->wb.wb[index/4];
drivers/gpu/drm/radeon/radeon_fence.c
790
rdev->fence_drv[ring].gpu_addr = rdev->wb.gpu_addr + index;
drivers/gpu/drm/radeon/radeon_fence.c
93
if (likely(rdev->wb.enabled || !drv->scratch_reg)) {
drivers/gpu/drm/radeon/radeon_ring.c
308
else if (rdev->wb.enabled)
drivers/gpu/drm/radeon/radeon_ring.c
422
if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/radeon_ring.c
424
ring->next_rptr_gpu_addr = rdev->wb.gpu_addr + index;
drivers/gpu/drm/radeon/radeon_ring.c
425
ring->next_rptr_cpu_addr = &rdev->wb.wb[index/4];
drivers/gpu/drm/radeon/si.c
3400
} else if (rdev->wb.enabled) {
drivers/gpu/drm/radeon/si.c
3644
WREG32(SCRATCH_ADDR, ((rdev->wb.gpu_addr + RADEON_WB_SCRATCH_OFFSET) >> 8) & 0xFFFFFFFF);
drivers/gpu/drm/radeon/si.c
3662
WREG32(CP_RB0_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFFFFFFFC);
drivers/gpu/drm/radeon/si.c
3663
WREG32(CP_RB0_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP_RPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/si.c
3665
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/si.c
3693
WREG32(CP_RB1_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFFFFFFFC);
drivers/gpu/drm/radeon/si.c
3694
WREG32(CP_RB1_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP1_RPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/si.c
3717
WREG32(CP_RB2_RPTR_ADDR, (rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFFFFFFFC);
drivers/gpu/drm/radeon/si.c
3718
WREG32(CP_RB2_RPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + RADEON_WB_CP2_RPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/si.c
5998
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/si.c
6002
WREG32(IH_RB_WPTR_ADDR_LO, (rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFFFFFFFC);
drivers/gpu/drm/radeon/si.c
6003
WREG32(IH_RB_WPTR_ADDR_HI, upper_32_bits(rdev->wb.gpu_addr + R600_WB_IH_WPTR_OFFSET) & 0xFF);
drivers/gpu/drm/radeon/si.c
6194
if (rdev->wb.enabled)
drivers/gpu/drm/radeon/si.c
6195
wptr = le32_to_cpu(rdev->wb.wb[R600_WB_IH_WPTR_OFFSET/4]);
drivers/gpu/drm/vkms/vkms_composer.c
471
static void blend(struct vkms_writeback_job *wb,
drivers/gpu/drm/vkms/vkms_composer.c
503
if (wb)
drivers/gpu/drm/vkms/vkms_composer.c
504
vkms_writeback_row(wb, output_buffer, y);
drivers/gpu/drm/vkms/vkms_drv.h
321
void vkms_writeback_row(struct vkms_writeback_job *wb, const struct line_buffer *src_buffer, int y);
drivers/gpu/drm/vkms/vkms_formats.c
668
void vkms_writeback_row(struct vkms_writeback_job *wb,
drivers/gpu/drm/vkms/vkms_formats.c
671
struct vkms_frame_info *frame_info = &wb->wb_frame_info;
drivers/gpu/drm/vkms/vkms_formats.c
681
wb->pixel_write(dst_pixels, &in_pixels[x]);
drivers/gpu/drm/vkms/vkms_writeback.c
170
struct drm_writeback_connector *wb = &vkms_output->wb_connector;
drivers/gpu/drm/vkms/vkms_writeback.c
181
drm_connector_helper_add(&wb->base, &vkms_wb_conn_helper_funcs);
drivers/gpu/drm/vkms/vkms_writeback.c
183
return drmm_writeback_connector_init(&vkmsdev->drm, wb,
drivers/hid/hid-wiimote-modules.c
2225
__u8 sx, sy, tb, wb, bd, bm, bp, bo, br, bb, bg, by, bu;
drivers/hid/hid-wiimote-modules.c
2261
wb = ext[3] & 0x1f;
drivers/hid/hid-wiimote-modules.c
2281
input_report_abs(wdata->extension.input, ABS_HAT1X, wb - 0x10);
drivers/md/bcache/sysfs.c
175
int wb = dc->writeback_running;
drivers/md/bcache/sysfs.c
204
wb ? atomic_long_read(&dc->writeback_rate.rate) << 9 : 0);
drivers/md/bcache/sysfs.c
230
wb ? atomic_long_read(&dc->writeback_rate.rate) << 9
drivers/md/bcache/sysfs.c
235
wb ? dc->writeback_rate_proportional << 9 : 0);
drivers/md/bcache/sysfs.c
237
wb ? dc->writeback_rate_integral_scaled << 9 : 0);
drivers/md/bcache/sysfs.c
238
bch_hprint(change, wb ? dc->writeback_rate_change << 9 : 0);
drivers/md/bcache/sysfs.c
239
next_io = wb ? div64_s64(dc->writeback_rate.next-local_clock(),
drivers/md/dm-writecache.c
1656
struct writeback_struct *wb = container_of(bio, struct writeback_struct, bio);
drivers/md/dm-writecache.c
1657
struct dm_writecache *wc = wb->wc;
drivers/md/dm-writecache.c
1663
list_add_tail(&wb->endio_entry, &wc->endio_list);
drivers/md/dm-writecache.c
1684
struct writeback_struct *wb;
drivers/md/dm-writecache.c
1689
wb = list_entry(list->next, struct writeback_struct, endio_entry);
drivers/md/dm-writecache.c
1690
list_del(&wb->endio_entry);
drivers/md/dm-writecache.c
1692
if (unlikely(wb->bio.bi_status != BLK_STS_OK))
drivers/md/dm-writecache.c
1693
writecache_error(wc, blk_status_to_errno(wb->bio.bi_status),
drivers/md/dm-writecache.c
1694
"write error %d", wb->bio.bi_status);
drivers/md/dm-writecache.c
1697
e = wb->wc_list[i];
drivers/md/dm-writecache.c
1712
} while (++i < wb->wc_list_n);
drivers/md/dm-writecache.c
1714
if (wb->wc_list != wb->wc_list_inline)
drivers/md/dm-writecache.c
1715
kfree(wb->wc_list);
drivers/md/dm-writecache.c
1716
bio_put(&wb->bio);
drivers/md/dm-writecache.c
1796
static bool wc_add_block(struct writeback_struct *wb, struct wc_entry *e)
drivers/md/dm-writecache.c
1798
struct dm_writecache *wc = wb->wc;
drivers/md/dm-writecache.c
1804
if (unlikely(bio_end_sector(&wb->bio) >= wc->data_device_sectors))
drivers/md/dm-writecache.c
1807
return bio_add_page(&wb->bio, persistent_memory_page(address),
drivers/md/dm-writecache.c
1833
struct writeback_struct *wb;
drivers/md/dm-writecache.c
1845
wb = container_of(bio, struct writeback_struct, bio);
drivers/md/dm-writecache.c
1846
wb->wc = wc;
drivers/md/dm-writecache.c
1851
wb->wc_list = kmalloc_objs(struct wc_entry *, max_pages,
drivers/md/dm-writecache.c
1854
if (likely(max_pages <= WB_LIST_INLINE) || unlikely(!wb->wc_list)) {
drivers/md/dm-writecache.c
1855
wb->wc_list = wb->wc_list_inline;
drivers/md/dm-writecache.c
1859
BUG_ON(!wc_add_block(wb, e));
drivers/md/dm-writecache.c
1861
wb->wc_list[0] = e;
drivers/md/dm-writecache.c
1862
wb->wc_list_n = 1;
drivers/md/dm-writecache.c
1864
while (wbl->size && wb->wc_list_n < max_pages) {
drivers/md/dm-writecache.c
1869
if (!wc_add_block(wb, f))
drivers/md/dm-writecache.c
1873
wb->wc_list[wb->wc_list_n++] = f;
drivers/media/dvb-frontends/dib3000mb.c
53
u8 wb[] = { ((reg >> 8) | 0x80) & 0xff, reg & 0xff };
drivers/media/dvb-frontends/dib3000mb.c
56
{ .addr = state->config.demod_address, .flags = 0, .buf = wb, .len = 2 },
drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c
142
static const unsigned short wb[][2] = {
drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c
152
for (i = 0; i < ARRAY_SIZE(wb); i++) {
drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c
153
if (wb[i][0] != val)
drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c
160
return s5c73m3_isp_command(state, COMM_AWB_MODE, wb[i][1]);
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
314
struct venc_h264_vpu_buf *wb = NULL;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
323
wb = inst->vsi->work_bufs;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
347
inst->work_bufs[i].size = wb[i].size;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
348
vpua = wb[i].vpua;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
349
wb_size = wb[i].size;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_h264_if.c
385
wb[i].iova = inst->work_bufs[i].dma_addr;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
155
struct venc_vp8_vpu_buf *wb = inst->vsi->work_bufs;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
158
if (wb[i].size == 0)
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
170
inst->work_bufs[i].size = wb[i].size;
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
189
wb[i].vpua);
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
190
memcpy(inst->work_bufs[i].va, tmp_va, wb[i].size);
drivers/media/platform/mediatek/vcodec/encoder/venc/venc_vp8_if.c
192
wb[i].iova = inst->work_bufs[i].dma_addr;
drivers/media/usb/dvb-usb/dib0700_devices.c
2216
u8 wb[4] = { 0xc >> 8, 0xc & 0xff, 0, 0 };
drivers/media/usb/dvb-usb/dib0700_devices.c
2219
{.addr = 0x1e >> 1, .flags = 0, .buf = wb, .len = 2},
drivers/media/usb/dvb-usb/dib0700_devices.c
2245
wb[2] = (data[index_data + 1] >> 8) & 0xff;
drivers/media/usb/dvb-usb/dib0700_devices.c
2246
wb[3] = (data[index_data + 1]) & 0xff;
drivers/media/usb/dvb-usb/dib0700_devices.c
2249
wb[0] = (data[index_data] >> 8) & 0xff;
drivers/media/usb/dvb-usb/dib0700_devices.c
2250
wb[1] = (data[index_data]) & 0xff;
drivers/media/usb/dvb-usb/dib0700_devices.c
2254
wb[2] |= rb[0];
drivers/media/usb/dvb-usb/dib0700_devices.c
2255
wb[3] |= rb[1] & ~(3 << 4);
drivers/media/usb/dvb-usb/dib0700_devices.c
2258
wb[0] = (data[index_data] >> 8)&0xff;
drivers/media/usb/dvb-usb/dib0700_devices.c
2259
wb[1] = (data[index_data])&0xff;
drivers/mtd/nand/raw/nandsim.c
835
struct weak_block *wb;
drivers/mtd/nand/raw/nandsim.c
854
wb = kzalloc_obj(*wb);
drivers/mtd/nand/raw/nandsim.c
855
if (!wb) {
drivers/mtd/nand/raw/nandsim.c
859
wb->erase_block_no = erase_block_no;
drivers/mtd/nand/raw/nandsim.c
860
wb->max_erases = max_erases;
drivers/mtd/nand/raw/nandsim.c
861
list_add(&wb->list, &weak_blocks);
drivers/mtd/nand/raw/nandsim.c
868
struct weak_block *wb;
drivers/mtd/nand/raw/nandsim.c
870
list_for_each_entry(wb, &weak_blocks, list)
drivers/mtd/nand/raw/nandsim.c
871
if (wb->erase_block_no == erase_block_no) {
drivers/mtd/nand/raw/nandsim.c
872
if (wb->erases_done >= wb->max_erases)
drivers/mtd/nand/raw/nandsim.c
874
wb->erases_done += 1;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
70
u8 wb)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
76
else if (wb && CHIP_IS_E1(bp))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
85
u32 len, u8 wb)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_init_ops.h
96
bnx2x_write_big_buf(bp, addr + i*4, cur_len, wb);
drivers/net/ethernet/freescale/enetc/enetc.c
1160
tstamp_lo = le32_to_cpu(txbd->wb.tstamp);
drivers/net/ethernet/freescale/enetc/enetc.c
1250
txbd->wb.status & ENETC_TXBD_STATS_WIN)
drivers/net/ethernet/freescale/enetc/enetc_hw.h
602
} wb; /* writeback descriptor */
drivers/net/ethernet/fungible/funcore/fun_queue.c
20
size_t hw_desc_sz, size_t sw_desc_sz, bool wb,
drivers/net/ethernet/fungible/funcore/fun_queue.c
33
if (wb)
drivers/net/ethernet/fungible/funcore/fun_queue.c
51
if (wb)
drivers/net/ethernet/fungible/funcore/fun_queue.c
58
bool wb, void *hw_va, dma_addr_t dma_addr, void *sw_va)
drivers/net/ethernet/fungible/funcore/fun_queue.c
63
if (wb)
drivers/net/ethernet/fungible/funcore/fun_queue.h
143
size_t hw_desc_sz, size_t sw_desc_size, bool wb,
drivers/net/ethernet/fungible/funcore/fun_queue.h
147
bool wb, void *hw_va, dma_addr_t dma_addr, void *sw_va);
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
122
#define MBOX_STATUS_FINISHED(wb) \
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
123
(((wb) & MBOX_WB_STATUS_MASK) != MBOX_WB_STATUS_NOT_FINISHED)
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
124
#define MBOX_STATUS_SUCCESS(wb) \
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
125
(((wb) & MBOX_WB_STATUS_MASK) == MBOX_WB_STATUS_FINISHED_SUCCESS)
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
126
#define MBOX_STATUS_ERRCODE(wb) \
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
127
((wb) & MBOX_WB_ERROR_CODE_MASK)
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
52
#define MBOX_STATUS_FINISHED(wb) \
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
53
((FIELD_PREP(MBOX_WB_STATUS_MASK, (wb))) != MBOX_WB_STATUS_NOT_FINISHED)
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
54
#define MBOX_STATUS_SUCCESS(wb) \
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
55
((FIELD_PREP(MBOX_WB_STATUS_MASK, (wb))) == \
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
57
#define MBOX_STATUS_ERRCODE(wb) \
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
58
((wb) & MBOX_WB_ERROR_CODE_MASK)
drivers/net/ethernet/intel/e1000/e1000_hw.h
525
} wb; /* writeback */
drivers/net/ethernet/intel/e1000/e1000_hw.h
556
} wb; /* writeback */
drivers/net/ethernet/intel/e1000e/hw.h
267
} wb; /* writeback */
drivers/net/ethernet/intel/e1000e/hw.h
303
} wb; /* writeback */
drivers/net/ethernet/intel/e1000e/netdev.c
1031
e1000_rx_hash(netdev, rx_desc->wb.lower.hi_dword.rss, skb);
drivers/net/ethernet/intel/e1000e/netdev.c
1034
rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/e1000e/netdev.c
1037
rx_desc->wb.upper.status_error &= cpu_to_le32(~0xFF);
drivers/net/ethernet/intel/e1000e/netdev.c
1050
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
1330
staterr = le32_to_cpu(rx_desc->wb.middle.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
1375
length = le16_to_cpu(rx_desc->wb.middle.length0);
drivers/net/ethernet/intel/e1000e/netdev.c
1390
int l1 = le16_to_cpu(rx_desc->wb.upper.length[0]);
drivers/net/ethernet/intel/e1000e/netdev.c
1423
length = le16_to_cpu(rx_desc->wb.upper.length[j]);
drivers/net/ethernet/intel/e1000e/netdev.c
1452
e1000_rx_hash(netdev, rx_desc->wb.lower.hi_dword.rss, skb);
drivers/net/ethernet/intel/e1000e/netdev.c
1454
if (rx_desc->wb.upper.header_status &
drivers/net/ethernet/intel/e1000e/netdev.c
1459
rx_desc->wb.middle.vlan);
drivers/net/ethernet/intel/e1000e/netdev.c
1462
rx_desc->wb.middle.status_error &= cpu_to_le32(~0xFF);
drivers/net/ethernet/intel/e1000e/netdev.c
1476
staterr = le32_to_cpu(rx_desc->wb.middle.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
1524
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
1552
length = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/e1000e/netdev.c
1625
e1000_rx_hash(netdev, rx_desc->wb.lower.hi_dword.rss, skb);
drivers/net/ethernet/intel/e1000e/netdev.c
1639
rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/e1000e/netdev.c
1642
rx_desc->wb.upper.status_error &= cpu_to_le32(~0xFF);
drivers/net/ethernet/intel/e1000e/netdev.c
1655
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
366
le32_to_cpu(rx_desc_ps->wb.middle.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
432
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
932
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/e1000e/netdev.c
962
length = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1716
rx_desc->wb.qword1.status_error_len = 0;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1754
qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1829
if ((rx_desc->wb.qword1.status_error_len & rss_mask) == rss_mask) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1830
hash = le32_to_cpu(rx_desc->wb.qword0.hi_dword.rss);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1848
u64 qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1864
__le16 vlan_tag = rx_desc->wb.qword0.lo_dword.l2tag1;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2476
qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len);
drivers/net/ethernet/intel/i40e/i40e_txrx.h
186
return !!(rx_desc->wb.qword1.status_error_len &
drivers/net/ethernet/intel/i40e/i40e_type.h
620
} wb; /* writeback */
drivers/net/ethernet/intel/i40e/i40e_type.h
672
} wb; /* writeback */
drivers/net/ethernet/intel/i40e/i40e_xsk.c
274
rx_desc->wb.qword1.status_error_len = 0;
drivers/net/ethernet/intel/i40e/i40e_xsk.c
435
qword = le64_to_cpu(rx_desc->wb.qword1.status_error_len);
drivers/net/ethernet/intel/ice/ice_ethtool.c
1246
if (!(rx_desc->wb.status_error0 &
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
208
} wb; /* writeback */
drivers/net/ethernet/intel/ice/ice_lan_tx_rx.h
40
} wb; /* writeback */
drivers/net/ethernet/intel/ice/ice_ptp.c
2308
if (!(rx_desc->wb.time_stamp_low & ICE_PTP_TS_VALID))
drivers/net/ethernet/intel/ice/ice_ptp.c
2323
ts_high = le32_to_cpu(rx_desc->wb.flex_ts.ts_high);
drivers/net/ethernet/intel/ice/ice_txrx.c
1004
size = le16_get_bits(rx_desc->wb.hdr_len_sph_flex_flags1,
drivers/net/ethernet/intel/ice/ice_txrx.c
1014
size = le16_to_cpu(rx_desc->wb.pkt_len) &
drivers/net/ethernet/intel/ice/ice_txrx.c
803
rx_desc->wb.status_error0 = 0;
drivers/net/ethernet/intel/ice/ice_txrx.c
884
rx_desc->wb.status_error0 = 0;
drivers/net/ethernet/intel/ice/ice_txrx.c
916
if (!ice_test_staterr(rx_desc->wb.status_error0, stat_err_bits))
drivers/net/ethernet/intel/ice/ice_txrx.c
985
if (!ice_test_staterr(rx_desc->wb.status_error0, stat_err_bits))
drivers/net/ethernet/intel/ice/ice_txrx.c
998
rxe = ice_test_staterr(rx_desc->wb.status_error0,
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
122
rx_status0 = le16_to_cpu(rx_desc->wb.status_error0);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
123
rx_status1 = le16_to_cpu(rx_desc->wb.status_error1);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
126
rx_desc->wb.rxdid == ICE_RXDID_FLEX_NIC &&
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
201
return le16_to_cpu(rx_desc->wb.ptype_flex_flags0) &
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
50
if (unlikely(rx_desc->wb.rxdid != ICE_RXDID_FLEX_NIC))
drivers/net/ethernet/intel/ice/ice_txrx_lib.h
38
if (likely(ice_test_staterr(rx_desc->wb.status_error0, ICE_RXD_EOF)))
drivers/net/ethernet/intel/ice/ice_txrx_lib.h
84
if (ice_test_staterr(rx_desc->wb.status_error0, stat_err_bits))
drivers/net/ethernet/intel/ice/ice_txrx_lib.h
85
return le16_to_cpu(rx_desc->wb.l2tag1);
drivers/net/ethernet/intel/ice/ice_txrx_lib.h
88
if (ice_test_staterr(rx_desc->wb.status_error1, stat_err_bits))
drivers/net/ethernet/intel/ice/ice_txrx_lib.h
89
return le16_to_cpu(rx_desc->wb.l2tag2_2nd);
drivers/net/ethernet/intel/ice/ice_xsk.c
271
rx_desc->wb.status_error0 = 0;
drivers/net/ethernet/intel/ice/ice_xsk.c
599
if (!ice_test_staterr(rx_desc->wb.status_error0, stat_err_bits))
drivers/net/ethernet/intel/ice/ice_xsk.c
613
size = le16_to_cpu(rx_desc->wb.pkt_len) &
drivers/net/ethernet/intel/ice/virt/fdir.c
1596
stat_err = le16_to_cpu(ctx->rx_desc.wb.status_error0);
drivers/net/ethernet/intel/igb/e1000_82575.h
106
} wb;
drivers/net/ethernet/intel/igb/e1000_82575.h
87
} wb; /* writeback */
drivers/net/ethernet/intel/igb/igb.h
448
return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits);
drivers/net/ethernet/intel/igb/igb_ethtool.c
1832
while (rx_desc->wb.upper.length) {
drivers/net/ethernet/intel/igb/igb_main.c
479
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/igb/igb_main.c
4837
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/igb/igb_main.c
8361
if (!(eop_desc->wb.status & cpu_to_le32(E1000_TXD_STAT_DD)))
drivers/net/ethernet/intel/igb/igb_main.c
8488
tx_buffer->next_to_watch->wb.status);
drivers/net/ethernet/intel/igb/igb_main.c
8817
le32_to_cpu(rx_desc->wb.upper.status_error));
drivers/net/ethernet/intel/igb/igb_main.c
8826
le32_to_cpu(rx_desc->wb.lower.hi_dword.rss),
drivers/net/ethernet/intel/igb/igb_main.c
8921
vid = be16_to_cpu((__force __be16)rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/igb/igb_main.c
8923
vid = le16_to_cpu(rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/igb/igb_main.c
9051
size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/igb/igb_main.c
9247
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/igb/igb_xsk.c
191
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/igb/igb_xsk.c
232
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/igb/igb_xsk.c
365
size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/igbvf/netdev.c
261
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/igbvf/netdev.c
276
hlen = le16_get_bits(rx_desc->wb.lower.lo_dword.hs_rss.hdr_info,
drivers/net/ethernet/intel/igbvf/netdev.c
281
length = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/igbvf/netdev.c
355
rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/igbvf/netdev.c
358
rx_desc->wb.upper.status_error = 0;
drivers/net/ethernet/intel/igbvf/netdev.c
370
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/igbvf/netdev.c
796
if (!(eop_desc->wb.status & cpu_to_le32(E1000_TXD_STAT_DD)))
drivers/net/ethernet/intel/igbvf/netdev.c
820
tx_desc->wb.status = 0;
drivers/net/ethernet/intel/igbvf/vf.h
64
} wb; /* writeback */
drivers/net/ethernet/intel/igbvf/vf.h
81
} wb;
drivers/net/ethernet/intel/igc/igc.h
434
return le32_get_bits(rx_desc->wb.lower.lo_dword.data, IGC_RSS_TYPE_MASK);
drivers/net/ethernet/intel/igc/igc.h
538
return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits);
drivers/net/ethernet/intel/igc/igc_base.h
24
} wb;
drivers/net/ethernet/intel/igc/igc_base.h
86
} wb; /* writeback */
drivers/net/ethernet/intel/igc/igc_dump.c
262
staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/igc/igc_main.c
1783
le32_to_cpu(rx_desc->wb.upper.status_error));
drivers/net/ethernet/intel/igc/igc_main.c
1811
u32 rss_hash = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss);
drivers/net/ethernet/intel/igc/igc_main.c
1829
vid = be16_to_cpu((__force __be16)rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/igc/igc_main.c
1831
vid = le16_to_cpu(rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/igc/igc_main.c
2297
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/igc/igc_main.c
2358
desc->wb.upper.length = 0;
drivers/net/ethernet/intel/igc/igc_main.c
2630
size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/igc/igc_main.c
2824
size = le16_to_cpu(desc->wb.upper.length);
drivers/net/ethernet/intel/igc/igc_main.c
3163
if (!(eop_desc->wb.status & cpu_to_le32(IGC_TXD_STAT_DD)))
drivers/net/ethernet/intel/igc/igc_main.c
3285
tx_buffer->next_to_watch->wb.status);
drivers/net/ethernet/intel/igc/igc_main.c
703
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/igc/igc_main.c
7056
*hash = le32_to_cpu(ctx->rx_desc->wb.lower.hi_dword.rss);
drivers/net/ethernet/intel/igc/igc_tsn.h
39
u32 status_error = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/ixgbe/ixgbe.h
560
return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits);
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2095
if (!(tx_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD)))
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2117
while (rx_desc->wb.upper.length) {
drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c
437
ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss);
drivers/net/ethernet/intel/ixgbe/ixgbe_fcoe.c
451
ddp->len = le32_to_cpu(rx_desc->wb.lower.hi_dword.rss);
drivers/net/ethernet/intel/ixgbe/ixgbe_ipsec.c
1151
__le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1378
if (!(eop_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD)))
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1639
rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) &
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1645
skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss),
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1661
__le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1680
__le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1829
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1907
u16 vid = le16_to_cpu(rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1955
__le32 rsc_enabled = rx_desc->wb.lower.lo_dword.data &
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1965
ntc = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2532
size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
4585
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
815
if (rx_desc->wb.upper.length) {
drivers/net/ethernet/intel/ixgbe/ixgbe_type.h
2898
} wb;
drivers/net/ethernet/intel/ixgbe/ixgbe_type.h
2929
} wb; /* writeback */
drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
191
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
272
size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
472
if (!(tx_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD)))
drivers/net/ethernet/intel/ixgbevf/defines.h
204
} wb;
drivers/net/ethernet/intel/ixgbevf/defines.h
235
} wb; /* writeback */
drivers/net/ethernet/intel/ixgbevf/ipsec.c
550
__le16 pkt_info = rx_desc->wb.lower.lo_dword.hs_rss.pkt_info;
drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
267
return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1146
size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1945
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
296
if (!(eop_desc->wb.status & cpu_to_le32(IXGBE_TXD_STAT_DD)))
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
392
eop_desc, (eop_desc ? eop_desc->wb.status : 0),
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
454
rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) &
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
460
skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss),
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
518
u16 vid = le16_to_cpu(rx_desc->wb.upper.vlan);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
694
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
164
return rx_desc->wb.upper.status_error & cpu_to_le32(stat_err_bits);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
188
size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
228
unsigned int size = le16_to_cpu(rx_desc->wb.upper.length);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
352
rx_desc->wb.upper.status_error = 0;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
354
rx_desc->wb.upper.length = 0;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
410
__le32 rsc_enabled = rx_desc->wb.lower.lo_dword.data &
drivers/net/ethernet/wangxun/libwx/wx_lib.c
420
ntc = le32_to_cpu(rx_desc->wb.upper.status_error);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
515
rss_type = le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) &
drivers/net/ethernet/wangxun/libwx/wx_lib.c
521
skb_set_hash(skb, le32_to_cpu(rx_desc->wb.lower.hi_dword.rss),
drivers/net/ethernet/wangxun/libwx/wx_lib.c
588
idx = (le16_to_cpu(rx_desc->wb.lower.lo_dword.hs_rss.pkt_info) &
drivers/net/ethernet/wangxun/libwx/wx_lib.c
592
le16_to_cpu(rx_desc->wb.upper.vlan));
drivers/net/ethernet/wangxun/libwx/wx_lib.c
790
} else if (!(eop_desc->wb.status & cpu_to_le32(WX_TXD_STAT_DD))) {
drivers/net/ethernet/wangxun/libwx/wx_type.h
652
((le32_to_cpu((_rxd)->wb.lower.lo_dword.data) >> 9) & 0xFF)
drivers/net/ethernet/wangxun/libwx/wx_type.h
954
} wb;
drivers/net/ethernet/wangxun/libwx/wx_type.h
985
} wb; /* writeback */
drivers/net/ethernet/wangxun/libwx/wx_vf_lib.c
265
rx_desc->wb.upper.length = 0;
drivers/net/wireless/microchip/wilc1000/spi.c
402
static int wilc_spi_tx_rx(struct wilc *wilc, u8 *wb, u8 *rb, u32 rlen)
drivers/net/wireless/microchip/wilc1000/spi.c
411
.tx_buf = wb,
drivers/net/wireless/microchip/wilc1000/spi.c
524
u8 wb[32], rb[32];
drivers/net/wireless/microchip/wilc1000/spi.c
531
memset(wb, 0x0, sizeof(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
533
c = (struct wilc_spi_cmd *)wb;
drivers/net/wireless/microchip/wilc1000/spi.c
554
c->u.simple_cmd.crc[0] = wilc_get_crc7(wb, cmd_len);
drivers/net/wireless/microchip/wilc1000/spi.c
559
if (cmd_len + resp_len > ARRAY_SIZE(wb)) {
drivers/net/wireless/microchip/wilc1000/spi.c
562
cmd_len, resp_len, ARRAY_SIZE(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
566
if (wilc_spi_tx_rx(wilc, wb, rb, cmd_len + resp_len)) {
drivers/net/wireless/microchip/wilc1000/spi.c
619
u8 wb[32], rb[32];
drivers/net/wireless/microchip/wilc1000/spi.c
624
memset(wb, 0x0, sizeof(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
626
c = (struct wilc_spi_cmd *)wb;
drivers/net/wireless/microchip/wilc1000/spi.c
637
c->u.internal_w_cmd.crc[0] = wilc_get_crc7(wb, cmd_len);
drivers/net/wireless/microchip/wilc1000/spi.c
645
c->u.w_cmd.crc[0] = wilc_get_crc7(wb, cmd_len);
drivers/net/wireless/microchip/wilc1000/spi.c
656
if (cmd_len + resp_len > ARRAY_SIZE(wb)) {
drivers/net/wireless/microchip/wilc1000/spi.c
659
cmd_len, resp_len, ARRAY_SIZE(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
663
if (wilc_spi_tx_rx(wilc, wb, rb, cmd_len + resp_len)) {
drivers/net/wireless/microchip/wilc1000/spi.c
694
u8 wb[32], rb[32];
drivers/net/wireless/microchip/wilc1000/spi.c
701
memset(wb, 0x0, sizeof(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
703
c = (struct wilc_spi_cmd *)wb;
drivers/net/wireless/microchip/wilc1000/spi.c
713
c->u.dma_cmd.crc[0] = wilc_get_crc7(wb, cmd_len);
drivers/net/wireless/microchip/wilc1000/spi.c
723
c->u.dma_cmd_ext.crc[0] = wilc_get_crc7(wb, cmd_len);
drivers/net/wireless/microchip/wilc1000/spi.c
734
if (cmd_len + resp_len > ARRAY_SIZE(wb)) {
drivers/net/wireless/microchip/wilc1000/spi.c
736
cmd_len, resp_len, ARRAY_SIZE(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
740
if (wilc_spi_tx_rx(wilc, wb, rb, cmd_len + resp_len)) {
drivers/net/wireless/microchip/wilc1000/spi.c
820
u8 wb[32], rb[32];
drivers/net/wireless/microchip/wilc1000/spi.c
828
memset(wb, 0x0, sizeof(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
830
c = (struct wilc_spi_cmd *)wb;
drivers/net/wireless/microchip/wilc1000/spi.c
840
c->u.simple_cmd.crc[0] = wilc_get_crc7(wb, cmd_len);
drivers/net/wireless/microchip/wilc1000/spi.c
843
if (cmd_len + resp_len > ARRAY_SIZE(wb)) {
drivers/net/wireless/microchip/wilc1000/spi.c
845
cmd_len, resp_len, ARRAY_SIZE(wb));
drivers/net/wireless/microchip/wilc1000/spi.c
849
if (wilc_spi_tx_rx(wilc, wb, rb, cmd_len + resp_len)) {
drivers/staging/media/atomisp/pci/ia_css_isp_params.c
1397
stage->binary->info->mem_offsets.offsets.param->dmem.wb.size;
drivers/staging/media/atomisp/pci/ia_css_isp_params.c
1400
stage->binary->info->mem_offsets.offsets.param->dmem.wb.offset;
drivers/staging/media/atomisp/pci/ia_css_isp_params.h
97
struct ia_css_isp_parameter wb;
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.c
50
const struct sh_css_isp_wb_params *wb,
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.c
53
if (!wb) return;
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.c
56
"wb_gain_shift", wb->gain_shift);
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.c
58
"wb_gain_gr", wb->gain_gr);
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.c
60
"wb_gain_r", wb->gain_r);
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.c
62
"wb_gain_b", wb->gain_b);
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.c
64
"wb_gain_gb", wb->gain_gb);
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.h
23
const struct sh_css_isp_wb_params *wb,
drivers/staging/media/atomisp/pci/isp/kernels/wb/wb_1.0/ia_css_wb.host.h
28
const struct ia_css_wb_config *wb,
drivers/staging/media/atomisp/pci/runtime/debug/src/ia_css_debug.c
907
ia_css_wb_dump(FIND_DMEM_PARAMS(stream, wb), IA_CSS_DEBUG_VERBOSE);
drivers/usb/class/cdc-acm.c
1145
struct acm_wb *wb;
drivers/usb/class/cdc-acm.c
1147
for (wb = &acm->wb[0], i = 0; i < ACM_NW; i++, wb++)
drivers/usb/class/cdc-acm.c
1148
usb_free_coherent(acm->dev, acm->writesize, wb->buf, wb->dmah);
drivers/usb/class/cdc-acm.c
1164
struct acm_wb *wb;
drivers/usb/class/cdc-acm.c
1166
for (wb = &acm->wb[0], i = 0; i < ACM_NW; i++, wb++) {
drivers/usb/class/cdc-acm.c
1167
wb->buf = usb_alloc_coherent(acm->dev, acm->writesize, GFP_KERNEL,
drivers/usb/class/cdc-acm.c
1168
&wb->dmah);
drivers/usb/class/cdc-acm.c
1169
if (!wb->buf) {
drivers/usb/class/cdc-acm.c
1172
--wb;
drivers/usb/class/cdc-acm.c
1174
wb->buf, wb->dmah);
drivers/usb/class/cdc-acm.c
1454
struct acm_wb *snd = &(acm->wb[i]);
drivers/usb/class/cdc-acm.c
1564
usb_free_urb(acm->wb[i].urb);
drivers/usb/class/cdc-acm.c
157
usb_poison_urb(acm->wb[i].urb);
drivers/usb/class/cdc-acm.c
1617
usb_free_urb(acm->wb[i].urb);
drivers/usb/class/cdc-acm.c
169
usb_unpoison_urb(acm->wb[i].urb);
drivers/usb/class/cdc-acm.c
182
struct acm_wb *wb;
drivers/usb/class/cdc-acm.c
187
wb = &acm->wb[wbn];
drivers/usb/class/cdc-acm.c
188
if (!wb->use) {
drivers/usb/class/cdc-acm.c
189
wb->use = true;
drivers/usb/class/cdc-acm.c
190
wb->len = 0;
drivers/usb/class/cdc-acm.c
207
if(acm->wb[i].use)
drivers/usb/class/cdc-acm.c
216
static void acm_write_done(struct acm *acm, struct acm_wb *wb)
drivers/usb/class/cdc-acm.c
218
wb->use = false;
drivers/usb/class/cdc-acm.c
229
static int acm_start_wb(struct acm *acm, struct acm_wb *wb)
drivers/usb/class/cdc-acm.c
235
wb->urb->transfer_buffer = wb->buf;
drivers/usb/class/cdc-acm.c
236
wb->urb->transfer_dma = wb->dmah;
drivers/usb/class/cdc-acm.c
237
wb->urb->transfer_buffer_length = wb->len;
drivers/usb/class/cdc-acm.c
238
wb->urb->dev = acm->dev;
drivers/usb/class/cdc-acm.c
240
rc = usb_submit_urb(wb->urb, GFP_ATOMIC);
drivers/usb/class/cdc-acm.c
246
acm_write_done(acm, wb);
drivers/usb/class/cdc-acm.c
592
struct acm_wb *wb = urb->context;
drivers/usb/class/cdc-acm.c
593
struct acm *acm = wb->instance;
drivers/usb/class/cdc-acm.c
604
acm_write_done(acm, wb);
drivers/usb/class/cdc-acm.c
766
struct acm_wb *wb;
drivers/usb/class/cdc-acm.c
784
wb = urb->context;
drivers/usb/class/cdc-acm.c
785
wb->use = false;
drivers/usb/class/cdc-acm.c
821
struct acm_wb *wb;
drivers/usb/class/cdc-acm.c
834
wb = &acm->wb[wbn];
drivers/usb/class/cdc-acm.c
837
wb->use = false;
drivers/usb/class/cdc-acm.c
844
memcpy(wb->buf, buf, count);
drivers/usb/class/cdc-acm.c
845
wb->len = count;
drivers/usb/class/cdc-acm.c
849
wb->use = false;
drivers/usb/class/cdc-acm.c
855
usb_anchor_urb(wb->urb, &acm->delayed);
drivers/usb/class/cdc-acm.c
860
stat = acm_start_wb(acm, wb);
drivers/usb/class/cdc-acm.c
886
if (acm->wb[i].use)
drivers/usb/class/cdc-acm.c
887
usb_unlink_urb(acm->wb[i].urb);
drivers/usb/class/cdc-acm.h
69
struct acm_wb wb[ACM_NW];
fs/afs/file.c
104
af->wb = wbk;
fs/afs/file.c
110
af->wb = p;
fs/afs/file.c
185
if (af->wb)
fs/afs/file.c
186
afs_put_wb_key(af->wb);
fs/afs/internal.h
235
struct afs_wb_key *wb; /* Writeback key record for this file */
fs/f2fs/node.c
122
if (!sbi->sb->s_bdi->wb.dirty_exceeded)
fs/f2fs/node.c
81
if (sbi->sb->s_bdi->wb.dirty_exceeded)
fs/f2fs/segment.h
995
if (sbi->sb->s_bdi->wb.dirty_exceeded)
fs/fs-writeback.c
100
&wb->bdi->tot_write_bandwidth) < 0);
fs/fs-writeback.c
1016
static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages)
fs/fs-writeback.c
1018
unsigned long this_bw = wb->avg_write_bandwidth;
fs/fs-writeback.c
1019
unsigned long tot_bw = atomic_long_read(&wb->bdi->tot_write_bandwidth);
fs/fs-writeback.c
1051
struct bdi_writeback *wb = list_entry(&bdi->wb_list,
fs/fs-writeback.c
1057
list_for_each_entry_continue_rcu(wb, &bdi->wb_list, bdi_node) {
fs/fs-writeback.c
1069
if (!wb_has_dirty_io(wb) &&
fs/fs-writeback.c
1071
list_empty(&wb->b_dirty_time)))
fs/fs-writeback.c
1073
if (skip_if_busy && writeback_in_progress(wb))
fs/fs-writeback.c
1076
nr_pages = wb_split_bdi_pages(wb, base_work->nr_pages);
fs/fs-writeback.c
1083
wb_queue_work(wb, work);
fs/fs-writeback.c
1094
if (!wb_tryget(wb))
fs/fs-writeback.c
1104
wb_queue_work(wb, work);
fs/fs-writeback.c
1105
last_wb = wb;
fs/fs-writeback.c
1132
struct bdi_writeback *wb;
fs/fs-writeback.c
115
struct bdi_writeback *wb,
fs/fs-writeback.c
1156
wb = wb_get_lookup(bdi, memcg_css);
fs/fs-writeback.c
1157
if (!wb) {
fs/fs-writeback.c
118
assert_spin_locked(&wb->list_lock);
fs/fs-writeback.c
1184
wb_queue_work(wb, work);
fs/fs-writeback.c
1190
wb_put(wb);
fs/fs-writeback.c
1246
struct bdi_writeback *wb)
fs/fs-writeback.c
1248
assert_spin_locked(&wb->list_lock);
fs/fs-writeback.c
125
if (head != &wb->b_dirty_time)
fs/fs-writeback.c
1254
wb_io_lists_depopulated(wb);
fs/fs-writeback.c
126
return wb_io_lists_populated(wb);
fs/fs-writeback.c
1260
__acquires(&wb->list_lock)
fs/fs-writeback.c
1262
struct bdi_writeback *wb = inode_to_wb(inode);
fs/fs-writeback.c
1265
spin_lock(&wb->list_lock);
fs/fs-writeback.c
1266
return wb;
fs/fs-writeback.c
1270
__acquires(&wb->list_lock)
fs/fs-writeback.c
1272
struct bdi_writeback *wb = inode_to_wb(inode);
fs/fs-writeback.c
1274
spin_lock(&wb->list_lock);
fs/fs-writeback.c
1275
return wb;
fs/fs-writeback.c
1278
static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages)
fs/fs-writeback.c
128
wb_io_lists_depopulated(wb);
fs/fs-writeback.c
1289
if (!skip_if_busy || !writeback_in_progress(&bdi->wb)) {
fs/fs-writeback.c
1291
wb_queue_work(&bdi->wb, base_work);
fs/fs-writeback.c
1314
static void wb_start_writeback(struct bdi_writeback *wb, enum wb_reason reason)
fs/fs-writeback.c
1316
if (!wb_has_dirty_io(wb))
fs/fs-writeback.c
132
static void wb_wakeup(struct bdi_writeback *wb)
fs/fs-writeback.c
1327
if (test_bit(WB_start_all, &wb->state) ||
fs/fs-writeback.c
1328
test_and_set_bit(WB_start_all, &wb->state))
fs/fs-writeback.c
1331
wb->start_all_reason = reason;
fs/fs-writeback.c
1332
wb_wakeup(wb);
fs/fs-writeback.c
134
spin_lock_irq(&wb->work_lock);
fs/fs-writeback.c
1345
void wb_start_background_writeback(struct bdi_writeback *wb)
fs/fs-writeback.c
135
if (test_bit(WB_registered, &wb->state))
fs/fs-writeback.c
1351
trace_writeback_wake_background(wb);
fs/fs-writeback.c
1352
wb_wakeup(wb);
fs/fs-writeback.c
136
mod_delayed_work(bdi_wq, &wb->dwork, 0);
fs/fs-writeback.c
1360
struct bdi_writeback *wb;
fs/fs-writeback.c
1369
wb = inode_to_wb_and_lock_list(inode);
fs/fs-writeback.c
137
spin_unlock_irq(&wb->work_lock);
fs/fs-writeback.c
1374
wb_io_lists_depopulated(wb);
fs/fs-writeback.c
1377
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
1426
static void redirty_tail_locked(struct inode *inode, struct bdi_writeback *wb)
fs/fs-writeback.c
1438
wb_io_lists_depopulated(wb);
fs/fs-writeback.c
1441
if (!list_empty(&wb->b_dirty)) {
fs/fs-writeback.c
1444
tail = wb_inode(wb->b_dirty.next);
fs/fs-writeback.c
1448
inode_io_list_move_locked(inode, wb, &wb->b_dirty);
fs/fs-writeback.c
1451
static void redirty_tail(struct inode *inode, struct bdi_writeback *wb)
fs/fs-writeback.c
1454
redirty_tail_locked(inode, wb);
fs/fs-writeback.c
1461
static void requeue_io(struct inode *inode, struct bdi_writeback *wb)
fs/fs-writeback.c
1463
inode_io_list_move_locked(inode, wb, &wb->b_more_io);
fs/fs-writeback.c
154
static void wb_wakeup_delayed(struct bdi_writeback *wb)
fs/fs-writeback.c
1558
static void queue_io(struct bdi_writeback *wb, struct wb_writeback_work *work,
fs/fs-writeback.c
1564
assert_spin_locked(&wb->list_lock);
fs/fs-writeback.c
1565
list_splice_init(&wb->b_more_io, &wb->b_io);
fs/fs-writeback.c
1566
moved = move_expired_inodes(&wb->b_dirty, &wb->b_io, dirtied_before);
fs/fs-writeback.c
1569
moved += move_expired_inodes(&wb->b_dirty_time, &wb->b_io,
fs/fs-writeback.c
1572
wb_io_lists_populated(wb);
fs/fs-writeback.c
1573
trace_writeback_queue_io(wb, work, dirtied_before, moved);
fs/fs-writeback.c
159
spin_lock_irq(&wb->work_lock);
fs/fs-writeback.c
160
if (test_bit(WB_registered, &wb->state))
fs/fs-writeback.c
161
queue_delayed_work(bdi_wq, &wb->dwork, timeout);
fs/fs-writeback.c
162
spin_unlock_irq(&wb->work_lock);
fs/fs-writeback.c
1648
static void requeue_inode(struct inode *inode, struct bdi_writeback *wb,
fs/fs-writeback.c
1672
redirty_tail_locked(inode, wb);
fs/fs-writeback.c
1674
inode_cgwb_move_to_attached(inode, wb);
fs/fs-writeback.c
1686
requeue_io(inode, wb);
fs/fs-writeback.c
1695
redirty_tail_locked(inode, wb);
fs/fs-writeback.c
1703
redirty_tail_locked(inode, wb);
fs/fs-writeback.c
1706
inode_io_list_move_locked(inode, wb, &wb->b_dirty_time);
fs/fs-writeback.c
1710
inode_cgwb_move_to_attached(inode, wb);
fs/fs-writeback.c
180
static void wb_queue_work(struct bdi_writeback *wb,
fs/fs-writeback.c
183
trace_writeback_queue(wb, work);
fs/fs-writeback.c
1846
struct bdi_writeback *wb;
fs/fs-writeback.c
188
spin_lock_irq(&wb->work_lock);
fs/fs-writeback.c
1885
wb = inode_to_wb_and_lock_list(inode);
fs/fs-writeback.c
1898
inode_cgwb_move_to_attached(inode, wb);
fs/fs-writeback.c
190
if (test_bit(WB_registered, &wb->state)) {
fs/fs-writeback.c
1901
redirty_tail_locked(inode, wb);
fs/fs-writeback.c
1905
wb,
fs/fs-writeback.c
1906
&wb->b_dirty_time);
fs/fs-writeback.c
191
list_add_tail(&work->list, &wb->work_list);
fs/fs-writeback.c
1911
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
1919
struct bdi_writeback *wb, struct wb_writeback_work *work)
fs/fs-writeback.c
192
mod_delayed_work(bdi_wq, &wb->dwork, 0);
fs/fs-writeback.c
1939
pages = min(wb->avg_write_bandwidth / 2,
fs/fs-writeback.c
1956
struct bdi_writeback *wb,
fs/fs-writeback.c
196
spin_unlock_irq(&wb->work_lock);
fs/fs-writeback.c
1979
while (!list_empty(&wb->b_io)) {
fs/fs-writeback.c
1980
struct inode *inode = wb_inode(wb->b_io.prev);
fs/fs-writeback.c
1991
redirty_tail(inode, wb);
fs/fs-writeback.c
2010
redirty_tail_locked(inode, wb);
fs/fs-writeback.c
2024
requeue_io(inode, wb);
fs/fs-writeback.c
2029
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
2040
spin_lock(&wb->list_lock);
fs/fs-writeback.c
2046
write_chunk = writeback_chunk_size(inode->i_sb, wb, work);
fs/fs-writeback.c
2092
if (unlikely(tmp_wb != wb)) {
fs/fs-writeback.c
2094
spin_lock(&wb->list_lock);
fs/fs-writeback.c
2111
static long __writeback_inodes_wb(struct bdi_writeback *wb,
fs/fs-writeback.c
2117
while (!list_empty(&wb->b_io)) {
fs/fs-writeback.c
2118
struct inode *inode = wb_inode(wb->b_io.prev);
fs/fs-writeback.c
2127
redirty_tail(inode, wb);
fs/fs-writeback.c
2130
wrote += writeback_sb_inodes(sb, wb, work);
fs/fs-writeback.c
2145
static long writeback_inodes_wb(struct bdi_writeback *wb, long nr_pages,
fs/fs-writeback.c
2157
spin_lock(&wb->list_lock);
fs/fs-writeback.c
2158
if (list_empty(&wb->b_io))
fs/fs-writeback.c
2159
queue_io(wb, &work, jiffies);
fs/fs-writeback.c
2160
__writeback_inodes_wb(wb, &work);
fs/fs-writeback.c
2161
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
2182
static long wb_writeback(struct bdi_writeback *wb,
fs/fs-writeback.c
2207
!list_empty(&wb->work_list))
fs/fs-writeback.c
2214
if (work->for_background && !wb_over_bg_thresh(wb))
fs/fs-writeback.c
2218
spin_lock(&wb->list_lock);
fs/fs-writeback.c
2220
trace_writeback_start(wb, work);
fs/fs-writeback.c
2221
if (list_empty(&wb->b_io)) {
fs/fs-writeback.c
2235
queue_io(wb, work, dirtied_before);
fs/fs-writeback.c
2239
progress = writeback_sb_inodes(work->sb, wb, work);
fs/fs-writeback.c
2241
progress = __writeback_inodes_wb(wb, work);
fs/fs-writeback.c
2242
trace_writeback_written(wb, work);
fs/fs-writeback.c
2253
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
2260
if (list_empty(&wb->b_more_io)) {
fs/fs-writeback.c
2261
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
2270
trace_writeback_wait(wb, work);
fs/fs-writeback.c
2271
inode = wb_inode(wb->b_more_io.prev);
fs/fs-writeback.c
2273
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
2285
static struct wb_writeback_work *get_next_work_item(struct bdi_writeback *wb)
fs/fs-writeback.c
2289
spin_lock_irq(&wb->work_lock);
fs/fs-writeback.c
2290
if (!list_empty(&wb->work_list)) {
fs/fs-writeback.c
2291
work = list_entry(wb->work_list.next,
fs/fs-writeback.c
2295
spin_unlock_irq(&wb->work_lock);
fs/fs-writeback.c
2299
static long wb_check_background_flush(struct bdi_writeback *wb)
fs/fs-writeback.c
2301
if (wb_over_bg_thresh(wb)) {
fs/fs-writeback.c
2311
return wb_writeback(wb, &work);
fs/fs-writeback.c
2317
static long wb_check_old_data_flush(struct bdi_writeback *wb)
fs/fs-writeback.c
2328
expired = wb->last_old_flush +
fs/fs-writeback.c
2333
wb->last_old_flush = jiffies;
fs/fs-writeback.c
2345
return wb_writeback(wb, &work);
fs/fs-writeback.c
2351
static long wb_check_start_all(struct bdi_writeback *wb)
fs/fs-writeback.c
2355
if (!test_bit(WB_start_all, &wb->state))
fs/fs-writeback.c
2361
.nr_pages = wb_split_bdi_pages(wb, nr_pages),
fs/fs-writeback.c
2364
.reason = wb->start_all_reason,
fs/fs-writeback.c
2367
nr_pages = wb_writeback(wb, &work);
fs/fs-writeback.c
2370
clear_bit(WB_start_all, &wb->state);
fs/fs-writeback.c
2378
static long wb_do_writeback(struct bdi_writeback *wb)
fs/fs-writeback.c
2383
set_bit(WB_writeback_running, &wb->state);
fs/fs-writeback.c
2384
while ((work = get_next_work_item(wb)) != NULL) {
fs/fs-writeback.c
2385
trace_writeback_exec(wb, work);
fs/fs-writeback.c
2386
wrote += wb_writeback(wb, work);
fs/fs-writeback.c
2393
wrote += wb_check_start_all(wb);
fs/fs-writeback.c
2398
wrote += wb_check_old_data_flush(wb);
fs/fs-writeback.c
2399
wrote += wb_check_background_flush(wb);
fs/fs-writeback.c
2400
clear_bit(WB_writeback_running, &wb->state);
fs/fs-writeback.c
2411
struct bdi_writeback *wb = container_of(to_delayed_work(work),
fs/fs-writeback.c
2415
set_worker_desc("flush-%s", bdi_dev_name(wb->bdi));
fs/fs-writeback.c
2418
!test_bit(WB_registered, &wb->state))) {
fs/fs-writeback.c
2426
pages_written = wb_do_writeback(wb);
fs/fs-writeback.c
2428
} while (!list_empty(&wb->work_list));
fs/fs-writeback.c
2435
pages_written = writeback_inodes_wb(wb, 1024,
fs/fs-writeback.c
2440
if (!list_empty(&wb->work_list))
fs/fs-writeback.c
2441
wb_wakeup(wb);
fs/fs-writeback.c
2442
else if (wb_has_dirty_io(wb) && dirty_writeback_interval)
fs/fs-writeback.c
2443
wb_wakeup_delayed(wb);
fs/fs-writeback.c
2452
struct bdi_writeback *wb;
fs/fs-writeback.c
2457
list_for_each_entry_rcu(wb, &bdi->wb_list, bdi_node)
fs/fs-writeback.c
2458
wb_start_writeback(wb, reason);
fs/fs-writeback.c
2511
struct bdi_writeback *wb;
fs/fs-writeback.c
2513
list_for_each_entry_rcu(wb, &bdi->wb_list, bdi_node)
fs/fs-writeback.c
2514
if (!list_empty(&wb->b_dirty_time))
fs/fs-writeback.c
2515
wb_wakeup(wb);
fs/fs-writeback.c
2590
struct bdi_writeback *wb = NULL;
fs/fs-writeback.c
2664
wb = locked_inode_to_wb_and_lock_list(inode);
fs/fs-writeback.c
2701
dirty_list = &wb->b_dirty;
fs/fs-writeback.c
2703
dirty_list = &wb->b_dirty_time;
fs/fs-writeback.c
2705
wakeup_bdi = inode_io_list_move_locked(inode, wb,
fs/fs-writeback.c
2715
(wb->bdi->capabilities & BDI_CAP_WRITEBACK))
fs/fs-writeback.c
2716
wb_wakeup_delayed(wb);
fs/fs-writeback.c
2718
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
2726
if (wb)
fs/fs-writeback.c
2727
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
278
struct bdi_writeback *wb = NULL;
fs/fs-writeback.c
285
wb = wb_get_create(bdi, memcg_css, GFP_ATOMIC);
fs/fs-writeback.c
289
wb = wb_get_create(bdi, memcg_css, GFP_ATOMIC);
fs/fs-writeback.c
294
if (!wb)
fs/fs-writeback.c
295
wb = &bdi->wb;
fs/fs-writeback.c
301
if (unlikely(cmpxchg(&inode->i_wb, NULL, wb)))
fs/fs-writeback.c
302
wb_put(wb);
fs/fs-writeback.c
314
struct bdi_writeback *wb)
fs/fs-writeback.c
316
assert_spin_locked(&wb->list_lock);
fs/fs-writeback.c
321
if (wb != &wb->bdi->wb)
fs/fs-writeback.c
322
list_move(&inode->i_io_list, &wb->b_attached);
fs/fs-writeback.c
325
wb_io_lists_depopulated(wb);
fs/fs-writeback.c
339
__acquires(&wb->list_lock)
fs/fs-writeback.c
342
struct bdi_writeback *wb = inode_to_wb(inode);
fs/fs-writeback.c
350
wb_get(wb);
fs/fs-writeback.c
352
spin_lock(&wb->list_lock);
fs/fs-writeback.c
355
if (likely(wb == inode->i_wb)) {
fs/fs-writeback.c
356
wb_put(wb); /* @inode already has ref */
fs/fs-writeback.c
357
return wb;
fs/fs-writeback.c
360
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
361
wb_put(wb);
fs/fs-writeback.c
375
__acquires(&wb->list_lock)
fs/fs-writeback.c
627
static void wb_queue_isw(struct bdi_writeback *wb,
fs/fs-writeback.c
630
if (llist_add(&isw->list, &wb->switch_wbs_ctxs))
fs/fs-writeback.c
631
queue_work(isw_wq, &wb->switch_work);
fs/fs-writeback.c
720
bool cleanup_offline_cgwb(struct bdi_writeback *wb)
fs/fs-writeback.c
734
for (memcg_css = wb->memcg_css->parent; memcg_css;
fs/fs-writeback.c
736
new_wb = wb_get_create(wb->bdi, memcg_css, GFP_KERNEL);
fs/fs-writeback.c
741
new_wb = &wb->bdi->wb; /* wb_get() is noop for bdi's wb */
fs/fs-writeback.c
744
spin_lock(&wb->list_lock);
fs/fs-writeback.c
753
restart = isw_prepare_wbs_switch(new_wb, isw, &wb->b_attached, &nr);
fs/fs-writeback.c
755
restart = isw_prepare_wbs_switch(new_wb, isw, &wb->b_dirty_time,
fs/fs-writeback.c
757
spin_unlock(&wb->list_lock);
fs/fs-writeback.c
767
trace_inode_switch_wbs_queue(wb, new_wb, nr);
fs/fs-writeback.c
792
wbc->wb = inode_to_wb(inode);
fs/fs-writeback.c
795
wbc->wb_id = wbc->wb->memcg_css->id;
fs/fs-writeback.c
802
wb_get(wbc->wb);
fs/fs-writeback.c
81
static bool wb_io_lists_populated(struct bdi_writeback *wb)
fs/fs-writeback.c
812
if (unlikely(wb_dying(wbc->wb) && !css_is_dying(wbc->wb->memcg_css)))
fs/fs-writeback.c
83
if (wb_has_dirty_io(wb)) {
fs/fs-writeback.c
86
set_bit(WB_has_dirty_io, &wb->state);
fs/fs-writeback.c
87
WARN_ON_ONCE(!wb->avg_write_bandwidth);
fs/fs-writeback.c
873
struct bdi_writeback *wb = wbc->wb;
fs/fs-writeback.c
879
if (!wb)
fs/fs-writeback.c
88
atomic_long_add(wb->avg_write_bandwidth,
fs/fs-writeback.c
89
&wb->bdi->tot_write_bandwidth);
fs/fs-writeback.c
906
wb->avg_write_bandwidth);
fs/fs-writeback.c
94
static void wb_io_lists_depopulated(struct bdi_writeback *wb)
fs/fs-writeback.c
952
wb_put(wbc->wb);
fs/fs-writeback.c
953
wbc->wb = NULL;
fs/fs-writeback.c
96
if (wb_has_dirty_io(wb) && list_empty(&wb->b_dirty) &&
fs/fs-writeback.c
97
list_empty(&wb->b_io) && list_empty(&wb->b_more_io)) {
fs/fs-writeback.c
979
if (!wbc->wb || wbc->no_cgroup_owner)
fs/fs-writeback.c
98
clear_bit(WB_has_dirty_io, &wb->state);
fs/fs-writeback.c
99
WARN_ON_ONCE(atomic_long_sub_return(wb->avg_write_bandwidth,
fs/gfs2/super.c
452
if (bdi->wb.dirty_exceeded)
fs/nfs/internal.h
862
wb_stat_mod(&inode_to_bdi(inode)->wb, WB_WRITEBACK, nr);
fs/nfs/write.c
875
wb_stat_mod(&inode_to_bdi(folio->mapping->host)->wb,
include/linux/backing-dev-defs.h
193
struct bdi_writeback wb; /* the root writeback info for this bdi */
include/linux/backing-dev-defs.h
222
static inline bool wb_tryget(struct bdi_writeback *wb)
include/linux/backing-dev-defs.h
224
if (wb != &wb->bdi->wb)
include/linux/backing-dev-defs.h
225
return percpu_ref_tryget(&wb->refcnt);
include/linux/backing-dev-defs.h
233
static inline void wb_get(struct bdi_writeback *wb)
include/linux/backing-dev-defs.h
235
if (wb != &wb->bdi->wb)
include/linux/backing-dev-defs.h
236
percpu_ref_get(&wb->refcnt);
include/linux/backing-dev-defs.h
244
static inline void wb_put_many(struct bdi_writeback *wb, unsigned long nr)
include/linux/backing-dev-defs.h
246
if (WARN_ON_ONCE(!wb->bdi)) {
include/linux/backing-dev-defs.h
254
if (wb != &wb->bdi->wb)
include/linux/backing-dev-defs.h
255
percpu_ref_put_many(&wb->refcnt, nr);
include/linux/backing-dev-defs.h
262
static inline void wb_put(struct bdi_writeback *wb)
include/linux/backing-dev-defs.h
264
wb_put_many(wb, 1);
include/linux/backing-dev-defs.h
273
static inline bool wb_dying(struct bdi_writeback *wb)
include/linux/backing-dev-defs.h
275
return percpu_ref_is_dying(&wb->refcnt);
include/linux/backing-dev-defs.h
280
static inline bool wb_tryget(struct bdi_writeback *wb)
include/linux/backing-dev-defs.h
285
static inline void wb_get(struct bdi_writeback *wb)
include/linux/backing-dev-defs.h
289
static inline void wb_put(struct bdi_writeback *wb)
include/linux/backing-dev-defs.h
293
static inline void wb_put_many(struct bdi_writeback *wb, unsigned long nr)
include/linux/backing-dev-defs.h
297
static inline bool wb_dying(struct bdi_writeback *wb)
include/linux/backing-dev.h
127
static inline bool writeback_in_progress(struct bdi_writeback *wb)
include/linux/backing-dev.h
129
return test_bit(WB_writeback_running, &wb->state);
include/linux/backing-dev.h
181
struct bdi_writeback *wb;
include/linux/backing-dev.h
185
return &bdi->wb;
include/linux/backing-dev.h
187
wb = radix_tree_lookup(&bdi->cgwb_tree, memcg_css->id);
include/linux/backing-dev.h
193
if (likely(wb && wb->blkcg_css == task_css(current, io_cgrp_id)))
include/linux/backing-dev.h
194
return wb;
include/linux/backing-dev.h
210
struct bdi_writeback *wb;
include/linux/backing-dev.h
213
wb = wb_find_current(bdi);
include/linux/backing-dev.h
214
if (wb && unlikely(!wb_tryget(wb)))
include/linux/backing-dev.h
215
wb = NULL;
include/linux/backing-dev.h
218
if (unlikely(!wb)) {
include/linux/backing-dev.h
222
wb = wb_get_create(bdi, memcg_css, gfp);
include/linux/backing-dev.h
225
return wb;
include/linux/backing-dev.h
256
return wbc->wb ? wbc->wb : &inode_to_bdi(inode)->wb;
include/linux/backing-dev.h
319
return &bdi->wb;
include/linux/backing-dev.h
325
return &bdi->wb;
include/linux/backing-dev.h
330
return &inode_to_bdi(inode)->wb;
include/linux/backing-dev.h
39
void wb_start_background_writeback(struct bdi_writeback *wb);
include/linux/backing-dev.h
49
static inline bool wb_has_dirty_io(struct bdi_writeback *wb)
include/linux/backing-dev.h
51
return test_bit(WB_has_dirty_io, &wb->state);
include/linux/backing-dev.h
63
static inline void wb_stat_mod(struct bdi_writeback *wb,
include/linux/backing-dev.h
66
percpu_counter_add_batch(&wb->stat[item], amount, WB_STAT_BATCH);
include/linux/backing-dev.h
69
static inline s64 wb_stat(struct bdi_writeback *wb, enum wb_stat_item item)
include/linux/backing-dev.h
71
return percpu_counter_read_positive(&wb->stat[item]);
include/linux/backing-dev.h
74
static inline s64 wb_stat_sum(struct bdi_writeback *wb, enum wb_stat_item item)
include/linux/backing-dev.h
76
return percpu_counter_sum_positive(&wb->stat[item]);
include/linux/backing-dev.h
79
extern void wb_writeout_inc(struct bdi_writeback *wb);
include/linux/fs.h
436
void (*is_dirty_writeback) (struct folio *, bool *dirty, bool *wb);
include/linux/memcontrol.h
1536
struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb);
include/linux/memcontrol.h
1537
void mem_cgroup_wb_stats(struct bdi_writeback *wb, unsigned long *pfilepages,
include/linux/memcontrol.h
1542
struct bdi_writeback *wb);
include/linux/memcontrol.h
1545
struct bdi_writeback *wb)
include/linux/memcontrol.h
1553
if (unlikely(memcg && &memcg->css != wb->memcg_css))
include/linux/memcontrol.h
1554
mem_cgroup_track_foreign_dirty_slowpath(folio, wb);
include/linux/memcontrol.h
1557
void mem_cgroup_flush_foreign(struct bdi_writeback *wb);
include/linux/memcontrol.h
1561
static inline struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb)
include/linux/memcontrol.h
1566
static inline void mem_cgroup_wb_stats(struct bdi_writeback *wb,
include/linux/memcontrol.h
1575
struct bdi_writeback *wb)
include/linux/memcontrol.h
1579
static inline void mem_cgroup_flush_foreign(struct bdi_writeback *wb)
include/linux/pagemap.h
1254
void folio_account_cleaned(struct folio *folio, struct bdi_writeback *wb);
include/linux/writeback.h
107
((wbc)->wb ? (wbc)->wb->blkcg_css : blkcg_root_css)
include/linux/writeback.h
211
bool cleanup_offline_cgwb(struct bdi_writeback *wb);
include/linux/writeback.h
264
if (wbc->wb)
include/linux/writeback.h
265
bio_associate_blkg_from_css(bio, wbc->wb->blkcg_css);
include/linux/writeback.h
313
struct bdi_writeback *wb;
include/linux/writeback.h
344
unsigned long wb_calc_thresh(struct bdi_writeback *wb, unsigned long thresh);
include/linux/writeback.h
345
unsigned long cgwb_calc_thresh(struct bdi_writeback *wb);
include/linux/writeback.h
347
void wb_update_bandwidth(struct bdi_writeback *wb);
include/linux/writeback.h
356
bool wb_over_bg_thresh(struct bdi_writeback *wb);
include/linux/writeback.h
80
struct bdi_writeback *wb; /* wb this writeback is issued under */
include/trace/events/writeback.h
158
static inline ino_t __trace_wb_assign_cgroup(struct bdi_writeback *wb)
include/trace/events/writeback.h
160
return cgroup_ino(wb->memcg_css->cgroup);
include/trace/events/writeback.h
165
if (wbc->wb)
include/trace/events/writeback.h
166
return __trace_wb_assign_cgroup(wbc->wb);
include/trace/events/writeback.h
172
static inline ino_t __trace_wb_assign_cgroup(struct bdi_writeback *wb)
include/trace/events/writeback.h
275
TP_PROTO(struct folio *folio, struct bdi_writeback *wb),
include/trace/events/writeback.h
277
TP_ARGS(folio, wb),
include/trace/events/writeback.h
292
strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
include/trace/events/writeback.h
293
__entry->bdi_id = wb->bdi->id;
include/trace/events/writeback.h
295
__entry->memcg_id = wb->memcg_css->id;
include/trace/events/writeback.h
296
__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
include/trace/events/writeback.h
312
TP_PROTO(struct bdi_writeback *wb, unsigned int frn_bdi_id,
include/trace/events/writeback.h
315
TP_ARGS(wb, frn_bdi_id, frn_memcg_id),
include/trace/events/writeback.h
325
strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
include/trace/events/writeback.h
326
__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
include/trace/events/writeback.h
384
TP_PROTO(struct bdi_writeback *wb, struct wb_writeback_work *work),
include/trace/events/writeback.h
385
TP_ARGS(wb, work),
include/trace/events/writeback.h
398
strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
include/trace/events/writeback.h
406
__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
include/trace/events/writeback.h
423
TP_PROTO(struct bdi_writeback *wb, struct wb_writeback_work *work), \
include/trace/events/writeback.h
424
TP_ARGS(wb, work))
include/trace/events/writeback.h
444
TP_PROTO(struct bdi_writeback *wb),
include/trace/events/writeback.h
445
TP_ARGS(wb),
include/trace/events/writeback.h
451
strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
include/trace/events/writeback.h
452
__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
include/trace/events/writeback.h
461
TP_PROTO(struct bdi_writeback *wb), \
include/trace/events/writeback.h
462
TP_ARGS(wb))
include/trace/events/writeback.h
531
TP_PROTO(struct bdi_writeback *wb,
include/trace/events/writeback.h
535
TP_ARGS(wb, work, dirtied_before, moved),
include/trace/events/writeback.h
545
strscpy_pad(__entry->name, bdi_dev_name(wb->bdi), 32);
include/trace/events/writeback.h
550
__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
include/trace/events/writeback.h
609
TP_PROTO(struct bdi_writeback *wb,
include/trace/events/writeback.h
613
TP_ARGS(wb, dirty_rate, task_ratelimit),
include/trace/events/writeback.h
627
strscpy_pad(__entry->bdi, bdi_dev_name(wb->bdi), 32);
include/trace/events/writeback.h
628
__entry->write_bw = KBps(wb->write_bandwidth);
include/trace/events/writeback.h
629
__entry->avg_write_bw = KBps(wb->avg_write_bandwidth);
include/trace/events/writeback.h
631
__entry->dirty_ratelimit = KBps(wb->dirty_ratelimit);
include/trace/events/writeback.h
634
KBps(wb->balanced_dirty_ratelimit);
include/trace/events/writeback.h
635
__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
include/trace/events/writeback.h
655
TP_PROTO(struct bdi_writeback *wb,
include/trace/events/writeback.h
664
TP_ARGS(wb, dtc,
include/trace/events/writeback.h
688
strscpy_pad(__entry->bdi, bdi_dev_name(wb->bdi), 32);
include/trace/events/writeback.h
705
__entry->cgroup_ino = __trace_wb_assign_cgroup(wb);
include/uapi/drm/lima_drm.h
84
__u32 wb[3 * LIMA_PP_WB_REG_NUM];
include/uapi/drm/lima_drm.h
93
__u32 wb[3 * LIMA_PP_WB_REG_NUM];
mm/backing-dev.c
1001
list_del_rcu(&wb->bdi_node);
mm/backing-dev.c
106
collect_wb_stats(stats, &bdi->wb);
mm/backing-dev.c
1106
set_bit(WB_registered, &bdi->wb.state);
mm/backing-dev.c
1160
wb_shutdown(&bdi->wb);
mm/backing-dev.c
1188
WARN_ON_ONCE(test_bit(WB_registered, &bdi->wb.state));
mm/backing-dev.c
1190
wb_exit(&bdi->wb);
mm/backing-dev.c
152
!list_empty(&bdi->bdi_list), bdi->wb.state);
mm/backing-dev.c
158
static void wb_stats_show(struct seq_file *m, struct bdi_writeback *wb,
mm/backing-dev.c
176
cgroup_ino(wb->memcg_css->cgroup),
mm/backing-dev.c
185
K(wb->avg_write_bandwidth),
mm/backing-dev.c
190
wb->state);
mm/backing-dev.c
198
struct bdi_writeback *wb;
mm/backing-dev.c
203
list_for_each_entry_rcu(wb, &bdi->wb_list, bdi_node) {
mm/backing-dev.c
206
if (!wb_tryget(wb))
mm/backing-dev.c
209
collect_wb_stats(&stats, wb);
mm/backing-dev.c
217
if (mem_cgroup_wb_domain(wb)) {
mm/backing-dev.c
219
stats.wb_thresh = min(stats.wb_thresh, cgwb_calc_thresh(wb));
mm/backing-dev.c
223
wb_stats_show(m, wb, &stats);
mm/backing-dev.c
225
wb_put(wb);
mm/backing-dev.c
504
struct bdi_writeback *wb = container_of(to_delayed_work(work),
mm/backing-dev.c
507
wb_update_bandwidth(wb);
mm/backing-dev.c
515
static int wb_init(struct bdi_writeback *wb, struct backing_dev_info *bdi,
mm/backing-dev.c
520
memset(wb, 0, sizeof(*wb));
mm/backing-dev.c
522
wb->bdi = bdi;
mm/backing-dev.c
523
wb->last_old_flush = jiffies;
mm/backing-dev.c
524
INIT_LIST_HEAD(&wb->b_dirty);
mm/backing-dev.c
525
INIT_LIST_HEAD(&wb->b_io);
mm/backing-dev.c
526
INIT_LIST_HEAD(&wb->b_more_io);
mm/backing-dev.c
527
INIT_LIST_HEAD(&wb->b_dirty_time);
mm/backing-dev.c
528
spin_lock_init(&wb->list_lock);
mm/backing-dev.c
530
atomic_set(&wb->writeback_inodes, 0);
mm/backing-dev.c
531
wb->bw_time_stamp = jiffies;
mm/backing-dev.c
532
wb->balanced_dirty_ratelimit = INIT_BW;
mm/backing-dev.c
533
wb->dirty_ratelimit = INIT_BW;
mm/backing-dev.c
534
wb->write_bandwidth = INIT_BW;
mm/backing-dev.c
535
wb->avg_write_bandwidth = INIT_BW;
mm/backing-dev.c
537
spin_lock_init(&wb->work_lock);
mm/backing-dev.c
538
INIT_LIST_HEAD(&wb->work_list);
mm/backing-dev.c
539
INIT_DELAYED_WORK(&wb->dwork, wb_workfn);
mm/backing-dev.c
540
INIT_DELAYED_WORK(&wb->bw_dwork, wb_update_bandwidth_workfn);
mm/backing-dev.c
542
err = fprop_local_init_percpu(&wb->completions, gfp);
mm/backing-dev.c
546
err = percpu_counter_init_many(wb->stat, 0, gfp, NR_WB_STAT_ITEMS);
mm/backing-dev.c
548
fprop_local_destroy_percpu(&wb->completions);
mm/backing-dev.c
553
static void cgwb_remove_from_bdi_list(struct bdi_writeback *wb);
mm/backing-dev.c
558
static void wb_shutdown(struct bdi_writeback *wb)
mm/backing-dev.c
561
spin_lock_irq(&wb->work_lock);
mm/backing-dev.c
562
if (!test_and_clear_bit(WB_registered, &wb->state)) {
mm/backing-dev.c
563
spin_unlock_irq(&wb->work_lock);
mm/backing-dev.c
566
spin_unlock_irq(&wb->work_lock);
mm/backing-dev.c
568
cgwb_remove_from_bdi_list(wb);
mm/backing-dev.c
574
mod_delayed_work(bdi_wq, &wb->dwork, 0);
mm/backing-dev.c
575
flush_delayed_work(&wb->dwork);
mm/backing-dev.c
576
WARN_ON(!list_empty(&wb->work_list));
mm/backing-dev.c
577
flush_delayed_work(&wb->bw_dwork);
mm/backing-dev.c
580
static void wb_exit(struct bdi_writeback *wb)
mm/backing-dev.c
582
WARN_ON(delayed_work_pending(&wb->dwork));
mm/backing-dev.c
583
percpu_counter_destroy_many(wb->stat, NR_WB_STAT_ITEMS);
mm/backing-dev.c
584
fprop_local_destroy_percpu(&wb->completions);
mm/backing-dev.c
604
struct bdi_writeback *wb = container_of(rcu_head,
mm/backing-dev.c
607
percpu_ref_exit(&wb->refcnt);
mm/backing-dev.c
608
kfree(wb);
mm/backing-dev.c
613
struct bdi_writeback *wb = container_of(work, struct bdi_writeback,
mm/backing-dev.c
615
struct backing_dev_info *bdi = wb->bdi;
mm/backing-dev.c
617
mutex_lock(&wb->bdi->cgwb_release_mutex);
mm/backing-dev.c
618
wb_shutdown(wb);
mm/backing-dev.c
620
css_put(wb->memcg_css);
mm/backing-dev.c
621
css_put(wb->blkcg_css);
mm/backing-dev.c
622
mutex_unlock(&wb->bdi->cgwb_release_mutex);
mm/backing-dev.c
625
blkcg_unpin_online(wb->blkcg_css);
mm/backing-dev.c
627
fprop_local_destroy_percpu(&wb->memcg_completions);
mm/backing-dev.c
63
struct bdi_writeback *wb)
mm/backing-dev.c
630
list_del(&wb->offline_node);
mm/backing-dev.c
633
wb_exit(wb);
mm/backing-dev.c
635
WARN_ON_ONCE(!list_empty(&wb->b_attached));
mm/backing-dev.c
636
WARN_ON_ONCE(work_pending(&wb->switch_work));
mm/backing-dev.c
637
call_rcu(&wb->rcu, cgwb_free_rcu);
mm/backing-dev.c
642
struct bdi_writeback *wb = container_of(refcnt, struct bdi_writeback,
mm/backing-dev.c
644
queue_work(cgwb_release_wq, &wb->release_work);
mm/backing-dev.c
647
static void cgwb_kill(struct bdi_writeback *wb)
mm/backing-dev.c
651
WARN_ON(!radix_tree_delete(&wb->bdi->cgwb_tree, wb->memcg_css->id));
mm/backing-dev.c
652
list_del(&wb->memcg_node);
mm/backing-dev.c
653
list_del(&wb->blkcg_node);
mm/backing-dev.c
654
list_add(&wb->offline_node, &offline_cgwbs);
mm/backing-dev.c
655
percpu_ref_kill(&wb->refcnt);
mm/backing-dev.c
658
static void cgwb_remove_from_bdi_list(struct bdi_writeback *wb)
mm/backing-dev.c
661
list_del_rcu(&wb->bdi_node);
mm/backing-dev.c
67
spin_lock(&wb->list_lock);
mm/backing-dev.c
671
struct bdi_writeback *wb;
mm/backing-dev.c
68
list_for_each_entry(inode, &wb->b_dirty, i_io_list)
mm/backing-dev.c
682
wb = radix_tree_lookup(&bdi->cgwb_tree, memcg_css->id);
mm/backing-dev.c
683
if (wb && wb->blkcg_css != blkcg_css) {
mm/backing-dev.c
684
cgwb_kill(wb);
mm/backing-dev.c
685
wb = NULL;
mm/backing-dev.c
688
if (wb)
mm/backing-dev.c
692
wb = kmalloc_obj(*wb, gfp);
mm/backing-dev.c
693
if (!wb) {
mm/backing-dev.c
698
ret = wb_init(wb, bdi, gfp);
mm/backing-dev.c
70
list_for_each_entry(inode, &wb->b_io, i_io_list)
mm/backing-dev.c
702
ret = percpu_ref_init(&wb->refcnt, cgwb_release, 0, gfp);
mm/backing-dev.c
706
ret = fprop_local_init_percpu(&wb->memcg_completions, gfp);
mm/backing-dev.c
710
wb->memcg_css = memcg_css;
mm/backing-dev.c
711
wb->blkcg_css = blkcg_css;
mm/backing-dev.c
712
INIT_LIST_HEAD(&wb->b_attached);
mm/backing-dev.c
713
INIT_WORK(&wb->switch_work, inode_switch_wbs_work_fn);
mm/backing-dev.c
714
init_llist_head(&wb->switch_wbs_ctxs);
mm/backing-dev.c
715
INIT_WORK(&wb->release_work, cgwb_release_workfn);
mm/backing-dev.c
716
set_bit(WB_registered, &wb->state);
mm/backing-dev.c
72
list_for_each_entry(inode, &wb->b_more_io, i_io_list)
mm/backing-dev.c
727
if (test_bit(WB_registered, &bdi->wb.state) &&
mm/backing-dev.c
730
ret = radix_tree_insert(&bdi->cgwb_tree, memcg_css->id, wb);
mm/backing-dev.c
732
list_add_tail_rcu(&wb->bdi_node, &bdi->wb_list);
mm/backing-dev.c
733
list_add(&wb->memcg_node, memcg_cgwb_list);
mm/backing-dev.c
734
list_add(&wb->blkcg_node, blkcg_cgwb_list);
mm/backing-dev.c
74
list_for_each_entry(inode, &wb->b_dirty_time, i_io_list)
mm/backing-dev.c
750
fprop_local_destroy_percpu(&wb->memcg_completions);
mm/backing-dev.c
752
percpu_ref_exit(&wb->refcnt);
mm/backing-dev.c
754
wb_exit(wb);
mm/backing-dev.c
756
kfree(wb);
mm/backing-dev.c
77
spin_unlock(&wb->list_lock);
mm/backing-dev.c
788
struct bdi_writeback *wb;
mm/backing-dev.c
79
stats->nr_writeback += wb_stat(wb, WB_WRITEBACK);
mm/backing-dev.c
791
return &bdi->wb;
mm/backing-dev.c
794
wb = radix_tree_lookup(&bdi->cgwb_tree, memcg_css->id);
mm/backing-dev.c
795
if (wb) {
mm/backing-dev.c
80
stats->nr_reclaimable += wb_stat(wb, WB_RECLAIMABLE);
mm/backing-dev.c
800
if (unlikely(wb->blkcg_css != blkcg_css || !wb_tryget(wb)))
mm/backing-dev.c
801
wb = NULL;
mm/backing-dev.c
806
return wb;
mm/backing-dev.c
81
stats->nr_dirtied += wb_stat(wb, WB_DIRTIED);
mm/backing-dev.c
82
stats->nr_written += wb_stat(wb, WB_WRITTEN);
mm/backing-dev.c
822
struct bdi_writeback *wb;
mm/backing-dev.c
827
wb = wb_get_lookup(bdi, memcg_css);
mm/backing-dev.c
828
} while (!wb && !cgwb_create(bdi, memcg_css, gfp));
mm/backing-dev.c
83
stats->wb_thresh += wb_calc_thresh(wb, stats->dirty_thresh);
mm/backing-dev.c
830
return wb;
mm/backing-dev.c
841
ret = wb_init(&bdi->wb, bdi, GFP_KERNEL);
mm/backing-dev.c
843
bdi->wb.memcg_css = &root_mem_cgroup->css;
mm/backing-dev.c
844
bdi->wb.blkcg_css = blkcg_root_css;
mm/backing-dev.c
845
INIT_WORK(&bdi->wb.switch_work, inode_switch_wbs_work_fn);
mm/backing-dev.c
846
init_llist_head(&bdi->wb.switch_wbs_ctxs);
mm/backing-dev.c
855
struct bdi_writeback *wb;
mm/backing-dev.c
857
WARN_ON(test_bit(WB_registered, &bdi->wb.state));
mm/backing-dev.c
867
wb = list_first_entry(&bdi->wb_list, struct bdi_writeback,
mm/backing-dev.c
870
wb_shutdown(wb);
mm/backing-dev.c
886
struct bdi_writeback *wb;
mm/backing-dev.c
892
wb = list_first_entry(&offline_cgwbs, struct bdi_writeback,
mm/backing-dev.c
894
list_move(&wb->offline_node, &processed);
mm/backing-dev.c
90
struct bdi_writeback *wb;
mm/backing-dev.c
905
if (wb_has_dirty_io(wb))
mm/backing-dev.c
908
if (!wb_tryget(wb))
mm/backing-dev.c
912
while (cleanup_offline_cgwb(wb))
mm/backing-dev.c
916
wb_put(wb);
mm/backing-dev.c
93
list_for_each_entry_rcu(wb, &bdi->wb_list, bdi_node) {
mm/backing-dev.c
934
struct bdi_writeback *wb, *next;
mm/backing-dev.c
937
list_for_each_entry_safe(wb, next, memcg_cgwb_list, memcg_node)
mm/backing-dev.c
938
cgwb_kill(wb);
mm/backing-dev.c
94
if (!wb_tryget(wb))
mm/backing-dev.c
953
struct bdi_writeback *wb, *next;
mm/backing-dev.c
957
list_for_each_entry_safe(wb, next, list, blkcg_node)
mm/backing-dev.c
958
cgwb_kill(wb);
mm/backing-dev.c
966
list_add_tail_rcu(&bdi->wb.bdi_node, &bdi->wb_list);
mm/backing-dev.c
97
collect_wb_stats(stats, wb);
mm/backing-dev.c
98
wb_put(wb);
mm/backing-dev.c
989
return wb_init(&bdi->wb, bdi, GFP_KERNEL);
mm/backing-dev.c
996
list_add_tail_rcu(&bdi->wb.bdi_node, &bdi->wb_list);
mm/backing-dev.c
999
static void cgwb_remove_from_bdi_list(struct bdi_writeback *wb)
mm/memcontrol.c
3386
struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb)
mm/memcontrol.c
3388
struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css);
mm/memcontrol.c
3414
void mem_cgroup_wb_stats(struct bdi_writeback *wb, unsigned long *pfilepages,
mm/memcontrol.c
3418
struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css);
mm/memcontrol.c
3484
struct bdi_writeback *wb)
mm/memcontrol.c
3493
trace_track_foreign_dirty(folio, wb);
mm/memcontrol.c
3502
if (frn->bdi_id == wb->bdi->id &&
mm/memcontrol.c
3503
frn->memcg_id == wb->memcg_css->id)
mm/memcontrol.c
3529
frn->bdi_id = wb->bdi->id;
mm/memcontrol.c
3530
frn->memcg_id = wb->memcg_css->id;
mm/memcontrol.c
3536
void mem_cgroup_flush_foreign(struct bdi_writeback *wb)
mm/memcontrol.c
3538
struct mem_cgroup *memcg = mem_cgroup_from_css(wb->memcg_css);
mm/memcontrol.c
3555
trace_flush_foreign(wb, frn->bdi_id, frn->memcg_id);
mm/page-writeback.c
1065
struct bdi_writeback *wb = dtc->wb;
mm/page-writeback.c
1066
unsigned long write_bw = READ_ONCE(wb->avg_write_bandwidth);
mm/page-writeback.c
1113
if (unlikely(wb->bdi->capabilities & BDI_CAP_STRICTLIMIT)) {
mm/page-writeback.c
1226
static void wb_update_write_bandwidth(struct bdi_writeback *wb,
mm/page-writeback.c
1231
unsigned long avg = wb->avg_write_bandwidth;
mm/page-writeback.c
1232
unsigned long old = wb->write_bandwidth;
mm/page-writeback.c
1245
bw = written - min(written, wb->written_stamp);
mm/page-writeback.c
125
#define GDTC_INIT(__wb) .wb = (__wb), \
mm/page-writeback.c
1252
bw += (u64)wb->write_bandwidth * (period - elapsed);
mm/page-writeback.c
1267
if (wb_has_dirty_io(wb)) {
mm/page-writeback.c
1268
long delta = avg - wb->avg_write_bandwidth;
mm/page-writeback.c
1270
&wb->bdi->tot_write_bandwidth) <= 0);
mm/page-writeback.c
1272
wb->write_bandwidth = bw;
mm/page-writeback.c
1273
WRITE_ONCE(wb->avg_write_bandwidth, avg);
mm/page-writeback.c
131
#define MDTC_INIT(__wb, __gdtc) .wb = (__wb), \
mm/page-writeback.c
1334
struct bdi_writeback *wb = dtc->wb;
mm/page-writeback.c
1339
unsigned long write_bw = wb->avg_write_bandwidth;
mm/page-writeback.c
1340
unsigned long dirty_ratelimit = wb->dirty_ratelimit;
mm/page-writeback.c
1352
dirty_rate = (dirtied - wb->dirtied_stamp) * HZ / elapsed;
mm/page-writeback.c
1442
if (unlikely(wb->bdi->capabilities & BDI_CAP_STRICTLIMIT)) {
mm/page-writeback.c
1448
x = min3(wb->balanced_dirty_ratelimit,
mm/page-writeback.c
1453
x = max3(wb->balanced_dirty_ratelimit,
mm/page-writeback.c
1475
WRITE_ONCE(wb->dirty_ratelimit, max(dirty_ratelimit, 1UL));
mm/page-writeback.c
1476
wb->balanced_dirty_ratelimit = balanced_dirty_ratelimit;
mm/page-writeback.c
1478
trace_bdi_dirty_ratelimit(wb, dirty_rate, task_ratelimit);
mm/page-writeback.c
1485
struct bdi_writeback *wb = gdtc->wb;
mm/page-writeback.c
1491
spin_lock(&wb->list_lock);
mm/page-writeback.c
1499
elapsed = max(now - wb->bw_time_stamp, 1UL);
mm/page-writeback.c
1500
dirtied = percpu_counter_read(&wb->stat[WB_DIRTIED]);
mm/page-writeback.c
1501
written = percpu_counter_read(&wb->stat[WB_WRITTEN]);
mm/page-writeback.c
151
static struct fprop_local_percpu *wb_memcg_completions(struct bdi_writeback *wb)
mm/page-writeback.c
1516
wb_update_write_bandwidth(wb, elapsed, written);
mm/page-writeback.c
1518
wb->dirtied_stamp = dirtied;
mm/page-writeback.c
1519
wb->written_stamp = written;
mm/page-writeback.c
1520
WRITE_ONCE(wb->bw_time_stamp, now);
mm/page-writeback.c
1521
spin_unlock(&wb->list_lock);
mm/page-writeback.c
1524
void wb_update_bandwidth(struct bdi_writeback *wb)
mm/page-writeback.c
1526
struct dirty_throttle_control gdtc = { GDTC_INIT(wb) };
mm/page-writeback.c
153
return &wb->memcg_completions;
mm/page-writeback.c
1534
static void wb_bandwidth_estimate_start(struct bdi_writeback *wb)
mm/page-writeback.c
1537
unsigned long elapsed = now - READ_ONCE(wb->bw_time_stamp);
mm/page-writeback.c
1540
!atomic_read(&wb->writeback_inodes)) {
mm/page-writeback.c
1541
spin_lock(&wb->list_lock);
mm/page-writeback.c
1542
wb->dirtied_stamp = wb_stat(wb, WB_DIRTIED);
mm/page-writeback.c
1543
wb->written_stamp = wb_stat(wb, WB_WRITTEN);
mm/page-writeback.c
1544
WRITE_ONCE(wb->bw_time_stamp, now);
mm/page-writeback.c
1545
spin_unlock(&wb->list_lock);
mm/page-writeback.c
156
static void wb_min_max_ratio(struct bdi_writeback *wb,
mm/page-writeback.c
1566
static unsigned long wb_max_pause(struct bdi_writeback *wb,
mm/page-writeback.c
1569
unsigned long bw = READ_ONCE(wb->avg_write_bandwidth);
mm/page-writeback.c
1585
static long wb_min_pause(struct bdi_writeback *wb,
mm/page-writeback.c
159
unsigned long this_bw = READ_ONCE(wb->avg_write_bandwidth);
mm/page-writeback.c
1591
long hi = ilog2(READ_ONCE(wb->avg_write_bandwidth));
mm/page-writeback.c
1592
long lo = ilog2(READ_ONCE(wb->dirty_ratelimit));
mm/page-writeback.c
160
unsigned long tot_bw = atomic_long_read(&wb->bdi->tot_write_bandwidth);
mm/page-writeback.c
161
unsigned long long min = wb->bdi->min_ratio;
mm/page-writeback.c
162
unsigned long long max = wb->bdi->max_ratio;
mm/page-writeback.c
1662
struct bdi_writeback *wb = dtc->wb;
mm/page-writeback.c
1693
wb_reclaimable = wb_stat_sum(wb, WB_RECLAIMABLE);
mm/page-writeback.c
1694
dtc->wb_dirty = wb_reclaimable + wb_stat_sum(wb, WB_WRITEBACK);
mm/page-writeback.c
1696
wb_reclaimable = wb_stat(wb, WB_RECLAIMABLE);
mm/page-writeback.c
1697
dtc->wb_dirty = wb_reclaimable + wb_stat(wb, WB_WRITEBACK);
mm/page-writeback.c
1802
static int balance_dirty_pages(struct bdi_writeback *wb,
mm/page-writeback.c
1805
struct dirty_throttle_control gdtc_stor = { GDTC_INIT(wb) };
mm/page-writeback.c
1806
struct dirty_throttle_control mdtc_stor = { MDTC_INIT(wb, &gdtc_stor) };
mm/page-writeback.c
1819
struct backing_dev_info *bdi = wb->bdi;
mm/page-writeback.c
1838
if (nr_dirty > gdtc->bg_thresh && !writeback_in_progress(wb))
mm/page-writeback.c
1839
wb_start_background_writeback(wb);
mm/page-writeback.c
185
#define GDTC_INIT(__wb) .wb = (__wb), \
mm/page-writeback.c
1879
if (unlikely(!writeback_in_progress(wb)))
mm/page-writeback.c
1880
wb_start_background_writeback(wb);
mm/page-writeback.c
1882
mem_cgroup_flush_foreign(wb);
mm/page-writeback.c
1907
wb->dirty_exceeded = gdtc->dirty_exceeded ||
mm/page-writeback.c
1909
if (time_is_before_jiffies(READ_ONCE(wb->bw_time_stamp) +
mm/page-writeback.c
1914
dirty_ratelimit = READ_ONCE(wb->dirty_ratelimit);
mm/page-writeback.c
1917
max_pause = wb_max_pause(wb, sdtc->wb_dirty);
mm/page-writeback.c
1918
min_pause = wb_min_pause(wb, max_pause,
mm/page-writeback.c
1939
trace_balance_dirty_pages(wb,
mm/page-writeback.c
1964
trace_balance_dirty_pages(wb,
mm/page-writeback.c
205
static struct fprop_local_percpu *wb_memcg_completions(struct bdi_writeback *wb)
mm/page-writeback.c
2050
struct bdi_writeback *wb = NULL;
mm/page-writeback.c
2059
wb = wb_get_create_current(bdi, GFP_KERNEL);
mm/page-writeback.c
2060
if (!wb)
mm/page-writeback.c
2061
wb = &bdi->wb;
mm/page-writeback.c
2064
if (wb->dirty_exceeded)
mm/page-writeback.c
2096
ret = balance_dirty_pages(wb, current->nr_dirtied, flags);
mm/page-writeback.c
2098
wb_put(wb);
mm/page-writeback.c
210
static void wb_min_max_ratio(struct bdi_writeback *wb,
mm/page-writeback.c
2127
struct bdi_writeback *wb = dtc->wb;
mm/page-writeback.c
213
*minp = wb->bdi->min_ratio;
mm/page-writeback.c
2131
dtc->wb_dirty = wb_stat_sum(wb, WB_RECLAIMABLE);
mm/page-writeback.c
2133
dtc->wb_dirty = wb_stat(wb, WB_RECLAIMABLE);
mm/page-writeback.c
214
*maxp = wb->bdi->max_ratio;
mm/page-writeback.c
2159
bool wb_over_bg_thresh(struct bdi_writeback *wb)
mm/page-writeback.c
2161
struct dirty_throttle_control gdtc = { GDTC_INIT(wb) };
mm/page-writeback.c
2162
struct dirty_throttle_control mdtc = { MDTC_INIT(wb, &gdtc) };
mm/page-writeback.c
2567
struct bdi_writeback *wb;
mm/page-writeback.c
2571
wb = inode_to_wb_wbc(mapping->host, wbc);
mm/page-writeback.c
2572
wb_bandwidth_estimate_start(wb);
mm/page-writeback.c
2596
if (time_is_before_jiffies(READ_ONCE(wb->bw_time_stamp) +
mm/page-writeback.c
2598
wb_update_bandwidth(wb);
mm/page-writeback.c
2626
struct bdi_writeback *wb;
mm/page-writeback.c
2630
wb = inode_to_wb(inode);
mm/page-writeback.c
2635
wb_stat_mod(wb, WB_RECLAIMABLE, nr);
mm/page-writeback.c
2636
wb_stat_mod(wb, WB_DIRTIED, nr);
mm/page-writeback.c
2641
mem_cgroup_track_foreign_dirty(folio, wb);
mm/page-writeback.c
2649
void folio_account_cleaned(struct folio *folio, struct bdi_writeback *wb)
mm/page-writeback.c
2655
wb_stat_mod(wb, WB_RECLAIMABLE, -nr);
mm/page-writeback.c
2752
struct bdi_writeback *wb;
mm/page-writeback.c
2755
wb = unlocked_inode_to_wb_begin(inode, &cookie);
mm/page-writeback.c
2758
wb_stat_mod(wb, WB_DIRTIED, -nr);
mm/page-writeback.c
2843
struct bdi_writeback *wb;
mm/page-writeback.c
2846
wb = unlocked_inode_to_wb_begin(inode, &cookie);
mm/page-writeback.c
2849
folio_account_cleaned(folio, wb);
mm/page-writeback.c
2881
struct bdi_writeback *wb;
mm/page-writeback.c
2919
wb = unlocked_inode_to_wb_begin(inode, &cookie);
mm/page-writeback.c
2924
wb_stat_mod(wb, WB_RECLAIMABLE, -nr);
mm/page-writeback.c
2934
static void wb_inode_writeback_start(struct bdi_writeback *wb)
mm/page-writeback.c
2936
atomic_inc(&wb->writeback_inodes);
mm/page-writeback.c
2939
static void wb_inode_writeback_end(struct bdi_writeback *wb)
mm/page-writeback.c
2942
atomic_dec(&wb->writeback_inodes);
mm/page-writeback.c
2950
spin_lock_irqsave(&wb->work_lock, flags);
mm/page-writeback.c
2951
if (test_bit(WB_registered, &wb->state))
mm/page-writeback.c
2952
queue_delayed_work(bdi_wq, &wb->bw_dwork, BANDWIDTH_INTERVAL);
mm/page-writeback.c
2953
spin_unlock_irqrestore(&wb->work_lock, flags);
mm/page-writeback.c
2964
struct bdi_writeback *wb;
mm/page-writeback.c
2972
wb = inode_to_wb(inode);
mm/page-writeback.c
2973
wb_stat_mod(wb, WB_WRITEBACK, -nr);
mm/page-writeback.c
2974
__wb_writeout_add(wb, nr);
mm/page-writeback.c
2976
wb_inode_writeback_end(wb);
mm/page-writeback.c
3005
struct bdi_writeback *wb;
mm/page-writeback.c
3016
wb = inode_to_wb(inode);
mm/page-writeback.c
3017
wb_stat_mod(wb, WB_WRITEBACK, nr);
mm/page-writeback.c
3019
wb_inode_writeback_start(wb);
mm/page-writeback.c
573
static inline void __wb_writeout_add(struct bdi_writeback *wb, long nr)
mm/page-writeback.c
577
wb_stat_mod(wb, WB_WRITTEN, nr);
mm/page-writeback.c
578
wb_domain_writeout_add(&global_wb_domain, &wb->completions,
mm/page-writeback.c
579
wb->bdi->max_prop_frac, nr);
mm/page-writeback.c
581
cgdom = mem_cgroup_wb_domain(wb);
mm/page-writeback.c
583
wb_domain_writeout_add(cgdom, wb_memcg_completions(wb),
mm/page-writeback.c
584
wb->bdi->max_prop_frac, nr);
mm/page-writeback.c
587
void wb_writeout_inc(struct bdi_writeback *wb)
mm/page-writeback.c
592
__wb_writeout_add(wb, 1);
mm/page-writeback.c
860
mem_cgroup_wb_stats(dtc->wb, &filepages, &headroom, &dtc->dirty,
mm/page-writeback.c
896
struct bdi_writeback *wb = dtc->wb;
mm/page-writeback.c
912
wb_min_max_ratio(wb, &wb_min_ratio, &wb_max_ratio);
mm/page-writeback.c
924
if (unlikely(wb->bdi->capabilities & BDI_CAP_STRICTLIMIT))
mm/page-writeback.c
937
unsigned long wb_calc_thresh(struct bdi_writeback *wb, unsigned long thresh)
mm/page-writeback.c
939
struct dirty_throttle_control gdtc = { GDTC_INIT(wb) };
mm/page-writeback.c
945
unsigned long cgwb_calc_thresh(struct bdi_writeback *wb)
mm/page-writeback.c
948
struct dirty_throttle_control mdtc = { MDTC_INIT(wb, &gdtc) };
scripts/kallsyms.c
650
int wa, wb;
scripts/kallsyms.c
660
wb = (sb->sym[0] == 'w') || (sb->sym[0] == 'W');
scripts/kallsyms.c
661
if (wa != wb)
scripts/kallsyms.c
662
return wa - wb;
scripts/kallsyms.c
666
wb = may_be_linker_script_provide_symbol(sb);
scripts/kallsyms.c
667
if (wa != wb)
scripts/kallsyms.c
668
return wa - wb;
scripts/kallsyms.c
672
wb = strspn(sym_name(sb), "_");
scripts/kallsyms.c
673
if (wa != wb)
scripts/kallsyms.c
674
return wa - wb;
security/apparmor/match.c
708
#define inc_wb_pos(wb) \
security/apparmor/match.c
711
wb->pos = (wb->pos + 1) & (WB_HISTORY_SIZE - 1); \
security/apparmor/match.c
712
wb->len = (wb->len + 1) > WB_HISTORY_SIZE ? WB_HISTORY_SIZE : \
security/apparmor/match.c
713
wb->len + 1; \
security/apparmor/match.c
718
static bool is_loop(struct match_workbuf *wb, aa_state_t state,
security/apparmor/match.c
721
int pos = wb->pos;
security/apparmor/match.c
724
if (wb->history[pos] < state)
security/apparmor/match.c
727
for (i = 0; i < wb->len; i++) {
security/apparmor/match.c
728
if (wb->history[pos] == state) {
security/apparmor/match.c
740
const char *str, struct match_workbuf *wb,
security/apparmor/match.c
751
AA_BUG(!wb);
security/apparmor/match.c
766
wb->history[wb->pos] = state;
security/apparmor/match.c
772
if (is_loop(wb, state, &adjust)) {
security/apparmor/match.c
777
inc_wb_pos(wb);
security/apparmor/match.c
785
wb->history[wb->pos] = state;
security/apparmor/match.c
791
if (is_loop(wb, state, &adjust)) {
security/apparmor/match.c
796
inc_wb_pos(wb);
security/apparmor/match.c
823
DEFINE_MATCH_WB(wb);
security/apparmor/match.c
827
return leftmatch_fb(dfa, start, str, &wb, count);
tools/testing/selftests/cgroup/test_zswap.c
322
static int test_zswap_writeback_one(const char *cgroup, bool wb)
tools/testing/selftests/cgroup/test_zswap.c
332
if (cg_run(cgroup, attempt_writeback, (void *) &wb))
tools/testing/selftests/cgroup/test_zswap.c
340
if (wb != !!zswpwb_after) {
tools/testing/selftests/cgroup/test_zswap.c
342
zswpwb_after, wb ? "enabled" : "disabled");
tools/testing/selftests/cgroup/test_zswap.c
350
static int test_zswap_writeback(const char *root, bool wb)
tools/testing/selftests/cgroup/test_zswap.c
363
if (cg_write(test_group, "memory.zswap.writeback", wb ? "1" : "0"))
tools/testing/selftests/cgroup/test_zswap.c
366
if (test_zswap_writeback_one(test_group, wb))
tools/testing/selftests/cgroup/test_zswap.c
385
if (test_zswap_writeback_one(test_group_child, wb))