Symbol: dirty
arch/arc/mm/tlb.c
491
int dirty = !test_and_set_bit(PG_dc_clean, &folio->flags.f);
arch/arc/mm/tlb.c
492
if (dirty) {
arch/arm/mm/cache-l2x0.c
1032
u32 dirty = 0;
arch/arm/mm/cache-l2x0.c
1052
of_property_read_u32(np, "arm,dirty-latency", &dirty);
arch/arm/mm/cache-l2x0.c
1053
if (dirty) {
arch/arm/mm/cache-l2x0.c
1055
val |= (dirty - 1) << L2X0_AUX_CTRL_DIRTY_LATENCY_SHIFT;
arch/arm/probes/kprobes/opt-arm.c
140
__arch_remove_optimized_kprobe(struct optimized_kprobe *op, int dirty)
arch/arm/probes/kprobes/opt-arm.c
143
free_optinsn_slot(op->optinsn.insn, dirty);
arch/arm64/include/asm/hugetlb.h
43
pte_t pte, int dirty);
arch/arm64/include/asm/pgtable.h
1252
pte_t entry, int dirty);
arch/arm64/include/asm/pgtable.h
1258
pmd_t entry, int dirty)
arch/arm64/include/asm/pgtable.h
1261
pmd_pte(entry), dirty);
arch/arm64/include/asm/pgtable.h
1664
pte_t entry, int dirty);
arch/arm64/include/asm/pgtable.h
1891
pte_t entry, int dirty)
arch/arm64/include/asm/pgtable.h
1898
return __ptep_set_access_flags(vma, addr, ptep, entry, dirty);
arch/arm64/include/asm/pgtable.h
1900
return contpte_ptep_set_access_flags(vma, addr, ptep, entry, dirty);
arch/arm64/mm/contpte.c
625
pte_t entry, int dirty)
arch/arm64/mm/contpte.c
687
if (dirty)
arch/arm64/mm/contpte.c
691
__ptep_set_access_flags(vma, addr, ptep, entry, dirty);
arch/arm64/mm/fault.c
209
pte_t entry, int dirty)
arch/arm64/mm/fault.c
241
if (dirty)
arch/arm64/mm/hugetlbpage.c
422
pte_t pte, int dirty)
arch/arm64/mm/hugetlbpage.c
432
return __ptep_set_access_flags(vma, addr, ptep, pte, dirty);
arch/loongarch/include/asm/hugetlb.h
59
int dirty)
arch/m68k/fpsp040/fpsp.h
128
.set FPR_DIRTY_BITS,LV-91 | fpr dirty bits
arch/mips/cavium-octeon/executive/cvmx-l2c.c
680
tag.s.D = l2c_tadx_tag.s.dirty;
arch/mips/include/asm/hugetlb.h
55
int dirty)
arch/mips/include/asm/octeon/cvmx-l2c-defs.h
213
__BITFIELD_FIELD(uint64_t dirty:1,
arch/mips/include/asm/pgtable.h
496
pte_t entry, int dirty)
arch/nios2/include/asm/pgtable.h
293
pte_t entry, int dirty)
arch/openrisc/mm/cache.c
86
int dirty = !test_and_set_bit(PG_dc_clean, &folio->flags.f);
arch/openrisc/mm/cache.c
93
if ((vma->vm_flags & VM_EXEC) && dirty) {
arch/parisc/include/asm/hugetlb.h
29
pte_t pte, int dirty);
arch/parisc/mm/hugetlbpage.c
151
pte_t pte, int dirty)
arch/powerpc/include/asm/book3s/64/pgtable.h
1296
pmd_t entry, int dirty);
arch/powerpc/include/asm/book3s/64/pgtable.h
1300
pud_t entry, int dirty);
arch/powerpc/include/asm/hugetlb.h
69
pte_t pte, int dirty);
arch/powerpc/include/asm/kvm_book3s.h
250
unsigned long gpa, bool dirty);
arch/powerpc/include/asm/kvm_host.h
373
bool dirty; /* true => area has been modified by kernel */
arch/powerpc/include/asm/pgtable.h
112
pte_t *ptep, pte_t entry, int dirty);
arch/powerpc/kernel/asm-offsets.c
430
OFFSET(VCPU_VPA_DIRTY, kvm_vcpu, arch.vpa.dirty);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1092
if (!vpa->dirty || !vpa->pinned_addr)
arch/powerpc/kvm/book3s_64_mmu_hv.c
1099
vpa->dirty = false;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1159
bool dirty)
arch/powerpc/kvm/book3s_64_mmu_hv.c
1168
if (!dirty)
arch/powerpc/kvm/book3s_64_mmu_hv.c
1630
int valid, dirty;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1633
dirty = hpte_dirty(revp, hptp);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1634
if (!first_pass && !dirty)
arch/powerpc/kvm/book3s_64_mmu_hv.c
1648
if (valid || dirty) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1662
dirty = !!(revp->guest_rpte & HPTE_GR_MODIFIED);
arch/powerpc/kvm/book3s_64_mmu_hv.c
1669
dirty = 1;
arch/powerpc/kvm/book3s_64_mmu_hv.c
1682
if (valid == want_valid && dirty) {
arch/powerpc/kvm/book3s_64_mmu_hv.c
1688
if (!(valid == want_valid && (first_pass || dirty)))
arch/powerpc/kvm/book3s_hv.c
3137
vpa->dirty);
arch/powerpc/kvm/book3s_hv.c
4107
vcpu->arch.vpa.dirty = 1;
arch/powerpc/kvm/book3s_hv.c
6381
vpa->dirty = false;
arch/powerpc/kvm/book3s_hv.c
724
vpap->dirty = false;
arch/powerpc/kvm/book3s_hv.c
746
old_vpa.dirty);
arch/powerpc/kvm/book3s_hv.c
758
old_vpa.dirty);
arch/powerpc/kvm/book3s_hv.c
766
old_vpa.dirty);
arch/powerpc/kvm/book3s_hv.c
853
vcpu->arch.vpa.dirty = true;
arch/powerpc/kvm/book3s_hv.c
876
vcpu->arch.vpa.dirty = true;
arch/powerpc/kvm/e500_mmu.c
857
struct kvm_dirty_tlb *dirty)
arch/powerpc/kvm/powerpc.c
2076
struct kvm_dirty_tlb dirty;
arch/powerpc/kvm/powerpc.c
2078
if (copy_from_user(&dirty, argp, sizeof(dirty)))
arch/powerpc/kvm/powerpc.c
2081
r = kvm_vcpu_ioctl_dirty_tlb(vcpu, &dirty);
arch/powerpc/mm/book3s64/pgtable.c
62
pmd_t *pmdp, pmd_t entry, int dirty)
arch/powerpc/mm/book3s64/pgtable.c
82
pud_t *pudp, pud_t entry, int dirty)
arch/powerpc/mm/pgtable.c
146
int dirty)
arch/powerpc/mm/pgtable.c
161
if (dirty || pte_exec(pte) || !is_exec_fault())
arch/powerpc/mm/pgtable.c
256
pte_t *ptep, pte_t entry, int dirty)
arch/powerpc/mm/pgtable.c
259
entry = set_access_flags_filter(entry, vma, dirty);
arch/powerpc/mm/pgtable.c
272
pte_t pte, int dirty)
arch/powerpc/mm/pgtable.c
280
ptep_set_access_flags(vma, addr, ptep, pte, dirty);
arch/powerpc/mm/pgtable.c
285
pte = set_access_flags_filter(pte, vma, dirty);
arch/riscv/include/asm/hugetlb.h
45
pte_t pte, int dirty);
arch/riscv/include/asm/pgtable.h
1012
pmd_t entry, int dirty)
arch/riscv/include/asm/pgtable.h
1014
return ptep_set_access_flags(vma, address, (pte_t *)pmdp, pmd_pte(entry), dirty);
arch/riscv/include/asm/pgtable.h
1107
pud_t entry, int dirty)
arch/riscv/include/asm/pgtable.h
1109
return ptep_set_access_flags(vma, address, (pte_t *)pudp, pud_pte(entry), dirty);
arch/riscv/include/asm/pgtable.h
660
pte_t *ptep, pte_t entry, int dirty);
arch/riscv/mm/hugetlbpage.c
278
int dirty)
arch/riscv/mm/hugetlbpage.c
286
return ptep_set_access_flags(vma, addr, ptep, pte, dirty);
arch/riscv/mm/pgtable.c
10
pte_t entry, int dirty)
arch/s390/include/asm/hugetlb.h
60
pte_t pte, int dirty)
arch/s390/include/asm/pgtable.h
1305
pte_t entry, int dirty)
arch/s390/include/asm/pgtable.h
1672
pmd_t entry, int dirty)
arch/s390/include/asm/pgtable.h
1677
if (dirty)
arch/s390/kvm/dat.c
1063
int dat_perform_essa(union asce asce, gfn_t gfn, int orc, union essa_state *state, bool *dirty)
arch/s390/kvm/dat.c
1135
*dirty = !pgste.cmma_d;
arch/s390/kvm/dat.h
393
static inline union pte _pte(kvm_pfn_t pfn, bool writable, bool dirty, bool special)
arch/s390/kvm/dat.h
397
res.h.p = !dirty;
arch/s390/kvm/dat.h
401
res.s.d = dirty;
arch/s390/kvm/dat.h
402
res.s.sd = dirty;
arch/s390/kvm/dat.h
431
static inline union crste _crste_fc1(kvm_pfn_t pfn, int tt, bool writable, bool dirty)
arch/s390/kvm/dat.h
436
res.h.p = !dirty;
arch/s390/kvm/dat.h
441
res.s.fc1.d = dirty;
arch/s390/kvm/dat.h
442
res.s.fc1.sd = dirty;
arch/s390/kvm/dat.h
544
int dat_perform_essa(union asce asce, gfn_t gfn, int orc, union essa_state *state, bool *dirty);
arch/sh/mm/cache.c
151
int dirty = !test_and_set_bit(PG_dcache_clean, &folio->flags.f);
arch/sh/mm/cache.c
152
if (dirty)
arch/sparc/include/asm/hugetlb.h
43
pte_t pte, int dirty)
arch/sparc/mm/init_64.c
477
bool dirty = test_bit(PG_dcache_dirty, &folio->flags.f);
arch/sparc/mm/init_64.c
478
if (dirty) {
arch/x86/events/core.c
1661
__set_bit(event->hw.idx, cpuc->dirty);
arch/x86/events/core.c
2552
__clear_bit(cpuc->assign[i], cpuc->dirty);
arch/x86/events/core.c
2554
if (bitmap_empty(cpuc->dirty, X86_PMC_IDX_MAX))
arch/x86/events/core.c
2557
for_each_set_bit(i, cpuc->dirty, X86_PMC_IDX_MAX) {
arch/x86/events/core.c
2569
bitmap_zero(cpuc->dirty, X86_PMC_IDX_MAX);
arch/x86/events/perf_event.h
271
unsigned long dirty[BITS_TO_LONGS(X86_PMC_IDX_MAX)];
arch/x86/include/asm/pgtable.h
1240
pte_t entry, int dirty);
arch/x86/include/asm/pgtable.h
1300
pmd_t entry, int dirty);
arch/x86/include/asm/pgtable.h
1303
pud_t entry, int dirty);
arch/x86/kernel/cpu/sgx/main.c
102
list_move_tail(&page->list, &dirty);
arch/x86/kernel/cpu/sgx/main.c
109
list_splice(&dirty, dirty_page_list);
arch/x86/kernel/cpu/sgx/main.c
64
LIST_HEAD(dirty);
arch/x86/kernel/kprobes/opt.c
361
void __arch_remove_optimized_kprobe(struct optimized_kprobe *op, int dirty)
arch/x86/kernel/kprobes/opt.c
368
if (dirty)
arch/x86/kernel/kprobes/opt.c
371
free_optinsn_slot(slot, dirty);
arch/x86/kvm/emulate.c
251
unsigned long dirty = ctxt->regs_dirty;
arch/x86/kvm/emulate.c
254
for_each_set_bit(reg, &dirty, NR_EMULATOR_GPRS)
arch/x86/kvm/x86.c
3811
goto dirty;
arch/x86/kvm/x86.c
3840
dirty:
arch/x86/mm/pgtable.c
393
pte_t entry, int dirty)
arch/x86/mm/pgtable.c
397
if (changed && dirty)
arch/x86/mm/pgtable.c
406
pmd_t entry, int dirty)
arch/x86/mm/pgtable.c
412
if (changed && dirty) {
arch/x86/mm/pgtable.c
426
pud_t *pudp, pud_t entry, int dirty)
arch/x86/mm/pgtable.c
432
if (changed && dirty) {
drivers/block/amiflop.c
1437
if (unit[drive].dirty == 1) {
drivers/block/amiflop.c
1496
floppy->dirty = 1;
drivers/block/amiflop.c
1716
if (unit[drive].dirty == 1) {
drivers/block/amiflop.c
1757
p->dirty = 0;
drivers/block/amiflop.c
202
int dirty; /* true when trackbuf is not on disk */
drivers/block/amiflop.c
763
unit[drive].dirty = 0;
drivers/char/ps3flash.c
220
priv->dirty = true;
drivers/char/ps3flash.c
28
bool dirty;
drivers/char/ps3flash.c
53
if (!priv->dirty || priv->tag < 0)
drivers/char/ps3flash.c
60
priv->dirty = false;
drivers/char/xilinx_hwicap/buffer_icap.c
272
bool dirty = false;
drivers/char/xilinx_hwicap/buffer_icap.c
281
dirty = true;
drivers/char/xilinx_hwicap/buffer_icap.c
300
dirty = false;
drivers/char/xilinx_hwicap/buffer_icap.c
304
if (dirty) {
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
554
.dirty = amdgpu_dirtyfb
drivers/gpu/drm/drm_client.c
446
if (!buffer || !buffer->fb || !buffer->fb->funcs->dirty)
drivers/gpu/drm/drm_client.c
457
return buffer->fb->funcs->dirty(buffer->fb, buffer->client->file,
drivers/gpu/drm/drm_client.c
461
return buffer->fb->funcs->dirty(buffer->fb, buffer->client->file,
drivers/gpu/drm/drm_fbdev_dma.c
178
if (helper->fb->funcs->dirty) {
drivers/gpu/drm/drm_fbdev_dma.c
183
ret = helper->fb->funcs->dirty(helper->fb, NULL, 0, 0, clip, 1);
drivers/gpu/drm/drm_fbdev_dma.c
306
if (fb->funcs->dirty)
drivers/gpu/drm/drm_fbdev_shmem.c
116
if (helper->fb->funcs->dirty) {
drivers/gpu/drm/drm_fbdev_shmem.c
117
ret = helper->fb->funcs->dirty(helper->fb, NULL, 0, 0, clip, 1);
drivers/gpu/drm/drm_fbdev_ttm.c
155
if (helper->fb->funcs->dirty) {
drivers/gpu/drm/drm_fbdev_ttm.c
156
ret = helper->fb->funcs->dirty(helper->fb, NULL, 0, 0, clip, 1);
drivers/gpu/drm/drm_framebuffer.c
764
if (fb->funcs->dirty) {
drivers/gpu/drm/drm_framebuffer.c
765
ret = fb->funcs->dirty(fb, file_priv, flags, r->color,
drivers/gpu/drm/drm_gem.c
741
bool dirty, bool accessed)
drivers/gpu/drm/drm_gem.c
766
if (dirty)
drivers/gpu/drm/drm_gem_framebuffer_helper.c
296
.dirty = drm_atomic_helper_dirtyfb,
drivers/gpu/drm/i915/display/i9xx_wm.c
1028
bool dirty = false;
drivers/gpu/drm/i915/display/i9xx_wm.c
1033
dirty |= raw->plane[plane_id] != value;
drivers/gpu/drm/i915/display/i9xx_wm.c
1037
return dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
1044
bool dirty = false;
drivers/gpu/drm/i915/display/i9xx_wm.c
1052
dirty |= raw->fbc != value;
drivers/gpu/drm/i915/display/i9xx_wm.c
1056
return dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
1069
bool dirty = false;
drivers/gpu/drm/i915/display/i9xx_wm.c
1073
dirty |= g4x_raw_plane_wm_set(crtc_state, 0, plane_id, 0);
drivers/gpu/drm/i915/display/i9xx_wm.c
1075
dirty |= g4x_raw_fbc_wm_set(crtc_state, 0, 0);
drivers/gpu/drm/i915/display/i9xx_wm.c
1089
dirty |= raw->plane[plane_id] != wm;
drivers/gpu/drm/i915/display/i9xx_wm.c
1107
dirty |= raw->fbc != wm;
drivers/gpu/drm/i915/display/i9xx_wm.c
1112
dirty |= g4x_raw_plane_wm_set(crtc_state, level, plane_id, USHRT_MAX);
drivers/gpu/drm/i915/display/i9xx_wm.c
1115
dirty |= g4x_raw_fbc_wm_set(crtc_state, level, USHRT_MAX);
drivers/gpu/drm/i915/display/i9xx_wm.c
1118
if (dirty) {
drivers/gpu/drm/i915/display/i9xx_wm.c
1133
return dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
1268
unsigned int dirty = 0;
drivers/gpu/drm/i915/display/i9xx_wm.c
1279
dirty |= BIT(plane->id);
drivers/gpu/drm/i915/display/i9xx_wm.c
1282
if (!dirty)
drivers/gpu/drm/i915/display/i9xx_wm.c
1670
bool dirty = false;
drivers/gpu/drm/i915/display/i9xx_wm.c
1675
dirty |= raw->plane[plane_id] != value;
drivers/gpu/drm/i915/display/i9xx_wm.c
1679
return dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
1689
bool dirty = false;
drivers/gpu/drm/i915/display/i9xx_wm.c
1692
dirty |= vlv_raw_plane_wm_set(crtc_state, 0, plane_id, 0);
drivers/gpu/drm/i915/display/i9xx_wm.c
1704
dirty |= raw->plane[plane_id] != wm;
drivers/gpu/drm/i915/display/i9xx_wm.c
1709
dirty |= vlv_raw_plane_wm_set(crtc_state, level, plane_id, USHRT_MAX);
drivers/gpu/drm/i915/display/i9xx_wm.c
1712
if (dirty)
drivers/gpu/drm/i915/display/i9xx_wm.c
1720
return dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
1807
unsigned int dirty = 0;
drivers/gpu/drm/i915/display/i9xx_wm.c
1818
dirty |= BIT(plane->id);
drivers/gpu/drm/i915/display/i9xx_wm.c
1831
dirty = ~0;
drivers/gpu/drm/i915/display/i9xx_wm.c
1833
if (!dirty)
drivers/gpu/drm/i915/display/i9xx_wm.c
1837
if (dirty & ~BIT(PLANE_CURSOR)) {
drivers/gpu/drm/i915/display/i9xx_wm.c
3280
unsigned int dirty = 0;
drivers/gpu/drm/i915/display/i9xx_wm.c
3286
dirty |= WM_DIRTY_PIPE(pipe);
drivers/gpu/drm/i915/display/i9xx_wm.c
3288
dirty |= WM_DIRTY_LP_ALL;
drivers/gpu/drm/i915/display/i9xx_wm.c
3293
dirty |= WM_DIRTY_FBC;
drivers/gpu/drm/i915/display/i9xx_wm.c
3295
dirty |= WM_DIRTY_LP_ALL;
drivers/gpu/drm/i915/display/i9xx_wm.c
3299
dirty |= WM_DIRTY_DDB;
drivers/gpu/drm/i915/display/i9xx_wm.c
3301
dirty |= WM_DIRTY_LP_ALL;
drivers/gpu/drm/i915/display/i9xx_wm.c
3305
if (dirty & WM_DIRTY_LP_ALL)
drivers/gpu/drm/i915/display/i9xx_wm.c
3306
return dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
3317
dirty |= WM_DIRTY_LP(wm_lp);
drivers/gpu/drm/i915/display/i9xx_wm.c
3319
return dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
3323
unsigned int dirty)
drivers/gpu/drm/i915/display/i9xx_wm.c
3328
if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] & WM_LP_ENABLE) {
drivers/gpu/drm/i915/display/i9xx_wm.c
3333
if (dirty & WM_DIRTY_LP(2) && previous->wm_lp[1] & WM_LP_ENABLE) {
drivers/gpu/drm/i915/display/i9xx_wm.c
3338
if (dirty & WM_DIRTY_LP(1) && previous->wm_lp[0] & WM_LP_ENABLE) {
drivers/gpu/drm/i915/display/i9xx_wm.c
3360
unsigned int dirty;
drivers/gpu/drm/i915/display/i9xx_wm.c
3362
dirty = ilk_compute_wm_dirty(display, previous, results);
drivers/gpu/drm/i915/display/i9xx_wm.c
3363
if (!dirty)
drivers/gpu/drm/i915/display/i9xx_wm.c
3366
_ilk_disable_lp_wm(display, dirty);
drivers/gpu/drm/i915/display/i9xx_wm.c
3368
if (dirty & WM_DIRTY_PIPE(PIPE_A))
drivers/gpu/drm/i915/display/i9xx_wm.c
3370
if (dirty & WM_DIRTY_PIPE(PIPE_B))
drivers/gpu/drm/i915/display/i9xx_wm.c
3372
if (dirty & WM_DIRTY_PIPE(PIPE_C))
drivers/gpu/drm/i915/display/i9xx_wm.c
3375
if (dirty & WM_DIRTY_DDB) {
drivers/gpu/drm/i915/display/i9xx_wm.c
3386
if (dirty & WM_DIRTY_FBC)
drivers/gpu/drm/i915/display/i9xx_wm.c
3390
if (dirty & WM_DIRTY_LP(1) &&
drivers/gpu/drm/i915/display/i9xx_wm.c
3395
if (dirty & WM_DIRTY_LP(2) && previous->wm_lp_spr[1] != results->wm_lp_spr[1])
drivers/gpu/drm/i915/display/i9xx_wm.c
3397
if (dirty & WM_DIRTY_LP(3) && previous->wm_lp_spr[2] != results->wm_lp_spr[2])
drivers/gpu/drm/i915/display/i9xx_wm.c
3401
if (dirty & WM_DIRTY_LP(1) && previous->wm_lp[0] != results->wm_lp[0])
drivers/gpu/drm/i915/display/i9xx_wm.c
3403
if (dirty & WM_DIRTY_LP(2) && previous->wm_lp[1] != results->wm_lp[1])
drivers/gpu/drm/i915/display/i9xx_wm.c
3405
if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] != results->wm_lp[2])
drivers/gpu/drm/i915/display/intel_dpt.c
171
dpt->obj->mm.dirty = true;
drivers/gpu/drm/i915/display/intel_fb.c
2205
.dirty = intel_user_framebuffer_dirty,
drivers/gpu/drm/i915/display/intel_fbdev.c
173
if (helper->fb->funcs->dirty)
drivers/gpu/drm/i915/display/intel_fbdev.c
174
return helper->fb->funcs->dirty(helper->fb, NULL, 0, 0, clip, 1);
drivers/gpu/drm/i915/gem/i915_gem_domain.c
174
obj->mm.dirty = true;
drivers/gpu/drm/i915/gem/i915_gem_domain.c
238
obj->mm.dirty = true;
drivers/gpu/drm/i915/gem/i915_gem_domain.c
763
obj->mm.dirty = true;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
1244
if (!obj->mm.dirty)
drivers/gpu/drm/i915/gem/i915_gem_internal.c
125
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
287
obj->mm.dirty = true;
drivers/gpu/drm/i915/gem/i915_gem_mman.c
485
obj->mm.dirty = true;
drivers/gpu/drm/i915/gem/i915_gem_object.h
852
bool dirty, bool backup);
drivers/gpu/drm/i915/gem/i915_gem_object_types.h
691
bool dirty:1;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
613
obj->mm.dirty = true;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
780
if (!obj->mm.dirty)
drivers/gpu/drm/i915/gem/i915_gem_phys.c
106
if (obj->mm.dirty) {
drivers/gpu/drm/i915/gem/i915_gem_phys.c
128
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
35
bool dirty, bool backup)
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
363
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
393
obj->mm.dirty, obj->mm.madv == I915_MADV_WILLNEED);
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
395
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/i915_gem_shmem.c
51
if (dirty)
drivers/gpu/drm/i915/gem/i915_gem_tiling.c
303
vma->fence->dirty = true;
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
177
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
180
if (obj->mm.dirty && trylock_page(page)) {
drivers/gpu/drm/i915/gem/i915_gem_userptr.c
205
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/selftests/huge_gem_object.c
91
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
161
obj->mm.dirty = false;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
310
obj->mm.dirty = false;
drivers/gpu/drm/i915/gt/intel_context.c
136
vma->obj->mm.dirty = true;
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
232
WRITE_ONCE(fence->dirty, false);
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.c
379
if (!fence->dirty) {
drivers/gpu/drm/i915/gt/intel_ggtt_fencing.h
37
bool dirty;
drivers/gpu/drm/i915/i915_debugfs.c
190
obj->mm.dirty ? " dirty" : "",
drivers/gpu/drm/i915/i915_gem.c
886
reg->dirty = true;
drivers/gpu/drm/i915/i915_vma.c
2039
obj->mm.dirty = true;
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
105
obj->mm.dirty = false;
drivers/gpu/drm/mediatek/mtk_crtc.c
583
if (plane_state->pending.dirty) {
drivers/gpu/drm/mediatek/mtk_crtc.c
585
plane_state->pending.dirty = false;
drivers/gpu/drm/mediatek/mtk_plane.c
273
mtk_plane_state->pending.dirty = true;
drivers/gpu/drm/mediatek/mtk_plane.c
296
mtk_plane_state->pending.dirty = true;
drivers/gpu/drm/mediatek/mtk_plane.h
32
bool dirty;
drivers/gpu/drm/msm/msm_fb.c
56
.dirty = msm_framebuffer_dirtyfb,
drivers/gpu/drm/msm/msm_fbdev.c
76
if (helper->fb->funcs->dirty) {
drivers/gpu/drm/msm/msm_fbdev.c
77
ret = helper->fb->funcs->dirty(helper->fb, NULL, 0, 0, clip, 1);
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1132
if (reg->dirty) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1320
if (!op->reg || op->reg->dirty) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1363
if (unlikely(reg->dirty)) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
433
reg->dirty = true;
drivers/gpu/drm/nouveau/nouveau_uvmm.h
28
bool dirty;
drivers/gpu/drm/omapdrm/omap_fb.c
82
.dirty = omap_framebuffer_dirty,
drivers/gpu/drm/omapdrm/omap_fbdev.c
132
if (helper->fb->funcs->dirty)
drivers/gpu/drm/omapdrm/omap_fbdev.c
133
return helper->fb->funcs->dirty(helper->fb, NULL, 0, 0, clip, 1);
drivers/gpu/drm/omapdrm/omap_gem.c
715
bool dirty = false;
drivers/gpu/drm/omapdrm/omap_gem.c
732
dirty = true;
drivers/gpu/drm/omapdrm/omap_gem.c
737
if (dirty) {
drivers/gpu/drm/qxl/qxl_display.c
471
.dirty = qxl_framebuffer_surface_dirty,
drivers/gpu/drm/rockchip/rockchip_drm_fb.c
24
.dirty = drm_atomic_helper_dirtyfb,
drivers/gpu/drm/tests/drm_buddy_test.c
375
LIST_HEAD(dirty);
drivers/gpu/drm/tests/drm_buddy_test.c
405
list = &dirty;
drivers/gpu/drm/tests/drm_buddy_test.c
421
list_for_each_entry(block, &dirty, link)
drivers/gpu/drm/tests/drm_buddy_test.c
436
drm_buddy_free_list(&mm, &dirty, 0);
drivers/gpu/drm/tests/drm_buddy_test.c
457
list = &dirty;
drivers/gpu/drm/tests/drm_buddy_test.c
467
drm_buddy_free_list(&mm, &dirty, 0);
drivers/gpu/drm/tiny/gm12u320.c
403
struct drm_rect *dirty)
drivers/gpu/drm/tiny/gm12u320.c
415
gm12u320->fb_update.rect = *dirty;
drivers/gpu/drm/tiny/gm12u320.c
421
rect->x1 = min(rect->x1, dirty->x1);
drivers/gpu/drm/tiny/gm12u320.c
422
rect->y1 = min(rect->y1, dirty->y1);
drivers/gpu/drm/tiny/gm12u320.c
423
rect->x2 = max(rect->x2, dirty->x2);
drivers/gpu/drm/tiny/gm12u320.c
424
rect->y2 = max(rect->y2, dirty->y2);
drivers/gpu/drm/virtio/virtgpu_display.c
67
.dirty = drm_atomic_helper_dirtyfb,
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1011
unsigned long *dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1020
i = find_first_bit(dirty, max_num);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1037
next_bit = find_next_bit(dirty, max_num, i + 1);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
110
unsigned long dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1160
while ((hit = find_next_bit(&cbs->dirty, VMW_BINDING_NUM_BITS, hit))
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1188
__clear_bit(hit, &cbs->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1211
&cbs->per_shader[biv->shader_slot].dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1212
__set_bit(VMW_BINDING_PS_BIT, &cbs->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1229
__set_bit(VMW_BINDING_RT_BIT, &cbs->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1246
__set_bit(VMW_BINDING_SO_T_BIT, &cbs->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1266
__set_bit(VMW_BINDING_VB_BIT, &cbs->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1312
__set_bit(VMW_BINDING_UAV_BIT, &cbs->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
1320
__set_bit(VMW_BINDING_CS_UAV_BIT, &cbs->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
784
unsigned long *dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
792
i = find_first_bit(dirty, max_num);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
803
next_bit = find_next_bit(dirty, max_num, i + 1);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
981
if (!test_bit(VMW_BINDING_PS_SR_BIT, &sb->dirty))
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
988
__clear_bit(VMW_BINDING_PS_SR_BIT, &sb->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_binding.h
194
unsigned long dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
59
if (res->dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_bo.c
73
WARN_ON(vbo->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_bo.h
99
struct vmw_bo_dirty *dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
490
bool dirty = false;
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
499
} else if (new_bo->dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
501
dirty = vmw_bo_is_dirty(new_bo);
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
502
if (dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_cursor_plane.c
509
return dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
151
struct vmw_resource_dirty *dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
806
bool dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1506
bool dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1534
dirty = (cmd->body.transfer == SVGA3D_WRITE_HOST_VRAM) ?
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
1536
ret = vmw_cmd_res_check(dev_priv, sw_context, vmw_res_surface, dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
294
u32 dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
307
if (dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
309
rcache->private, dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
314
ret = vmw_validation_add_resource(sw_context->ctx, res, 0, dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
322
dirty, (void **)&ctx_info,
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
617
u32 dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
640
if (dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
642
rcache->private, dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c
659
ret = vmw_execbuf_res_val_add(sw_context, res, dirty, vmw_val_add_flag_none);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1512
struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1519
dirty->dev_priv = dev_priv;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1522
if (dirty->crtc) {
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1523
units[num_units++] = vmw_crtc_to_du(dirty->crtc);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1543
dirty->unit = unit;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1544
if (dirty->fifo_reserve_size > 0) {
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1545
dirty->cmd = VMW_CMD_RESERVE(dev_priv,
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1546
dirty->fifo_reserve_size);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1547
if (!dirty->cmd)
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1550
memset(dirty->cmd, 0, dirty->fifo_reserve_size);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1552
dirty->num_hits = 0;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1564
dirty->fb_x = (s32) clips_ptr->x1;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1565
dirty->fb_y = (s32) clips_ptr->y1;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1566
dirty->unit_x2 = (s32) clips_ptr->x2 + dest_x -
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1568
dirty->unit_y2 = (s32) clips_ptr->y2 + dest_y -
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1571
dirty->fb_x = vclips_ptr->x;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1572
dirty->fb_y = vclips_ptr->y;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1573
dirty->unit_x2 = dirty->fb_x + vclips_ptr->w +
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1575
dirty->unit_y2 = dirty->fb_y + vclips_ptr->h +
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1579
dirty->unit_x1 = dirty->fb_x + dest_x - crtc_x;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1580
dirty->unit_y1 = dirty->fb_y + dest_y - crtc_y;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1583
if (dirty->unit_x1 >= crtc_width ||
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1584
dirty->unit_y1 >= crtc_height ||
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1585
dirty->unit_x2 <= 0 || dirty->unit_y2 <= 0)
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1589
dirty->unit_x2 = min_t(s32, dirty->unit_x2,
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1591
dirty->unit_y2 = min_t(s32, dirty->unit_y2,
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1595
clip_left = min_t(s32, dirty->unit_x1, 0);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1596
clip_top = min_t(s32, dirty->unit_y1, 0);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1597
dirty->unit_x1 -= clip_left;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1598
dirty->unit_y1 -= clip_top;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1599
dirty->fb_x -= clip_left;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1600
dirty->fb_y -= clip_top;
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1602
dirty->clip(dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
1605
dirty->fifo_commit(dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
433
if (!bo->dirty && surf && surf->res.dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
497
.dirty = drm_atomic_helper_dirtyfb,
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
603
.dirty = drm_atomic_helper_dirtyfb,
drivers/gpu/drm/vmwgfx/vmwgfx_kms.c
774
if (surface->res.dirty == NULL)
drivers/gpu/drm/vmwgfx/vmwgfx_kms.h
382
struct vmw_kms_dirty *dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
108
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
113
if (dirty->end <= dirty->start)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
117
dirty->start + offset,
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
118
dirty->end - dirty->start);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
120
if (100UL * num_marked / dirty->bitmap_size >
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
122
dirty->change_count++;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
124
dirty->change_count = 0;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
126
if (dirty->change_count > VMW_DIRTY_NUM_CHANGE_TRIGGERS) {
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
128
pgoff_t end = dirty->bitmap_size;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
130
dirty->method = VMW_BO_DIRTY_PAGETABLE;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
132
&dirty->bitmap[0],
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
134
bitmap_clear(&dirty->bitmap[0], 0, dirty->bitmap_size);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
135
if (dirty->start < dirty->end)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
136
bitmap_set(&dirty->bitmap[0], dirty->start,
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
137
dirty->end - dirty->start);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
138
dirty->change_count = 0;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
151
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
153
if (dirty->method == VMW_BO_DIRTY_PAGETABLE)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
173
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
177
if (dirty->method != VMW_BO_DIRTY_PAGETABLE || start >= end)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
183
&dirty->bitmap[0], &dirty->start,
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
184
&dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
218
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
223
if (dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
224
kref_get(&dirty->ref_count);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
228
size = sizeof(*dirty) + BITS_TO_LONGS(num_pages) * sizeof(long);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
229
dirty = kvzalloc(size, GFP_KERNEL);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
230
if (!dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
235
dirty->bitmap_size = num_pages;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
236
dirty->start = dirty->bitmap_size;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
237
dirty->end = 0;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
238
kref_init(&dirty->ref_count);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
240
dirty->method = VMW_BO_DIRTY_PAGETABLE;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
245
dirty->method = VMW_BO_DIRTY_MKWRITE;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
251
&dirty->bitmap[0],
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
252
&dirty->start, &dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
255
vbo->dirty = dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
265
struct vmw_bo_dirty *dirty = container_of(kref, struct vmw_bo_dirty, ref_count);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
267
kvfree(dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
282
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
284
if (dirty && kref_put(&dirty->ref_count, vmw_bo_dirty_free))
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
285
vbo->dirty = NULL;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
301
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
310
if (res_start >= dirty->end || res_end <= dirty->start)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
313
cur = max(res_start, dirty->start);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
314
res_end = max(res_end, dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
318
start = find_next_bit(&dirty->bitmap[0], res_end, cur);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
322
end = find_next_zero_bit(&dirty->bitmap[0], res_end, start + 1);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
325
bitmap_clear(&dirty->bitmap[0], start, num);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
329
if (res_start <= dirty->start && res_end > dirty->start)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
330
dirty->start = res_end;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
331
if (res_start < dirty->end && res_end >= dirty->end)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
332
dirty->end = res_start;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
337
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
346
if (res_start >= dirty->end || res_end <= dirty->start)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
349
cur = max(res_start, dirty->start);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
350
res_end = max(res_end, dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
354
start = find_next_bit(&dirty->bitmap[0], res_end, cur);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
358
end = find_next_zero_bit(&dirty->bitmap[0], res_end, start + 1);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
361
bitmap_clear(&dirty->bitmap[0], start, num);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
364
if (res_start <= dirty->start && res_end > dirty->start)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
365
dirty->start = res_end;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
366
if (res_start < dirty->end && res_end >= dirty->end)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
367
dirty->end = res_start;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
383
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
388
if (res_start >= dirty->end || res_end <= dirty->start)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
391
res_start = max(res_start, dirty->start);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
392
res_end = min(res_end, dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
393
bitmap_clear(&dirty->bitmap[0], res_start, res_end - res_start);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
395
if (res_start <= dirty->start && res_end > dirty->start)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
396
dirty->start = res_end;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
397
if (res_start < dirty->end && res_end >= dirty->end)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
398
dirty->end = res_start;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
428
if (vbo->dirty && vbo->dirty->method == VMW_BO_DIRTY_MKWRITE &&
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
429
!test_bit(page_offset, &vbo->dirty->bitmap[0])) {
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
430
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
432
__set_bit(page_offset, &dirty->bitmap[0]);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
433
dirty->start = min(dirty->start, page_offset);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
434
dirty->end = max(dirty->end, page_offset + 1);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
459
if (vbo->dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
481
if (vbo->dirty && vbo->dirty->method == VMW_BO_DIRTY_MKWRITE)
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
57
return vbo->dirty && (vbo->dirty->start < vbo->dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
70
struct vmw_bo_dirty *dirty = vbo->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
77
offset, dirty->bitmap_size,
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
78
offset, &dirty->bitmap[0],
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
79
&dirty->start, &dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
81
dirty->change_count++;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
83
dirty->change_count = 0;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
85
if (dirty->change_count > VMW_DIRTY_NUM_CHANGE_TRIGGERS) {
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
86
dirty->change_count = 0;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
87
dirty->method = VMW_BO_DIRTY_MKWRITE;
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
89
offset, dirty->bitmap_size);
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
91
offset, dirty->bitmap_size,
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
92
offset, &dirty->bitmap[0],
drivers/gpu/drm/vmwgfx/vmwgfx_page_dirty.c
93
&dirty->start, &dirty->end);
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
1071
if (res->dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
141
if (res->dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
235
res->dirty = NULL;
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
402
if (res->guest_memory_bo->dirty && !res->dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
406
} else if (!res->guest_memory_bo->dirty && res->dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
415
if (res->dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
454
bool dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
479
WARN_ON(res->coherent && !new_guest_memory_bo->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_resource.c
493
res->res_dirty = dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1001
if (!dirty->num_hits) {
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1002
vmw_cmd_commit(dirty->dev_priv, 0);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1024
cmd->body.destScreenId = dirty->unit->unit;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1027
for (i = 0; i < dirty->num_hits; ++i, ++blit) {
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1034
vmw_cmd_commit(dirty->dev_priv, region_size + sizeof(*cmd));
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1048
static void vmw_sou_surface_clip(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1051
container_of(dirty, typeof(*sdirty), base);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1052
struct vmw_kms_sou_dirty_cmd *cmd = dirty->cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1056
blit += dirty->num_hits;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1057
blit->left = dirty->unit_x1;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1058
blit->top = dirty->unit_y1;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1059
blit->right = dirty->unit_x2;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1060
blit->bottom = dirty->unit_y2;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1063
sdirty->left = min_t(s32, sdirty->left, dirty->unit_x1);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1064
sdirty->top = min_t(s32, sdirty->top, dirty->unit_y1);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1065
sdirty->right = max_t(s32, sdirty->right, dirty->unit_x2);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1066
sdirty->bottom = max_t(s32, sdirty->bottom, dirty->unit_y2);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1068
dirty->num_hits++;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1155
static void vmw_sou_bo_fifo_commit(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1157
if (!dirty->num_hits) {
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1158
vmw_cmd_commit(dirty->dev_priv, 0);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1162
vmw_cmd_commit(dirty->dev_priv,
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1164
dirty->num_hits);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1174
static void vmw_sou_bo_clip(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1176
struct vmw_kms_sou_bo_blit *blit = dirty->cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1178
blit += dirty->num_hits;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1180
blit->body.destScreenId = dirty->unit->unit;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1181
blit->body.srcOrigin.x = dirty->fb_x;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1182
blit->body.srcOrigin.y = dirty->fb_y;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1183
blit->body.destRect.left = dirty->unit_x1;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1184
blit->body.destRect.top = dirty->unit_y1;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1185
blit->body.destRect.right = dirty->unit_x2;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1186
blit->body.destRect.bottom = dirty->unit_y2;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1187
dirty->num_hits++;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1221
struct vmw_kms_dirty dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1239
dirty.crtc = crtc;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1240
dirty.fifo_commit = vmw_sou_bo_fifo_commit;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1241
dirty.clip = vmw_sou_bo_clip;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1242
dirty.fifo_reserve_size = sizeof(struct vmw_kms_sou_bo_blit) *
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1245
0, 0, num_clips, increment, &dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1267
static void vmw_sou_readback_fifo_commit(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1269
if (!dirty->num_hits) {
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1270
vmw_cmd_commit(dirty->dev_priv, 0);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1274
vmw_cmd_commit(dirty->dev_priv,
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1276
dirty->num_hits);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1286
static void vmw_sou_readback_clip(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1288
struct vmw_kms_sou_readback_blit *blit = dirty->cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1290
blit += dirty->num_hits;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1292
blit->body.srcScreenId = dirty->unit->unit;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1293
blit->body.destOrigin.x = dirty->fb_x;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1294
blit->body.destOrigin.y = dirty->fb_y;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1295
blit->body.srcRect.left = dirty->unit_x1;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1296
blit->body.srcRect.top = dirty->unit_y1;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1297
blit->body.srcRect.right = dirty->unit_x2;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1298
blit->body.srcRect.bottom = dirty->unit_y2;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1299
dirty->num_hits++;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1329
struct vmw_kms_dirty dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1347
dirty.crtc = crtc;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1348
dirty.fifo_commit = vmw_sou_readback_fifo_commit;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1349
dirty.clip = vmw_sou_readback_clip;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1350
dirty.fifo_reserve_size = sizeof(struct vmw_kms_sou_readback_blit) *
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1353
0, 0, num_clips, 1, &dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
990
static void vmw_sou_surface_fifo_commit(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
993
container_of(dirty, typeof(*sdirty), base);
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
994
struct vmw_kms_sou_dirty_cmd *cmd = dirty->cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
995
s32 trans_x = dirty->unit->crtc.x - sdirty->dst_x;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
996
s32 trans_y = dirty->unit->crtc.y - sdirty->dst_y;
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
997
size_t region_size = dirty->num_hits * sizeof(SVGASignedRect);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
466
static void vmw_stdu_bo_cpu_clip(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
469
container_of(dirty, struct vmw_stdu_dirty, base);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
471
dirty->num_hits = 1;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
474
ddirty->left = min_t(s32, ddirty->left, dirty->unit_x1);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
475
ddirty->top = min_t(s32, ddirty->top, dirty->unit_y1);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
476
ddirty->right = max_t(s32, ddirty->right, dirty->unit_x2);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
477
ddirty->bottom = max_t(s32, ddirty->bottom, dirty->unit_y2);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
484
ddirty->fb_left = min_t(s32, ddirty->fb_left, dirty->fb_x);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
485
ddirty->fb_top = min_t(s32, ddirty->fb_top, dirty->fb_y);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
497
static void vmw_stdu_bo_cpu_commit(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
500
container_of(dirty, struct vmw_stdu_dirty, base);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
502
container_of(dirty->unit, typeof(*stdu), base);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
509
if (!dirty->num_hits)
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
621
static void vmw_kms_stdu_surface_clip(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
624
container_of(dirty, struct vmw_stdu_dirty, base);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
625
struct vmw_stdu_surface_copy *cmd = dirty->cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
627
container_of(dirty->unit, typeof(*stdu), base);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
632
blit += dirty->num_hits;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
633
blit->srcx = dirty->fb_x;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
634
blit->srcy = dirty->fb_y;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
635
blit->x = dirty->unit_x1;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
636
blit->y = dirty->unit_y1;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
638
blit->w = dirty->unit_x2 - dirty->unit_x1;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
639
blit->h = dirty->unit_y2 - dirty->unit_y1;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
642
dirty->num_hits++;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
645
sdirty->left = min_t(s32, sdirty->left, dirty->unit_x1);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
646
sdirty->top = min_t(s32, sdirty->top, dirty->unit_y1);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
647
sdirty->right = max_t(s32, sdirty->right, dirty->unit_x2);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
648
sdirty->bottom = max_t(s32, sdirty->bottom, dirty->unit_y2);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
660
static void vmw_kms_stdu_surface_fifo_commit(struct vmw_kms_dirty *dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
663
container_of(dirty, struct vmw_stdu_dirty, base);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
665
container_of(dirty->unit, typeof(*stdu), base);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
666
struct vmw_stdu_surface_copy *cmd = dirty->cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
668
size_t blit_size = sizeof(SVGA3dCopyBox) * dirty->num_hits;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
671
if (!dirty->num_hits) {
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
672
vmw_cmd_commit(dirty->dev_priv, 0);
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
683
update = (struct vmw_stdu_update *) &blit[dirty->num_hits];
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
687
update = dirty->cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
694
vmw_cmd_commit(dirty->dev_priv, commit_size);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1291
if (res->guest_memory_bo->dirty && res->guest_memory_dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1766
static void vmw_subres_dirty_add(struct vmw_surface_dirty *dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1770
const struct vmw_surface_cache *cache = &dirty->cache;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1771
SVGA3dBox *box = &dirty->boxes[loc_start->sub_resource];
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1776
if (WARN_ON(loc_start->sub_resource >= dirty->num_subres))
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1814
static void vmw_subres_dirty_full(struct vmw_surface_dirty *dirty, u32 subres)
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1816
const struct vmw_surface_cache *cache = &dirty->cache;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1819
SVGA3dBox *box = &dirty->boxes[subres];
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1836
struct vmw_surface_dirty *dirty =
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1837
(struct vmw_surface_dirty *) res->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1844
cache = &dirty->cache;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1858
for (sub_res = 0; sub_res < dirty->num_subres; ++sub_res)
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1859
vmw_subres_dirty_full(dirty, sub_res);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1864
vmw_subres_dirty_add(dirty, &loc1, &loc2);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1871
vmw_subres_dirty_add(dirty, &loc1, &loc_max);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1873
vmw_subres_dirty_add(dirty, &loc_min, &loc2);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1876
vmw_subres_dirty_full(dirty, sub_res);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1887
struct vmw_surface_dirty *dirty =
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1888
(struct vmw_surface_dirty *) res->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1889
const struct vmw_surface_cache *cache = &dirty->cache;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1891
SVGA3dBox *box = &dirty->boxes[0];
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1929
struct vmw_surface_dirty *dirty =
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1930
(struct vmw_surface_dirty *) res->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1932
const struct vmw_surface_cache *cache = &dirty->cache;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1944
for (i = 0; i < dirty->num_subres; ++i) {
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1945
const SVGA3dBox *box = &dirty->boxes[i];
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1962
for (i = 0; i < dirty->num_subres; ++i) {
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1963
const SVGA3dBox *box = &dirty->boxes[i];
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1993
memset(&dirty->boxes[0], 0, sizeof(dirty->boxes[0]) *
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
1994
dirty->num_subres);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2006
struct vmw_surface_dirty *dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2024
dirty_size = struct_size(dirty, boxes, num_subres);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2026
dirty = kvzalloc(dirty_size, GFP_KERNEL);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2027
if (!dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2035
&dirty->cache);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2039
dirty->num_subres = num_subres;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2040
res->dirty = (struct vmw_resource_dirty *) dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2045
kvfree(dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2055
struct vmw_surface_dirty *dirty =
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2056
(struct vmw_surface_dirty *) res->dirty;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2058
kvfree(dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
2059
res->dirty = NULL;
drivers/gpu/drm/vmwgfx/vmwgfx_surface.c
643
WARN_ON(res->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
288
u32 dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
335
if (dirty) {
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
338
node->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
358
void *val_private, u32 dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
362
if (!dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
368
val->dirty = (dirty & VMW_RES_DIRTY_SET) ? 1 : 0;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
481
val->dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
565
if (vbo->dirty)
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
593
val->dirty);
drivers/gpu/drm/vmwgfx/vmwgfx_validation.c
82
u32 dirty : 1;
drivers/gpu/drm/vmwgfx/vmwgfx_validation.h
161
u32 dirty,
drivers/gpu/drm/vmwgfx/vmwgfx_validation.h
187
void *val_private, u32 dirty);
drivers/infiniband/core/umem.c
50
static void __ib_umem_release(struct ib_device *dev, struct ib_umem *umem, int dirty)
drivers/infiniband/core/umem.c
52
bool make_dirty = umem->writable && dirty;
drivers/infiniband/core/umem.c
56
if (dirty)
drivers/infiniband/hw/hfi1/hfi.h
2101
size_t npages, bool dirty);
drivers/infiniband/hw/hfi1/user_pages.c
96
size_t npages, bool dirty)
drivers/infiniband/hw/hfi1/user_pages.c
98
unpin_user_pages_dirty_lock(p, npages, dirty);
drivers/infiniband/hw/mlx4/mlx4_ib.h
293
bool dirty;
drivers/infiniband/hw/mlx4/qp.c
793
if (!range || (range->refcount == range->size) || range->dirty) {
drivers/infiniband/hw/mlx4/qp.c
853
range->dirty = true;
drivers/infiniband/hw/usnic/usnic_uiom.c
210
int dirty)
drivers/infiniband/hw/usnic/usnic_uiom.c
234
usnic_uiom_put_pages(&uiomr->chunk_list, dirty & writable);
drivers/infiniband/hw/usnic/usnic_uiom.c
65
static void usnic_uiom_put_pages(struct list_head *chunk_list, int dirty)
drivers/infiniband/hw/usnic/usnic_uiom.c
77
unpin_user_pages_dirty_lock(&page, 1, dirty);
drivers/input/joystick/grip_mp.c
415
port->dirty = 1;
drivers/input/joystick/grip_mp.c
426
port->dirty = 0;
drivers/input/joystick/grip_mp.c
48
int dirty; /* has the state been updated? */
drivers/input/joystick/grip_mp.c
529
port->dirty = 0;
drivers/input/joystick/grip_mp.c
553
if (grip->port[i]->dirty)
drivers/input/joystick/grip_mp.c
624
if (port->dirty) /* report initial state, if any */
drivers/interconnect/qcom/icc-rpmh.h
136
bool dirty;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3661
struct iommu_dirty_bitmap *dirty)
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3666
return ops->read_and_clear_dirty(ops, iova, size, flags, dirty);
drivers/iommu/generic_pt/iommu_pt.h
192
struct iommu_dirty_bitmap *dirty;
drivers/iommu/generic_pt/iommu_pt.h
197
struct pt_iommu_dirty_args *dirty,
drivers/iommu/generic_pt/iommu_pt.h
215
if (dirty->dirty->bitmap)
drivers/iommu/generic_pt/iommu_pt.h
216
iova_bitmap_set(dirty->dirty->bitmap, pts->range->va,
drivers/iommu/generic_pt/iommu_pt.h
219
if (!(dirty->flags & IOMMU_DIRTY_NO_CLEAR)) {
drivers/iommu/generic_pt/iommu_pt.h
225
iommu_iotlb_gather_add_range(dirty->dirty->gather,
drivers/iommu/generic_pt/iommu_pt.h
235
struct pt_iommu_dirty_args *dirty = arg;
drivers/iommu/generic_pt/iommu_pt.h
246
record_dirty(&pts, dirty,
drivers/iommu/generic_pt/iommu_pt.h
272
struct iommu_dirty_bitmap *dirty)
drivers/iommu/generic_pt/iommu_pt.h
277
.dirty = dirty,
drivers/iommu/io-pgtable-arm.c
831
struct iommu_dirty_bitmap *dirty = walk_data->data;
drivers/iommu/io-pgtable-arm.c
837
iommu_dirty_bitmap_record(dirty, walk_data->addr, size);
drivers/iommu/io-pgtable-arm.c
848
struct iommu_dirty_bitmap *dirty)
drivers/iommu/io-pgtable-arm.c
854
.data = dirty,
drivers/iommu/iommufd/io_pagetable.c
533
struct iommu_dirty_bitmap *dirty;
drivers/iommu/iommufd/io_pagetable.c
544
struct iommu_dirty_bitmap *dirty = arg->dirty;
drivers/iommu/iommufd/io_pagetable.c
555
dirty);
drivers/iommu/iommufd/io_pagetable.c
572
struct iommu_dirty_bitmap dirty;
drivers/iommu/iommufd/io_pagetable.c
586
iommu_dirty_bitmap_init(&dirty, iter, &gather);
drivers/iommu/iommufd/io_pagetable.c
591
arg.dirty = &dirty;
drivers/iommu/iommufd/io_pagetable.c
651
struct iommu_dirty_bitmap dirty;
drivers/iommu/iommufd/io_pagetable.c
657
iommu_dirty_bitmap_init(&dirty, NULL, &gather);
drivers/iommu/iommufd/io_pagetable.c
666
&dirty);
drivers/iommu/iommufd/iommufd_test.h
151
} dirty;
drivers/iommu/iommufd/selftest.c
1759
cmd->dirty.out_nr_dirty = count;
drivers/iommu/iommufd/selftest.c
2153
return iommufd_test_dirty(ucmd, cmd->id, cmd->dirty.iova,
drivers/iommu/iommufd/selftest.c
2154
cmd->dirty.length,
drivers/iommu/iommufd/selftest.c
2155
cmd->dirty.page_size,
drivers/iommu/iommufd/selftest.c
2156
u64_to_user_ptr(cmd->dirty.uptr),
drivers/iommu/iommufd/selftest.c
2157
cmd->dirty.flags);
drivers/md/bcache/btree.h
166
BTREE_FLAG(dirty);
drivers/md/bcache/journal.c
633
j->cur->dirty = false;
drivers/md/bcache/journal.c
835
if (c->journal.cur->dirty)
drivers/md/bcache/journal.c
872
} else if (!w->dirty) {
drivers/md/bcache/journal.c
873
w->dirty = true;
drivers/md/bcache/journal.h
99
bool dirty;
drivers/md/bcache/sysfs.c
1056
size_t unused = 0, available = 0, dirty = 0, meta = 0;
drivers/md/bcache/sysfs.c
1074
dirty++;
drivers/md/bcache/sysfs.c
1117
dirty * 100 / (size_t) ca->sb.nbuckets,
drivers/md/bcache/sysfs.c
218
char dirty[20];
drivers/md/bcache/sysfs.c
232
bch_hprint(dirty, bcache_dev_sectors_dirty(&dc->disk) << 9);
drivers/md/bcache/sysfs.c
250
rate, dirty, target, proportional,
drivers/md/bcache/writeback.c
106
c->gc_stats.in_use > BCH_WRITEBACK_FRAGMENT_THRESHOLD_LOW && dirty > 0) {
drivers/md/bcache/writeback.c
108
div_s64((dirty_buckets * c->cache->sb.bucket_size), dirty);
drivers/md/bcache/writeback.c
122
fps = div_s64(dirty, dirty_buckets) * fp_term;
drivers/md/bcache/writeback.c
84
int64_t dirty = bcache_dev_sectors_dirty(&dc->disk);
drivers/md/bcache/writeback.c
85
int64_t error = dirty - target;
drivers/md/dm-cache-metadata.c
1332
bool dirty = true;
drivers/md/dm-cache-metadata.c
1344
dirty = flags & M_DIRTY;
drivers/md/dm-cache-metadata.c
1346
r = fn(context, oblock, to_cblock(cb), dirty,
drivers/md/dm-cache-metadata.c
1374
bool dirty = true;
drivers/md/dm-cache-metadata.c
1386
dirty = dm_bitset_cursor_get_value(dirty_cursor);
drivers/md/dm-cache-metadata.c
1388
r = fn(context, oblock, to_cblock(cb), dirty,
drivers/md/dm-cache-metadata.c
1512
static int __dirty(struct dm_cache_metadata *cmd, dm_cblock_t cblock, bool dirty)
drivers/md/dm-cache-metadata.c
1525
if (((flags & M_DIRTY) && dirty) || (!(flags & M_DIRTY) && !dirty))
drivers/md/dm-cache-metadata.c
1529
value = pack_value(oblock, (flags & ~M_DIRTY) | (dirty ? M_DIRTY : 0));
drivers/md/dm-cache-metadata.h
91
dm_cblock_t cblock, bool dirty,
drivers/md/dm-cache-policy-internal.h
60
bool dirty, uint32_t hint, bool hint_valid)
drivers/md/dm-cache-policy-internal.h
62
return p->load_mapping(p, oblock, cblock, dirty, hint, hint_valid);
drivers/md/dm-cache-policy-smq.c
1009
if (!e->dirty) {
drivers/md/dm-cache-policy-smq.c
1014
q_requeue(&mq->dirty, e, 1u,
drivers/md/dm-cache-policy-smq.c
1120
q_redistribute(&mq->dirty);
drivers/md/dm-cache-policy-smq.c
1151
return q_size(&mq->dirty) == 0u;
drivers/md/dm-cache-policy-smq.c
1191
e = q_peek(&mq->dirty, mq->dirty.nr_levels, idle);
drivers/md/dm-cache-policy-smq.c
1194
q_del(&mq->dirty, e);
drivers/md/dm-cache-policy-smq.c
1203
q_push_front(&mq->dirty, e);
drivers/md/dm-cache-policy-smq.c
1533
e->dirty = set;
drivers/md/dm-cache-policy-smq.c
1536
e->dirty = set;
drivers/md/dm-cache-policy-smq.c
1568
bool dirty, uint32_t hint, bool hint_valid)
drivers/md/dm-cache-policy-smq.c
1575
e->dirty = dirty;
drivers/md/dm-cache-policy-smq.c
1798
q_init(&mq->dirty, &mq->es, NR_CACHE_LEVELS);
drivers/md/dm-cache-policy-smq.c
44
bool dirty:1;
drivers/md/dm-cache-policy-smq.c
725
e->dirty = true; /* FIXME: audit */
drivers/md/dm-cache-policy-smq.c
828
struct queue dirty;
drivers/md/dm-cache-policy-smq.c
889
struct queue *q = &mq->dirty;
drivers/md/dm-cache-policy-smq.c
935
q_push(&mq->dirty, sentinel);
drivers/md/dm-cache-policy-smq.c
961
q_del(e->dirty ? &mq->dirty : &mq->clean, e);
drivers/md/dm-cache-policy-smq.c
966
if (e->dirty)
drivers/md/dm-cache-policy-smq.c
967
q_push(&mq->dirty, e);
drivers/md/dm-cache-policy-smq.c
982
if (e->dirty)
drivers/md/dm-cache-policy-smq.c
983
q_push_front(&mq->dirty, e);
drivers/md/dm-cache-policy.h
97
dm_cblock_t cblock, bool dirty,
drivers/md/dm-cache-target.c
2808
bool dirty, uint32_t hint, bool hint_valid)
drivers/md/dm-cache-target.c
2812
if (dirty) {
drivers/md/dm-cache-target.c
2818
return policy_load_mapping(cache->policy, oblock, cblock, dirty, hint, hint_valid);
drivers/md/dm-cache-target.c
2822
bool dirty, uint32_t hint, bool hint_valid)
drivers/md/dm-cache-target.c
2827
if (dirty) {
drivers/md/dm-cache-target.c
2836
return load_mapping(context, oblock, cblock, dirty, hint, hint_valid);
drivers/md/dm-vdo/block-map.c
1574
bool dirty;
drivers/md/dm-vdo/block-map.c
1585
dirty = (page->writing_generation != page->generation);
drivers/md/dm-vdo/block-map.c
1597
if (dirty && attempt_increment(zone)) {
drivers/md/dm-vdo/block-map.c
1605
if (dirty) {
drivers/md/dm-vdo/block-map.c
2005
static void update_period(struct dirty_lists *dirty, sequence_number_t period)
drivers/md/dm-vdo/block-map.c
2007
while (dirty->next_period <= period) {
drivers/md/dm-vdo/block-map.c
2008
if ((dirty->next_period - dirty->oldest_period) == dirty->maximum_age)
drivers/md/dm-vdo/block-map.c
2009
expire_oldest_list(dirty);
drivers/md/dm-vdo/block-map.c
2010
dirty->next_period++;
drivers/md/md-bitmap.c
1236
int dirty, need_write;
drivers/md/md-bitmap.c
1245
dirty = test_and_clear_page_attr(bitmap, i, BITMAP_PAGE_DIRTY);
drivers/md/md-bitmap.c
1248
if (dirty || need_write) {
drivers/md/md-llbitmap.c
1202
bitmap_fill(pctl->dirty, llbitmap->blocks_per_page);
drivers/md/md-llbitmap.c
1349
bitmap_fill(llbitmap->pctl[0]->dirty, nr_blocks);
drivers/md/md-llbitmap.c
1415
bitmap_fill(pctl->dirty, llbitmap->blocks_per_page);
drivers/md/md-llbitmap.c
253
unsigned long dirty[];
drivers/md/md-llbitmap.c
402
set_bit(block, pctl->dirty);
drivers/md/md-llbitmap.c
413
if (test_and_set_bit(block, pctl->dirty)) {
drivers/md/md-llbitmap.c
489
if (!test_and_clear_bit(block, pctl->dirty))
drivers/md/md-llbitmap.c
544
unsigned int size = struct_size(pctl, dirty, BITS_TO_LONGS(
drivers/md/persistent-data/dm-bitset.c
112
if (!info->current_index_set || !info->dirty)
drivers/md/persistent-data/dm-bitset.c
124
info->dirty = false;
drivers/md/persistent-data/dm-bitset.c
143
info->dirty = false;
drivers/md/persistent-data/dm-bitset.c
177
info->dirty = true;
drivers/md/persistent-data/dm-bitset.c
194
info->dirty = true;
drivers/md/persistent-data/dm-bitset.h
75
bool dirty:1;
drivers/md/persistent-data/dm-space-map-common.c
1097
iec->dirty = false;
drivers/md/persistent-data/dm-space-map-common.c
1121
if (iec->dirty) {
drivers/md/persistent-data/dm-space-map-common.c
1131
iec->dirty = false;
drivers/md/persistent-data/dm-space-map-common.c
1150
iec->dirty = true;
drivers/md/persistent-data/dm-space-map-common.c
1154
if (iec->dirty) {
drivers/md/persistent-data/dm-space-map-common.c
1162
iec->dirty = true;
drivers/md/persistent-data/dm-space-map-common.c
1176
iec->dirty = false;
drivers/md/persistent-data/dm-space-map-common.c
1199
if (iec->valid && iec->dirty)
drivers/md/persistent-data/dm-space-map-common.h
67
bool dirty;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1027
instance->dirty[uh->frame_context_idx],
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1030
if (instance->dirty[uh->frame_context_idx])
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1554
if (!instance->dirty[uh->frame_context_idx])
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1583
instance->dirty[uh->frame_context_idx] = 1;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
481
unsigned char dirty[4];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
825
instance->dirty[i] = 0;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
827
instance->dirty[uh->frame_context_idx] = 0;
drivers/media/usb/uvc/uvc_ctrl.c
2259
if (!ctrl->dirty)
drivers/media/usb/uvc/uvc_ctrl.c
2276
ctrl->dirty = 0;
drivers/media/usb/uvc/uvc_ctrl.c
2651
if (!ctrl->dirty) {
drivers/media/usb/uvc/uvc_ctrl.c
2661
ctrl->dirty = 1;
drivers/media/usb/uvc/uvc_ctrl.c
2985
ctrl->dirty = 1;
drivers/media/usb/uvc/uvcvideo.h
157
u8 dirty:1,
drivers/misc/sgi-gru/gruhandles.c
152
unsigned long vaddr, int asid, int dirty,
drivers/misc/sgi-gru/gruhandles.c
159
tfh->dirty = dirty;
drivers/misc/sgi-gru/gruhandles.c
168
unsigned long vaddr, int asid, int dirty,
drivers/misc/sgi-gru/gruhandles.c
175
tfh->dirty = dirty;
drivers/misc/sgi-gru/gruhandles.h
285
unsigned int dirty:1;
drivers/misc/sgi-gru/gruhandles.h
511
int gaa, unsigned long vaddr, int asid, int dirty, int pagesize);
drivers/misc/sgi-gru/gruhandles.h
513
int gaa, unsigned long vaddr, int asid, int dirty, int pagesize);
drivers/misc/vmw_vmci/vmci_queue_pair.c
632
u64 num_pages, bool dirty)
drivers/misc/vmw_vmci/vmci_queue_pair.c
637
if (dirty)
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
314
ring->dirty = 0;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
358
ring->dirty = 0;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
156
return (ring->rdesc_count - (ring->cur - ring->dirty));
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
161
return (ring->cur - ring->dirty);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2058
while (ring->dirty != ring->cur) {
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2059
rdata = XGBE_GET_DESC_DATA(ring, ring->dirty);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2067
hw_if->rx_desc_reset(pdata, rdata, ring->dirty);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2069
ring->dirty++;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2077
rdata = XGBE_GET_DESC_DATA(ring, ring->dirty - 1);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2174
(ring->dirty != cur)) {
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2175
rdata = XGBE_GET_DESC_DATA(ring, ring->dirty);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2186
xgbe_dump_tx_desc(pdata, ring, ring->dirty, 1, 0);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2198
ring->dirty++;
drivers/net/ethernet/amd/xgbe/xgbe.h
399
unsigned int dirty;
drivers/net/ethernet/atheros/ag71xx.c
1125
while (ring->curr != ring->dirty) {
drivers/net/ethernet/atheros/ag71xx.c
1127
u32 i = ring->dirty & ring_mask;
drivers/net/ethernet/atheros/ag71xx.c
1141
ring->dirty++;
drivers/net/ethernet/atheros/ag71xx.c
1171
ring->dirty = 0;
drivers/net/ethernet/atheros/ag71xx.c
1261
ring->dirty = 0;
drivers/net/ethernet/atheros/ag71xx.c
1274
for (; ring->curr - ring->dirty > 0; ring->dirty++) {
drivers/net/ethernet/atheros/ag71xx.c
1278
i = ring->dirty & ring_mask;
drivers/net/ethernet/atheros/ag71xx.c
1551
if (ring->curr - ring->dirty >= ring_size - ring_min) {
drivers/net/ethernet/atheros/ag71xx.c
1620
limit, ring->curr, ring->dirty);
drivers/net/ethernet/atheros/ag71xx.c
1630
if ((ring->dirty + ring_size) == ring->curr) {
drivers/net/ethernet/atheros/ag71xx.c
1673
ring->curr, ring->dirty, done);
drivers/net/ethernet/atheros/ag71xx.c
1692
if (!rx_ring->buf[rx_ring->dirty % rx_ring_size].rx.rx_buf)
drivers/net/ethernet/atheros/ag71xx.c
320
unsigned int dirty;
drivers/net/ethernet/atheros/ag71xx.c
782
while (ring->dirty + n != ring->curr) {
drivers/net/ethernet/atheros/ag71xx.c
787
i = (ring->dirty + n) & ring_mask;
drivers/net/ethernet/atheros/ag71xx.c
814
ring->dirty += n;
drivers/net/ethernet/atheros/ag71xx.c
831
if ((ring->curr - ring->dirty) < (ring_size * 3) / 4)
drivers/net/ethernet/atheros/ag71xx.c
963
ag->tx_ring.dirty = 0;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
457
bool dirty;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
461
dirty = memcmp(&mlxsw_sp_port->hdroom->bufs, &hdroom->bufs, sizeof(hdroom->bufs));
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
462
if (!dirty && !force)
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
488
bool dirty;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
492
dirty = memcmp(&mlxsw_sp_port->hdroom->prios, &hdroom->prios, sizeof(hdroom->prios));
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
493
if (!dirty && !force)
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
512
bool dirty;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
515
dirty = memcmp(&mlxsw_sp_port->hdroom->int_buf, &hdroom->int_buf, sizeof(hdroom->int_buf));
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
516
if (!dirty && !force)
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
130
group->dirty = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
375
entry->dirty = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
384
if (slaves != entry->slave_cnt || !entry->dirty) {
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
393
entry->dirty = false;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
487
group_entry->dirty = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
609
group->dirty = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
65
bool dirty;
drivers/net/ethernet/renesas/rswitch.h
973
unsigned int dirty;
drivers/net/ethernet/renesas/rswitch_main.c
1080
rswitch_gwca_ts_queue_fill(priv, gq->dirty, num);
drivers/net/ethernet/renesas/rswitch_main.c
1081
gq->dirty = rswitch_next_queue_index(gq, false, num);
drivers/net/ethernet/renesas/rswitch_main.c
245
unsigned int index = cur ? gq->cur : gq->dirty;
drivers/net/ethernet/renesas/rswitch_main.c
257
if (gq->cur >= gq->dirty)
drivers/net/ethernet/renesas/rswitch_main.c
258
return gq->cur - gq->dirty;
drivers/net/ethernet/renesas/rswitch_main.c
260
return gq->ring_size - gq->dirty + gq->cur;
drivers/net/ethernet/renesas/rswitch_main.c
265
struct rswitch_ext_ts_desc *desc = &gq->rx_ring[gq->dirty];
drivers/net/ethernet/renesas/rswitch_main.c
874
ret = rswitch_gwca_queue_alloc_rx_buf(gq, gq->dirty, num);
drivers/net/ethernet/renesas/rswitch_main.c
877
ret = rswitch_gwca_queue_ext_ts_fill(ndev, gq, gq->dirty, num);
drivers/net/ethernet/renesas/rswitch_main.c
880
gq->dirty = rswitch_next_queue_index(gq, false, num);
drivers/net/ethernet/renesas/rswitch_main.c
899
desc = &gq->tx_ring[gq->dirty];
drivers/net/ethernet/renesas/rswitch_main.c
903
skb = gq->skbs[gq->dirty];
drivers/net/ethernet/renesas/rswitch_main.c
908
gq->unmap_addrs[gq->dirty],
drivers/net/ethernet/renesas/rswitch_main.c
910
dev_kfree_skb_any(gq->skbs[gq->dirty]);
drivers/net/ethernet/renesas/rswitch_main.c
911
gq->skbs[gq->dirty] = NULL;
drivers/net/ethernet/renesas/rswitch_main.c
915
gq->dirty = rswitch_next_queue_index(gq, false, 1);
drivers/net/ethernet/renesas/rswitch_main.c
916
desc = &gq->tx_ring[gq->dirty];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
377
u32 dirty;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
380
dirty = rx_q->cur_rx - rx_q->dirty_rx;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
382
dirty = priv->dma_conf.dma_rx_size - rx_q->dirty_rx + rx_q->cur_rx;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
384
return dirty;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4940
int dirty = stmmac_rx_dirty(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4947
while (dirty-- > 0) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5405
int dirty = stmmac_rx_dirty(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5454
if (dirty >= STMMAC_RX_FILL_BATCH) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5456
!stmmac_rx_refill_zc(priv, queue, dirty);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5457
dirty = 0;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5493
dirty++;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5510
dirty++;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5555
dirty++;
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
447
ring->dirty = 0;
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
487
ring->dirty = 0;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1063
(ring->dirty != cur)) {
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1064
desc_data = XLGMAC_GET_DESC_DATA(ring, ring->dirty);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1076
xlgmac_dump_tx_desc(pdata, ring, ring->dirty, 1, 0);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1088
ring->dirty++;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
30
return (ring->dma_desc_count - (ring->cur - ring->dirty));
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
35
return (ring->cur - ring->dirty);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
962
while (ring->dirty != ring->cur) {
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
963
desc_data = XLGMAC_GET_DESC_DATA(ring, ring->dirty);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
971
hw_ops->rx_desc_reset(pdata, desc_data, ring->dirty);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
973
ring->dirty++;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
982
desc_data = XLGMAC_GET_DESC_DATA(ring, ring->dirty - 1);
drivers/net/ethernet/synopsys/dwc-xlgmac.h
341
unsigned int dirty;
drivers/net/ethernet/via/via-velocity.c
1449
int avail, dirty, unusable;
drivers/net/ethernet/via/via-velocity.c
1461
dirty = vptr->rx.dirty - unusable;
drivers/net/ethernet/via/via-velocity.c
1463
dirty = (dirty > 0) ? dirty - 1 : vptr->options.numrx - 1;
drivers/net/ethernet/via/via-velocity.c
1464
vptr->rx.ring[dirty].rdesc0.len |= OWNED_BY_NIC;
drivers/net/ethernet/via/via-velocity.c
1564
int dirty = vptr->rx.dirty, done = 0;
drivers/net/ethernet/via/via-velocity.c
1567
struct rx_desc *rd = vptr->rx.ring + dirty;
drivers/net/ethernet/via/via-velocity.c
1573
if (!vptr->rx.info[dirty].skb) {
drivers/net/ethernet/via/via-velocity.c
1574
if (velocity_alloc_rx_buf(vptr, dirty) < 0)
drivers/net/ethernet/via/via-velocity.c
1578
dirty = (dirty < vptr->options.numrx - 1) ? dirty + 1 : 0;
drivers/net/ethernet/via/via-velocity.c
1579
} while (dirty != vptr->rx.curr);
drivers/net/ethernet/via/via-velocity.c
1582
vptr->rx.dirty = dirty;
drivers/net/ethernet/via/via-velocity.c
555
vptr->rx.dirty = vptr->rx.filled = vptr->rx.curr = 0;
drivers/net/ethernet/via/via-velocity.h
1411
int dirty;
drivers/perf/xgene_pmu.c
310
XGENE_PMU_EVENT_ATTR(csw-inbound-dirty, 0x16),
drivers/perf/xgene_pmu.c
385
XGENE_PMU_EVENT_ATTR(dirty-eviction, 0x07),
drivers/power/supply/bq27xxx_battery.c
1082
bool has_data, dirty;
drivers/power/supply/bq27xxx_battery.c
1303
buf->dirty = false;
drivers/power/supply/bq27xxx_battery.c
1358
buf->dirty = true;
drivers/power/supply/bq27xxx_battery.c
1413
if (!buf->dirty)
drivers/power/supply/bq27xxx_battery.c
1459
buf->dirty = false;
drivers/power/supply/bq27xxx_battery.c
1502
updated = bd.dirty || bt.dirty;
drivers/s390/char/con3270.c
1281
line->dirty = 1;
drivers/s390/char/con3270.c
1306
tty3270_get_view_line(tp, i)->dirty = 1;
drivers/s390/char/con3270.c
1311
line->dirty = 1;
drivers/s390/char/con3270.c
1461
line->dirty = 1;
drivers/s390/char/con3270.c
294
line->dirty = 1;
drivers/s390/char/con3270.c
480
if (!line->dirty)
drivers/s390/char/con3270.c
485
line->dirty = 0;
drivers/s390/char/con3270.c
489
tp->screen[i].dirty = 0;
drivers/s390/char/con3270.c
502
if (!line->dirty)
drivers/s390/char/con3270.c
507
line->dirty = 0;
drivers/s390/char/con3270.c
631
tty3270_get_view_line(tp, i)->dirty = 1;
drivers/s390/char/con3270.c
64
int dirty;
drivers/scsi/st.c
1333
STp->dirty = 0;
drivers/scsi/st.c
1782
STp->dirty = 1;
drivers/scsi/st.c
1812
STp->dirty = !(STbp->writing ==
drivers/scsi/st.c
1898
STp->dirty = 0;
drivers/scsi/st.c
2925
if (STp->dirty || (STp->buffer)->buffer_bytes != 0)
drivers/scsi/st.c
4394
tpnt->dirty = 0;
drivers/scsi/st.c
790
if (STp->dirty == 1) {
drivers/scsi/st.c
819
STp->dirty = 0;
drivers/scsi/st.c
832
STp->dirty = 0;
drivers/scsi/st.h
158
unsigned char dirty;
drivers/soc/qcom/rpmh-internal.h
83
bool dirty;
drivers/soc/qcom/rpmh.c
148
ctrlr->dirty |= (req->sleep_val != old_sleep_val ||
drivers/soc/qcom/rpmh.c
280
ctrlr->dirty = true;
drivers/soc/qcom/rpmh.c
444
if (!ctrlr->dirty) {
drivers/soc/qcom/rpmh.c
473
ctrlr->dirty = false;
drivers/soc/qcom/rpmh.c
499
ctrlr->dirty = true;
drivers/vfio/pci/mlx5/cmd.c
1614
struct iova_bitmap *dirty)
drivers/vfio/pci/mlx5/cmd.c
1647
iova_bitmap_set(dirty, addr, qp->tracked_page_size);
drivers/vfio/pci/mlx5/cmd.c
1657
struct iova_bitmap *dirty, int *tracker_status)
drivers/vfio/pci/mlx5/cmd.c
1670
set_report_output(size, ix, qp, dirty);
drivers/vfio/pci/mlx5/cmd.c
1698
struct iova_bitmap *dirty, int *tracker_status)
drivers/vfio/pci/mlx5/cmd.c
1716
mlx5vf_rq_cqe(qp, cqe, dirty, tracker_status);
drivers/vfio/pci/mlx5/cmd.c
1725
struct iova_bitmap *dirty)
drivers/vfio/pci/mlx5/cmd.c
1759
poll_err = mlx5vf_cq_poll_one(cq, tracker->host_qp, dirty,
drivers/vfio/pci/mlx5/cmd.c
1765
dirty, &tracker->status);
drivers/vfio/pci/mlx5/cmd.h
246
unsigned long length, struct iova_bitmap *dirty);
drivers/vfio/pci/pds/dirty.c
100
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
128
struct pds_vfio_dirty *dirty = &pds_vfio->dirty;
drivers/vfio/pci/pds/dirty.c
130
if (!dirty->regions)
drivers/vfio/pci/pds/dirty.c
133
for (int i = 0; i < dirty->num_regions; i++) {
drivers/vfio/pci/pds/dirty.c
134
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
172
static void pds_vfio_dirty_free_regions(struct pds_vfio_dirty *dirty)
drivers/vfio/pci/pds/dirty.c
174
vfree(dirty->regions);
drivers/vfio/pci/pds/dirty.c
175
dirty->regions = NULL;
drivers/vfio/pci/pds/dirty.c
176
dirty->num_regions = 0;
drivers/vfio/pci/pds/dirty.c
184
struct pds_vfio_dirty *dirty = &pds_vfio->dirty;
drivers/vfio/pci/pds/dirty.c
188
dirty->regions = vcalloc(num_regions, sizeof(struct pds_vfio_region));
drivers/vfio/pci/pds/dirty.c
189
if (!dirty->regions)
drivers/vfio/pci/pds/dirty.c
191
dirty->num_regions = num_regions;
drivers/vfio/pci/pds/dirty.c
195
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
21
return pds_vfio->dirty.is_enabled;
drivers/vfio/pci/pds/dirty.c
235
pds_vfio_dirty_free_bitmaps(dirty);
drivers/vfio/pci/pds/dirty.c
237
pds_vfio_dirty_free_regions(dirty);
drivers/vfio/pci/pds/dirty.c
26
pds_vfio->dirty.is_enabled = true;
drivers/vfio/pci/pds/dirty.c
31
pds_vfio->dirty.is_enabled = false;
drivers/vfio/pci/pds/dirty.c
361
pds_vfio_dirty_free_bitmaps(&pds_vfio->dirty);
drivers/vfio/pci/pds/dirty.c
362
pds_vfio_dirty_free_regions(&pds_vfio->dirty);
drivers/vfio/pci/pds/dirty.c
509
struct pds_vfio_dirty *dirty = &pds_vfio->dirty;
drivers/vfio/pci/pds/dirty.c
511
for (int i = 0; i < dirty->num_regions; i++) {
drivers/vfio/pci/pds/dirty.c
512
struct pds_vfio_region *region = &dirty->regions[i];
drivers/vfio/pci/pds/dirty.c
603
unsigned long length, struct iova_bitmap *dirty)
drivers/vfio/pci/pds/dirty.c
611
err = pds_vfio_dirty_sync(pds_vfio, dirty, iova, length);
drivers/vfio/pci/pds/dirty.c
94
static void pds_vfio_dirty_free_bitmaps(struct pds_vfio_dirty *dirty)
drivers/vfio/pci/pds/dirty.c
96
if (!dirty->regions)
drivers/vfio/pci/pds/dirty.c
99
for (int i = 0; i < dirty->num_regions; i++) {
drivers/vfio/pci/pds/dirty.h
36
struct iova_bitmap *dirty);
drivers/vfio/pci/pds/vfio_dev.h
18
struct pds_vfio_dirty dirty;
drivers/vfio/vfio_iommu_type1.c
2971
struct vfio_iommu_type1_dirty_bitmap dirty;
drivers/vfio/vfio_iommu_type1.c
2983
if (copy_from_user(&dirty, (void __user *)arg, minsz))
drivers/vfio/vfio_iommu_type1.c
2986
if (dirty.argsz < minsz || dirty.flags & ~mask)
drivers/vfio/vfio_iommu_type1.c
2990
if (__ffs(dirty.flags) != __fls(dirty.flags))
drivers/vfio/vfio_iommu_type1.c
2993
if (dirty.flags & VFIO_IOMMU_DIRTY_PAGES_FLAG_START) {
drivers/vfio/vfio_iommu_type1.c
3005
} else if (dirty.flags & VFIO_IOMMU_DIRTY_PAGES_FLAG_STOP) {
drivers/vfio/vfio_iommu_type1.c
3013
} else if (dirty.flags & VFIO_IOMMU_DIRTY_PAGES_FLAG_GET_BITMAP) {
drivers/vfio/vfio_iommu_type1.c
3016
size_t data_size = dirty.argsz - minsz;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
499
bool shadow_dirty, dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
502
dirty = mp->info_dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
506
if (!dirty && !shadow_dirty) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
576
bool shadow_dirty, dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
579
dirty = op->info_dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
583
if (!dirty && !shadow_dirty) {
fs/btrfs/block-group.c
3516
LIST_HEAD(dirty);
fs/btrfs/block-group.c
3525
list_splice_init(&cur_trans->dirty_bgs, &dirty);
fs/btrfs/block-group.c
3546
while (!list_empty(&dirty)) {
fs/btrfs/block-group.c
3549
cache = list_first_entry(&dirty, struct btrfs_block_group,
fs/btrfs/block-group.c
3651
list_splice_init(&cur_trans->dirty_bgs, &dirty);
fs/btrfs/block-group.c
3656
if (!list_empty(&dirty)) {
fs/btrfs/block-group.c
3665
list_splice_init(&dirty, &cur_trans->dirty_bgs);
fs/btrfs/inode.c
6472
int dirty;
fs/btrfs/inode.c
6479
dirty = inode_update_time(inode, type, flags);
fs/btrfs/inode.c
6480
if (dirty <= 0)
fs/btrfs/inode.c
6481
return dirty;
fs/btrfs/qgroup.c
1431
if (list_empty(&qgroup->dirty))
fs/btrfs/qgroup.c
1432
list_add(&qgroup->dirty, &fs_info->dirty_qgroups);
fs/btrfs/qgroup.c
219
INIT_LIST_HEAD(&prealloc->dirty);
fs/btrfs/qgroup.c
230
list_del(&qgroup->dirty);
fs/btrfs/qgroup.c
2829
bool dirty = false;
fs/btrfs/qgroup.c
2841
dirty = true;
fs/btrfs/qgroup.c
2846
dirty = true;
fs/btrfs/qgroup.c
2857
dirty = true;
fs/btrfs/qgroup.c
2868
dirty = true;
fs/btrfs/qgroup.c
2882
dirty = true;
fs/btrfs/qgroup.c
2892
dirty = true;
fs/btrfs/qgroup.c
2898
if (dirty)
fs/btrfs/qgroup.c
3125
struct btrfs_qgroup, dirty);
fs/btrfs/qgroup.c
3126
list_del_init(&qgroup->dirty);
fs/btrfs/qgroup.h
247
struct list_head dirty; /* dirty groups */
fs/btrfs/subpage.c
395
dirty, start, len);
fs/btrfs/subpage.c
419
dirty, start, len);
fs/btrfs/subpage.c
425
if (subpage_test_bitmap_all_zero(fs_info, folio, dirty))
fs/btrfs/subpage.c
561
IMPLEMENT_BTRFS_SUBPAGE_TEST_OP(dirty);
fs/btrfs/subpage.c
656
IMPLEMENT_BTRFS_PAGE_OPS(dirty, folio_mark_dirty, folio_clear_dirty_for_io,
fs/btrfs/subpage.c
706
start_bit = subpage_calc_start_bit(fs_info, folio, dirty, start, len);
fs/btrfs/subpage.c
712
SUBPAGE_DUMP_BITMAP(fs_info, folio, dirty, start, len);
fs/btrfs/subpage.c
795
GET_SUBPAGE_BITMAP(fs_info, folio, dirty, &dirty_bitmap);
fs/btrfs/subpage.c
826
GET_SUBPAGE_BITMAP(fs_info, folio, dirty, ret_bitmap);
fs/btrfs/subpage.h
182
DECLARE_BTRFS_SUBPAGE_OPS(dirty);
fs/buffer.c
111
*dirty = true;
fs/buffer.c
90
bool *dirty, bool *writeback)
fs/buffer.c
93
*dirty = false;
fs/ceph/addr.c
2121
int dirty;
fs/ceph/addr.c
2123
dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR,
fs/ceph/addr.c
2126
if (dirty)
fs/ceph/addr.c
2127
__mark_inode_dirty(inode, dirty);
fs/ceph/addr.c
2314
int dirty;
fs/ceph/addr.c
2320
dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR, &prealloc_cf);
fs/ceph/addr.c
2323
if (dirty)
fs/ceph/addr.c
2324
__mark_inode_dirty(inode, dirty);
fs/ceph/caps.c
1230
int op, caps, wanted, dirty;
fs/ceph/caps.c
1257
ceph_cap_string(arg->dirty), arg->seq, arg->issue_seq,
fs/ceph/caps.c
1275
fc->dirty = cpu_to_le32(arg->dirty);
fs/ceph/caps.c
1458
arg->dirty = flushing;
fs/ceph/caps.c
1533
ceph_vinop(inode), ceph_cap_string(arg->dirty),
fs/ceph/caps.c
1579
arg.dirty = capsnap->dirty;
fs/ceph/caps.c
1701
ceph_cap_string(capsnap->dirty));
fs/ceph/caps.c
1791
int dirty = 0;
fs/ceph/caps.c
1827
dirty |= I_DIRTY_SYNC;
fs/ceph/caps.c
1835
dirty |= I_DIRTY_DATASYNC;
fs/ceph/caps.c
1837
return dirty;
fs/ceph/caps.c
2484
int dirty;
fs/ceph/caps.c
2497
dirty = try_flush_caps(inode, &flush_tid);
fs/ceph/caps.c
2498
doutc(cl, "dirty caps are %s\n", ceph_cap_string(dirty));
fs/ceph/caps.c
2507
if (!err && (dirty & ~CEPH_CAP_ANY_FILE_WR)) {
fs/ceph/caps.c
2536
int dirty;
fs/ceph/caps.c
2545
dirty = try_flush_caps(inode, &flush_tid);
fs/ceph/caps.c
2546
if (dirty)
fs/ceph/caps.c
2623
ceph_cap_string(capsnap->dirty));
fs/ceph/caps.c
3498
int used, wanted, dirty;
fs/ceph/caps.c
3678
dirty = __ceph_caps_dirty(ci);
fs/ceph/caps.c
3681
ceph_cap_string(dirty));
fs/ceph/caps.c
3823
int dirty = le32_to_cpu(m->dirty);
fs/ceph/caps.c
3858
ceph_cap_string(dirty), ceph_cap_string(cleaned),
fs/ceph/caps.c
4034
int dirty = __ceph_caps_dirty(ci);
fs/ceph/caps.c
4040
issued |= implemented | dirty;
fs/ceph/caps.c
4851
int used, dirty;
fs/ceph/caps.c
4856
dirty = __ceph_caps_dirty(ci);
fs/ceph/caps.c
4859
inode, ceph_vinop(inode), mds, ceph_cap_string(used|dirty),
fs/ceph/caps.c
4863
drop &= ~(used | dirty);
fs/ceph/file.c
1299
int dirty;
fs/ceph/file.c
1308
dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR,
fs/ceph/file.c
1311
if (dirty)
fs/ceph/file.c
1312
__mark_inode_dirty(inode, dirty);
fs/ceph/file.c
2469
int dirty;
fs/ceph/file.c
2472
dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR,
fs/ceph/file.c
2475
if (dirty)
fs/ceph/file.c
2476
__mark_inode_dirty(inode, dirty);
fs/ceph/file.c
2677
int dirty;
fs/ceph/file.c
2733
dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR,
fs/ceph/file.c
2736
if (dirty)
fs/ceph/file.c
2737
__mark_inode_dirty(inode, dirty);
fs/ceph/file.c
2977
int src_got = 0, dst_got = 0, err, dirty;
fs/ceph/file.c
3135
dirty = __ceph_mark_dirty_caps(dst_ci, CEPH_CAP_FILE_WR, &prealloc_cf);
fs/ceph/file.c
3137
if (dirty)
fs/ceph/file.c
3138
__mark_inode_dirty(dst_inode, dirty);
fs/ceph/inode.c
651
ci->i_xattrs.dirty = false;
fs/ceph/snap.c
538
int used, dirty;
fs/ceph/snap.c
542
dirty = __ceph_caps_dirty(ci);
fs/ceph/snap.c
553
dirty |= CEPH_CAP_FILE_WR;
fs/ceph/snap.c
565
!(dirty & (CEPH_CAP_ANY_EXCL|CEPH_CAP_FILE_WR))) {
fs/ceph/snap.c
582
if (dirty & (CEPH_CAP_ANY_EXCL|CEPH_CAP_FILE_WR))
fs/ceph/snap.c
595
ceph_cap_string(dirty), capsnap->need_flush ? "" : "no_flush");
fs/ceph/snap.c
600
capsnap->dirty = dirty;
fs/ceph/snap.c
606
if (dirty & CEPH_CAP_XATTR_EXCL) {
fs/ceph/snap.c
684
ceph_cap_string(capsnap->dirty),
fs/ceph/snap.c
697
capsnap->context->seq, ceph_cap_string(capsnap->dirty),
fs/ceph/snap.c
706
capsnap->context->seq, ceph_cap_string(capsnap->dirty),
fs/ceph/super.h
255
int issued, dirty;
fs/ceph/super.h
320
int dirty;
fs/ceph/super.h
363
bool dirty;
fs/ceph/xattr.c
1186
int dirty = 0;
fs/ceph/xattr.c
1280
dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_XATTR_EXCL,
fs/ceph/xattr.c
1282
ci->i_xattrs.dirty = true;
fs/ceph/xattr.c
1290
if (dirty)
fs/ceph/xattr.c
1291
__mark_inode_dirty(inode, dirty);
fs/ceph/xattr.c
652
xattr->dirty = update_xattr;
fs/ceph/xattr.c
863
ci->i_xattrs.dirty = false;
fs/ceph/xattr.c
915
if (ci->i_xattrs.dirty) {
fs/ceph/xattr.c
945
ci->i_xattrs.dirty = false;
fs/dax.c
1049
bool dirty = write && !dax_fault_is_synchronous(iter, vmf->vma);
fs/dax.c
1052
if (dirty)
fs/dax.c
1090
if (dirty)
fs/dlm/lowcomms.c
1117
e->dirty = true;
fs/dlm/lowcomms.c
1182
entry->dirty = false;
fs/dlm/lowcomms.c
133
bool dirty;
fs/dlm/lowcomms.c
786
if (e->dirty)
fs/erofs/data.c
251
void erofs_onlinefolio_end(struct folio *folio, int err, bool dirty)
fs/erofs/data.c
258
v = dirty << EROFS_ONLINEFOLIO_DIRTY;
fs/erofs/internal.h
451
void erofs_onlinefolio_end(struct folio *folio, int err, bool dirty);
fs/ext4/orphan.c
107
bool dirty = false;
fs/ext4/orphan.c
159
dirty = true;
fs/ext4/orphan.c
164
if (dirty) {
fs/ext4/xattr.c
1142
struct buffer_head *bh, bool block_csum, bool dirty)
fs/ext4/xattr.c
1146
if (bh && dirty) {
fs/ext4/xattr.c
1169
bool dirty = false;
fs/ext4/xattr.c
1211
dirty));
fs/ext4/xattr.c
1248
dirty = true;
fs/ext4/xattr.c
1251
if (dirty) {
fs/f2fs/inline.c
158
int dirty, err;
fs/f2fs/inline.c
191
dirty = folio_clear_dirty_for_io(folio);
fs/f2fs/inline.c
199
if (dirty) {
fs/fat/fat.h
106
unsigned int dirty; /* fs state before mount */
fs/fat/inode.c
1745
sbi->dirty = bpb.fat32_state & FAT_STATE_DIRTY;
fs/fat/inode.c
1747
sbi->dirty = bpb.fat16_state & FAT_STATE_DIRTY;
fs/fat/inode.c
681
if (sbi->dirty) {
fs/fs-writeback.c
1134
unsigned long dirty;
fs/fs-writeback.c
1172
dirty = memcg_page_state(mem_cgroup_from_css(memcg_css), NR_FILE_DIRTY);
fs/fs-writeback.c
1173
dirty = dirty * 10 / 8;
fs/fs-writeback.c
1178
work->nr_pages = dirty;
fs/fs-writeback.c
1755
unsigned dirty;
fs/fs-writeback.c
1795
dirty = inode_state_read(inode) & I_DIRTY;
fs/fs-writeback.c
1796
inode_state_clear(inode, dirty);
fs/fs-writeback.c
1817
dirty |= I_PINNING_NETFS_WB; /* Cause write_inode */
fs/fs-writeback.c
1824
if (dirty & ~I_DIRTY_PAGES) {
fs/hpfs/hpfs.h
117
u8 dirty: 1; /* 0 clean, 1 "improperly stopped" */
fs/hpfs/hpfs.h
133
u8 dirty: 1; /* 0 clean, 1 "improperly stopped" */
fs/hpfs/super.c
31
sb->dirty = 1;
fs/hpfs/super.c
50
sb->dirty = hpfs_sb(s)->sb_chkdsk > 1 - hpfs_sb(s)->sb_was_error;
fs/hpfs/super.c
594
if (spareblock->dirty && !spareblock->old_wrote) {
fs/hpfs/super.c
603
spareblock->dirty = 1;
fs/inode.c
2113
unsigned int dirty = 0;
fs/inode.c
2118
dirty = inode_time_dirty_flag(inode);
fs/inode.c
2133
if (inode_maybe_inc_iversion(inode, !!dirty))
fs/inode.c
2134
dirty |= I_DIRTY_SYNC;
fs/inode.c
2140
return dirty;
fs/inode.c
2183
int dirty;
fs/inode.c
2195
dirty = inode_update_time(inode, type, flags);
fs/inode.c
2196
if (dirty <= 0)
fs/inode.c
2197
return dirty;
fs/inode.c
2198
__mark_inode_dirty(inode, dirty);
fs/jffs2/debug.c
163
uint32_t free = 0, dirty = 0, used = 0, wasted = 0,
fs/jffs2/debug.c
171
dirty += c->gcblock->dirty_size;
fs/jffs2/debug.c
179
dirty += c->nextblock->dirty_size;
fs/jffs2/debug.c
187
dirty += jeb->dirty_size;
fs/jffs2/debug.c
195
dirty += jeb->dirty_size;
fs/jffs2/debug.c
203
dirty += jeb->dirty_size;
fs/jffs2/debug.c
211
dirty += jeb->dirty_size;
fs/jffs2/debug.c
219
dirty += jeb->dirty_size;
fs/jffs2/debug.c
227
dirty += jeb->dirty_size;
fs/jffs2/debug.c
235
dirty += jeb->dirty_size;
fs/jffs2/debug.c
243
dirty += jeb->dirty_size;
fs/jffs2/debug.c
276
check(dirty);
fs/jffs2/debug.c
503
uint32_t dirty = 0;
fs/jffs2/debug.c
508
dirty += jeb->wasted_size;
fs/jffs2/debug.c
517
numblocks, dirty, dirty / numblocks);
fs/jffs2/debug.c
525
uint32_t dirty = 0;
fs/jffs2/debug.c
531
dirty += jeb->dirty_size;
fs/jffs2/debug.c
540
numblocks, dirty, dirty / numblocks);
fs/jffs2/debug.c
548
uint32_t dirty = 0;
fs/jffs2/debug.c
554
dirty += jeb->dirty_size;
fs/jffs2/debug.c
563
numblocks, dirty, dirty / numblocks);
fs/jffs2/nodemgmt.c
105
uint32_t dirty, avail;
fs/jffs2/nodemgmt.c
119
dirty = c->dirty_size + c->erasing_size - c->nr_erasing_blocks * c->sector_size + c->unchecked_size;
fs/jffs2/nodemgmt.c
120
if (dirty < c->nospc_dirty_size) {
fs/jffs2/nodemgmt.c
127
dirty, c->unchecked_size,
fs/jffs2/nodemgmt.c
847
uint32_t dirty;
fs/jffs2/nodemgmt.c
869
dirty = c->dirty_size + c->erasing_size - c->nr_erasing_blocks * c->sector_size;
fs/jffs2/nodemgmt.c
872
(dirty > c->nospc_dirty_size))
fs/jfs/jfs_txnmgr.c
2594
void txAbort(tid_t tid, int dirty)
fs/jfs/jfs_txnmgr.c
2637
if (dirty)
fs/nfs/file.c
522
bool *dirty, bool *writeback)
fs/nfs/file.c
544
*dirty = true;
fs/nilfs2/sufile.c
439
int clean, dirty;
fs/nilfs2/sufile.c
449
dirty = nilfs_segment_usage_dirty(su);
fs/nilfs2/sufile.c
457
nilfs_sufile_mod_counter(header_bh, clean ? (u64)-1 : 0, dirty ? 0 : 1);
fs/ntfs3/attrib.c
1240
mi_b->dirty = true;
fs/ntfs3/attrib.c
1345
ni->attr_list.dirty = true;
fs/ntfs3/attrib.c
1346
mi->dirty = true;
fs/ntfs3/attrib.c
1403
mi->dirty = true;
fs/ntfs3/attrib.c
1850
mi_b->dirty = true;
fs/ntfs3/attrib.c
1938
ni->attr_list.dirty = true;
fs/ntfs3/attrib.c
1939
mi->dirty = true;
fs/ntfs3/attrib.c
1961
mi_b->dirty = true;
fs/ntfs3/attrib.c
2110
ni->attr_list.dirty = true;
fs/ntfs3/attrib.c
2112
mi->dirty = true;
fs/ntfs3/attrib.c
2141
ni->attr_list.dirty = true;
fs/ntfs3/attrib.c
2236
mi_b->dirty = true;
fs/ntfs3/attrib.c
2447
mi_b->dirty = true;
fs/ntfs3/attrib.c
2624
ni->attr_list.dirty = true;
fs/ntfs3/attrib.c
2626
mi->dirty = true;
fs/ntfs3/attrib.c
2662
mi_b->dirty = true;
fs/ntfs3/attrib.c
2717
ni->attr_list.dirty = true;
fs/ntfs3/attrib.c
2719
mi->dirty = true;
fs/ntfs3/attrib.c
341
mi->dirty = true;
fs/ntfs3/attrib.c
368
mi->dirty = true;
fs/ntfs3/attrib.c
415
mi->dirty = true;
fs/ntfs3/attrib.c
448
bool dirty = false;
fs/ntfs3/attrib.c
473
dirty = true;
fs/ntfs3/attrib.c
501
mi_b->dirty = dirty = true;
fs/ntfs3/attrib.c
578
mi_b->dirty = dirty = true;
fs/ntfs3/attrib.c
669
mi_b->dirty = dirty = true;
fs/ntfs3/attrib.c
756
mi_b->dirty = dirty = true;
fs/ntfs3/attrib.c
798
mi->dirty = true;
fs/ntfs3/attrib.c
818
mi_b->dirty = dirty = true;
fs/ntfs3/attrib.c
872
mi_b->dirty = true;
fs/ntfs3/attrib.c
886
dirty = true;
fs/ntfs3/attrib.c
892
if (dirty) {
fs/ntfs3/attrlist.c
35
ni->attr_list.dirty = false;
fs/ntfs3/attrlist.c
357
al->dirty = true;
fs/ntfs3/attrlist.c
364
al->dirty = false;
fs/ntfs3/attrlist.c
389
al->dirty = true;
fs/ntfs3/attrlist.c
400
if (!al->dirty || !al->size)
fs/ntfs3/attrlist.c
423
ni->mi.dirty = true;
fs/ntfs3/attrlist.c
424
al->dirty = false;
fs/ntfs3/frecord.c
1494
ni->mi.dirty = true;
fs/ntfs3/frecord.c
1579
mi->dirty = true;
fs/ntfs3/frecord.c
1590
ni->mi.dirty = true;
fs/ntfs3/frecord.c
1745
mi->dirty = true;
fs/ntfs3/frecord.c
2137
mi->dirty = true;
fs/ntfs3/frecord.c
2825
if (ni->mi.dirty || ni->attr_list.dirty ||
fs/ntfs3/frecord.c
2830
if (rb_entry(node, struct mft_inode, node)->dirty)
fs/ntfs3/frecord.c
3013
mi->dirty = true;
fs/ntfs3/frecord.c
3083
mi->dirty = true;
fs/ntfs3/frecord.c
3167
ni->mi.dirty = true;
fs/ntfs3/frecord.c
3184
if (ni->attr_list.size && ni->attr_list.dirty) {
fs/ntfs3/frecord.c
3203
if (!mi->dirty)
fs/ntfs3/frecord.c
3222
if (ni->mi.dirty) {
fs/ntfs3/frecord.c
457
ni->attr_list.dirty = true;
fs/ntfs3/frecord.c
570
mi->dirty = true;
fs/ntfs3/frecord.c
571
ni->attr_list.dirty = true;
fs/ntfs3/frecord.c
621
if (!ni->attr_list.dirty)
fs/ntfs3/frecord.c
737
ni->attr_list.dirty = false;
fs/ntfs3/frecord.c
878
ni->attr_list.dirty = false;
fs/ntfs3/fslog.c
3231
mi->dirty = true;
fs/ntfs3/fslog.c
3237
mi->dirty = true;
fs/ntfs3/fslog.c
3248
mi->dirty = true;
fs/ntfs3/fslog.c
3278
mi->dirty = true;
fs/ntfs3/fslog.c
3294
mi->dirty = true;
fs/ntfs3/fslog.c
3338
mi->dirty = true;
fs/ntfs3/fslog.c
3370
mi->dirty = true;
fs/ntfs3/fslog.c
3389
mi->dirty = true;
fs/ntfs3/fslog.c
3418
mi->dirty = true;
fs/ntfs3/fslog.c
3446
mi->dirty = true;
fs/ntfs3/fslog.c
3461
mi->dirty = true;
fs/ntfs3/fslog.c
3476
mi->dirty = true;
fs/ntfs3/fslog.c
3492
mi->dirty = true;
fs/ntfs3/fslog.c
3500
mi->dirty = true;
fs/ntfs3/fslog.c
3701
if (mi && mi->dirty) {
fs/ntfs3/fsntfs.c
1001
mi->dirty = true;
fs/ntfs3/fsntfs.c
2484
bool dirty = false;
fs/ntfs3/fsntfs.c
2489
dirty = true;
fs/ntfs3/fsntfs.c
2543
if (dirty)
fs/ntfs3/fsntfs.c
545
ni->mi.dirty = true;
fs/ntfs3/fsntfs.c
944
int ntfs_set_state(struct ntfs_sb_info *sbi, enum NTFS_DIRTY_FLAGS dirty)
fs/ntfs3/fsntfs.c
962
if ((dirty == NTFS_DIRTY_CLEAR ? 0 : VOLUME_FLAG_DIRTY) ==
fs/ntfs3/fsntfs.c
986
switch (dirty) {
fs/ntfs3/index.c
1701
mi->dirty = true;
fs/ntfs3/index.c
1730
mi->dirty = true;
fs/ntfs3/index.c
2649
mi->dirty = true;
fs/ntfs3/index.c
2719
mi->dirty = true;
fs/ntfs3/index.c
284
static void bmp_buf_put(struct bmp_buf *bbuf, bool dirty)
fs/ntfs3/index.c
290
if (b && !b->non_res && dirty)
fs/ntfs3/index.c
291
bbuf->mi->dirty = true;
fs/ntfs3/index.c
295
if (!dirty)
fs/ntfs3/index.c
300
bbuf->mi->dirty = true;
fs/ntfs3/inode.c
1755
ni->mi.dirty = false;
fs/ntfs3/inode.c
432
ni->mi.dirty = true;
fs/ntfs3/inode.c
951
mi->dirty = true;
fs/ntfs3/ntfs_fs.h
343
bool dirty;
fs/ntfs3/ntfs_fs.h
420
bool dirty;
fs/ntfs3/ntfs_fs.h
654
int ntfs_set_state(struct ntfs_sb_info *sbi, enum NTFS_DIRTY_FLAGS dirty);
fs/ntfs3/record.c
140
mi->dirty = true;
fs/ntfs3/record.c
167
mi->dirty = true;
fs/ntfs3/record.c
396
if (!mi->dirty)
fs/ntfs3/record.c
409
mi->dirty = false;
fs/ntfs3/record.c
450
mi->dirty = true;
fs/ntfs3/record.c
530
mi->dirty = true;
fs/ntfs3/record.c
557
ni->mi.dirty = true;
fs/ntfs3/record.c
564
mi->dirty = true;
fs/ntfs3/record.c
58
mi->dirty = true;
fs/ntfs3/record.c
614
mi->dirty = true;
fs/ntfs3/record.c
659
mi->dirty = true;
fs/ntfs3/xattr.c
484
mi->dirty = true;
fs/ntfs3/xattr.c
513
mi->dirty = true;
fs/ntfs3/xattr.c
908
ni->mi.dirty = true;
fs/ntfs3/xattr.c
951
ni->mi.dirty = true;
fs/ocfs2/dlm/dlmcommon.h
295
struct list_head dirty;
fs/ocfs2/dlm/dlmdebug.c
451
!list_empty(&res->dirty),
fs/ocfs2/dlm/dlmdebug.c
97
str_no_yes(list_empty(&res->dirty)),
fs/ocfs2/dlm/dlmmaster.c
485
!list_empty(&res->dirty) ||
fs/ocfs2/dlm/dlmmaster.c
496
!list_empty(&res->dirty) ? 'D' : ' ',
fs/ocfs2/dlm/dlmmaster.c
509
BUG_ON(!list_empty(&res->dirty));
fs/ocfs2/dlm/dlmmaster.c
545
INIT_LIST_HEAD(&res->dirty);
fs/ocfs2/dlm/dlmthread.c
520
if (list_empty(&res->dirty)) {
fs/ocfs2/dlm/dlmthread.c
523
list_add_tail(&res->dirty, &dlm->dirty_list);
fs/ocfs2/dlm/dlmthread.c
708
struct dlm_lock_resource, dirty);
fs/ocfs2/dlm/dlmthread.c
717
list_del_init(&res->dirty);
fs/ocfs2/dlm/dlmthread.c
87
if (!list_empty(&res->dirty) || res->state & DLM_LOCK_RES_DIRTY)
fs/ocfs2/journal.c
1011
int dirty, int replayed)
fs/ocfs2/journal.c
1027
if (dirty)
fs/ocfs2/journal.c
55
int dirty, int replayed);
fs/ocfs2/journal.c
906
int ocfs2_journal_init(struct ocfs2_super *osb, int *dirty)
fs/ocfs2/journal.c
970
*dirty = (le32_to_cpu(di->id1.journal1.ij_flags) &
fs/ocfs2/journal.h
174
int ocfs2_journal_init(struct ocfs2_super *osb, int *dirty);
fs/ocfs2/super.c
2338
int dirty;
fs/ocfs2/super.c
2345
status = ocfs2_journal_init(osb, &dirty);
fs/ocfs2/super.c
2360
if (!dirty) {
fs/ocfs2/super.c
2374
status = ocfs2_journal_load(osb->journal, local, dirty);
fs/ocfs2/super.c
2389
if (dirty) {
fs/ocfs2/super.c
2408
if (dirty) {
fs/omfs/dir.c
159
struct inode *dirty;
fs/omfs/dir.c
202
dirty = omfs_iget(dir->i_sb, prev);
fs/omfs/dir.c
203
if (!IS_ERR(dirty)) {
fs/omfs/dir.c
204
mark_inode_dirty(dirty);
fs/omfs/dir.c
205
iput(dirty);
fs/orangefs/inode.c
879
int dirty;
fs/orangefs/inode.c
893
dirty = inode_update_time(inode, type, flags);
fs/orangefs/inode.c
894
if (dirty <= 0)
fs/orangefs/inode.c
895
return dirty;
fs/proc/task_mmu.c
1025
bool present = false, young = false, dirty = false;
fs/proc/task_mmu.c
1031
dirty = pte_dirty(ptent);
fs/proc/task_mmu.c
1061
smaps_account(mss, page, false, young, dirty, locked, present);
fs/proc/task_mmu.c
3126
unsigned long dirty;
fs/proc/task_mmu.c
3149
md->dirty += nr_pages;
fs/proc/task_mmu.c
3347
if (md->dirty)
fs/proc/task_mmu.c
3348
seq_printf(m, " dirty=%lu", md->dirty);
fs/proc/task_mmu.c
3350
if (md->pages != md->anon && md->pages != md->dirty)
fs/proc/task_mmu.c
888
bool dirty, bool locked, bool private)
fs/proc/task_mmu.c
902
if (dirty || folio_test_dirty(folio)) {
fs/proc/task_mmu.c
917
bool compound, bool young, bool dirty, bool locked,
fs/proc/task_mmu.c
932
if (!folio_test_swapbacked(folio) && !dirty &&
fs/proc/task_mmu.c
962
dirty, locked, present);
fs/proc/task_mmu.c
987
dirty, locked, exclusive);
fs/pstore/zone.c
1185
atomic_set(&zone->dirty, 0);
fs/pstore/zone.c
219
goto dirty;
fs/pstore/zone.c
223
goto dirty;
fs/pstore/zone.c
228
goto dirty;
fs/pstore/zone.c
234
goto dirty;
fs/pstore/zone.c
240
goto dirty;
fs/pstore/zone.c
246
goto dirty;
fs/pstore/zone.c
251
dirty:
fs/pstore/zone.c
255
atomic_set(&zone->dirty, true);
fs/pstore/zone.c
272
if (!atomic_xchg(&zone->dirty, false))
fs/pstore/zone.c
277
atomic_set(&zone->dirty, true);
fs/pstore/zone.c
308
atomic_set(&new->dirty, false);
fs/pstore/zone.c
347
if (atomic_read(&zone->dirty)) {
fs/pstore/zone.c
86
atomic_t dirty;
fs/pstore/zone.c
968
atomic_set(&zone->dirty, 0);
fs/quota/dquot.c
686
struct list_head dirty;
fs/quota/dquot.c
703
list_replace_init(&dqopt->info[cnt].dqi_dirty_list, &dirty);
fs/quota/dquot.c
704
while (!list_empty(&dirty)) {
fs/quota/dquot.c
705
dquot = list_first_entry(&dirty, struct dquot,
fs/ubifs/debug.c
1101
if (ui->ui_size != ui->synced_i_size && !ui->dirty) {
fs/ubifs/debug.c
252
pr_err("\tdirty %u\n", ui->dirty);
fs/ubifs/debug.c
686
spc = lp->free + lp->dirty;
fs/ubifs/debug.c
694
lp->lnum, lp->free, lp->dirty, c->leb_size - spc, spc,
fs/ubifs/debug.c
698
lp->lnum, lp->free, lp->dirty, c->leb_size - spc, spc,
fs/ubifs/debug.c
826
i + c->lpt_first, c->ltab[i].free, c->ltab[i].dirty,
fs/ubifs/debug.c
918
lprops->dirty, lprops->flags);
fs/ubifs/debug.c
937
i, lp->free, lp->dirty, lp->flags, lp->lnum);
fs/ubifs/dir.c
1423
ubifs_assert(c, !whiteout_ui->dirty);
fs/ubifs/dir.c
1530
release = old_inode_ui->dirty;
fs/ubifs/dir.c
498
ubifs_assert(c, ui->dirty);
fs/ubifs/file.c
1236
release = ui->dirty;
fs/ubifs/file.c
1389
release = ui->dirty;
fs/ubifs/file.c
1424
release = ui->dirty;
fs/ubifs/file.c
1562
release = ui->dirty;
fs/ubifs/file.c
285
if (ui->dirty)
fs/ubifs/file.c
325
if (ui->dirty)
fs/ubifs/file.c
362
if (!ui->dirty)
fs/ubifs/file.c
506
if (!ui->dirty)
fs/ubifs/find.c
103
if (lprops->free + lprops->dirty == c->leb_size) {
fs/ubifs/find.c
107
} else if (lprops->dirty < c->dead_wm)
fs/ubifs/find.c
138
if (lprops->free + lprops->dirty < min_space)
fs/ubifs/find.c
140
if (lprops->dirty < c->dead_wm)
fs/ubifs/find.c
154
if (lprops->free + lprops->dirty < min_space)
fs/ubifs/find.c
158
if (lprops->dirty < c->dead_wm)
fs/ubifs/find.c
180
ubifs_assert(c, lprops->free + lprops->dirty >= min_space);
fs/ubifs/find.c
181
ubifs_assert(c, lprops->dirty >= c->dead_wm ||
fs/ubifs/find.c
183
lprops->free + lprops->dirty == c->leb_size));
fs/ubifs/find.c
277
sum = idx_lp->free + idx_lp->dirty;
fs/ubifs/find.c
294
if (lp->dirty + lp->free < min_space)
fs/ubifs/find.c
300
if (idx_lp->free + idx_lp->dirty >= lp->free + lp->dirty)
fs/ubifs/find.c
306
ubifs_assert(c, lp->free + lp->dirty >= c->dead_wm);
fs/ubifs/find.c
317
ubifs_assert(c, lp->dirty >= c->dead_wm ||
fs/ubifs/find.c
318
(pick_free && lp->free + lp->dirty == c->leb_size));
fs/ubifs/find.c
322
lp->lnum, lp->free, lp->dirty, lp->flags);
fs/ubifs/find.c
378
if (lprops->free + lprops->dirty == c->leb_size && lprops->dirty > 0)
fs/ubifs/find.c
54
if (lprops->free + lprops->dirty >= c->dark_wm)
fs/ubifs/find.c
608
if (lprops->free + lprops->dirty != c->leb_size)
fs/ubifs/find.c
640
ubifs_assert(c, lprops->free + lprops->dirty == c->leb_size);
fs/ubifs/find.c
700
lnum, lprops->free, lprops->dirty, lprops->flags);
fs/ubifs/find.c
735
return lpa->dirty + lpa->free - lpb->dirty - lpb->free;
fs/ubifs/find.c
762
c->dirty_idx.arr[c->dirty_idx.cnt - 1]->dirty,
fs/ubifs/find.c
800
if (lprops->free + lprops->dirty < c->min_idx_node_sz)
fs/ubifs/find.c
857
ubifs_assert(c, lprops->free + lprops->dirty >= c->min_idx_node_sz);
fs/ubifs/find.c
862
lprops->lnum, lprops->free, lprops->dirty, lprops->flags);
fs/ubifs/find.c
897
lp->lnum, lp->dirty, lp->free, lp->flags);
fs/ubifs/find.c
926
dbg_find("LEB %d, dirty %d and free %d flags %#x", lp->lnum, lp->dirty,
fs/ubifs/find.c
97
if (lprops->free + lprops->dirty < data->min_space)
fs/ubifs/gc.c
503
if (lp->free + lp->dirty == c->leb_size) {
fs/ubifs/gc.c
549
lnum, lp->free, lp->dirty);
fs/ubifs/gc.c
585
lnum, lp->free, lp->dirty);
fs/ubifs/gc.c
742
lp.lnum, lp.free, lp.dirty, lp.free + lp.dirty,
fs/ubifs/ioctl.c
122
release = ui->dirty;
fs/ubifs/journal.c
1397
ubifs_assert(c, !whiteout_ui->dirty);
fs/ubifs/journal.c
625
if (ui->dirty)
fs/ubifs/journal.c
627
ui->dirty = 0;
fs/ubifs/lprops.c
1022
int cat, lnum = lp->lnum, is_idx = 0, used = 0, free, dirty, ret;
fs/ubifs/lprops.c
1091
if (lp->free + lp->dirty == c->leb_size &&
fs/ubifs/lprops.c
1094
lst->total_dirty += lp->dirty;
fs/ubifs/lprops.c
1145
dirty = sleb->endpt - used;
fs/ubifs/lprops.c
1147
if (free > c->leb_size || free < 0 || dirty > c->leb_size ||
fs/ubifs/lprops.c
1148
dirty < 0) {
fs/ubifs/lprops.c
1150
lnum, free, dirty);
fs/ubifs/lprops.c
1154
if (lp->free + lp->dirty == c->leb_size &&
fs/ubifs/lprops.c
1155
free + dirty == c->leb_size)
fs/ubifs/lprops.c
1167
dirty = lp->dirty;
fs/ubifs/lprops.c
1171
if (is_idx && lp->free + lp->dirty == free + dirty &&
fs/ubifs/lprops.c
1185
dirty = lp->dirty;
fs/ubifs/lprops.c
1188
if (lp->free != free || lp->dirty != dirty)
fs/ubifs/lprops.c
1213
lst->total_used += c->leb_size - free - dirty;
fs/ubifs/lprops.c
1215
lst->total_dirty += dirty;
fs/ubifs/lprops.c
1218
int spc = free + dirty;
fs/ubifs/lprops.c
1232
lnum, lp->free, lp->dirty, lp->flags, free, dirty);
fs/ubifs/lprops.c
32
return lprops->free + lprops->dirty;
fs/ubifs/lprops.c
34
return lprops->dirty;
fs/ubifs/lprops.c
407
if (lprops->free + lprops->dirty == c->leb_size) {
fs/ubifs/lprops.c
415
if (lprops->dirty + lprops->free >= c->min_idx_node_sz)
fs/ubifs/lprops.c
418
if (lprops->dirty >= c->dead_wm &&
fs/ubifs/lprops.c
419
lprops->dirty > lprops->free)
fs/ubifs/lprops.c
522
int free, int dirty, int flags,
fs/ubifs/lprops.c
532
lprops->lnum, free, dirty, flags);
fs/ubifs/lprops.c
545
ubifs_assert(c, dirty == LPROPS_NC || dirty >= 0);
fs/ubifs/lprops.c
554
ubifs_assert(c, !(lprops->free & 7) && !(lprops->dirty & 7));
fs/ubifs/lprops.c
563
old_spc = lprops->free + lprops->dirty;
fs/ubifs/lprops.c
585
if (dirty != LPROPS_NC) {
fs/ubifs/lprops.c
586
dirty = ALIGN(dirty, 8);
fs/ubifs/lprops.c
587
c->lst.total_dirty += dirty - lprops->dirty;
fs/ubifs/lprops.c
588
lprops->dirty = dirty;
fs/ubifs/lprops.c
604
new_spc = lprops->free + lprops->dirty;
fs/ubifs/lprops.c
649
int ubifs_change_one_lp(struct ubifs_info *c, int lnum, int free, int dirty,
fs/ubifs/lprops.c
664
lp = ubifs_change_lp(c, lp, free, dirty, flags, idx_gc_cnt);
fs/ubifs/lprops.c
688
int ubifs_update_one_lp(struct ubifs_info *c, int lnum, int free, int dirty,
fs/ubifs/lprops.c
703
lp = ubifs_change_lp(c, lp, free, lp->dirty + dirty, flags, 0);
fs/ubifs/lprops.c
813
ubifs_assert(c, lprops->free + lprops->dirty == c->leb_size);
fs/ubifs/lprops.c
837
ubifs_assert(c, lprops->free + lprops->dirty == c->leb_size);
fs/ubifs/lprops.c
863
lprops->lnum, lprops->free, lprops->dirty,
fs/ubifs/lprops.c
869
lprops->lnum, lprops->free, lprops->dirty,
fs/ubifs/lprops.c
877
if (lprops->free + lprops->dirty != c->leb_size) {
fs/ubifs/lprops.c
879
lprops->lnum, lprops->free, lprops->dirty,
fs/ubifs/lprops.c
885
lprops->lnum, lprops->free, lprops->dirty,
fs/ubifs/lprops.c
907
if (lprops->free + lprops->dirty != c->leb_size) {
fs/ubifs/lprops.c
909
lprops->lnum, lprops->free, lprops->dirty,
fs/ubifs/lprops.c
915
lprops->lnum, lprops->free, lprops->dirty,
fs/ubifs/lprops.c
921
lprops->lnum, lprops->free, lprops->dirty,
fs/ubifs/lpt.c
1054
int dirty = ubifs_unpack_bits(c, &addr, &pos, c->lpt_spc_bits);
fs/ubifs/lpt.c
1056
if (free < 0 || free > c->leb_size || dirty < 0 ||
fs/ubifs/lpt.c
1057
dirty > c->leb_size || free + dirty > c->leb_size)
fs/ubifs/lpt.c
1061
c->ltab[i].dirty = dirty;
fs/ubifs/lpt.c
1161
int dirty = pnode->lprops[i].dirty;
fs/ubifs/lpt.c
1166
if (dirty < 0 || dirty > c->leb_size || (dirty & 7))
fs/ubifs/lpt.c
1168
if (dirty + free > c->leb_size)
fs/ubifs/lpt.c
1499
pnode->lprops[iip].free, pnode->lprops[iip].dirty,
fs/ubifs/lpt.c
1639
pnode->lprops[iip].free, pnode->lprops[iip].dirty,
fs/ubifs/lpt.c
2353
lprops->dirty);
fs/ubifs/lpt.c
2359
if (lprops->free + lprops->dirty != c->leb_size) {
fs/ubifs/lpt.c
2362
lprops->dirty);
fs/ubifs/lpt.c
353
pack_bits(c, &addr, &pos, pnode->lprops[i].dirty >> 3,
fs/ubifs/lpt.c
415
pack_bits(c, &addr, &pos, ltab[i].dirty, c->lpt_spc_bits);
fs/ubifs/lpt.c
452
void ubifs_add_lpt_dirt(struct ubifs_info *c, int lnum, int dirty)
fs/ubifs/lpt.c
454
if (!dirty || !lnum)
fs/ubifs/lpt.c
457
lnum, dirty, c->ltab[lnum - c->lpt_first].dirty);
fs/ubifs/lpt.c
459
c->ltab[lnum - c->lpt_first].dirty += dirty;
fs/ubifs/lpt.c
469
static void set_ltab(struct ubifs_info *c, int lnum, int free, int dirty)
fs/ubifs/lpt.c
473
c->ltab[lnum - c->lpt_first].dirty, free, dirty);
fs/ubifs/lpt.c
476
c->ltab[lnum - c->lpt_first].dirty = dirty;
fs/ubifs/lpt.c
644
ltab[i].dirty = 0;
fs/ubifs/lpt.c
661
pnode->lprops[0].dirty = iopos - node_sz;
fs/ubifs/lpt.c
667
pnode->lprops[1].dirty = iopos - node_sz;
fs/ubifs/lpt.c
684
pnode->lprops[0].dirty = 0;
fs/ubifs/lpt.c
688
pnode->lprops[1].dirty = 0;
fs/ubifs/lpt.c
990
lprops->dirty = ubifs_unpack_bits(c, &addr, &pos, c->space_bits);
fs/ubifs/lpt.c
991
lprops->dirty <<= 3;
fs/ubifs/lpt_commit.c
1152
int i, lnum = -1, dirty = 0;
fs/ubifs/lpt_commit.c
1158
c->ltab[i].free + c->ltab[i].dirty == c->leb_size)
fs/ubifs/lpt_commit.c
1160
if (c->ltab[i].dirty > dirty) {
fs/ubifs/lpt_commit.c
1161
dirty = c->ltab[i].dirty;
fs/ubifs/lpt_commit.c
129
static void upd_ltab(struct ubifs_info *c, int lnum, int free, int dirty)
fs/ubifs/lpt_commit.c
133
c->ltab[lnum - c->lpt_first].dirty, free, dirty);
fs/ubifs/lpt_commit.c
136
c->ltab[lnum - c->lpt_first].dirty += dirty;
fs/ubifs/lpt_commit.c
1592
int err, len = c->leb_size, dirty = 0, node_type, node_num, node_len;
fs/ubifs/lpt_commit.c
1619
dirty += pad_len;
fs/ubifs/lpt_commit.c
1633
if (dirty != c->ltab[i].dirty) {
fs/ubifs/lpt_commit.c
1635
lnum, dirty, c->ltab[i].dirty);
fs/ubifs/lpt_commit.c
1644
dirty += node_len;
fs/ubifs/lpt_commit.c
691
else if (c->ltab[i].free + c->ltab[i].dirty == c->leb_size)
fs/ubifs/lpt_commit.c
715
if (c->ltab[i].dirty > 0 &&
fs/ubifs/lpt_commit.c
716
c->ltab[i].free + c->ltab[i].dirty == c->leb_size) {
fs/ubifs/lpt_commit.c
719
c->ltab[i].dirty = 0;
fs/ubifs/misc.h
161
static inline int ubifs_add_dirt(struct ubifs_info *c, int lnum, int dirty)
fs/ubifs/misc.h
163
return ubifs_update_one_lp(c, lnum, LPROPS_NC, dirty, 0, 0);
fs/ubifs/recovery.c
1180
ubifs_assert(c, lp.free + lp.dirty >= wbuf->offs);
fs/ubifs/replay.c
101
dirty = lp->dirty;
fs/ubifs/replay.c
102
if (b->bud->start == 0 && (lp->free != c->leb_size || lp->dirty != 0)) {
fs/ubifs/replay.c
123
lp->free, lp->dirty);
fs/ubifs/replay.c
125
lp->free, lp->dirty);
fs/ubifs/replay.c
126
dirty -= c->leb_size - lp->free;
fs/ubifs/replay.c
134
if (dirty != 0)
fs/ubifs/replay.c
136
b->bud->lnum, lp->free, lp->dirty, b->free,
fs/ubifs/replay.c
137
b->dirty);
fs/ubifs/replay.c
139
lp = ubifs_change_lp(c, lp, b->free, dirty + b->dirty,
fs/ubifs/replay.c
76
int dirty;
fs/ubifs/replay.c
824
b->dirty = sleb->endpt - offs - used;
fs/ubifs/replay.c
827
lnum, b->dirty, b->free);
fs/ubifs/replay.c
91
int err = 0, dirty;
fs/ubifs/super.c
310
if (!ui->dirty) {
fs/ubifs/super.c
330
ui->dirty = 0;
fs/ubifs/super.c
382
if (ui->dirty)
fs/ubifs/super.c
400
if (!ui->dirty) {
fs/ubifs/super.c
401
ui->dirty = 1;
fs/ubifs/ubifs.h
1978
void ubifs_add_lpt_dirt(struct ubifs_info *c, int lnum, int dirty);
fs/ubifs/ubifs.h
1996
int free, int dirty, int flags,
fs/ubifs/ubifs.h
2006
int ubifs_change_one_lp(struct ubifs_info *c, int lnum, int free, int dirty,
fs/ubifs/ubifs.h
2008
int ubifs_update_one_lp(struct ubifs_info *c, int lnum, int free, int dirty,
fs/ubifs/ubifs.h
406
unsigned int dirty:1;
fs/ubifs/ubifs.h
481
int dirty;
fs/ubifs/ubifs.h
499
int dirty;
fs/xfs/scrub/inode_repair.c
1721
bool dirty = false;
fs/xfs/scrub/inode_repair.c
1727
dirty = true;
fs/xfs/scrub/inode_repair.c
1734
dirty = true;
fs/xfs/scrub/inode_repair.c
1741
dirty = true;
fs/xfs/scrub/inode_repair.c
1747
if (dirty)
fs/xfs/scrub/nlinks_repair.c
216
dirty = true;
fs/xfs/scrub/nlinks_repair.c
227
dirty = true;
fs/xfs/scrub/nlinks_repair.c
238
dirty = true;
fs/xfs/scrub/nlinks_repair.c
247
dirty = true;
fs/xfs/scrub/nlinks_repair.c
250
if (!dirty) {
fs/xfs/scrub/nlinks_repair.c
99
bool dirty = false;
fs/xfs/scrub/quota_repair.c
109
bool *dirty)
fs/xfs/scrub/quota_repair.c
147
*dirty = true;
fs/xfs/scrub/quota_repair.c
156
bool *dirty)
fs/xfs/scrub/quota_repair.c
161
*dirty = true;
fs/xfs/scrub/quota_repair.c
164
*dirty = true;
fs/xfs/scrub/quota_repair.c
177
bool dirty = false;
fs/xfs/scrub/quota_repair.c
190
error = xrep_quota_item_bmap(sc, dq, &dirty);
fs/xfs/scrub/quota_repair.c
198
dirty = true;
fs/xfs/scrub/quota_repair.c
203
dirty = true;
fs/xfs/scrub/quota_repair.c
208
dirty = true;
fs/xfs/scrub/quota_repair.c
222
dirty = true;
fs/xfs/scrub/quota_repair.c
230
dirty = true;
fs/xfs/scrub/quota_repair.c
237
dirty = true;
fs/xfs/scrub/quota_repair.c
240
xrep_quota_item_timer(sc, &dq->q_blk, &dirty);
fs/xfs/scrub/quota_repair.c
241
xrep_quota_item_timer(sc, &dq->q_ino, &dirty);
fs/xfs/scrub/quota_repair.c
242
xrep_quota_item_timer(sc, &dq->q_rtb, &dirty);
fs/xfs/scrub/quota_repair.c
244
if (!dirty)
fs/xfs/scrub/quotacheck_repair.c
106
if (error || !dirty)
fs/xfs/scrub/quotacheck_repair.c
52
bool dirty = false;
fs/xfs/scrub/quotacheck_repair.c
77
dirty = true;
fs/xfs/scrub/quotacheck_repair.c
84
dirty = true;
fs/xfs/scrub/quotacheck_repair.c
91
dirty = true;
fs/xfs/scrub/repair.c
1197
bool dirty = false;
fs/xfs/scrub/repair.c
1223
dirty = true;
fs/xfs/scrub/repair.c
1235
if (!dirty) {
fs/xfs/scrub/repair.c
1236
dirty = true;
fs/xfs/scrub/repair.c
1248
if (dirty) {
fs/xfs/scrub/repair.c
1252
dirty = false;
fs/xfs/xfs_buf_item.c
686
bool dirty = bip->bli_flags & XFS_BLI_DIRTY;
fs/xfs/xfs_buf_item.c
699
ASSERT((!ordered && dirty == xfs_buf_item_dirty_format(bip)) ||
fs/xfs/xfs_buf_item.c
700
(ordered && dirty && !xfs_buf_item_dirty_format(bip)));
fs/xfs/xfs_buf_item.c
747
if (!dirty)
fs/xfs/xfs_buf_item_recover.c
1084
bool dirty;
fs/xfs/xfs_buf_item_recover.c
1086
dirty = xlog_recover_do_dquot_buffer(mp, log, item, bp, buf_f);
fs/xfs/xfs_buf_item_recover.c
1087
if (!dirty)
fs/xfs/xfs_iops.c
1199
int dirty;
fs/xfs/xfs_iops.c
1201
dirty = inode_update_time(inode, type, flags);
fs/xfs/xfs_iops.c
1202
if (dirty <= 0)
fs/xfs/xfs_iops.c
1203
return dirty;
fs/xfs/xfs_iops.c
1204
if (dirty == I_DIRTY_TIME) {
fs/xfs/xfs_trans.c
951
bool dirty = (tp->t_flags & XFS_TRANS_DIRTY);
fs/xfs/xfs_trans.c
963
dirty = true;
fs/xfs/xfs_trans.c
973
if (dirty && !xfs_is_shutdown(mp)) {
fs/xfs/xfs_trans.c
979
if (!dirty && !xlog_is_shutdown(log)) {
fs/xfs/xfs_trans.c
994
xfs_trans_free_items(tp, dirty);
include/asm-generic/hugetlb.h
111
pte_t pte, int dirty)
include/asm-generic/hugetlb.h
113
return ptep_set_access_flags(vma, addr, ptep, pte, dirty);
include/drm/drm_framebuffer.h
98
int (*dirty)(struct drm_framebuffer *framebuffer,
include/drm/drm_gem.h
593
bool dirty, bool accessed);
include/hyperv/hvhdk.h
43
u32 dirty;
include/hyperv/hvhdk_mini.h
324
u8 dirty : 1;
include/linux/buffer_head.h
123
BUFFER_FNS(Dirty, dirty)
include/linux/buffer_head.h
124
TAS_BUFFER_FNS(Dirty, dirty)
include/linux/buffer_head.h
189
bool *dirty, bool *writeback);
include/linux/ceph/ceph_fs.h
779
__le32 caps, wanted, dirty; /* latest issued/wanted/dirty */
include/linux/ceph/libceph.h
317
bool dirty);
include/linux/fs.h
436
void (*is_dirty_writeback) (struct folio *, bool *dirty, bool *wb);
include/linux/generic_pt/iommu.h
207
unsigned long flags, struct iommu_dirty_bitmap *dirty); \
include/linux/io-pgtable.h
227
struct iommu_dirty_bitmap *dirty);
include/linux/iommu.h
1066
static inline void iommu_dirty_bitmap_init(struct iommu_dirty_bitmap *dirty,
include/linux/iommu.h
1073
dirty->bitmap = bitmap;
include/linux/iommu.h
1074
dirty->gather = gather;
include/linux/iommu.h
1077
static inline void iommu_dirty_bitmap_record(struct iommu_dirty_bitmap *dirty,
include/linux/iommu.h
1081
if (dirty->bitmap)
include/linux/iommu.h
1082
iova_bitmap_set(dirty->bitmap, iova, length);
include/linux/iommu.h
1084
if (dirty->gather)
include/linux/iommu.h
1085
iommu_iotlb_gather_add_range(dirty->gather, iova, length);
include/linux/iommu.h
1414
static inline void iommu_dirty_bitmap_init(struct iommu_dirty_bitmap *dirty,
include/linux/iommu.h
1420
static inline void iommu_dirty_bitmap_record(struct iommu_dirty_bitmap *dirty,
include/linux/iommu.h
389
struct iommu_dirty_bitmap *dirty);
include/linux/kprobes.h
298
kprobe_opcode_t *slot, int dirty);
include/linux/kprobes.h
311
static inline void free_##__name##_slot(kprobe_opcode_t *slot, int dirty)\
include/linux/kprobes.h
313
__free_insn_slot(&kprobe_##__name##_slots, slot, dirty); \
include/linux/kvm_host.h
1278
bool unused, bool dirty)
include/linux/kvm_host.h
1297
else if (dirty)
include/linux/page-flags.h
556
PAGEFLAG(Dirty, dirty, PF_HEAD) TESTSCFLAG(Dirty, dirty, PF_HEAD)
include/linux/page-flags.h
557
__CLEARPAGEFLAG(Dirty, dirty, PF_HEAD)
include/linux/pgtable.h
429
pte_t entry, int dirty);
include/linux/pgtable.h
436
pmd_t entry, int dirty);
include/linux/pgtable.h
439
pud_t entry, int dirty);
include/linux/pgtable.h
443
pmd_t entry, int dirty)
include/linux/pgtable.h
450
pud_t entry, int dirty)
include/linux/vfio.h
255
struct iova_bitmap *dirty);
include/linux/writeback.h
317
unsigned long dirty; /* file_dirty + write + nfs */
include/sound/emu10k1.h
1487
unsigned char dirty;
include/sound/soc-dapm.h
564
struct list_head dirty;
include/trace/events/bcache.h
342
__field(u8, dirty )
include/trace/events/bcache.h
353
__entry->dirty = KEY_DIRTY(k);
include/trace/events/bcache.h
362
__entry->size, __entry->dirty)
include/trace/events/bcache.h
49
__field(bool, dirty )
include/trace/events/bcache.h
56
__entry->dirty = KEY_DIRTY(k);
include/trace/events/bcache.h
60
__entry->offset, __entry->size, __entry->dirty)
include/trace/events/f2fs.h
1371
__field(int, dirty)
include/trace/events/f2fs.h
1382
__entry->dirty = folio_test_dirty(folio);
include/trace/events/f2fs.h
1393
__entry->dirty,
include/trace/events/f2fs.h
313
__field(int, dirty)
include/trace/events/f2fs.h
319
__entry->dirty = is_sbi_flag_set(F2FS_SB(sb), SBI_IS_DIRTY);
include/trace/events/f2fs.h
325
__entry->dirty ? "dirty" : "not dirty",
include/trace/events/mmflags.h
147
DEF_PAGEFLAG_NAME(dirty), \
include/trace/events/writeback.h
672
__field(unsigned long, dirty)
include/trace/events/writeback.h
692
__entry->dirty = dtc->dirty;
include/trace/events/writeback.h
718
__entry->dirty,
include/uapi/drm/radeon_drm.h
431
unsigned int dirty;
include/uapi/drm/radeon_drm.h
441
unsigned int dirty;
include/uapi/linux/nilfs2_api.h
87
NILFS_SUINFO_FNS(DIRTY, dirty)
include/uapi/linux/nilfs2_ondisk.h
648
NILFS_SEGMENT_USAGE_FNS(DIRTY, dirty)
kernel/kprobes.c
273
kprobe_opcode_t *slot, int dirty)
kernel/kprobes.c
284
if (dirty) {
kernel/trace/ring_buffer.c
799
size_t dirty;
kernel/trace/ring_buffer.c
810
dirty = ring_buffer_nr_dirty_pages(buffer, cpu) + 1;
kernel/trace/ring_buffer.c
812
return (dirty * 100) >= (full * nr_pages);
mm/huge_memory.c
2997
bool anon_exclusive = false, dirty = false;
mm/huge_memory.c
3068
dirty = softleaf_is_migration_dirty(entry);
mm/huge_memory.c
3129
dirty = true;
mm/huge_memory.c
3200
if (dirty)
mm/huge_memory.c
3246
if (dirty)
mm/memory-failure.c
1225
{ sc|dirty, sc|dirty, MF_MSG_DIRTY_SWAPCACHE, me_swapcache_dirty },
mm/memory-failure.c
1226
{ sc|dirty, sc, MF_MSG_CLEAN_SWAPCACHE, me_swapcache_clean },
mm/memory-failure.c
1228
{ mlock|dirty, mlock|dirty, MF_MSG_DIRTY_MLOCKED_LRU, me_pagecache_dirty },
mm/memory-failure.c
1229
{ mlock|dirty, mlock, MF_MSG_CLEAN_MLOCKED_LRU, me_pagecache_clean },
mm/memory-failure.c
1231
{ unevict|dirty, unevict|dirty, MF_MSG_DIRTY_UNEVICTABLE_LRU, me_pagecache_dirty },
mm/memory-failure.c
1232
{ unevict|dirty, unevict, MF_MSG_CLEAN_UNEVICTABLE_LRU, me_pagecache_clean },
mm/memory-failure.c
1234
{ lru|dirty, lru|dirty, MF_MSG_DIRTY_LRU, me_pagecache_dirty },
mm/memory-failure.c
1235
{ lru|dirty, lru, MF_MSG_CLEAN_LRU, me_pagecache_clean },
mm/migrate.c
577
int dirty;
mm/migrate.c
637
dirty = folio_test_dirty(folio);
mm/migrate.c
638
if (dirty) {
mm/migrate.c
696
if (dirty && mapping_can_writeback(mapping)) {
mm/page-writeback.c
1079
if (unlikely(dtc->dirty >= limit))
mm/page-writeback.c
1088
pos_ratio = pos_ratio_polynom(setpoint, dtc->dirty, limit);
mm/page-writeback.c
1295
thresh = max(thresh, dtc->dirty);
mm/page-writeback.c
1335
unsigned long dirty = dtc->dirty;
mm/page-writeback.c
1443
dirty = dtc->wb_dirty;
mm/page-writeback.c
1447
if (dirty < setpoint) {
mm/page-writeback.c
1557
static unsigned long dirty_poll_interval(unsigned long dirty,
mm/page-writeback.c
1560
if (thresh > dirty)
mm/page-writeback.c
1561
return 1UL << (ilog2(thresh - dirty) >> 1);
mm/page-writeback.c
1704
unsigned long dirty, thresh;
mm/page-writeback.c
1707
dirty = dtc->wb_dirty;
mm/page-writeback.c
1710
dirty = dtc->dirty;
mm/page-writeback.c
1714
return dirty_poll_interval(dirty, thresh);
mm/page-writeback.c
1729
unsigned long dirty, thresh, bg_thresh;
mm/page-writeback.c
1733
dirty = dtc->wb_dirty;
mm/page-writeback.c
1737
dirty = dtc->dirty;
mm/page-writeback.c
1741
dtc->freerun = dirty <= dirty_freerun_ceiling(thresh, bg_thresh);
mm/page-writeback.c
1777
((dtc->dirty > dtc->thresh) || strictlimit);
mm/page-writeback.c
2140
if (dtc->dirty > dtc->bg_thresh)
mm/page-writeback.c
442
unsigned long dirty;
mm/page-writeback.c
445
dirty = DIV_ROUND_UP(vm_dirty_bytes, PAGE_SIZE) *
mm/page-writeback.c
448
dirty = vm_dirty_ratio * node_memory / 100;
mm/page-writeback.c
451
dirty += dirty / 4;
mm/page-writeback.c
457
return min_t(unsigned long, dirty, UINT_MAX);
mm/page-writeback.c
833
unsigned long clean = filepages - min(filepages, mdtc->dirty);
mm/page-writeback.c
834
unsigned long global_clean = gdtc->avail - min(gdtc->avail, gdtc->dirty);
mm/page-writeback.c
854
dtc->dirty = global_node_page_state(NR_FILE_DIRTY);
mm/page-writeback.c
856
dtc->dirty += global_node_page_state(NR_WRITEBACK);
mm/page-writeback.c
860
mem_cgroup_wb_stats(dtc->wb, &filepages, &headroom, &dtc->dirty,
mm/page-writeback.c
863
dtc->dirty += writeback;
mm/page-writeback.c
923
if (thresh > dtc->dirty) {
mm/page-writeback.c
925
wb_thresh = max(wb_thresh, (thresh - dtc->dirty) / 100);
mm/page-writeback.c
927
wb_thresh = max(wb_thresh, (thresh - dtc->dirty) / 8);
mm/page-writeback.c
972
unsigned long dirty,
mm/page-writeback.c
978
x = div64_s64(((s64)setpoint - (s64)dirty) << RATELIMIT_CALC_SHIFT,
mm/pgtable-generic.c
113
pmd_t entry, int dirty)
mm/pgtable-generic.c
72
pte_t entry, int dirty)
mm/vmalloc.c
2608
unsigned long free, dirty;
mm/vmalloc.c
2735
vb->dirty = 0;
mm/vmalloc.c
2789
if (vb->free + vb->dirty != VMAP_BBMAP_BITS ||
mm/vmalloc.c
2790
vb->dirty == VMAP_BBMAP_BITS)
mm/vmalloc.c
2800
WRITE_ONCE(vb->dirty, VMAP_BBMAP_BITS);
mm/vmalloc.c
2829
unsigned long dirty = READ_ONCE(vb->dirty);
mm/vmalloc.c
2831
if (free + dirty != VMAP_BBMAP_BITS ||
mm/vmalloc.c
2832
dirty == VMAP_BBMAP_BITS)
mm/vmalloc.c
2940
WRITE_ONCE(vb->dirty, vb->dirty + (1UL << order));
mm/vmalloc.c
2941
if (vb->dirty == VMAP_BBMAP_BITS) {
mm/vmalloc.c
2974
vb->dirty_max && vb->dirty != VMAP_BBMAP_BITS) {
mm/vmscan.c
1106
bool dirty, writeback;
mm/vmscan.c
1150
folio_check_dirty_writeback(folio, &dirty, &writeback);
mm/vmscan.c
1151
if (dirty || writeback)
mm/vmscan.c
1154
if (dirty && !writeback)
mm/vmscan.c
171
unsigned int dirty;
mm/vmscan.c
2067
sc->nr.dirty += stat.nr_dirty;
mm/vmscan.c
3470
int new_gen, bool dirty)
mm/vmscan.c
3477
if (dirty && !folio_test_dirty(folio) &&
mm/vmscan.c
3497
bool dirty;
mm/vmscan.c
3547
walk_update_folio(walk, last, gen, dirty);
mm/vmscan.c
3550
dirty = false;
mm/vmscan.c
3554
dirty = true;
mm/vmscan.c
3560
walk_update_folio(walk, last, gen, dirty);
mm/vmscan.c
3576
bool dirty;
mm/vmscan.c
3638
walk_update_folio(walk, last, gen, dirty);
mm/vmscan.c
3641
dirty = false;
mm/vmscan.c
3645
dirty = true;
mm/vmscan.c
3652
walk_update_folio(walk, last, gen, dirty);
mm/vmscan.c
4204
bool dirty;
mm/vmscan.c
4274
walk_update_folio(walk, last, gen, dirty);
mm/vmscan.c
4277
dirty = false;
mm/vmscan.c
4281
dirty = true;
mm/vmscan.c
4286
walk_update_folio(walk, last, gen, dirty);
mm/vmscan.c
4439
bool dirty, writeback;
mm/vmscan.c
4489
dirty = folio_test_dirty(folio);
mm/vmscan.c
4491
if (type == LRU_GEN_FILE && dirty) {
mm/vmscan.c
4498
if (writeback || (type == LRU_GEN_FILE && dirty)) {
mm/vmscan.c
6114
if (sc->nr.dirty && sc->nr.dirty == sc->nr.congested) {
mm/vmscan.c
955
bool *dirty, bool *writeback)
mm/vmscan.c
968
*dirty = false;
mm/vmscan.c
974
*dirty = folio_test_dirty(folio);
mm/vmscan.c
983
mapping->a_ops->is_dirty_writeback(folio, dirty, writeback);
net/ceph/pagevec.c
13
void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty)
net/ceph/pagevec.c
18
if (dirty)
samples/vfio-mdev/mtty.c
1301
struct iova_bitmap *dirty)
scripts/kconfig/gconf.c
58
static void conf_changed(bool dirty)
scripts/kconfig/gconf.c
60
gtk_widget_set_sensitive(save_btn, dirty);
scripts/kconfig/gconf.c
61
gtk_widget_set_sensitive(save_menu_item, dirty);
scripts/kconfig/qconf.cc
1792
void ConfigMainWindow::conf_changed(bool dirty)
scripts/kconfig/qconf.cc
1795
saveAction->setEnabled(dirty);
sound/hda/common/codec.c
1147
p->dirty = 0;
sound/hda/common/codec.c
1155
p->dirty = 1;
sound/hda/common/codec.c
1222
if (p->dirty)
sound/hda/common/codec.c
670
unsigned char dirty; /* setups should be cleared */
sound/pci/ctxfi/cthw20k1.c
1007
ctl->dirty.bf.spoctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
101
union src_dirty dirty;
sound/pci/ctxfi/cthw20k1.c
1012
ctl->dirty.bf.i2soctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
1024
ctl->dirty.bf.spoctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
1029
ctl->dirty.bf.i2soctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
1063
ctl->dirty.bf.spoctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
1076
ctl->dirty.bf.daoimap = 1;
sound/pci/ctxfi/cthw20k1.c
1085
ctl->dirty.bf.daoimap = 1;
sound/pci/ctxfi/cthw20k1.c
1094
ctl->dirty.bf.daoimap = 1;
sound/pci/ctxfi/cthw20k1.c
1103
if (ctl->dirty.bf.i2sictl || ctl->dirty.bf.i2soctl) {
sound/pci/ctxfi/cthw20k1.c
1105
if ((ctl->dirty.bf.i2sictl & (0x1 << i)))
sound/pci/ctxfi/cthw20k1.c
1106
ctl->dirty.bf.i2sictl &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k1.c
1108
if ((ctl->dirty.bf.i2soctl & (0x1 << i)))
sound/pci/ctxfi/cthw20k1.c
1109
ctl->dirty.bf.i2soctl &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k1.c
1114
if (ctl->dirty.bf.spoctl) {
sound/pci/ctxfi/cthw20k1.c
1116
if ((ctl->dirty.bf.spoctl & (0x1 << i)))
sound/pci/ctxfi/cthw20k1.c
1117
ctl->dirty.bf.spoctl &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k1.c
1122
if (ctl->dirty.bf.spictl) {
sound/pci/ctxfi/cthw20k1.c
1124
if ((ctl->dirty.bf.spictl & (0x1 << i)))
sound/pci/ctxfi/cthw20k1.c
1125
ctl->dirty.bf.spictl &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k1.c
1130
if (ctl->dirty.bf.daoimap) {
sound/pci/ctxfi/cthw20k1.c
1133
ctl->dirty.bf.daoimap = 0;
sound/pci/ctxfi/cthw20k1.c
124
union src_mgr_dirty dirty;
sound/pci/ctxfi/cthw20k1.c
148
union srcimp_mgr_dirty dirty;
sound/pci/ctxfi/cthw20k1.c
181
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
190
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
199
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
208
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
217
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
226
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
235
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
244
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
253
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
262
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
271
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
280
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k1.c
289
ctl->dirty.bf.ccr = 1;
sound/pci/ctxfi/cthw20k1.c
298
ctl->dirty.bf.ca = 1;
sound/pci/ctxfi/cthw20k1.c
307
ctl->dirty.bf.sa = 1;
sound/pci/ctxfi/cthw20k1.c
316
ctl->dirty.bf.la = 1;
sound/pci/ctxfi/cthw20k1.c
325
ctl->dirty.bf.mpr = 1;
sound/pci/ctxfi/cthw20k1.c
331
((struct src_rsc_ctrl_blk *)blk)->dirty.bf.czbfs = (clear ? 1 : 0);
sound/pci/ctxfi/cthw20k1.c
337
((struct src_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff);
sound/pci/ctxfi/cthw20k1.c
343
((struct src_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0);
sound/pci/ctxfi/cthw20k1.c
364
if (ctl->dirty.bf.czbfs) {
sound/pci/ctxfi/cthw20k1.c
375
ctl->dirty.bf.czbfs = 0;
sound/pci/ctxfi/cthw20k1.c
377
if (ctl->dirty.bf.mpr) {
sound/pci/ctxfi/cthw20k1.c
386
ctl->dirty.bf.mpr = 0;
sound/pci/ctxfi/cthw20k1.c
388
if (ctl->dirty.bf.sa) {
sound/pci/ctxfi/cthw20k1.c
390
ctl->dirty.bf.sa = 0;
sound/pci/ctxfi/cthw20k1.c
392
if (ctl->dirty.bf.la) {
sound/pci/ctxfi/cthw20k1.c
394
ctl->dirty.bf.la = 0;
sound/pci/ctxfi/cthw20k1.c
396
if (ctl->dirty.bf.ca) {
sound/pci/ctxfi/cthw20k1.c
398
ctl->dirty.bf.ca = 0;
sound/pci/ctxfi/cthw20k1.c
404
if (ctl->dirty.bf.ccr) {
sound/pci/ctxfi/cthw20k1.c
406
ctl->dirty.bf.ccr = 0;
sound/pci/ctxfi/cthw20k1.c
408
if (ctl->dirty.bf.ctl) {
sound/pci/ctxfi/cthw20k1.c
410
ctl->dirty.bf.ctl = 0;
sound/pci/ctxfi/cthw20k1.c
421
ctl->dirty.bf.ca = 0;
sound/pci/ctxfi/cthw20k1.c
428
return ((struct src_rsc_ctrl_blk *)blk)->dirty.data;
sound/pci/ctxfi/cthw20k1.c
439
((struct src_mgr_ctrl_blk *)blk)->dirty.bf.enbsa = 1;
sound/pci/ctxfi/cthw20k1.c
447
((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32));
sound/pci/ctxfi/cthw20k1.c
454
((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32));
sound/pci/ctxfi/cthw20k1.c
464
if (ctl->dirty.bf.enbsa) {
sound/pci/ctxfi/cthw20k1.c
469
ctl->dirty.bf.enbsa = 0;
sound/pci/ctxfi/cthw20k1.c
472
if ((ctl->dirty.data & (0x1 << i))) {
sound/pci/ctxfi/cthw20k1.c
474
ctl->dirty.data &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k1.c
528
ctl->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k1.c
537
ctl->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k1.c
546
ctl->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k1.c
555
ctl->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k1.c
563
if (ctl->dirty.bf.srcimap) {
sound/pci/ctxfi/cthw20k1.c
566
ctl->dirty.bf.srcimap = 0;
sound/pci/ctxfi/cthw20k1.c
597
union amixer_dirty dirty;
sound/pci/ctxfi/cthw20k1.c
605
ctl->dirty.bf.amoplo = 1;
sound/pci/ctxfi/cthw20k1.c
620
ctl->dirty.bf.amoplo = 1;
sound/pci/ctxfi/cthw20k1.c
629
ctl->dirty.bf.amoplo = 1;
sound/pci/ctxfi/cthw20k1.c
638
ctl->dirty.bf.amophi = 1;
sound/pci/ctxfi/cthw20k1.c
647
ctl->dirty.bf.amophi = 1;
sound/pci/ctxfi/cthw20k1.c
653
((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff);
sound/pci/ctxfi/cthw20k1.c
659
((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0);
sound/pci/ctxfi/cthw20k1.c
667
if (ctl->dirty.bf.amoplo || ctl->dirty.bf.amophi) {
sound/pci/ctxfi/cthw20k1.c
669
ctl->dirty.bf.amoplo = 0;
sound/pci/ctxfi/cthw20k1.c
671
ctl->dirty.bf.amophi = 0;
sound/pci/ctxfi/cthw20k1.c
686
return ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data;
sound/pci/ctxfi/cthw20k1.c
757
union dai_dirty dirty;
sound/pci/ctxfi/cthw20k1.c
772
union dao_dirty dirty;
sound/pci/ctxfi/cthw20k1.c
821
union daio_mgr_dirty dirty;
sound/pci/ctxfi/cthw20k1.c
829
ctl->dirty.bf.srtctl = 1;
sound/pci/ctxfi/cthw20k1.c
838
ctl->dirty.bf.srtctl = 1;
sound/pci/ctxfi/cthw20k1.c
847
ctl->dirty.bf.srtctl = 1;
sound/pci/ctxfi/cthw20k1.c
856
ctl->dirty.bf.srtctl = 1;
sound/pci/ctxfi/cthw20k1.c
865
ctl->dirty.bf.srtctl = 1;
sound/pci/ctxfi/cthw20k1.c
874
ctl->dirty.bf.srtctl = 1;
sound/pci/ctxfi/cthw20k1.c
882
if (ctl->dirty.bf.srtctl) {
sound/pci/ctxfi/cthw20k1.c
890
ctl->dirty.bf.srtctl = 0;
sound/pci/ctxfi/cthw20k1.c
920
((struct dao_ctrl_blk *)blk)->dirty.bf.spos = 1;
sound/pci/ctxfi/cthw20k1.c
928
if (ctl->dirty.bf.spos) {
sound/pci/ctxfi/cthw20k1.c
933
ctl->dirty.bf.spos = 0;
sound/pci/ctxfi/cthw20k1.c
973
ctl->dirty.bf.spictl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
978
ctl->dirty.bf.i2sictl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
990
ctl->dirty.bf.spictl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k1.c
995
ctl->dirty.bf.i2sictl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k2.c
101
union src_dirty dirty;
sound/pci/ctxfi/cthw20k2.c
1019
ctl->dirty.bf.atxctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k2.c
1032
ctl->dirty.bf.daoimap = 1;
sound/pci/ctxfi/cthw20k2.c
1041
ctl->dirty.bf.daoimap = 1;
sound/pci/ctxfi/cthw20k2.c
1048
((struct daio_mgr_ctrl_blk *)blk)->dirty.bf.daoimap = 1;
sound/pci/ctxfi/cthw20k2.c
1059
if ((ctl->dirty.bf.atxctl & (0x1 << i))) {
sound/pci/ctxfi/cthw20k2.c
1062
ctl->dirty.bf.atxctl &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k2.c
1065
if ((ctl->dirty.bf.arxctl & (0x1 << i))) {
sound/pci/ctxfi/cthw20k2.c
1068
ctl->dirty.bf.arxctl &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k2.c
1072
if (ctl->dirty.bf.daoimap) {
sound/pci/ctxfi/cthw20k2.c
1075
ctl->dirty.bf.daoimap = 0;
sound/pci/ctxfi/cthw20k2.c
124
union src_mgr_dirty dirty;
sound/pci/ctxfi/cthw20k2.c
148
union srcimp_mgr_dirty dirty;
sound/pci/ctxfi/cthw20k2.c
181
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
190
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
199
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
208
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
217
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
226
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
235
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
244
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
253
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
262
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
271
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
280
ctl->dirty.bf.ctl = 1;
sound/pci/ctxfi/cthw20k2.c
289
ctl->dirty.bf.ccr = 1;
sound/pci/ctxfi/cthw20k2.c
298
ctl->dirty.bf.ca = 1;
sound/pci/ctxfi/cthw20k2.c
307
ctl->dirty.bf.sa = 1;
sound/pci/ctxfi/cthw20k2.c
316
ctl->dirty.bf.la = 1;
sound/pci/ctxfi/cthw20k2.c
325
ctl->dirty.bf.mpr = 1;
sound/pci/ctxfi/cthw20k2.c
331
((struct src_rsc_ctrl_blk *)blk)->dirty.bf.czbfs = (clear ? 1 : 0);
sound/pci/ctxfi/cthw20k2.c
337
((struct src_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff);
sound/pci/ctxfi/cthw20k2.c
343
((struct src_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0);
sound/pci/ctxfi/cthw20k2.c
364
if (ctl->dirty.bf.czbfs) {
sound/pci/ctxfi/cthw20k2.c
375
ctl->dirty.bf.czbfs = 0;
sound/pci/ctxfi/cthw20k2.c
377
if (ctl->dirty.bf.mpr) {
sound/pci/ctxfi/cthw20k2.c
386
ctl->dirty.bf.mpr = 0;
sound/pci/ctxfi/cthw20k2.c
388
if (ctl->dirty.bf.sa) {
sound/pci/ctxfi/cthw20k2.c
390
ctl->dirty.bf.sa = 0;
sound/pci/ctxfi/cthw20k2.c
392
if (ctl->dirty.bf.la) {
sound/pci/ctxfi/cthw20k2.c
394
ctl->dirty.bf.la = 0;
sound/pci/ctxfi/cthw20k2.c
396
if (ctl->dirty.bf.ca) {
sound/pci/ctxfi/cthw20k2.c
398
ctl->dirty.bf.ca = 0;
sound/pci/ctxfi/cthw20k2.c
404
if (ctl->dirty.bf.ccr) {
sound/pci/ctxfi/cthw20k2.c
406
ctl->dirty.bf.ccr = 0;
sound/pci/ctxfi/cthw20k2.c
408
if (ctl->dirty.bf.ctl) {
sound/pci/ctxfi/cthw20k2.c
410
ctl->dirty.bf.ctl = 0;
sound/pci/ctxfi/cthw20k2.c
421
ctl->dirty.bf.ca = 0;
sound/pci/ctxfi/cthw20k2.c
428
return ((struct src_rsc_ctrl_blk *)blk)->dirty.data;
sound/pci/ctxfi/cthw20k2.c
439
((struct src_mgr_ctrl_blk *)blk)->dirty.bf.enbsa = 1;
sound/pci/ctxfi/cthw20k2.c
447
((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32));
sound/pci/ctxfi/cthw20k2.c
454
((struct src_mgr_ctrl_blk *)blk)->dirty.data |= (0x1 << (idx/32));
sound/pci/ctxfi/cthw20k2.c
464
if (ctl->dirty.bf.enbsa) {
sound/pci/ctxfi/cthw20k2.c
469
ctl->dirty.bf.enbsa = 0;
sound/pci/ctxfi/cthw20k2.c
472
if ((ctl->dirty.data & (0x1 << i))) {
sound/pci/ctxfi/cthw20k2.c
474
ctl->dirty.data &= ~(0x1 << i);
sound/pci/ctxfi/cthw20k2.c
528
ctl->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k2.c
537
ctl->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k2.c
546
ctl->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k2.c
553
((struct srcimp_mgr_ctrl_blk *)blk)->dirty.bf.srcimap = 1;
sound/pci/ctxfi/cthw20k2.c
561
if (ctl->dirty.bf.srcimap) {
sound/pci/ctxfi/cthw20k2.c
564
ctl->dirty.bf.srcimap = 0;
sound/pci/ctxfi/cthw20k2.c
596
union amixer_dirty dirty;
sound/pci/ctxfi/cthw20k2.c
604
ctl->dirty.bf.amoplo = 1;
sound/pci/ctxfi/cthw20k2.c
613
ctl->dirty.bf.amoplo = 1;
sound/pci/ctxfi/cthw20k2.c
622
ctl->dirty.bf.amoplo = 1;
sound/pci/ctxfi/cthw20k2.c
631
ctl->dirty.bf.amoplo = 1;
sound/pci/ctxfi/cthw20k2.c
640
ctl->dirty.bf.amophi = 1;
sound/pci/ctxfi/cthw20k2.c
649
ctl->dirty.bf.amophi = 1;
sound/pci/ctxfi/cthw20k2.c
655
((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = (flags & 0xffff);
sound/pci/ctxfi/cthw20k2.c
661
((struct amixer_rsc_ctrl_blk *)blk)->dirty.data = ~(0x0);
sound/pci/ctxfi/cthw20k2.c
669
if (ctl->dirty.bf.amoplo || ctl->dirty.bf.amophi) {
sound/pci/ctxfi/cthw20k2.c
671
ctl->dirty.bf.amoplo = 0;
sound/pci/ctxfi/cthw20k2.c
673
ctl->dirty.bf.amophi = 0;
sound/pci/ctxfi/cthw20k2.c
688
return ((struct amixer_rsc_ctrl_blk *)blk)->dirty.data;
sound/pci/ctxfi/cthw20k2.c
748
union dai_dirty dirty;
sound/pci/ctxfi/cthw20k2.c
789
union dao_dirty dirty;
sound/pci/ctxfi/cthw20k2.c
811
union daio_mgr_dirty dirty;
sound/pci/ctxfi/cthw20k2.c
819
ctl->dirty.bf.srt = 1;
sound/pci/ctxfi/cthw20k2.c
828
ctl->dirty.bf.srt = 1;
sound/pci/ctxfi/cthw20k2.c
837
ctl->dirty.bf.srt = 1;
sound/pci/ctxfi/cthw20k2.c
846
ctl->dirty.bf.srt = 1;
sound/pci/ctxfi/cthw20k2.c
855
ctl->dirty.bf.srt = 1;
sound/pci/ctxfi/cthw20k2.c
864
ctl->dirty.bf.srt = 1;
sound/pci/ctxfi/cthw20k2.c
872
if (ctl->dirty.bf.srt) {
sound/pci/ctxfi/cthw20k2.c
874
ctl->dirty.bf.srt = 0;
sound/pci/ctxfi/cthw20k2.c
904
((struct dao_ctrl_blk *)blk)->dirty.bf.atxcsl = 1;
sound/pci/ctxfi/cthw20k2.c
912
if (ctl->dirty.bf.atxcsl) {
sound/pci/ctxfi/cthw20k2.c
918
ctl->dirty.bf.atxcsl = 0;
sound/pci/ctxfi/cthw20k2.c
956
ctl->dirty.bf.arxctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k2.c
966
ctl->dirty.bf.arxctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k2.c
975
ctl->dirty.bf.atxctl |= (0x1 << idx);
sound/pci/ctxfi/cthw20k2.c
984
ctl->dirty.bf.atxctl |= (0x1 << idx);
sound/pci/ctxfi/ctsrc.c
178
unsigned int dirty = 0;
sound/pci/ctxfi/ctsrc.c
184
dirty = hw->src_get_dirty(src->rsc.ctrl_blk) & conj_mask;
sound/pci/ctxfi/ctsrc.c
195
hw->src_set_dirty(src->rsc.ctrl_blk, dirty);
sound/pci/ctxfi/ctsrc.c
297
unsigned int dirty;
sound/pci/ctxfi/ctsrc.c
321
dirty = hw->src_get_dirty(src->rsc.ctrl_blk);
sound/pci/ctxfi/ctsrc.c
324
hw->src_set_dirty(src->rsc.ctrl_blk, dirty);
sound/pci/emu10k1/emu10k1_callback.c
444
hw->voices[ch].dirty = 1;
sound/pci/emu10k1/emupcm.c
312
emu->voices[voice].dirty = 1;
sound/pci/emu10k1/emuproc.c
442
voice->dirty,
sound/pci/emu10k1/voice.c
70
if (pvoice->dirty)
sound/pci/emu10k1/voice.c
73
pvoice->use = pvoice->dirty = pvoice->last = 0;
sound/soc/soc-dapm.c
217
return !list_empty(&w->dirty);
sound/soc/soc-dapm.c
2283
list_for_each_entry(w, &card->dapm_dirty, dirty) {
sound/soc/soc-dapm.c
229
list_add_tail(&w->dirty, &w->dapm->card->dapm_dirty);
sound/soc/soc-dapm.c
2294
list_del_init(&w->dirty);
sound/soc/soc-dapm.c
2859
list_del(&w->dirty);
sound/soc/soc-dapm.c
3876
INIT_LIST_HEAD(&w->dirty);
tools/testing/selftests/iommu/iommufd_utils.h
432
__u64 *bitmap, __u64 *dirty)
tools/testing/selftests/iommu/iommufd_utils.h
438
.dirty = {
tools/testing/selftests/iommu/iommufd_utils.h
450
if (dirty)
tools/testing/selftests/iommu/iommufd_utils.h
451
*dirty = cmd.dirty.out_nr_dirty;
tools/testing/selftests/kvm/include/x86/kvm_util_arch.h
20
uint64_t dirty;
tools/testing/selftests/kvm/include/x86/processor.h
1474
#define PTE_DIRTY_MASK(mmu) ((mmu)->arch.pte_masks.dirty)
tools/testing/selftests/kvm/lib/x86/processor.c
187
.dirty = BIT_ULL(6),
tools/testing/selftests/kvm/lib/x86/vmx.c
62
.dirty = BIT_ULL(9),