DIRTY
if (__riscv_v_vstate_check(regs->status, DIRTY)) {
regs->status = __riscv_v_vstate_or(regs->status, DIRTY);
if (__riscv_v_vstate_check(task_pt_regs(current)->status, DIRTY)) {
if (depth == 0 && __riscv_v_vstate_check(regs->status, DIRTY))
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
DIRTY, UPDATE_IN_PROGRESS) == CLEAN) {
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
atomic_set_release(&apic->vcpu->kvm->arch.apic_map_dirty, DIRTY);
(background || (TREE_EMPTY(d, CLEAN) && TREE_EMPTY(d, DIRTY))))
TREE_EMPTY(d, DIRTY) && TREE_EMPTY(d, FAILING))
results &= !DIRTY(regvals.clock_params[i]);
results &= !DIRTY(regvals.regs[i]);
return (!DIRTY(cmode) && !DIRTY(xres) && !DIRTY(yres)
&& !DIRTY(vxres) && !DIRTY(vyres));
PRINT_ONE_FLAG(state, dest, cur, DIRTY);
CHECK_FLAG_VALUE(DIRTY);
p->dirty_bitmap = dirty_i->dirty_segmap[DIRTY];
p->max_search = dirty_i->nr_dirty[DIRTY];
__remove_dirty_segment(sbi, new_segno, DIRTY);
__locate_dirty_segment(sbi, segno, DIRTY);
if (dirty_type == DIRTY) {
if (unlikely(t >= DIRTY)) {
if (dirty_type == DIRTY) {
__remove_dirty_segment(sbi, segno, DIRTY);
__locate_dirty_segment(sbi, segno, DIRTY);
__remove_dirty_segment(sbi, segno, DIRTY);
for_each_set_bit(segno, dirty_i->dirty_segmap[DIRTY], MAIN_SEGS(sbi)) {
__remove_dirty_segment(sbi, segno, DIRTY);
for_each_set_bit(segno, dirty_i->dirty_segmap[DIRTY], MAIN_SEGS(sbi)) {
for_each_set_bit(segno, dirty_i->dirty_segmap[DIRTY], MAIN_SEGS(sbi)) {
NILFS_SUINFO_FNS(DIRTY, dirty)
NILFS_SEGMENT_USAGE_FNS(DIRTY, dirty)