Symbol: flush
arch/arm/boot/compressed/misc.c
100
flush();
arch/arm64/include/asm/fpsimd.h
201
extern void sve_alloc(struct task_struct *task, bool flush);
arch/arm64/include/asm/fpsimd.h
317
static inline void sve_alloc(struct task_struct *task, bool flush) { }
arch/arm64/include/asm/fpsimd.h
405
extern void sme_alloc(struct task_struct *task, bool flush);
arch/arm64/include/asm/fpsimd.h
440
static inline void sme_alloc(struct task_struct *task, bool flush) { }
arch/arm64/kernel/fpsimd.c
1176
void sme_alloc(struct task_struct *task, bool flush)
arch/arm64/kernel/fpsimd.c
1179
if (flush)
arch/arm64/kernel/fpsimd.c
749
void sve_alloc(struct task_struct *task, bool flush)
arch/arm64/kernel/fpsimd.c
752
if (flush)
arch/microblaze/kernel/cpu/cache.c
480
wdc.flush);
arch/microblaze/kernel/cpu/cache.c
500
CACHE_RANGE_LOOP_2(start, end, cpuinfo.dcache_line_length, wdc.flush);
arch/mips/include/asm/octeon/cvmx-pci-defs.h
570
uint32_t flush:1;
arch/mips/include/asm/octeon/cvmx-pci-defs.h
580
uint32_t flush:1;
arch/mips/mm/tlbex.c
725
unsigned int flush)
arch/mips/mm/tlbex.c
734
if (cpu_has_ftlb && flush) {
arch/mips/pci/pci-octeon.c
522
cfg22.s.flush = 1;
arch/powerpc/boot/decompress.c
132
ret = __decompress(inbuf, input_size, NULL, flush, outbuf,
arch/powerpc/include/asm/iommu.h
71
void (*flush)(struct iommu_table *tbl);
arch/powerpc/include/asm/mmu.h
304
unsigned long dw1, bool flush);
arch/powerpc/kernel/iommu.c
1057
if (tbl->it_ops->flush)
arch/powerpc/kernel/iommu.c
1058
tbl->it_ops->flush(tbl);
arch/powerpc/kernel/iommu.c
381
if (tbl->it_ops->flush)
arch/powerpc/kernel/iommu.c
382
tbl->it_ops->flush(tbl);
arch/powerpc/kernel/iommu.c
468
if (tbl->it_ops->flush)
arch/powerpc/kernel/iommu.c
469
tbl->it_ops->flush(tbl);
arch/powerpc/kernel/iommu.c
579
if (tbl->it_ops->flush)
arch/powerpc/kernel/iommu.c
580
tbl->it_ops->flush(tbl);
arch/powerpc/kernel/iommu.c
643
if (tbl->it_ops->flush)
arch/powerpc/kernel/iommu.c
644
tbl->it_ops->flush(tbl);
arch/powerpc/mm/book3s64/pgtable.c
354
unsigned long dw1, bool flush)
arch/powerpc/mm/book3s64/pgtable.c
381
} else if (flush) {
arch/powerpc/platforms/cell/spufs/file.c
1719
.flush = spufs_mfc_flush,
arch/powerpc/platforms/pseries/papr_scm.c
1269
ndr_desc.flush = papr_scm_pmem_flush;
arch/powerpc/sysdev/dart_iommu.c
327
.flush = dart_flush,
arch/x86/hyperv/mmu.c
128
&flush->processor_mask);
arch/x86/hyperv/mmu.c
132
if (!flush->processor_mask) {
arch/x86/hyperv/mmu.c
142
max_gvas = (PAGE_SIZE - sizeof(*flush)) / sizeof(flush->gva_list[0]);
arch/x86/hyperv/mmu.c
145
flush->flags |= HV_FLUSH_NON_GLOBAL_MAPPINGS_ONLY;
arch/x86/hyperv/mmu.c
147
flush, NULL);
arch/x86/hyperv/mmu.c
151
flush, NULL);
arch/x86/hyperv/mmu.c
153
gva_n = fill_gva_list(flush->gva_list, 0,
arch/x86/hyperv/mmu.c
156
gva_n, 0, flush, NULL);
arch/x86/hyperv/mmu.c
176
struct hv_tlb_flush_ex *flush;
arch/x86/hyperv/mmu.c
182
flush = *this_cpu_ptr(hyperv_pcpu_input_arg);
arch/x86/hyperv/mmu.c
189
flush->address_space = virt_to_phys(info->mm->pgd);
arch/x86/hyperv/mmu.c
190
flush->address_space &= CR3_ADDR_MASK;
arch/x86/hyperv/mmu.c
191
flush->flags = 0;
arch/x86/hyperv/mmu.c
193
flush->address_space = 0;
arch/x86/hyperv/mmu.c
194
flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES;
arch/x86/hyperv/mmu.c
197
flush->hv_vp_set.valid_bank_mask = 0;
arch/x86/hyperv/mmu.c
199
flush->hv_vp_set.format = HV_GENERIC_SET_SPARSE_4K;
arch/x86/hyperv/mmu.c
200
nr_bank = cpumask_to_vpset_skip(&flush->hv_vp_set, cpus,
arch/x86/hyperv/mmu.c
214
(PAGE_SIZE - sizeof(*flush) - nr_bank *
arch/x86/hyperv/mmu.c
215
sizeof(flush->hv_vp_set.bank_contents[0])) /
arch/x86/hyperv/mmu.c
216
sizeof(flush->gva_list[0]);
arch/x86/hyperv/mmu.c
219
flush->flags |= HV_FLUSH_NON_GLOBAL_MAPPINGS_ONLY;
arch/x86/hyperv/mmu.c
222
0, nr_bank, flush, NULL);
arch/x86/hyperv/mmu.c
227
0, nr_bank, flush, NULL);
arch/x86/hyperv/mmu.c
229
gva_n = fill_gva_list(flush->gva_list, nr_bank,
arch/x86/hyperv/mmu.c
233
gva_n, nr_bank, flush, NULL);
arch/x86/hyperv/mmu.c
63
struct hv_tlb_flush *flush;
arch/x86/hyperv/mmu.c
75
flush = *this_cpu_ptr(hyperv_pcpu_input_arg);
arch/x86/hyperv/mmu.c
77
if (unlikely(!flush)) {
arch/x86/hyperv/mmu.c
87
flush->address_space = virt_to_phys(info->mm->pgd);
arch/x86/hyperv/mmu.c
88
flush->address_space &= CR3_ADDR_MASK;
arch/x86/hyperv/mmu.c
89
flush->flags = 0;
arch/x86/hyperv/mmu.c
91
flush->address_space = 0;
arch/x86/hyperv/mmu.c
92
flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES;
arch/x86/hyperv/mmu.c
95
flush->processor_mask = 0;
arch/x86/hyperv/mmu.c
97
flush->flags |= HV_FLUSH_ALL_PROCESSORS;
arch/x86/hyperv/nested.c
101
flush = *this_cpu_ptr(hyperv_pcpu_input_arg);
arch/x86/hyperv/nested.c
103
if (unlikely(!flush)) {
arch/x86/hyperv/nested.c
108
flush->address_space = as;
arch/x86/hyperv/nested.c
109
flush->flags = 0;
arch/x86/hyperv/nested.c
111
gpa_n = fill_flush_list_func(flush, data);
arch/x86/hyperv/nested.c
118
gpa_n, 0, flush, NULL);
arch/x86/hyperv/nested.c
23
struct hv_guest_mapping_flush *flush;
arch/x86/hyperv/nested.c
33
flush = *this_cpu_ptr(hyperv_pcpu_input_arg);
arch/x86/hyperv/nested.c
35
if (unlikely(!flush)) {
arch/x86/hyperv/nested.c
40
flush->address_space = as;
arch/x86/hyperv/nested.c
41
flush->flags = 0;
arch/x86/hyperv/nested.c
44
flush, NULL);
arch/x86/hyperv/nested.c
57
struct hv_guest_mapping_flush_list *flush,
arch/x86/hyperv/nested.c
74
flush->gpa_list[gpa_n].page.additional_pages = additional_pages;
arch/x86/hyperv/nested.c
75
flush->gpa_list[gpa_n].page.largepage = false;
arch/x86/hyperv/nested.c
76
flush->gpa_list[gpa_n].page.basepfn = cur;
arch/x86/hyperv/nested.c
90
struct hv_guest_mapping_flush_list *flush;
arch/x86/include/asm/mshyperv.h
177
struct hv_guest_mapping_flush_list *flush,
arch/x86/include/asm/mshyperv.h
33
struct hv_guest_mapping_flush_list *flush,
arch/x86/include/asm/paravirt.h
507
PVOP_VCALL0(pv_ops, mmu.lazy_mode.flush);
arch/x86/include/asm/paravirt_types.h
27
void (*flush)(void);
arch/x86/kernel/paravirt.c
211
.flush = paravirt_nop,
arch/x86/kvm/hyperv.c
2014
struct hv_tlb_flush flush;
arch/x86/kvm/hyperv.c
2052
flush.address_space = hc->ingpa;
arch/x86/kvm/hyperv.c
2053
flush.flags = hc->outgpa;
arch/x86/kvm/hyperv.c
2054
flush.processor_mask = sse128_lo(hc->xmm[0]);
arch/x86/kvm/hyperv.c
2058
&flush, sizeof(flush))))
arch/x86/kvm/hyperv.c
2060
hc->data_offset = sizeof(flush);
arch/x86/kvm/hyperv.c
2063
trace_kvm_hv_flush_tlb(flush.processor_mask,
arch/x86/kvm/hyperv.c
2064
flush.address_space, flush.flags,
arch/x86/kvm/hyperv.c
2068
sparse_banks[0] = flush.processor_mask;
arch/x86/kvm/hyperv.c
2077
all_cpus = (flush.flags & HV_FLUSH_ALL_PROCESSORS) ||
arch/x86/kvm/hyperv.c
2078
flush.processor_mask == 0;
arch/x86/kvm/kvm_onhyperv.c
18
static int kvm_fill_hv_flush_list_func(struct hv_guest_mapping_flush_list *flush,
arch/x86/kvm/kvm_onhyperv.c
23
return hyperv_fill_flush_guest_mapping_list(flush, range->start_gfn,
arch/x86/kvm/mmu/mmu.c
1290
static void drop_large_spte(struct kvm *kvm, u64 *sptep, bool flush)
arch/x86/kvm/mmu/mmu.c
1299
if (flush)
arch/x86/kvm/mmu/mmu.c
1336
bool flush = false;
arch/x86/kvm/mmu/mmu.c
1339
flush |= spte_write_protect(sptep, pt_protect);
arch/x86/kvm/mmu/mmu.c
1341
return flush;
arch/x86/kvm/mmu/mmu.c
1364
bool flush = false;
arch/x86/kvm/mmu/mmu.c
1368
flush |= test_and_clear_bit(PT_WRITABLE_SHIFT,
arch/x86/kvm/mmu/mmu.c
1371
flush |= spte_clear_dirty(sptep);
arch/x86/kvm/mmu/mmu.c
1374
return flush;
arch/x86/kvm/mmu/mmu.c
1591
bool flush)
arch/x86/kvm/mmu/mmu.c
1600
flush |= fn(kvm, iterator.rmap, slot);
arch/x86/kvm/mmu/mmu.c
1606
if (flush && flush_on_yield) {
arch/x86/kvm/mmu/mmu.c
1609
flush = false;
arch/x86/kvm/mmu/mmu.c
1615
return flush;
arch/x86/kvm/mmu/mmu.c
1640
bool flush)
arch/x86/kvm/mmu/mmu.c
1644
start, end - 1, can_yield, true, flush);
arch/x86/kvm/mmu/mmu.c
1649
bool flush = false;
arch/x86/kvm/mmu/mmu.c
1663
flush = __kvm_rmap_zap_gfn_range(kvm, range->slot,
arch/x86/kvm/mmu/mmu.c
1665
range->may_block, flush);
arch/x86/kvm/mmu/mmu.c
1668
flush = kvm_tdp_mmu_unmap_gfn_range(kvm, range, flush);
arch/x86/kvm/mmu/mmu.c
1674
return flush;
arch/x86/kvm/mmu/mmu.c
2065
int flush = 0;
arch/x86/kvm/mmu/mmu.c
2076
flush |= ret;
arch/x86/kvm/mmu/mmu.c
2088
return flush;
arch/x86/kvm/mmu/mmu.c
2205
bool flush = false;
arch/x86/kvm/mmu/mmu.c
2215
flush = false;
arch/x86/kvm/mmu/mmu.c
2220
flush |= kvm_sync_page(vcpu, sp, &invalid_list) > 0;
arch/x86/kvm/mmu/mmu.c
2224
kvm_mmu_remote_flush_or_zap(vcpu->kvm, &invalid_list, flush);
arch/x86/kvm/mmu/mmu.c
2231
flush = false;
arch/x86/kvm/mmu/mmu.c
2235
kvm_mmu_remote_flush_or_zap(vcpu->kvm, &invalid_list, flush);
arch/x86/kvm/mmu/mmu.c
2541
struct kvm_mmu_page *sp, bool flush)
arch/x86/kvm/mmu/mmu.c
2553
drop_large_spte(kvm, sptep, flush);
arch/x86/kvm/mmu/mmu.c
3038
bool flush = false;
arch/x86/kvm/mmu/mmu.c
3062
flush = true;
arch/x86/kvm/mmu/mmu.c
3066
flush = true;
arch/x86/kvm/mmu/mmu.c
3074
if (flush)
arch/x86/kvm/mmu/mmu.c
3085
flush |= mmu_spte_update(sptep, spte);
arch/x86/kvm/mmu/mmu.c
3092
if (flush)
arch/x86/kvm/mmu/mmu.c
6252
bool flush = false;
arch/x86/kvm/mmu/mmu.c
6290
flush = true;
arch/x86/kvm/mmu/mmu.c
6294
kvm_mmu_remote_flush_or_zap(vcpu->kvm, &invalid_list, flush);
arch/x86/kvm/mmu/mmu.c
6893
bool flush = false;
arch/x86/kvm/mmu/mmu.c
6898
return flush;
arch/x86/kvm/mmu/mmu.c
6910
flush = __kvm_rmap_zap_gfn_range(kvm, memslot, start,
arch/x86/kvm/mmu/mmu.c
6911
end, true, flush);
arch/x86/kvm/mmu/mmu.c
6915
return flush;
arch/x86/kvm/mmu/mmu.c
6924
bool flush;
arch/x86/kvm/mmu/mmu.c
6935
flush = kvm_rmap_zap_gfn_range(kvm, gfn_start, gfn_end);
arch/x86/kvm/mmu/mmu.c
6938
flush = kvm_tdp_mmu_zap_leafs(kvm, gfn_start, gfn_end, flush);
arch/x86/kvm/mmu/mmu.c
6940
if (flush)
arch/x86/kvm/mmu/mmu.c
7062
bool flush = false;
arch/x86/kvm/mmu/mmu.c
7088
flush |= !is_last_spte(*sptep, sp->role.level);
arch/x86/kvm/mmu/mmu.c
7097
__link_shadow_page(kvm, cache, huge_sptep, sp, flush);
arch/x86/kvm/mmu/mmu.c
7379
bool flush)
arch/x86/kvm/mmu/mmu.c
7401
kvm_mmu_remote_flush_or_zap(kvm, &invalid_list, flush);
arch/x86/kvm/mmu/mmu.c
7402
flush = false;
arch/x86/kvm/mmu/mmu.c
7408
kvm_mmu_remote_flush_or_zap(kvm, &invalid_list, flush);
arch/x86/kvm/mmu/mmu.c
7421
bool flush;
arch/x86/kvm/mmu/mmu.c
7424
flush = kvm_unmap_gfn_range(kvm, &range);
arch/x86/kvm/mmu/mmu.c
7425
kvm_mmu_zap_memslot_pages_and_flush(kvm, slot, flush);
arch/x86/kvm/mmu/mmu.c
7742
bool flush = false;
arch/x86/kvm/mmu/mmu.c
7797
flush |= kvm_tdp_mmu_zap_possible_nx_huge_page(kvm, sp);
arch/x86/kvm/mmu/mmu.c
7806
kvm_mmu_remote_flush_or_zap(kvm, &invalid_list, flush);
arch/x86/kvm/mmu/mmu.c
7814
flush = false;
arch/x86/kvm/mmu/mmu.c
7818
kvm_mmu_remote_flush_or_zap(kvm, &invalid_list, flush);
arch/x86/kvm/mmu/tdp_mmu.c
1009
flush = true;
arch/x86/kvm/mmu/tdp_mmu.c
1018
return flush;
arch/x86/kvm/mmu/tdp_mmu.c
1026
bool kvm_tdp_mmu_zap_leafs(struct kvm *kvm, gfn_t start, gfn_t end, bool flush)
arch/x86/kvm/mmu/tdp_mmu.c
1032
flush = tdp_mmu_zap_leafs(kvm, root, start, end, true, flush);
arch/x86/kvm/mmu/tdp_mmu.c
1034
return flush;
arch/x86/kvm/mmu/tdp_mmu.c
1354
bool flush)
arch/x86/kvm/mmu/tdp_mmu.c
1362
flush = tdp_mmu_zap_leafs(kvm, root, range->start, range->end,
arch/x86/kvm/mmu/tdp_mmu.c
1363
range->may_block, flush);
arch/x86/kvm/mmu/tdp_mmu.c
1365
return flush;
arch/x86/kvm/mmu/tdp_mmu.c
1804
bool flush = false;
arch/x86/kvm/mmu/tdp_mmu.c
1815
if (tdp_mmu_iter_cond_resched(kvm, &iter, flush, true)) {
arch/x86/kvm/mmu/tdp_mmu.c
1816
flush = false;
arch/x86/kvm/mmu/tdp_mmu.c
1855
flush = true;
arch/x86/kvm/mmu/tdp_mmu.c
1858
if (flush)
arch/x86/kvm/mmu/tdp_mmu.c
817
bool flush, bool shared)
arch/x86/kvm/mmu/tdp_mmu.c
824
if (flush)
arch/x86/kvm/mmu/tdp_mmu.c
981
gfn_t start, gfn_t end, bool can_yield, bool flush)
arch/x86/kvm/mmu/tdp_mmu.c
993
tdp_mmu_iter_cond_resched(kvm, &iter, flush, false)) {
arch/x86/kvm/mmu/tdp_mmu.c
994
flush = false;
arch/x86/kvm/mmu/tdp_mmu.h
66
bool kvm_tdp_mmu_zap_leafs(struct kvm *kvm, gfn_t start, gfn_t end, bool flush);
arch/x86/kvm/mmu/tdp_mmu.h
77
bool flush);
arch/x86/xen/mmu_pv.c
2230
pv_ops.mmu.lazy_mode.flush = xen_flush_lazy_mmu;
block/blk-flush.c
105
return 1 << ffz(rq->flush.seq);
block/blk-flush.c
121
rq->end_io = rq->flush.saved_end_io;
block/blk-flush.c
156
BUG_ON(rq->flush.seq & seq);
block/blk-flush.c
157
rq->flush.seq |= seq;
block/blk-flush.c
373
rq->flush.seq = 0;
block/blk-flush.c
375
rq->flush.saved_end_io = rq->end_io; /* Usually NULL */
block/blk-flush.c
445
rq->flush.seq |= REQ_FSEQ_PREFLUSH;
crypto/deflate.c
84
int flush = Z_FINISH;
crypto/deflate.c
93
flush = Z_NO_FLUSH;
crypto/deflate.c
98
ret = zlib_deflate(stream, flush);
drivers/accel/habanalabs/common/habanalabs.h
2944
void (*flush)(struct hl_ctx *ctx);
drivers/accel/habanalabs/common/mmu/mmu.c
263
mmu_funcs->flush(ctx);
drivers/accel/habanalabs/common/mmu/mmu.c
351
mmu_funcs->flush(ctx);
drivers/accel/habanalabs/common/mmu/mmu.c
367
mmu_funcs->flush(ctx);
drivers/accel/habanalabs/common/mmu/mmu_v1.c
546
mmu->flush = hl_mmu_dr_flush;
drivers/accel/habanalabs/common/mmu/mmu_v2.c
334
mmu->flush = hl_mmu_dr_flush;
drivers/accel/habanalabs/common/mmu/mmu_v2_hr.c
391
mmu->flush = hl_mmu_hr_flush;
drivers/accessibility/speakup/speakup_acntpc.c
122
.flush = synth_flush,
drivers/accessibility/speakup/speakup_acntpc.c
198
synth->flush(synth);
drivers/accessibility/speakup/speakup_acntsa.c
109
.flush = spk_synth_flush,
drivers/accessibility/speakup/speakup_apollo.c
121
.flush = spk_synth_flush,
drivers/accessibility/speakup/speakup_apollo.c
166
synth->flush(synth);
drivers/accessibility/speakup/speakup_audptr.c
112
.flush = synth_flush,
drivers/accessibility/speakup/speakup_bns.c
105
.flush = spk_synth_flush,
drivers/accessibility/speakup/speakup_decext.c
136
.flush = synth_flush,
drivers/accessibility/speakup/speakup_decext.c
177
synth->flush(synth);
drivers/accessibility/speakup/speakup_decpc.c
232
.flush = synth_flush,
drivers/accessibility/speakup/speakup_decpc.c
397
synth->flush(synth);
drivers/accessibility/speakup/speakup_dectlk.c
145
.flush = synth_flush,
drivers/accessibility/speakup/speakup_dectlk.c
191
wake_up_interruptible(&flush);
drivers/accessibility/speakup/speakup_dectlk.c
239
prepare_to_wait(&flush, &wait, TASK_INTERRUPTIBLE);
drivers/accessibility/speakup/speakup_dectlk.c
244
finish_wait(&flush, &wait);
drivers/accessibility/speakup/speakup_dectlk.c
252
synth->flush(synth);
drivers/accessibility/speakup/speakup_dectlk.c
41
static DECLARE_WAIT_QUEUE_HEAD(flush);
drivers/accessibility/speakup/speakup_dtlk.c
138
.flush = synth_flush,
drivers/accessibility/speakup/speakup_dtlk.c
212
synth->flush(synth);
drivers/accessibility/speakup/speakup_dummy.c
125
.flush = spk_synth_flush,
drivers/accessibility/speakup/speakup_keypc.c
113
.flush = synth_flush,
drivers/accessibility/speakup/speakup_keypc.c
199
synth->flush(synth);
drivers/accessibility/speakup/speakup_ltlk.c
121
.flush = spk_synth_flush,
drivers/accessibility/speakup/speakup_soft.c
146
.flush = NULL,
drivers/accessibility/speakup/speakup_spkout.c
112
.flush = synth_flush,
drivers/accessibility/speakup/speakup_txprt.c
110
.flush = spk_synth_flush,
drivers/accessibility/speakup/spk_types.h
196
void (*flush)(struct spk_synth *synth);
drivers/accessibility/speakup/synth.c
81
synth->flush(synth);
drivers/acpi/nfit/core.c
1089
struct acpi_nfit_flush_address *flush;
drivers/acpi/nfit/core.c
1092
if (nfit_flush->flush->device_handle != device_handle)
drivers/acpi/nfit/core.c
1095
flush = nfit_flush->flush;
drivers/acpi/nfit/core.c
1097
flush->hint_count,
drivers/acpi/nfit/core.c
1102
for (i = 0; i < flush->hint_count; i++) {
drivers/acpi/nfit/core.c
1105
res->start = flush->hint_address[i];
drivers/acpi/nfit/core.c
2004
struct acpi_nfit_flush_address *flush;
drivers/acpi/nfit/core.c
2058
flush = nfit_mem->nfit_flush ? nfit_mem->nfit_flush->flush
drivers/acpi/nfit/core.c
2062
flags, cmd_mask, flush ? flush->hint_count : 0,
drivers/acpi/nfit/core.c
906
static size_t sizeof_flush(struct acpi_nfit_flush_address *flush)
drivers/acpi/nfit/core.c
908
if (flush->header.length < sizeof(*flush))
drivers/acpi/nfit/core.c
910
return struct_size(flush, hint_address, flush->hint_count);
drivers/acpi/nfit/core.c
915
struct acpi_nfit_flush_address *flush)
drivers/acpi/nfit/core.c
920
if (!sizeof_flush(flush))
drivers/acpi/nfit/core.c
924
if (sizeof_flush(nfit_flush->flush) != sizeof_flush(flush))
drivers/acpi/nfit/core.c
927
if (memcmp(nfit_flush->flush, flush,
drivers/acpi/nfit/core.c
928
sizeof_flush(flush)) == 0) {
drivers/acpi/nfit/core.c
935
+ sizeof_flush(flush), GFP_KERNEL);
drivers/acpi/nfit/core.c
939
memcpy(nfit_flush->flush, flush, sizeof_flush(flush));
drivers/acpi/nfit/core.c
942
flush->device_handle, flush->hint_count);
drivers/acpi/nfit/nfit.h
189
struct acpi_nfit_flush_address flush[];
drivers/android/binder.c
6993
.flush = binder_flush,
drivers/block/aoe/aoedev.c
407
return flush(str, cnt, NOT_EXITING);
drivers/block/aoe/aoedev.c
536
flush(NULL, 0, EXITING);
drivers/bluetooth/bfusb.c
670
hdev->flush = bfusb_flush;
drivers/bluetooth/bluecard_cs.c
711
hdev->flush = bluecard_hci_flush;
drivers/bluetooth/bpa10x.c
398
hdev->flush = bpa10x_flush;
drivers/bluetooth/bt3c_cs.c
567
hdev->flush = bt3c_hci_flush;
drivers/bluetooth/btmrvl_main.c
676
hdev->flush = btmrvl_flush;
drivers/bluetooth/btmtksdio.c
1380
hdev->flush = btmtksdio_flush;
drivers/bluetooth/btmtkuart.c
869
hdev->flush = btmtkuart_flush;
drivers/bluetooth/btnxpuart.c
1887
hdev->flush = btnxpuart_flush;
drivers/bluetooth/btrsi.c
139
hdev->flush = rsi_hci_flush;
drivers/bluetooth/btsdio.c
326
hdev->flush = btsdio_flush;
drivers/bluetooth/btusb.c
4169
hdev->flush = btusb_flush;
drivers/bluetooth/dtl1_cs.c
466
hdev->flush = dtl1_hci_flush;
drivers/bluetooth/hci_ag6xx.c
306
.flush = ag6xx_flush,
drivers/bluetooth/hci_aml.c
674
.flush = aml_flush,
drivers/bluetooth/hci_ath.c
253
.flush = ath_flush,
drivers/bluetooth/hci_bcm.c
1301
.flush = bcm_flush,
drivers/bluetooth/hci_bcsp.c
770
.flush = bcsp_flush
drivers/bluetooth/hci_h4.c
138
.flush = h4_flush,
drivers/bluetooth/hci_h5.c
849
.flush = h5_flush,
drivers/bluetooth/hci_intel.c
1050
.flush = intel_flush,
drivers/bluetooth/hci_ldisc.c
245
hu->proto->flush(hu);
drivers/bluetooth/hci_ldisc.c
258
hdev->flush = hci_uart_flush;
drivers/bluetooth/hci_ldisc.c
269
hdev->flush = NULL;
drivers/bluetooth/hci_ldisc.c
664
hdev->flush = hci_uart_flush;
drivers/bluetooth/hci_ll.c
809
.flush = ll_flush,
drivers/bluetooth/hci_mrvl.c
441
.flush = mrvl_flush,
drivers/bluetooth/hci_mrvl.c
455
.flush = mrvl_flush,
drivers/bluetooth/hci_nokia.c
673
.flush = nokia_flush,
drivers/bluetooth/hci_qca.c
2092
.flush = qca_flush,
drivers/bluetooth/hci_serdev.c
108
hu->proto->flush(hu);
drivers/bluetooth/hci_serdev.c
133
hdev->flush = hci_uart_flush;
drivers/bluetooth/hci_serdev.c
149
hdev->flush = NULL;
drivers/bluetooth/hci_serdev.c
353
hdev->flush = hci_uart_flush;
drivers/bluetooth/hci_uart.h
56
int (*flush)(struct hci_uart *hu);
drivers/bluetooth/hci_vhci.c
434
hdev->flush = vhci_flush;
drivers/bluetooth/virtio_bt.c
304
hdev->flush = virtbt_flush;
drivers/char/ps3flash.c
315
.flush = ps3flash_flush,
drivers/char/xillybus/xillybus_core.c
1798
.flush = xillybus_flush,
drivers/char/xillybus/xillyusb.c
1898
.flush = xillyusb_flush,
drivers/cxl/core/port.c
2465
static BUS_ATTR_WO(flush);
drivers/firmware/efi/libstub/efistub.h
683
efi_status_t (__efiapi *flush) (efi_file_protocol_t *);
drivers/firmware/efi/libstub/efistub.h
696
u32 flush;
drivers/firmware/efi/libstub/efistub.h
779
void *flush;
drivers/firmware/efi/libstub/efistub.h
802
u32 flush;
drivers/gpib/agilent_82357a/agilent_82357a.c
369
static int agilent_82357a_abort(struct agilent_82357a_priv *a_priv, int flush)
drivers/gpib/agilent_82357a/agilent_82357a.c
382
if (flush)
drivers/gpib/agilent_82357a/agilent_82357a.c
406
if (flush) {
drivers/gpib/lpvo_usb_gpib/lpvo_usb_gpib.c
1840
.flush = skel_flush,
drivers/gpu/drm/amd/amdgpu/amdgpu_drv.c
3037
.flush = amdgpu_flush,
drivers/gpu/drm/arm/display/komeda/d71/d71_dev.c
612
.flush = d71_flush,
drivers/gpu/drm/arm/display/komeda/komeda_crtc.c
277
mdev->funcs->flush(mdev, master->id, kcrtc_st->active_pipes);
drivers/gpu/drm/arm/display/komeda/komeda_crtc.c
311
mdev->funcs->flush(mdev, kcrtc->master->id, 0);
drivers/gpu/drm/arm/display/komeda/komeda_dev.h
133
void (*flush)(struct komeda_dev *mdev,
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
175
u32 link_target, flush = 0;
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
182
flush = VIVS_GL_FLUSH_CACHE_PE2D;
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
184
flush = VIVS_GL_FLUSH_CACHE_DEPTH |
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
190
if (flush) {
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
206
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush);
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
27
u32 flush = 0;
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
353
u32 flush = VIVS_MMUv2_CONFIGURATION_MODE_MASK |
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
366
flush |= etnaviv_iommuv2_get_mtlb_addr(gpu->mmu_context);
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
369
flush);
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
38
flush = VIVS_GL_FLUSH_CACHE_PE2D;
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
40
flush = VIVS_GL_FLUSH_CACHE_DEPTH | VIVS_GL_FLUSH_CACHE_COLOR;
drivers/gpu/drm/etnaviv/etnaviv_buffer.c
42
CMD_LOAD_STATE(buffer, VIVS_GL_FLUSH_CACHE, flush);
drivers/gpu/drm/i915/display/intel_fb.c
2169
goto flush;
drivers/gpu/drm/i915/display/intel_fb.c
2174
goto flush;
drivers/gpu/drm/i915/display/intel_fb.c
2180
goto flush;
drivers/gpu/drm/i915/display/intel_fb.c
2197
flush:
drivers/gpu/drm/i915/gem/i915_gem_pm.c
147
bool flush = false;
drivers/gpu/drm/i915/gem/i915_gem_pm.c
179
flush |= (obj->read_domains & I915_GEM_DOMAIN_CPU) == 0;
drivers/gpu/drm/i915/gem/i915_gem_pm.c
184
if (flush)
drivers/gpu/drm/i915/gt/gen2_engine_cs.c
145
int flush, int post)
drivers/gpu/drm/i915/gt/gen2_engine_cs.c
152
while (flush--) {
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
159
mutex_lock(&ppgtt->flush);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
169
mutex_unlock(&ppgtt->flush);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
179
bool flush = false;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
204
flush = true;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
211
if (flush && i915_vma_is_bound(ppgtt->vma, I915_VMA_GLOBAL_BIND)) {
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
277
mutex_destroy(&ppgtt->flush);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
438
mutex_init(&ppgtt->flush);
drivers/gpu/drm/i915/gt/gen6_ppgtt.h
16
struct mutex flush;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1662
bool flush;
drivers/gpu/drm/i915/gt/intel_ggtt.c
1667
flush = i915_ggtt_resume_vm(&ggtt->vm, false);
drivers/gpu/drm/i915/gt/intel_ggtt.c
1678
if (flush)
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
1321
guc->log.stats[GUC_CAPTURE_LOG_BUFFER].flush += log_buf_state_local.flush_to_file;
drivers/gpu/drm/i915/gt/uc/intel_guc_log.c
427
log->stats[type].flush += log_buf_state_local.flush_to_file;
drivers/gpu/drm/i915/gt/uc/intel_guc_log.c
873
log->stats[type].flush,
drivers/gpu/drm/i915/gt/uc/intel_guc_log.h
80
u32 flush;
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
81
uint32_t flush = 0;
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
85
flush |= pipe2flush(pipe_id);
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
88
flush |= ovlp2flush(mdp4_crtc->ovlp);
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
90
DBG("%s: flush=%08x", mdp4_crtc->name, flush);
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
92
mdp4_crtc->flushed_mask = flush;
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
94
mdp4_write(mdp4_kms, REG_MDP4_OVERLAY_FLUSH, flush);
drivers/gpu/drm/msm/msm_gpu.h
63
void (*flush)(struct msm_gpu *gpu, struct msm_ringbuffer *ring);
drivers/gpu/drm/nouveau/dispnv50/head.c
41
struct nv50_head_atom *asyh, bool flush)
drivers/gpu/drm/nouveau/dispnv50/head.c
44
.mask = asyh->clr.mask & ~(flush ? 0 : asyh->set.mask),
drivers/gpu/drm/nouveau/dispnv50/head.h
28
struct nv50_head_atom *asyh, bool flush);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
133
nv50_wndw_flush_clr(struct nv50_wndw *wndw, u32 *interlock, bool flush,
drivers/gpu/drm/nouveau/dispnv50/wndw.c
137
.mask = asyw->clr.mask & ~(flush ? 0 : asyw->set.mask),
drivers/gpu/drm/nouveau/dispnv50/wndw.h
43
void nv50_wndw_flush_clr(struct nv50_wndw *, u32 *interlock, bool flush,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/base.c
29
if (bar && bar->func->flush)
drivers/gpu/drm/nouveau/nvkm/subdev/bar/base.c
30
bar->func->flush(bar);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/g84.c
55
.flush = g84_bar_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/gf100.c
188
.flush = g84_bar_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/gk20a.c
31
.flush = g84_bar_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/gm107.c
58
.flush = g84_bar_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/gm20b.c
31
.flush = g84_bar_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/nv50.c
247
.flush = nv50_bar_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/bar/priv.h
25
void (*flush)(struct nvkm_bar *);
drivers/gpu/drm/nouveau/nvkm/subdev/bar/tu102.c
92
.flush = g84_bar_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/fb/rammcp77.c
39
u32 flush = ((ram->base.size - (ram->poller_base + 0x40)) >> 5) - 1;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/rammcp77.c
48
nvkm_wr32(device, 0x100c24, flush);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/bar.c
185
rm->flush = r535_bar_flush;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
46
vmm->func->flush(vmm, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
55
vmm->func->flush(vmm, 0);
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/base.c
76
if (ltc->func->flush)
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/base.c
77
ltc->func->flush(ltc);
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/ga102.c
51
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/gf100.c
249
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/gk104.c
50
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/gm107.c
145
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/gm200.c
57
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/gp100.c
69
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/gp102.c
47
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/gp10b.c
54
.flush = gf100_ltc_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/ltc/priv.h
26
void (*flush)(struct nvkm_ltc *);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
122
it->flush = min(it->flush, it->max - it->lvl);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
128
if (it->flush != NVKM_VMM_LEVELS_MAX) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
129
if (it->vmm->func->flush) {
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
130
TRA(it, "flush: %d", it->flush);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
131
it->vmm->func->flush(it->vmm, it->flush);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
133
it->flush = NVKM_VMM_LEVELS_MAX;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
539
it.flush = NVKM_VMM_LEVELS_MAX;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c
80
int flush;
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h
156
void (*flush)(struct nvkm_vmm *, int depth);
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
377
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgf100.c
392
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgh100.c
286
.flush = tu102_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c
73
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk104.c
88
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c
42
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgk20a.c
57
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
115
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm200.c
131
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c
30
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgm20b.c
46
.flush = gf100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp100.c
599
.flush = gp100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgp10b.c
30
.flush = gp100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmgv100.c
68
.flush = gp100_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmmcp77.c
29
.flush = nv50_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv41.c
98
.flush = nv41_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv44.c
200
.flush = nv44_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmnv50.c
371
.flush = nv50_vmm_flush,
drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmmtu102.c
59
.flush = tu102_vmm_flush,
drivers/gpu/drm/qxl/qxl_cmd.c
202
bool qxl_queue_garbage_collect(struct qxl_device *qdev, bool flush)
drivers/gpu/drm/qxl/qxl_cmd.c
206
if (flush)
drivers/gpu/drm/qxl/qxl_drv.h
406
bool qxl_queue_garbage_collect(struct qxl_device *qdev, bool flush);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
1049
size_t size, bool flush)
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
1059
if (flush)
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
1106
struct vmw_cmdbuf_header *header, bool flush)
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
1109
vmw_cmdbuf_commit_cur(man, size, flush);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
1120
if (flush)
drivers/gpu/drm/vmwgfx/vmwgfx_drv.h
1294
bool flush);
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
117
fifo_size = sizeof(*cmds) + sizeof(*flush) + sizeof(*items) * num_items;
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
125
flush = (struct vmw_escape_video_flush *)&items[num_items];
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
164
fill_flush(flush, arg->stream_id);
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
184
struct vmw_escape_video_flush flush;
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
206
fill_flush(&cmds->flush, stream_id);
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
63
SVGAEscapeVideoFlush flush;
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
77
fill_escape(&cmd->escape, sizeof(cmd->flush));
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
78
cmd->flush.cmdType = SVGA_ESCAPE_VMWARE_VIDEO_FLUSH;
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
79
cmd->flush.streamId = stream_id;
drivers/gpu/drm/vmwgfx/vmwgfx_overlay.c
93
struct vmw_escape_video_flush *flush;
drivers/gpu/drm/xe/xe_guc_capture.c
1402
guc->log.stats[GUC_LOG_TYPE_STATE_CAPTURE].flush += tmp;
drivers/gpu/drm/xe/xe_guc_log_types.h
53
u32 flush;
drivers/gpu/drm/xe/xe_guc_tlb_inval.c
225
.flush = tlb_inval_flush,
drivers/gpu/drm/xe/xe_tlb_inval.c
74
tlb_inval->ops->flush(tlb_inval);
drivers/gpu/drm/xe/xe_tlb_inval_types.h
64
void (*flush)(struct xe_tlb_inval *tlb_inval);
drivers/gpu/host1x/dev.h
292
host->cdma_op->flush(cdma);
drivers/gpu/host1x/dev.h
41
void (*flush)(struct host1x_cdma *cdma);
drivers/gpu/host1x/hw/cdma_hw.c
366
.flush = cdma_flush,
drivers/hid/hid-haptic.c
416
int (*flush)(struct input_dev *dev, struct file *file);
drivers/hid/hid-haptic.c
511
flush = dev->flush;
drivers/hid/hid-haptic.c
543
dev->flush = flush;
drivers/hid/wacom_sys.c
109
bool flush = false;
drivers/hid/wacom_sys.c
141
flush = true;
drivers/hid/wacom_sys.c
162
if (flush)
drivers/hid/wacom_sys.c
168
return insert && !flush;
drivers/hsi/controllers/omap_ssi_port.c
1168
port->flush = ssi_flush;
drivers/hsi/controllers/omap_ssi_port.c
1233
port->flush = hsi_dummy_cl;
drivers/hsi/hsi_core.c
491
port[i]->flush = hsi_dummy_cl;
drivers/hwtracing/intel_th/gth.c
246
OUTPUT_PARM(flush, BIT(7), 0, 1, output),
drivers/i2c/busses/i2c-diolan-u2c.c
156
static int diolan_write_cmd(struct i2c_diolan_u2c *dev, bool flush)
drivers/i2c/busses/i2c-diolan-u2c.c
158
if (flush || dev->olen >= DIOLAN_FLUSH_LEN)
drivers/i2c/busses/i2c-diolan-u2c.c
164
static int diolan_usb_cmd(struct i2c_diolan_u2c *dev, u8 command, bool flush)
drivers/i2c/busses/i2c-diolan-u2c.c
168
return diolan_write_cmd(dev, flush);
drivers/i2c/busses/i2c-diolan-u2c.c
173
bool flush)
drivers/i2c/busses/i2c-diolan-u2c.c
178
return diolan_write_cmd(dev, flush);
drivers/i2c/busses/i2c-diolan-u2c.c
183
u8 d2, bool flush)
drivers/i2c/busses/i2c-diolan-u2c.c
189
return diolan_write_cmd(dev, flush);
drivers/infiniband/hw/hfi1/pio.c
38
int flush = 0; /* re-read sendctrl to make sure it is flushed */
drivers/infiniband/hw/hfi1/pio.c
73
flush = 1;
drivers/infiniband/hw/hfi1/pio.c
82
if (flush)
drivers/infiniband/hw/ionic/ionic_controlpath.c
2395
if (cq->flush && cq->vcq->ibcq.comp_handler)
drivers/infiniband/hw/ionic/ionic_controlpath.c
2423
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_controlpath.c
2438
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
1285
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
1362
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
150
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
194
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
327
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
336
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
592
if (likely(!cq->flush))
drivers/infiniband/hw/ionic/ionic_datapath.c
595
cq->flush = false;
drivers/infiniband/hw/ionic/ionic_datapath.c
611
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
628
cq->flush = true;
drivers/infiniband/hw/ionic/ionic_ibdev.h
202
bool flush;
drivers/infiniband/hw/mlx4/mad.c
2090
struct mlx4_ib_demux_pv_ctx *ctx, int flush)
drivers/infiniband/hw/mlx4/mad.c
2096
if (flush)
drivers/infiniband/sw/rxe/rxe_req.c
462
reth_set_rkey(pkt, ibwr->wr.flush.rkey);
drivers/infiniband/sw/rxe/rxe_req.c
471
feth_init(pkt, ibwr->wr.flush.type, ibwr->wr.flush.level);
drivers/infiniband/sw/rxe/rxe_resp.c
636
res->flush.va = qp->resp.va + qp->resp.offset;
drivers/infiniband/sw/rxe/rxe_resp.c
637
res->flush.length = qp->resp.length;
drivers/infiniband/sw/rxe/rxe_resp.c
638
res->flush.type = feth_plt(pkt);
drivers/infiniband/sw/rxe/rxe_resp.c
639
res->flush.level = feth_sel(pkt);
drivers/infiniband/sw/rxe/rxe_resp.c
660
if (res->flush.level == IB_FLUSH_RANGE) {
drivers/infiniband/sw/rxe/rxe_resp.c
661
start = res->flush.va;
drivers/infiniband/sw/rxe/rxe_resp.c
662
length = res->flush.length;
drivers/infiniband/sw/rxe/rxe_resp.c
668
if (res->flush.type & IB_FLUSH_PERSISTENT) {
drivers/infiniband/sw/rxe/rxe_resp.c
673
} else if (res->flush.type & IB_FLUSH_GLOBAL) {
drivers/infiniband/sw/rxe/rxe_verbs.h
196
} flush;
drivers/input/ff-core.c
317
dev->flush = input_ff_flush;
drivers/input/input.c
627
if (dev->flush)
drivers/input/input.c
628
return dev->flush(dev, file);
drivers/input/misc/uinput.c
358
dev->flush = uinput_dev_flush;
drivers/iommu/intel/iommu.c
1111
iommu->flush.flush_context(iommu, did_old,
drivers/iommu/intel/iommu.c
1115
iommu->flush.flush_iotlb(iommu, did_old, 0, 0,
drivers/iommu/intel/iommu.c
1132
iommu->flush.flush_context(iommu, 0,
drivers/iommu/intel/iommu.c
1136
iommu->flush.flush_iotlb(iommu, did, 0, 0, DMA_TLB_DSI_FLUSH);
drivers/iommu/intel/iommu.c
1430
iommu->flush.flush_context = __iommu_flush_context;
drivers/iommu/intel/iommu.c
1431
iommu->flush.flush_iotlb = __iommu_flush_iotlb;
drivers/iommu/intel/iommu.c
1435
iommu->flush.flush_context = qi_flush_context;
drivers/iommu/intel/iommu.c
1436
iommu->flush.flush_iotlb = qi_flush_iotlb;
drivers/iommu/intel/iommu.c
1818
iommu->flush.flush_context(iommu, 0, 0, 0,
drivers/iommu/intel/iommu.c
1820
iommu->flush.flush_iotlb(iommu, 0, 0, 0,
drivers/iommu/intel/iommu.c
717
iommu->flush.flush_context(iommu, 0, 0, 0, DMA_CCMD_GLOBAL_INVL);
drivers/iommu/intel/iommu.c
720
iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_GLOBAL_FLUSH);
drivers/iommu/intel/iommu.h
712
struct iommu_flush flush;
drivers/iommu/intel/pasid.c
287
iommu->flush.flush_iotlb(iommu, did, 0, 0, DMA_TLB_DSI_FLUSH);
drivers/iommu/intel/pasid.c
548
iommu->flush.flush_iotlb(iommu, did, 0, 0, DMA_TLB_DSI_FLUSH);
drivers/iommu/intel/pasid.c
855
iommu->flush.flush_context(iommu, 0,
drivers/iommu/intel/pasid.c
860
iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_GLOBAL_FLUSH);
drivers/iommu/intel/pasid.c
884
iommu->flush.flush_context(iommu, 0,
drivers/iommu/intel/pasid.c
888
iommu->flush.flush_iotlb(iommu, 0, 0, 0, DMA_TLB_DSI_FLUSH);
drivers/iommu/intel/pasid.c
968
iommu->flush.flush_context(iommu, did, PCI_DEVID(info->bus, info->devfn),
drivers/iommu/intel/pasid.c
977
iommu->flush.flush_iotlb(iommu, did, 0, 0, DMA_TLB_DSI_FLUSH);
drivers/mailbox/mailbox.c
298
if (!chan->mbox->ops->flush)
drivers/mailbox/mailbox.c
301
ret = chan->mbox->ops->flush(chan, timeout);
drivers/mailbox/mtk-cmdq-mailbox.c
617
.flush = cmdq_mbox_flush,
drivers/mailbox/sprd-mailbox.c
325
.flush = sprd_mbox_flush,
drivers/mailbox/tegra-hsp.c
585
.flush = tegra_hsp_mailbox_flush,
drivers/md/bcache/btree.c
606
static int mca_reap(struct btree *b, unsigned int min_order, bool flush)
drivers/md/bcache/btree.c
621
if (!flush) {
drivers/md/bcache/request.c
1038
struct bio *flush;
drivers/md/bcache/request.c
1040
flush = bio_alloc_bioset(bio->bi_bdev, 0,
drivers/md/bcache/request.c
1043
if (!flush) {
drivers/md/bcache/request.c
1047
flush->bi_end_io = backing_request_endio;
drivers/md/bcache/request.c
1048
flush->bi_private = cl;
drivers/md/bcache/request.c
1050
closure_bio_submit(s->iop.c, flush, cl);
drivers/md/dm-delay.c
172
if (dc->flush.dev)
drivers/md/dm-delay.c
173
dm_put_device(ti, dc->flush.dev);
drivers/md/dm-delay.c
249
ret = delay_class_ctr(ti, &dc->flush, argv);
drivers/md/dm-delay.c
262
ret = delay_class_ctr(ti, &dc->flush, argv + 3);
drivers/md/dm-delay.c
268
ret = delay_class_ctr(ti, &dc->flush, argv + 6);
drivers/md/dm-delay.c
271
max_delay = max(max_delay, dc->flush.delay);
drivers/md/dm-delay.c
272
min_delay = min_not_zero(min_delay, dc->flush.delay);
drivers/md/dm-delay.c
372
c = &dc->flush;
drivers/md/dm-delay.c
411
DMEMIT("%u %u %u", dc->read.ops, dc->write.ops, dc->flush.ops);
drivers/md/dm-delay.c
422
DMEMIT_DELAY_CLASS(&dc->flush);
drivers/md/dm-delay.c
444
ret = fn(ti, dc->flush.dev, dc->flush.start, ti->len, data);
drivers/md/dm-delay.c
45
struct delay_class flush;
drivers/md/dm-log-userspace-base.c
878
.flush = userspace_flush,
drivers/md/dm-log.c
856
.flush = core_flush,
drivers/md/dm-log.c
875
.flush = disk_flush,
drivers/md/dm-region-hash.c
502
rh->log->type->flush(rh->log);
drivers/md/dm-region-hash.c
689
return rh->log->type->flush(rh->log);
drivers/md/dm-vdo/flush.c
103
struct vdo_flush *flush = NULL;
drivers/md/dm-vdo/flush.c
106
flush = vdo_allocate_memory_nowait(sizeof(struct vdo_flush), __func__);
drivers/md/dm-vdo/flush.c
108
int result = vdo_allocate(1, struct vdo_flush, __func__, &flush);
drivers/md/dm-vdo/flush.c
114
if (flush != NULL) {
drivers/md/dm-vdo/flush.c
117
vdo_initialize_completion(&flush->completion, flusher->vdo,
drivers/md/dm-vdo/flush.c
121
return flush;
drivers/md/dm-vdo/flush.c
182
static void vdo_complete_flush(struct vdo_flush *flush);
drivers/md/dm-vdo/flush.c
251
struct vdo_flush *flush =
drivers/md/dm-vdo/flush.c
254
flusher->notify_generation = flush->flush_generation;
drivers/md/dm-vdo/flush.c
269
struct vdo_flush *flush = completion_as_vdo_flush(completion);
drivers/md/dm-vdo/flush.c
279
vdo_complete_flush(flush);
drivers/md/dm-vdo/flush.c
283
flush->flush_generation = flusher->flush_generation++;
drivers/md/dm-vdo/flush.c
285
vdo_waitq_enqueue_waiter(&flusher->notifiers, &flush->waiter);
drivers/md/dm-vdo/flush.c
327
struct vdo_flush *flush =
drivers/md/dm-vdo/flush.c
330
if (flush->flush_generation >= oldest_active_generation)
drivers/md/dm-vdo/flush.c
333
VDO_ASSERT_LOG_ONLY((flush->flush_generation ==
drivers/md/dm-vdo/flush.c
337
(unsigned long long) flush->flush_generation);
drivers/md/dm-vdo/flush.c
339
vdo_complete_flush(flush);
drivers/md/dm-vdo/flush.c
369
static void initialize_flush(struct vdo_flush *flush, struct vdo *vdo)
drivers/md/dm-vdo/flush.c
371
bio_list_init(&flush->bios);
drivers/md/dm-vdo/flush.c
372
bio_list_merge_init(&flush->bios, &vdo->flusher->waiting_flush_bios);
drivers/md/dm-vdo/flush.c
375
static void launch_flush(struct vdo_flush *flush)
drivers/md/dm-vdo/flush.c
377
struct vdo_completion *completion = &flush->completion;
drivers/md/dm-vdo/flush.c
398
struct vdo_flush *flush = mempool_alloc(vdo->flusher->flush_pool, GFP_NOWAIT);
drivers/md/dm-vdo/flush.c
410
if (flush == NULL) {
drivers/md/dm-vdo/flush.c
416
initialize_flush(flush, vdo);
drivers/md/dm-vdo/flush.c
420
launch_flush(flush);
drivers/md/dm-vdo/flush.c
431
static void release_flush(struct vdo_flush *flush)
drivers/md/dm-vdo/flush.c
434
struct flusher *flusher = flush->completion.vdo->flusher;
drivers/md/dm-vdo/flush.c
441
initialize_flush(flush, flusher->vdo);
drivers/md/dm-vdo/flush.c
448
launch_flush(flush);
drivers/md/dm-vdo/flush.c
452
mempool_free(flush, flusher->flush_pool);
drivers/md/dm-vdo/flush.c
462
struct vdo_flush *flush = completion_as_vdo_flush(completion);
drivers/md/dm-vdo/flush.c
466
while ((bio = bio_list_pop(&flush->bios)) != NULL) {
drivers/md/dm-vdo/flush.c
484
release_flush(flush);
drivers/md/dm-vdo/flush.c
515
static void vdo_complete_flush(struct vdo_flush *flush)
drivers/md/dm-vdo/flush.c
517
struct vdo_completion *completion = &flush->completion;
drivers/md/dm-vdo/message-stats.c
275
write_u64("flush : ", stats->flush, ", ", buf, maxlen);
drivers/md/dm-vdo/statistics.h
166
u64 flush;
drivers/md/dm-vdo/vdo.c
1436
b->flush = atomic64_read(&a->flush);
drivers/md/dm-vdo/vdo.c
1448
.flush = minuend.flush - subtrahend.flush,
drivers/md/dm-vdo/vdo.h
81
atomic64_t flush; /* Number of REQ_FLUSH bios */
drivers/md/dm-vdo/vio.c
463
atomic64_inc(&bio_stats->flush);
drivers/md/dm-vdo/vio.c
487
atomic64_inc(&bio_stats->flush);
drivers/md/md-bitmap.c
228
static bool bitmap_enabled(void *data, bool flush)
drivers/md/md-bitmap.c
232
if (!flush)
drivers/md/md-bitmap.c
2987
.flush = bitmap_flush,
drivers/md/md-bitmap.h
137
static inline bool md_bitmap_enabled(struct mddev *mddev, bool flush)
drivers/md/md-bitmap.h
146
return mddev->bitmap_ops->enabled(mddev->bitmap, flush);
drivers/md/md-bitmap.h
86
bool (*enabled)(void *data, bool flush);
drivers/md/md-bitmap.h
92
void (*flush)(struct mddev *mddev);
drivers/md/md-llbitmap.c
1332
static bool llbitmap_enabled(void *data, bool flush)
drivers/md/md-llbitmap.c
1582
.flush = llbitmap_flush,
drivers/md/md.c
6861
mddev->bitmap_ops->flush(mddev);
drivers/media/dvb-core/dvb_vb2.c
235
bool flush)
drivers/media/dvb-core/dvb_vb2.c
292
if (flush && ctx->buf) {
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1066
static void s5p_mfc_set_flush(struct s5p_mfc_ctx *ctx, int flush)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v5.c
1071
if (flush)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v6.c
1816
static inline void s5p_mfc_set_flush(struct s5p_mfc_ctx *ctx, int flush)
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_opr_v6.c
1821
if (flush) {
drivers/media/platform/st/sti/delta/delta-v4l2.c
1541
call_dec_op(dec, flush, ctx);
drivers/media/platform/st/sti/delta/delta.h
382
int (*flush)(struct delta_ctx *ctx);
drivers/memory/tegra/tegra20-emc.c
299
static int emc_complete_timing_change(struct tegra_emc *emc, bool flush)
drivers/memory/tegra/tegra20-emc.c
304
dev_dbg(emc->dev, "%s: flush %d\n", __func__, flush);
drivers/memory/tegra/tegra20-emc.c
306
if (flush) {
drivers/mtd/mtd_blkdevs.c
339
if (tr->flush)
drivers/mtd/mtd_blkdevs.c
59
if (tr->flush(dev))
drivers/mtd/mtdblock.c
348
.flush = mtdblock_flush,
drivers/mtd/mtdswap.c
1476
.flush = mtdswap_flush,
drivers/mtd/sm_ftl.c
1264
.flush = sm_flush,
drivers/net/ethernet/atheros/ag71xx.c
769
static int ag71xx_tx_packets(struct ag71xx *ag, bool flush, int budget)
drivers/net/ethernet/atheros/ag71xx.c
791
if (!flush && !ag71xx_desc_empty(desc)) {
drivers/net/ethernet/atheros/ag71xx.c
801
if (flush)
drivers/net/ethernet/emulex/benet/be_cmds.h
1222
u8 flush;
drivers/net/ethernet/emulex/benet/be_main.c
1375
bool flush = !netdev_xmit_more();
drivers/net/ethernet/emulex/benet/be_main.c
1405
if (flush || __netif_subqueue_stopped(netdev, q_idx))
drivers/net/ethernet/emulex/benet/be_main.c
1414
if (flush && txo->pend_wrb_cnt)
drivers/net/ethernet/engleder/tsnep_tc.c
39
u32 properties, u32 interval, bool flush)
drivers/net/ethernet/engleder/tsnep_tc.c
50
if (flush) {
drivers/net/ethernet/intel/iavf/iavf.h
587
void iavf_irq_enable(struct iavf_adapter *adapter, bool flush);
drivers/net/ethernet/intel/iavf/iavf_main.c
372
void iavf_irq_enable(struct iavf_adapter *adapter, bool flush)
drivers/net/ethernet/intel/iavf/iavf_main.c
379
if (flush)
drivers/net/ethernet/intel/idpf/xdp.h
87
static inline void idpf_xdp_tx_finalize(void *_xdpsq, bool sent, bool flush)
drivers/net/ethernet/intel/idpf/xdp.h
91
if ((!flush || unlikely(!sent)) &&
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3407
bool flush)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3464
if (flush)
drivers/net/ethernet/mediatek/mtk_wed_wo.c
100
bool flush)
drivers/net/ethernet/mediatek/mtk_wed_wo.c
111
if (flush)
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
128
flush = true;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
131
if (flush) {
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
72
bool flush = false;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2187
bool flush = cqe->shampo.flush;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
2271
if (flush && rq->hw_gro_data->skb)
drivers/net/ethernet/sfc/efx.h
238
bool flush);
drivers/net/ethernet/sfc/falcon/selftest.c
298
if ((state == NULL) || state->flush)
drivers/net/ethernet/sfc/falcon/selftest.c
551
state->flush = false;
drivers/net/ethernet/sfc/falcon/selftest.c
641
state->flush = true;
drivers/net/ethernet/sfc/falcon/selftest.c
650
state->flush = true;
drivers/net/ethernet/sfc/falcon/selftest.c
683
state->flush = true;
drivers/net/ethernet/sfc/falcon/selftest.c
83
bool flush;
drivers/net/ethernet/sfc/selftest.c
295
if ((state == NULL) || state->flush)
drivers/net/ethernet/sfc/selftest.c
548
state->flush = false;
drivers/net/ethernet/sfc/selftest.c
634
state->flush = true;
drivers/net/ethernet/sfc/selftest.c
643
state->flush = true;
drivers/net/ethernet/sfc/selftest.c
676
state->flush = true;
drivers/net/ethernet/sfc/selftest.c
85
bool flush;
drivers/net/ethernet/sfc/siena/efx.h
216
struct xdp_frame **xdpfs, bool flush);
drivers/net/ethernet/sfc/siena/selftest.c
296
if ((state == NULL) || state->flush)
drivers/net/ethernet/sfc/siena/selftest.c
549
state->flush = false;
drivers/net/ethernet/sfc/siena/selftest.c
639
state->flush = true;
drivers/net/ethernet/sfc/siena/selftest.c
648
state->flush = true;
drivers/net/ethernet/sfc/siena/selftest.c
681
state->flush = true;
drivers/net/ethernet/sfc/siena/selftest.c
86
bool flush;
drivers/net/ethernet/sfc/siena/tx.c
209
bool flush)
drivers/net/ethernet/sfc/siena/tx.c
283
if (flush && i > 0)
drivers/net/ethernet/sfc/tx.c
414
bool flush)
drivers/net/ethernet/sfc/tx.c
488
if (flush && i > 0)
drivers/net/ethernet/ti/cpsw.c
223
if (sync_ctx->flush) {
drivers/net/ethernet/ti/cpsw.c
243
sync_ctx.flush = 0;
drivers/net/ethernet/ti/cpsw.c
259
sync_ctx.flush = 1;
drivers/net/ethernet/ti/cpsw_new.c
166
if (sync_ctx->flush) {
drivers/net/ethernet/ti/cpsw_new.c
186
sync_ctx.flush = 0;
drivers/net/ethernet/ti/cpsw_new.c
202
sync_ctx.flush = 1;
drivers/net/ethernet/ti/cpsw_priv.h
407
int flush; /* flush flag */
drivers/net/geneve.c
864
NAPI_GRO_CB(skb)->flush = 1;
drivers/net/geneve.c
878
int flush = 1;
drivers/net/geneve.c
938
flush = 0;
drivers/net/geneve.c
941
skb_gro_flush_final(skb, pp, flush);
drivers/net/ppp/ppp_async.c
668
goto flush; /* error, e.g. loss of CD */
drivers/net/ppp/ppp_async.c
703
flush:
drivers/net/ppp/ppp_synctty.c
602
goto flush; /* error, e.g. loss of CD */
drivers/net/ppp/ppp_synctty.c
623
flush:
drivers/net/tun.c
2378
struct xdp_buff *xdp, int *flush,
drivers/net/tun.c
2416
*flush = true;
drivers/net/tun.c
2525
int flush = 0, queued = 0;
drivers/net/tun.c
2535
ret = tun_xdp_one(tun, tfile, xdp, &flush, &tpage);
drivers/net/tun.c
2540
if (flush)
drivers/net/vxlan/vxlan_core.c
712
int flush = 1;
drivers/net/vxlan/vxlan_core.c
716
flush = 0;
drivers/net/vxlan/vxlan_core.c
718
skb_gro_flush_final_remcsum(skb, pp, flush, &grc);
drivers/net/vxlan/vxlan_core.c
731
int flush = 1;
drivers/net/vxlan/vxlan_core.c
741
flush = 0;
drivers/net/vxlan/vxlan_core.c
744
skb_gro_flush_final_remcsum(skb, pp, flush, &grc);
drivers/net/wireless/ath/ar5523/ar5523.c
1376
.flush = ar5523_flush,
drivers/net/wireless/ath/ath10k/mac.c
9495
.flush = ath10k_flush,
drivers/net/wireless/ath/ath11k/mac.c
10114
.flush = ath11k_mac_op_flush,
drivers/net/wireless/ath/ath12k/mac.c
10609
goto flush;
drivers/net/wireless/ath/ath12k/mac.c
10617
flush:
drivers/net/wireless/ath/ath12k/wifi7/hw.c
1002
.flush = ath12k_mac_op_flush,
drivers/net/wireless/ath/ath9k/ath9k.h
581
int ath_rx_tasklet(struct ath_softc *sc, int flush, bool hp);
drivers/net/wireless/ath/ath9k/main.c
2008
bool flush = false;
drivers/net/wireless/ath/ath9k/main.c
2039
flush = true;
drivers/net/wireless/ath/ath9k/main.c
2044
if (!flush)
drivers/net/wireless/ath/ath9k/main.c
2173
goto flush;
drivers/net/wireless/ath/ath9k/main.c
2188
flush:
drivers/net/wireless/ath/ath9k/main.c
2815
.flush = ath9k_flush,
drivers/net/wireless/ath/ath9k/recv.c
1060
int ath_rx_tasklet(struct ath_softc *sc, int flush, bool hp)
drivers/net/wireless/ath/ath9k/recv.c
1223
ath_rx_buf_relink(sc, bf, flush);
drivers/net/wireless/ath/ath9k/recv.c
1224
if (!flush)
drivers/net/wireless/ath/ath9k/recv.c
1226
} else if (!flush) {
drivers/net/wireless/ath/ath9k/recv.c
38
bool flush)
drivers/net/wireless/ath/ath9k/recv.c
65
else if (!flush)
drivers/net/wireless/ath/ath9k/recv.c
72
bool flush)
drivers/net/wireless/ath/ath9k/recv.c
75
ath_rx_buf_link(sc, sc->rx.buf_hold, flush);
drivers/net/wireless/ath/ath9k/xmit.c
502
bool flush = !!(ts->ts_status & ATH9K_TX_FLUSH);
drivers/net/wireless/ath/ath9k/xmit.c
592
} else if (flush) {
drivers/net/wireless/ath/ath9k/xmit.c
734
bool txok, flush;
drivers/net/wireless/ath/ath9k/xmit.c
737
flush = !!(ts->ts_status & ATH9K_TX_FLUSH);
drivers/net/wireless/ath/ath9k/xmit.c
758
if (!flush) {
drivers/net/wireless/ath/ath9k/xmit.c
770
if (!flush)
drivers/net/wireless/ath/carl9170/main.c
1722
.flush = carl9170_op_flush,
drivers/net/wireless/ath/carl9170/main.c
855
static int carl9170_update_survey(struct ar9170 *ar, bool flush, bool noise)
drivers/net/wireless/ath/carl9170/main.c
871
if (flush)
drivers/net/wireless/ath/wcn36xx/main.c
1383
.flush = wcn36xx_flush,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
983
.flush = brcms_ops_flush,
drivers/net/wireless/intel/iwlegacy/3945-mac.c
3420
.flush = il_mac_flush,
drivers/net/wireless/intel/iwlegacy/4965-mac.c
6333
.flush = il_mac_flush,
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
2028
int flush;
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
2034
if (sscanf(buf, "%d", &flush) != 1)
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
1601
.flush = iwlagn_mac_flush,
drivers/net/wireless/intel/iwlwifi/mld/mac80211.c
2762
.flush = iwl_mld_mac80211_flush,
drivers/net/wireless/intel/iwlwifi/mld/sta.c
1108
bool flush, u8 tid)
drivers/net/wireless/intel/iwlwifi/mld/sta.c
1114
if (flush)
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
6373
.flush = iwl_mvm_mac_flush,
drivers/net/wireless/intel/iwlwifi/mvm/mld-mac80211.c
1046
.flush = iwl_mvm_mac_flush,
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
321
bool flush, u8 tid, u16 *queuptr)
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
330
if (flush)
drivers/net/wireless/intersil/p54/main.c
726
.flush = p54_flush,
drivers/net/wireless/marvell/mwifiex/fw.h
1675
u8 flush;
drivers/net/wireless/marvell/mwifiex/scan.c
2735
bg_query->flush = 1;
drivers/net/wireless/mediatek/mt76/dma.c
407
mt76_dma_tx_cleanup(struct mt76_dev *dev, struct mt76_queue *q, bool flush)
drivers/net/wireless/mediatek/mt76/dma.c
416
if (flush)
drivers/net/wireless/mediatek/mt76/dma.c
431
if (!flush && q->tail == last)
drivers/net/wireless/mediatek/mt76/dma.c
436
if (flush) {
drivers/net/wireless/mediatek/mt76/dma.c
492
int *len, u32 *info, bool *more, bool *drop, bool flush)
drivers/net/wireless/mediatek/mt76/dma.c
499
if (mt76_queue_is_wed_rro_rxdmad_c(q) && !flush)
drivers/net/wireless/mediatek/mt76/dma.c
548
mt76_dma_dequeue(struct mt76_dev *dev, struct mt76_queue *q, bool flush,
drivers/net/wireless/mediatek/mt76/dma.c
564
if (flush)
drivers/net/wireless/mediatek/mt76/dma.c
579
if (flush)
drivers/net/wireless/mediatek/mt76/dma.c
591
if (flush)
drivers/net/wireless/mediatek/mt76/dma.c
600
return mt76_dma_get_buf(dev, q, idx, len, info, more, drop, flush);
drivers/net/wireless/mediatek/mt76/mt76.h
1564
void mt76_tx_status_check(struct mt76_dev *dev, bool flush);
drivers/net/wireless/mediatek/mt76/mt76.h
305
void *(*dequeue)(struct mt76_dev *dev, struct mt76_queue *q, bool flush,
drivers/net/wireless/mediatek/mt76/mt76.h
311
bool flush);
drivers/net/wireless/mediatek/mt76/mt7603/main.c
714
.flush = mt7603_flush,
drivers/net/wireless/mediatek/mt76/mt76x0/pci.c
85
.flush = mt76x0e_flush,
drivers/net/wireless/mediatek/mt76/mt76x2/pci_main.c
142
.flush = mt76x2_flush,
drivers/net/wireless/mediatek/mt76/mt7921/main.c
1568
.flush = mt792x_flush,
drivers/net/wireless/mediatek/mt76/mt7925/main.c
2208
.flush = mt792x_flush,
drivers/net/wireless/mediatek/mt76/tx.c
213
mt76_tx_status_check(struct mt76_dev *dev, bool flush)
drivers/net/wireless/mediatek/mt76/tx.c
220
mt76_tx_status_skb_get(dev, wcid, flush ? -1 : 0, &list);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1728
.flush = rt2x00mac_flush,
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2026
.flush = rt2x00mac_flush,
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1817
.flush = rt2x00mac_flush,
drivers/net/wireless/ralink/rt2x00/rt2800pci.c
318
.flush = rt2x00mac_flush,
drivers/net/wireless/ralink/rt2x00/rt2800soc.c
182
.flush = rt2x00mac_flush,
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
658
.flush = rt2x00mac_flush,
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2895
.flush = rt2x00mac_flush,
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2315
.flush = rt2x00mac_flush,
drivers/net/wireless/realtek/rtlwifi/core.c
1724
if (rtlpriv->intf_ops->flush)
drivers/net/wireless/realtek/rtlwifi/core.c
1725
rtlpriv->intf_ops->flush(hw, queues, drop);
drivers/net/wireless/realtek/rtlwifi/core.c
1910
.flush = rtl_op_flush,
drivers/net/wireless/realtek/rtlwifi/pci.c
2357
.flush = rtl_pci_flush,
drivers/net/wireless/realtek/rtlwifi/wifi.h
2278
void (*flush)(struct ieee80211_hw *hw, u32 queues, bool drop);
drivers/net/wireless/realtek/rtw88/mac80211.c
976
.flush = rtw_ops_flush,
drivers/net/wireless/realtek/rtw89/mac80211.c
1976
.flush = rtw89_ops_flush,
drivers/net/wireless/silabs/wfx/main.c
154
.flush = wfx_flush,
drivers/net/wireless/st/cw1200/main.c
231
.flush = cw1200_flush,
drivers/net/wireless/ti/wlcore/main.c
6022
.flush = wlcore_op_flush,
drivers/net/wireless/virtual/mac80211_hwsim.c
4182
.flush = mac80211_hwsim_flush, \
drivers/net/wwan/iosm/iosm_ipc_mux.h
197
flush:1; /* flush net interface ? */
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
484
ipc_mux->session[if_id].flush = 1;
drivers/net/wwan/iosm/iosm_ipc_mux_codec.c
656
ipc_mux->session[if_id].flush = 1;
drivers/nfc/fdp/i2c.c
147
goto flush;
drivers/nfc/fdp/i2c.c
163
goto flush;
drivers/nfc/fdp/i2c.c
175
goto flush;
drivers/nfc/fdp/i2c.c
187
flush:
drivers/nfc/microread/i2c.c
156
goto flush;
drivers/nfc/microread/i2c.c
162
goto flush;
drivers/nfc/microread/i2c.c
179
goto flush;
drivers/nfc/microread/i2c.c
189
flush:
drivers/nfc/pn544/i2c.c
371
goto flush;
drivers/nfc/pn544/i2c.c
377
goto flush;
drivers/nfc/pn544/i2c.c
394
goto flush;
drivers/nfc/pn544/i2c.c
404
flush:
drivers/nvdimm/nd.h
320
int hint, void __iomem *flush)
drivers/nvdimm/nd.h
325
ndrd->flush_wpq[dimm * num + (hint & mask)] = flush;
drivers/nvdimm/nd.h
424
int (*flush)(struct nd_region *nd_region, struct bio *bio);
drivers/nvdimm/region_devs.c
1067
if (ndr_desc->flush)
drivers/nvdimm/region_devs.c
1068
nd_region->flush = ndr_desc->flush;
drivers/nvdimm/region_devs.c
1070
nd_region->flush = NULL;
drivers/nvdimm/region_devs.c
1115
if (!nd_region->flush)
drivers/nvdimm/region_devs.c
1118
if (nd_region->flush(nd_region, bio))
drivers/nvdimm/region_devs.c
1175
if (test_bit(ND_REGION_ASYNC, &nd_region->flags) && nd_region->flush)
drivers/nvdimm/region_devs.c
291
bool flush;
drivers/nvdimm/region_devs.c
292
int rc = kstrtobool(buf, &flush);
drivers/nvdimm/region_devs.c
297
if (!flush)
drivers/nvdimm/virtio_pmem.c
112
ndr_desc.flush = async_pmem_flush;
drivers/perf/xgene_pmu.c
381
XGENE_PMU_EVENT_ATTR(index-flush-eviction, 0x03),
drivers/s390/net/qeth_core_main.c
4150
bool flush;
drivers/s390/net/qeth_core_main.c
4161
flush = !qeth_iqd_may_bulk(queue, skb, hdr);
drivers/s390/net/qeth_core_main.c
4163
if (flush ||
drivers/s390/net/qeth_core_main.c
4171
flush = true;
drivers/s390/net/qeth_core_main.c
4173
if (flush)
drivers/s390/net/qeth_core_main.c
4199
flush = __netdev_tx_sent_queue(txq, bytes,
drivers/s390/net/qeth_core_main.c
4202
if (flush || next_element >= queue->max_elements) {
drivers/s390/net/qeth_core_main.c
4207
flush = true;
drivers/s390/net/qeth_core_main.c
4209
if (flush)
drivers/scsi/libiscsi_tcp.c
1065
flush:
drivers/scsi/libiscsi_tcp.c
1106
goto flush;
drivers/scsi/st.c
4256
.flush = st_flush,
drivers/spi/spi-pl022.c
1304
flush(pl022);
drivers/spi/spi-pl022.c
1338
flush(pl022);
drivers/staging/greybus/camera.c
859
.flush = gb_camera_op_flush,
drivers/staging/greybus/gb-camera.h
98
int (*flush)(void *priv, u32 *request_id);
drivers/staging/media/atomisp/pci/ia_css_env.h
35
void (*flush)(struct ia_css_acc_fw *fw);
drivers/staging/media/atomisp/pci/sh_css.c
1320
if (my_css.flush != env->cpu_mem_env.flush) {
drivers/staging/media/atomisp/pci/sh_css.c
1322
my_css.flush = env->cpu_mem_env.flush;
drivers/staging/media/atomisp/pci/sh_css.c
1349
flush_func = env->cpu_mem_env.flush;
drivers/staging/media/atomisp/pci/sh_css.c
1366
my_css.flush = flush_func;
drivers/staging/media/atomisp/pci/sh_css_struct.h
37
void (*flush)(struct ia_css_acc_fw *fw);
drivers/tty/hvc/hvc_console.c
118
if (ops->flush)
drivers/tty/hvc/hvc_console.c
119
return ops->flush(vtermno, wait);
drivers/tty/hvc/hvc_console.h
57
int (*flush)(uint32_t vtermno, bool wait);
drivers/tty/hvc/hvc_opal.c
146
.flush = opal_flush_chars,
drivers/tty/hvc/hvc_opal.c
55
.flush = opal_flush_chars,
drivers/tty/mips_ejtag_fdc.c
1262
.flush = kgdbfdc_flush,
drivers/usb/cdns3/cdns3-trace.h
28
TP_PROTO(struct cdns3_endpoint *ep_priv, u8 halt, u8 flush),
drivers/usb/cdns3/cdns3-trace.h
29
TP_ARGS(ep_priv, halt, flush),
drivers/usb/cdns3/cdns3-trace.h
33
__field(u8, flush)
drivers/usb/cdns3/cdns3-trace.h
38
__entry->flush = flush;
drivers/usb/cdns3/cdns3-trace.h
40
TP_printk("Halt %s for %s: %s", __entry->flush ? " and flush" : "",
drivers/usb/class/cdc-wdm.c
810
.flush = wdm_flush,
drivers/usb/class/usbtmc.c
2283
.flush = usbtmc_flush,
drivers/usb/gadget/udc/cdns2/cdns2-trace.h
83
TP_PROTO(struct cdns2_endpoint *ep_priv, u8 halt, u8 flush),
drivers/usb/gadget/udc/cdns2/cdns2-trace.h
84
TP_ARGS(ep_priv, halt, flush),
drivers/usb/gadget/udc/cdns2/cdns2-trace.h
88
__field(u8, flush)
drivers/usb/gadget/udc/cdns2/cdns2-trace.h
93
__entry->flush = flush;
drivers/usb/gadget/udc/cdns2/cdns2-trace.h
95
TP_printk("Halt %s for %s: %s", __entry->flush ? " and flush" : "",
drivers/usb/usb-skeleton.c
474
.flush = skel_flush,
drivers/vdpa/vdpa.c
1103
u8 flush;
drivers/vdpa/vdpa.c
1105
flush = ((features & BIT_ULL(VIRTIO_BLK_F_FLUSH)) == 0) ? 0 : 1;
drivers/vdpa/vdpa.c
1106
if (nla_put_u8(msg, VDPA_ATTR_DEV_BLK_FLUSH, flush))
drivers/vhost/vhost.c
282
struct vhost_flush_struct flush;
drivers/vhost/vhost.c
287
init_completion(&flush.wait_event);
drivers/vhost/vhost.c
288
vhost_work_init(&flush.work, vhost_flush_work);
drivers/vhost/vhost.c
290
vhost_worker_queue(worker, &flush.work);
drivers/vhost/vhost.c
296
wait_for_completion(&flush.wait_event);
fs/btrfs/block-rsv.c
214
enum btrfs_reserve_flush_enum flush)
fs/btrfs/block-rsv.c
221
ret = btrfs_reserve_metadata_bytes(block_rsv->space_info, num_bytes, flush);
fs/btrfs/block-rsv.c
244
enum btrfs_reserve_flush_enum flush)
fs/btrfs/block-rsv.c
261
ret = btrfs_reserve_metadata_bytes(block_rsv->space_info, num_bytes, flush);
fs/btrfs/block-rsv.h
74
enum btrfs_reserve_flush_enum flush);
fs/btrfs/block-rsv.h
78
enum btrfs_reserve_flush_enum flush);
fs/btrfs/delalloc-space.c
130
enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA;
fs/btrfs/delalloc-space.c
136
flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE;
fs/btrfs/delalloc-space.c
138
return btrfs_reserve_data_bytes(data_sinfo_for_inode(inode), bytes, flush);
fs/btrfs/delalloc-space.c
146
enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_DATA;
fs/btrfs/delalloc-space.c
155
flush = BTRFS_RESERVE_NO_FLUSH;
fs/btrfs/delalloc-space.c
157
flush = BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE;
fs/btrfs/delalloc-space.c
159
ret = btrfs_reserve_data_bytes(data_sinfo_for_inode(inode), len, flush);
fs/btrfs/delalloc-space.c
323
enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_FLUSH_ALL;
fs/btrfs/delalloc-space.c
336
flush = BTRFS_RESERVE_NO_FLUSH;
fs/btrfs/delalloc-space.c
339
flush = BTRFS_RESERVE_FLUSH_LIMIT;
fs/btrfs/delalloc-space.c
362
flush);
fs/btrfs/delayed-ref.c
211
enum btrfs_reserve_flush_enum flush)
fs/btrfs/delayed-ref.c
231
ret = btrfs_reserve_metadata_bytes(space_info, num_bytes, flush);
fs/btrfs/delayed-ref.h
418
enum btrfs_reserve_flush_enum flush);
fs/btrfs/space-info.c
1178
space_info->flush = false;
fs/btrfs/space-info.c
1190
space_info->flush = false;
fs/btrfs/space-info.c
1232
space_info->flush = false;
fs/btrfs/space-info.c
1289
enum btrfs_flush_state flush;
fs/btrfs/space-info.c
1325
flush = FLUSH_DELALLOC;
fs/btrfs/space-info.c
1330
flush = COMMIT_TRANS;
fs/btrfs/space-info.c
1334
flush = FLUSH_DELAYED_ITEMS_NR;
fs/btrfs/space-info.c
1337
flush = FLUSH_DELAYED_REFS_NR;
fs/btrfs/space-info.c
1350
flush_space(space_info, to_reclaim, flush, true);
fs/btrfs/space-info.c
1415
space_info->flush = false;
fs/btrfs/space-info.c
1426
space_info->flush = false;
fs/btrfs/space-info.c
1443
space_info->flush = false;
fs/btrfs/space-info.c
1460
space_info->flush = false;
fs/btrfs/space-info.c
1476
space_info->flush = false;
fs/btrfs/space-info.c
1655
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
1659
switch (flush) {
fs/btrfs/space-info.c
1679
ASSERT(0, "flush=%d", flush);
fs/btrfs/space-info.c
1694
orig_bytes, start_ns, flush, ticket->error);
fs/btrfs/space-info.c
1702
static inline bool is_normal_flushing(enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
1704
return (flush == BTRFS_RESERVE_FLUSH_ALL) ||
fs/btrfs/space-info.c
1705
(flush == BTRFS_RESERVE_FLUSH_ALL_STEAL);
fs/btrfs/space-info.c
1726
static inline bool can_steal(enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
1728
return (flush == BTRFS_RESERVE_FLUSH_ALL_STEAL ||
fs/btrfs/space-info.c
1729
flush == BTRFS_RESERVE_FLUSH_EVICT);
fs/btrfs/space-info.c
1736
static inline bool can_ticket(enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
1738
return (flush != BTRFS_RESERVE_NO_FLUSH &&
fs/btrfs/space-info.c
1739
flush != BTRFS_RESERVE_FLUSH_EMERGENCY);
fs/btrfs/space-info.c
1757
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
1776
ASSERT(flush != BTRFS_RESERVE_FLUSH_ALL, "flush=%d", flush);
fs/btrfs/space-info.c
1777
ASSERT(flush != BTRFS_RESERVE_FLUSH_ALL_STEAL, "flush=%d", flush);
fs/btrfs/space-info.c
1778
ASSERT(flush != BTRFS_RESERVE_FLUSH_EVICT, "flush=%d", flush);
fs/btrfs/space-info.c
1781
if (flush == BTRFS_RESERVE_FLUSH_DATA)
fs/btrfs/space-info.c
1794
if (is_normal_flushing(flush) || (flush == BTRFS_RESERVE_NO_FLUSH))
fs/btrfs/space-info.c
1806
can_overcommit(space_info, used, orig_bytes, flush))) {
fs/btrfs/space-info.c
1816
if (ret && unlikely(flush == BTRFS_RESERVE_FLUSH_EMERGENCY)) {
fs/btrfs/space-info.c
1831
if (ret && can_ticket(flush)) {
fs/btrfs/space-info.c
1837
ticket.steal = can_steal(flush);
fs/btrfs/space-info.c
1841
if (flush == BTRFS_RESERVE_FLUSH_ALL ||
fs/btrfs/space-info.c
1842
flush == BTRFS_RESERVE_FLUSH_ALL_STEAL ||
fs/btrfs/space-info.c
1843
flush == BTRFS_RESERVE_FLUSH_DATA) {
fs/btrfs/space-info.c
1845
if (!space_info->flush) {
fs/btrfs/space-info.c
1855
space_info->flush = true;
fs/btrfs/space-info.c
1858
orig_bytes, flush,
fs/btrfs/space-info.c
1876
orig_bytes, flush, "preempt");
fs/btrfs/space-info.c
1882
if (!ret || !can_ticket(flush))
fs/btrfs/space-info.c
1885
return handle_reserve_ticket(space_info, &ticket, start_ns, orig_bytes, flush);
fs/btrfs/space-info.c
1904
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
1908
ret = reserve_bytes(space_info, orig_bytes, flush);
fs/btrfs/space-info.c
1932
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
1937
ASSERT(flush == BTRFS_RESERVE_FLUSH_DATA ||
fs/btrfs/space-info.c
1938
flush == BTRFS_RESERVE_FLUSH_FREE_SPACE_INODE ||
fs/btrfs/space-info.c
1939
flush == BTRFS_RESERVE_NO_FLUSH, "flush=%d", flush);
fs/btrfs/space-info.c
1940
ASSERT(!current->journal_info || flush != BTRFS_RESERVE_FLUSH_DATA,
fs/btrfs/space-info.c
1942
(unsigned long)current->journal_info, flush);
fs/btrfs/space-info.c
1944
ret = reserve_bytes(space_info, bytes, flush);
fs/btrfs/space-info.c
444
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
494
if (flush == BTRFS_RESERVE_FLUSH_ALL)
fs/btrfs/space-info.c
514
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
516
const u64 avail = calc_available_free_space(space_info, flush);
fs/btrfs/space-info.c
523
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
529
return check_can_overcommit(space_info, space_info_used_bytes, bytes, flush);
fs/btrfs/space-info.c
533
enum btrfs_reserve_flush_enum flush)
fs/btrfs/space-info.c
543
return check_can_overcommit(space_info, used, bytes, flush);
fs/btrfs/space-info.c
581
enum btrfs_reserve_flush_enum flush = BTRFS_RESERVE_NO_FLUSH;
fs/btrfs/space-info.c
597
can_overcommit(space_info, used, ticket->bytes, flush)) {
fs/btrfs/space-info.c
609
flush = BTRFS_RESERVE_FLUSH_ALL;
fs/btrfs/space-info.h
149
bool flush; /* set if we are trying to make space */
fs/btrfs/space-info.h
284
enum btrfs_reserve_flush_enum flush);
fs/btrfs/space-info.h
287
enum btrfs_reserve_flush_enum flush);
fs/btrfs/space-info.h
299
enum btrfs_reserve_flush_enum flush);
fs/btrfs/transaction.c
569
enum btrfs_reserve_flush_enum flush,
fs/btrfs/transaction.c
581
ret = btrfs_reserve_metadata_bytes(si, bytes, flush);
fs/btrfs/transaction.c
588
if (ret && flush == BTRFS_RESERVE_FLUSH_ALL_STEAL) {
fs/btrfs/transaction.c
591
ret = btrfs_reserve_metadata_bytes(si, bytes, flush);
fs/btrfs/transaction.c
599
unsigned int type, enum btrfs_reserve_flush_enum flush,
fs/btrfs/transaction.c
662
ret = btrfs_reserve_trans_metadata(fs_info, flush, num_bytes,
fs/btrfs/transaction.c
671
} else if (num_items == 0 && flush == BTRFS_RESERVE_FLUSH_ALL &&
fs/btrfs/transaction.c
680
ret = btrfs_delayed_refs_rsv_refill(fs_info, flush);
fs/btrfs/volumes.c
476
int flush, struct file **bdev_file,
fs/btrfs/volumes.c
492
if (flush)
fs/ecryptfs/file.c
312
if (lower_file->f_op->flush) {
fs/ecryptfs/file.c
314
return lower_file->f_op->flush(lower_file, td);
fs/ecryptfs/file.c
437
.flush = ecryptfs_flush,
fs/f2fs/file.c
5467
.flush = f2fs_file_flush,
fs/fat/fat.h
50
flush:1, /* write things quickly */
fs/fat/file.c
178
MSDOS_SB(inode->i_sb)->options.flush) {
fs/fat/inode.c
1014
if (opts->flush)
fs/fat/inode.c
1233
opts->flush = 1;
fs/fat/inode.c
1892
if (!MSDOS_SB(sb)->options.flush)
fs/fuse/file.c
3167
.flush = fuse_flush,
fs/gfs2/glops.c
117
goto flush;
fs/gfs2/glops.c
128
goto flush;
fs/gfs2/glops.c
133
flush:
fs/gfs2/lops.c
294
bio_end_io_t *end_io, bool flush)
fs/gfs2/lops.c
303
if (blkno == nblk && !flush)
fs/nfs/file.c
958
.flush = nfs_file_flush,
fs/nfs/nfs4file.c
441
.flush = nfs4_file_flush,
fs/nfsd/export.c
279
.flush = expkey_flush,
fs/open.c
1468
if (filp->f_op->flush)
fs/open.c
1469
retval = filp->f_op->flush(filp, id);
fs/orangefs/file.c
583
.flush = orangefs_flush,
fs/overlayfs/file.c
627
if (realfile->f_op->flush) {
fs/overlayfs/file.c
629
err = realfile->f_op->flush(realfile, id);
fs/overlayfs/file.c
645
.flush = ovl_flush,
fs/smb/client/cifsfs.c
1603
.flush = cifs_flush,
fs/smb/client/cifsfs.c
1623
.flush = cifs_flush,
fs/smb/client/cifsfs.c
1643
.flush = cifs_flush,
fs/smb/client/cifsfs.c
1661
.flush = cifs_flush,
fs/smb/client/cifsfs.c
1679
.flush = cifs_flush,
fs/smb/client/cifsfs.c
1697
.flush = cifs_flush,
fs/smb/client/cifsglob.h
487
int (*flush)(const unsigned int, struct cifs_tcon *, struct cifs_fid *);
fs/smb/client/file.c
2809
if (server->ops->flush == NULL) {
fs/smb/client/file.c
2817
rc = server->ops->flush(xid, tcon, &smbfile->fid);
fs/smb/client/file.c
2822
rc = server->ops->flush(xid, tcon, &smbfile->fid);
fs/smb/client/file.c
956
return tcon->ses->server->ops->flush(xid, tcon,
fs/smb/client/file.c
962
rc = tcon->ses->server->ops->flush(xid, tcon, &cfile->fid);
fs/smb/client/smb1ops.c
1446
.flush = cifs_flush_file,
fs/smb/client/smb2ops.c
5409
.flush = smb2_flush_file,
fs/smb/client/smb2ops.c
5512
.flush = smb2_flush_file,
fs/smb/client/smb2ops.c
5620
.flush = smb2_flush_file,
fs/smb/client/smb2ops.c
5736
.flush = smb2_flush_file,
fs/xfs/scrub/dir_repair.c
1299
bool flush;
fs/xfs/scrub/dir_repair.c
1308
flush = xrep_dir_want_flush_stashed(rd);
fs/xfs/scrub/dir_repair.c
1310
if (flush) {
fs/xfs/scrub/parent_repair.c
575
bool flush;
fs/xfs/scrub/parent_repair.c
584
flush = xrep_parent_want_flush_stashed(rp);
fs/xfs/scrub/parent_repair.c
586
if (flush) {
include/linux/decompress/bunzip2.h
7
long (*flush)(void*, unsigned long),
include/linux/decompress/generic.h
7
long (*flush)(void*, unsigned long),
include/linux/decompress/inflate.h
7
long (*flush)(void*, unsigned long),
include/linux/decompress/unlz4.h
7
long (*flush)(void*, unsigned long),
include/linux/decompress/unlzma.h
7
long (*flush)(void*, unsigned long),
include/linux/decompress/unlzo.h
7
long (*flush)(void*, unsigned long),
include/linux/decompress/unxz.h
14
long (*flush)(void *src, unsigned long size),
include/linux/decompress/unzstd.h
7
long (*flush)(void*, unsigned long),
include/linux/dm-dirty-log.h
78
int (*flush)(struct dm_dirty_log *log);
include/linux/filter.h
1788
void *buf, unsigned long len, bool flush);
include/linux/filter.h
1823
unsigned long len, bool flush)
include/linux/fs.h
1942
int (*flush) (struct file *, fl_owner_t id);
include/linux/hsi/hsi.h
231
int (*flush)(struct hsi_client *cl);
include/linux/hsi/hsi.h
374
return hsi_get_port(cl)->flush(cl);
include/linux/input.h
187
int (*flush)(struct input_dev *dev, struct file *file);
include/linux/kgdb.h
292
void (*flush) (void);
include/linux/libnvdimm.h
142
int (*flush)(struct nd_region *nd_region, struct bio *bio);
include/linux/mailbox_controller.h
48
int (*flush)(struct mbox_chan *chan, unsigned long timeout);
include/linux/mlx5/device.h
837
u8 flush:1;
include/linux/mtd/blktrans.h
58
int (*flush)(struct mtd_blktrans_dev *dev);
include/linux/netfilter/ipset/ip_set.h
174
void (*flush)(struct ip_set *set);
include/linux/pagemap.h
46
int filemap_invalidate_inode(struct inode *inode, bool flush,
include/linux/sunrpc/cache.h
96
void (*flush)(void);
include/linux/swap.h
316
bool flush);
include/linux/zlib.h
223
extern int zlib_deflate (z_streamp strm, int flush);
include/linux/zlib.h
343
extern int zlib_inflate (z_streamp strm, int flush);
include/media/dvb_vb2.h
124
#define dvb_vb2_fill_buffer(ctx, file, wait, flags, flush) (0)
include/media/dvb_vb2.h
173
bool flush);
include/net/bluetooth/hci_core.h
647
int (*flush)(struct hci_dev *hdev);
include/net/gro.h
114
NAPI_GRO_CB(skb)->flush |= 1;
include/net/gro.h
129
NAPI_GRO_CB(skb)->flush |= 1;
include/net/gro.h
365
static inline void skb_gro_flush_final(struct sk_buff *skb, struct sk_buff *pp, int flush)
include/net/gro.h
368
NAPI_GRO_CB(skb)->flush |= flush;
include/net/gro.h
372
int flush,
include/net/gro.h
376
NAPI_GRO_CB(skb)->flush |= flush;
include/net/gro.h
382
static inline void skb_gro_flush_final(struct sk_buff *skb, struct sk_buff *pp, int flush)
include/net/gro.h
384
NAPI_GRO_CB(skb)->flush |= flush;
include/net/gro.h
388
int flush,
include/net/gro.h
391
NAPI_GRO_CB(skb)->flush |= flush;
include/net/gro.h
40
u16 flush;
include/net/gro.h
414
NAPI_GRO_CB(skb)->flush |= 1, NULL : \
include/net/gro.h
492
int flush;
include/net/gro.h
494
flush = __gro_receive_network_flush(th, th2, p, off - NAPI_GRO_CB(p)->network_offset, false);
include/net/gro.h
496
flush |= __gro_receive_network_flush(th, th2, p, off - NAPI_GRO_CB(p)->inner_network_offset, true);
include/net/gro.h
498
return flush;
include/net/inet_common.h
79
NAPI_GRO_CB(skb)->flush |= 1, NULL : \
include/net/libeth/xdp.h
1056
void (*finalize)(void *xdpsq, bool sent, bool flush))
include/net/libeth/xdp.h
1546
#define libeth_xdp_finalize_rx(bq, flush, finalize) \
include/net/libeth/xdp.h
1547
__libeth_xdp_finalize_rx(bq, 0, flush, finalize)
include/net/libeth/xdp.h
1553
void (*finalize)(void *xdpsq, bool sent, bool flush))
include/net/libeth/xdp.h
1639
#define LIBETH_XDP_DEFINE_RUN_PROG(name, flush) \
include/net/libeth/xdp.h
1640
bool __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, xdp)
include/net/libeth/xdp.h
1642
#define __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, pfx) \
include/net/libeth/xdp.h
1645
return libeth_##pfx##_run_prog(xdp, bq, flush); \
include/net/libeth/xdp.h
1673
#define LIBETH_XDP_DEFINE_RUN(name, run, flush, populate) \
include/net/libeth/xdp.h
1674
__LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, XDP)
include/net/libeth/xdp.h
1676
#define __LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, pfx) \
include/net/libeth/xdp.h
1677
LIBETH_##pfx##_DEFINE_RUN_PROG(static run, flush); \
include/net/libeth/xdp.h
1686
#define LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize) \
include/net/libeth/xdp.h
1687
__LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, xdp)
include/net/libeth/xdp.h
1689
#define __LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, pfx) \
include/net/libeth/xdp.h
1692
libeth_##pfx##_finalize_rx(bq, flush, finalize); \
include/net/libeth/xsk.h
290
void (*finalize)(void *xdpsq, bool sent, bool flush))
include/net/libeth/xsk.h
513
#define libeth_xsk_finalize_rx(bq, flush, finalize) \
include/net/libeth/xsk.h
514
__libeth_xdp_finalize_rx(bq, LIBETH_XDP_TX_XSK, flush, finalize)
include/net/libeth/xsk.h
556
#define LIBETH_XSK_DEFINE_RUN_PROG(name, flush) \
include/net/libeth/xsk.h
557
u32 __LIBETH_XDP_DEFINE_RUN_PROG(name, flush, xsk)
include/net/libeth/xsk.h
575
#define LIBETH_XSK_DEFINE_RUN(name, run, flush, populate) \
include/net/libeth/xsk.h
576
__LIBETH_XDP_DEFINE_RUN(name, run, flush, populate, XSK)
include/net/libeth/xsk.h
584
#define LIBETH_XSK_DEFINE_FINALIZE(name, flush, finalize) \
include/net/libeth/xsk.h
585
__LIBETH_XDP_DEFINE_FINALIZE(name, flush, finalize, xsk)
include/net/mac80211.h
4752
void (*flush)(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
include/net/netfilter/nf_tables.h
484
void (*flush)(const struct net *net,
include/trace/events/btrfs.h
1145
int flush, const char *reason),
include/trace/events/btrfs.h
1147
TP_ARGS(fs_info, flags, bytes, flush, reason),
include/trace/events/btrfs.h
1152
__field( int, flush )
include/trace/events/btrfs.h
1159
__entry->flush = flush;
include/trace/events/btrfs.h
1164
__get_str(reason), __entry->flush,
include/trace/events/btrfs.h
1165
__print_symbolic(__entry->flush, FLUSH_ACTIONS),
include/trace/events/btrfs.h
2237
u64 start_ns, int flush, int error),
include/trace/events/btrfs.h
2239
TP_ARGS(fs_info, flags, bytes, start_ns, flush, error),
include/trace/events/btrfs.h
2245
__field( int, flush )
include/trace/events/btrfs.h
2253
__entry->flush = flush;
include/trace/events/btrfs.h
2260
__print_symbolic(__entry->flush, FLUSH_ACTIONS),
include/uapi/drm/radeon_drm.h
594
int flush;
include/uapi/rdma/rdma_user_rxe.h
91
} flush;
io_uring/openclose.c
256
if (file->f_op->flush && (issue_flags & IO_URING_F_NONBLOCK)) {
ipc/mqueue.c
1595
.flush = mqueue_flush_file,
kernel/acct.c
209
if (file->f_op->flush)
kernel/acct.c
210
file->f_op->flush(file, NULL);
kernel/debug/gdbstub.c
1156
if (dbg_io_ops->flush)
kernel/debug/gdbstub.c
1157
dbg_io_ops->flush();
kernel/debug/gdbstub.c
133
if (dbg_io_ops->flush)
kernel/debug/gdbstub.c
134
dbg_io_ops->flush();
kernel/debug/gdbstub.c
167
if (dbg_io_ops->flush)
kernel/debug/gdbstub.c
168
dbg_io_ops->flush();
kernel/debug/gdbstub.c
188
if (dbg_io_ops->flush)
kernel/debug/gdbstub.c
189
dbg_io_ops->flush();
kernel/gcov/clang.c
82
void llvm_gcov_init(llvm_gcov_callback writeout, llvm_gcov_callback flush)
kernel/signal.c
875
sigset_t flush;
kernel/signal.c
888
siginitset(&flush, sigmask(SIGCONT));
kernel/signal.c
889
flush_sigqueue_mask(p, &flush, &signal->shared_pending);
kernel/signal.c
891
flush_sigqueue_mask(p, &flush, &t->pending);
kernel/signal.c
897
siginitset(&flush, SIG_KERNEL_STOP_MASK);
kernel/signal.c
898
flush_sigqueue_mask(p, &flush, &signal->shared_pending);
kernel/signal.c
900
flush_sigqueue_mask(p, &flush, &t->pending);
kernel/trace/trace.c
8297
.flush = tracing_buffers_flush,
lib/decompress_bunzip2.c
682
long (*flush)(void*, unsigned long),
lib/decompress_bunzip2.c
691
if (flush)
lib/decompress_bunzip2.c
713
if (!flush)
lib/decompress_bunzip2.c
716
if (i != flush(outbuf, i)) {
lib/decompress_bunzip2.c
742
if (flush)
lib/decompress_bunzip2.c
750
long (*flush)(void*, unsigned long),
lib/decompress_bunzip2.c
755
return bunzip2(buf, len - 4, fill, flush, outbuf, pos, error);
lib/decompress_inflate.c
138
if (!flush) {
lib/decompress_inflate.c
159
if (flush && strm->next_out > out_buf) {
lib/decompress_inflate.c
161
if (l != flush(out_buf, l)) {
lib/decompress_inflate.c
193
if (flush)
lib/decompress_inflate.c
202
long (*flush)(void*, unsigned long),
lib/decompress_inflate.c
207
return __gunzip(buf, len, fill, flush, out_buf, 0, pos, error);
lib/decompress_inflate.c
212
long (*flush)(void*, unsigned long),
lib/decompress_inflate.c
217
return __gunzip(buf, len, fill, flush, out_buf, out_len, pos, error);
lib/decompress_inflate.c
44
long (*flush)(void*, unsigned long),
lib/decompress_inflate.c
53
if (flush) {
lib/decompress_inflate.c
82
strm->workspace = malloc(flush ? zlib_inflate_workspacesize() :
lib/decompress_unlz4.c
174
if (flush && flush(outp, dest_len) != dest_len)
lib/decompress_unlz4.c
209
long (*flush)(void*, unsigned long),
lib/decompress_unlz4.c
215
return unlz4(buf, in_len - 4, fill, flush, output, posp, error);
lib/decompress_unlz4.c
33
long (*flush)(void *, unsigned long),
lib/decompress_unlz4.c
52
} else if (!flush) {
lib/decompress_unlzma.c
285
long (*flush)(void*, unsigned long);
lib/decompress_unlzma.c
303
if (!wr->flush) {
lib/decompress_unlzma.c
321
if (wr->flush && wr->buffer_pos == wr->header->dict_size) {
lib/decompress_unlzma.c
324
if (wr->flush((char *)wr->buffer, wr->header->dict_size)
lib/decompress_unlzma.c
541
long (*flush)(void*, unsigned long),
lib/decompress_unlzma.c
575
wr.flush = flush;
lib/decompress_unlzma.c
657
if (!wr.flush || wr.flush(wr.buffer, wr.buffer_pos) == wr.buffer_pos)
lib/decompress_unlzma.c
674
long (*flush)(void*, unsigned long),
lib/decompress_unlzma.c
679
return unlzma(buf, in_len - 4, fill, flush, output, posp, error);
lib/decompress_unlzo.c
112
} else if (!flush) {
lib/decompress_unlzo.c
243
if (flush && flush(out_buf, dst_len) != dst_len)
lib/decompress_unlzo.c
279
long (*flush)(void*, unsigned long),
lib/decompress_unlzo.c
284
return unlzo(buf, len, fill, flush, out_buf, pos, error);
lib/decompress_unlzo.c
99
long (*flush)(void *, unsigned long),
lib/decompress_unxz.c
260
long (*flush)(void *src, unsigned long size),
lib/decompress_unxz.c
276
if (fill == NULL && flush == NULL)
lib/decompress_unxz.c
284
if (flush == NULL) {
lib/decompress_unxz.c
306
if (fill == NULL && flush == NULL) {
lib/decompress_unxz.c
332
if (flush != NULL && (b.out_pos == b.out_size
lib/decompress_unxz.c
339
if (flush(b.out, b.out_pos) != (long)b.out_pos)
lib/decompress_unxz.c
349
if (flush != NULL)
lib/decompress_unxz.c
389
if (flush != NULL)
lib/decompress_unxz.c
407
long (*flush)(void *src, unsigned long size),
lib/decompress_unxz.c
412
return unxz(in, in_size, fill, flush, out, in_used, error);
lib/decompress_unzstd.c
167
long (*flush)(void*, unsigned long),
lib/decompress_unzstd.c
191
if (fill == NULL && flush == NULL)
lib/decompress_unzstd.c
227
if (flush != NULL) {
lib/decompress_unzstd.c
308
if (flush != NULL && out.pos > 0) {
lib/decompress_unzstd.c
309
if (out.pos != flush(out.dst, out.pos)) {
lib/decompress_unzstd.c
335
long (*flush)(void*, unsigned long),
lib/decompress_unzstd.c
340
return __unzstd(buf, len, fill, flush, out_buf, 0, pos, error);
lib/decompress_unzstd.c
345
long (*flush)(void*, unsigned long),
lib/decompress_unzstd.c
350
return __unzstd(buf, len, fill, flush, out_buf, out_len, pos, error);
lib/zlib_deflate/deflate.c
1004
FLUSH_BLOCK(s, flush == Z_FINISH);
lib/zlib_deflate/deflate.c
1005
return flush == Z_FINISH ? finish_done : block_done;
lib/zlib_deflate/deflate.c
1015
int flush
lib/zlib_deflate/deflate.c
1030
if (s->lookahead < MIN_LOOKAHEAD && flush == Z_NO_FLUSH) {
lib/zlib_deflate/deflate.c
1120
Assert (flush != Z_NO_FLUSH, "no flush?");
lib/zlib_deflate/deflate.c
1126
FLUSH_BLOCK(s, flush == Z_FINISH);
lib/zlib_deflate/deflate.c
1127
return flush == Z_FINISH ? finish_done : block_done;
lib/zlib_deflate/deflate.c
328
int flush
lib/zlib_deflate/deflate.c
335
flush > Z_FINISH || flush < 0) {
lib/zlib_deflate/deflate.c
341
(s->status == FINISH_STATE && flush != Z_FINISH)) {
lib/zlib_deflate/deflate.c
348
s->last_flush = flush;
lib/zlib_deflate/deflate.c
390
} else if (strm->avail_in == 0 && flush <= old_flush &&
lib/zlib_deflate/deflate.c
391
flush != Z_FINISH) {
lib/zlib_deflate/deflate.c
403
(flush != Z_NO_FLUSH && s->status != FINISH_STATE)) {
lib/zlib_deflate/deflate.c
406
bstate = DEFLATE_HOOK(strm, flush, &bstate) ? bstate :
lib/zlib_deflate/deflate.c
407
(*(configuration_table[s->level].func))(s, flush);
lib/zlib_deflate/deflate.c
426
if (flush == Z_PARTIAL_FLUSH) {
lib/zlib_deflate/deflate.c
428
} else if (flush == Z_PACKET_FLUSH) {
lib/zlib_deflate/deflate.c
437
if (flush == Z_FULL_FLUSH) {
lib/zlib_deflate/deflate.c
450
if (flush != Z_FINISH) return Z_OK;
lib/zlib_deflate/deflate.c
60
#define DEFLATE_HOOK(strm, flush, bstate) 0
lib/zlib_deflate/deflate.c
69
typedef block_state (*compress_func) (deflate_state *s, int flush);
lib/zlib_deflate/deflate.c
73
static block_state deflate_stored (deflate_state *s, int flush);
lib/zlib_deflate/deflate.c
74
static block_state deflate_fast (deflate_state *s, int flush);
lib/zlib_deflate/deflate.c
75
static block_state deflate_slow (deflate_state *s, int flush);
lib/zlib_deflate/deflate.c
862
int flush
lib/zlib_deflate/deflate.c
884
if (s->lookahead == 0 && flush == Z_NO_FLUSH) return need_more;
lib/zlib_deflate/deflate.c
908
FLUSH_BLOCK(s, flush == Z_FINISH);
lib/zlib_deflate/deflate.c
909
return flush == Z_FINISH ? finish_done : block_done;
lib/zlib_deflate/deflate.c
921
int flush
lib/zlib_deflate/deflate.c
935
if (s->lookahead < MIN_LOOKAHEAD && flush == Z_NO_FLUSH) {
lib/zlib_dfltcc/dfltcc_deflate.c
121
int flush,
lib/zlib_dfltcc/dfltcc_deflate.c
136
if (flush == Z_FULL_FLUSH)
lib/zlib_dfltcc/dfltcc_deflate.c
144
no_flush = flush == Z_NO_FLUSH;
lib/zlib_dfltcc/dfltcc_deflate.c
159
if (flush == Z_FINISH)
lib/zlib_dfltcc/dfltcc_deflate.c
162
if (flush == Z_FULL_FLUSH)
lib/zlib_dfltcc/dfltcc_deflate.c
217
need_empty_block = flush == Z_FINISH && param->bcf && !param->bhf;
lib/zlib_dfltcc/dfltcc_deflate.c
225
if (flush == Z_FINISH && !param->bcf)
lib/zlib_dfltcc/dfltcc_deflate.c
292
if (flush == Z_FINISH) {
lib/zlib_dfltcc/dfltcc_deflate.c
301
if (flush == Z_FULL_FLUSH)
lib/zlib_dfltcc/dfltcc_deflate.c
303
*result = flush == Z_NO_FLUSH ? need_more : block_done;
lib/zlib_dfltcc/dfltcc_deflate.h
10
int flush,
lib/zlib_dfltcc/dfltcc_inflate.c
89
int flush,
lib/zlib_dfltcc/dfltcc_inflate.c
98
if (flush == Z_BLOCK || flush == Z_PACKET_FLUSH) {
lib/zlib_dfltcc/dfltcc_inflate.h
16
int flush, int *ret);
lib/zlib_dfltcc/dfltcc_inflate.h
20
#define INFLATE_TYPEDO_HOOK(strm, flush) \
lib/zlib_dfltcc/dfltcc_inflate.h
25
action = dfltcc_inflate((strm), (flush), &ret); \
lib/zlib_inflate/inflate.c
23
#define INFLATE_TYPEDO_HOOK(strm, flush) do {} while (0)
lib/zlib_inflate/inflate.c
331
int zlib_inflate(z_streamp strm, int flush)
lib/zlib_inflate/inflate.c
409
if (flush == Z_BLOCK) goto inf_leave;
lib/zlib_inflate/inflate.c
412
INFLATE_TYPEDO_HOOK(strm, flush);
lib/zlib_inflate/inflate.c
760
if (flush == Z_PACKET_FLUSH && ret == Z_OK &&
lib/zlib_inflate/inflate.c
764
if (((in == 0 && out == 0) || flush == Z_FINISH) && ret == Z_OK)
mm/filemap.c
4522
int filemap_invalidate_inode(struct inode *inode, bool flush,
mm/filemap.c
4542
if (flush)
mm/memcontrol.c
2000
bool flush = false;
mm/memcontrol.c
2011
flush = true;
mm/memcontrol.c
2016
return flush;
mm/memcontrol.c
3067
bool flush = false;
mm/memcontrol.c
3073
flush = true;
mm/memcontrol.c
3077
return flush;
mm/migrate_device.c
1067
flush = true;
mm/migrate_device.c
1084
if (flush) {
mm/migrate_device.c
807
bool flush = false;
mm/migrate_device.c
871
flush = true;
mm/migrate_device.c
881
if (flush) {
mm/migrate_device.c
977
bool flush = false;
mm/slub.c
5812
goto flush;
mm/slub.c
5816
goto flush;
mm/slub.c
5836
flush:
mm/vmalloc.c
2949
static void _vm_unmap_aliases(unsigned long start, unsigned long end, int flush)
mm/vmalloc.c
2988
flush = 1;
mm/vmalloc.c
2996
if (!__purge_vmap_area_lazy(start, end, false) && flush)
mm/workingset.c
419
bool flush)
mm/workingset.c
476
if (flush)
net/8021q/vlan_core.c
473
int flush = 1;
net/8021q/vlan_core.c
489
flush = 0;
net/8021q/vlan_core.c
510
skb_gro_flush_final(skb, pp, flush);
net/bluetooth/hci_core.c
570
if (hdev->flush)
net/bluetooth/hci_core.c
571
hdev->flush(hdev);
net/bluetooth/hci_sync.c
5227
if (hdev->flush)
net/bluetooth/hci_sync.c
5228
hdev->flush(hdev);
net/bluetooth/hci_sync.c
5380
if (hdev->flush)
net/bluetooth/hci_sync.c
5381
hdev->flush(hdev);
net/bridge/br_sysfs_br.c
362
static DEVICE_ATTR_WO(flush);
net/bridge/br_sysfs_if.c
181
static BRPORT_ATTR(flush, 0200, NULL, store_flush);
net/core/filter.c
4013
void *buf, unsigned long len, bool flush)
net/core/filter.c
4022
src = flush ? buf : xdp->data + off;
net/core/filter.c
4023
dst = flush ? xdp->data + off : buf;
net/core/filter.c
4040
src = flush ? buf : ptr_buf + copy_off;
net/core/filter.c
4041
dst = flush ? ptr_buf + copy_off : buf;
net/core/gro.c
113
NAPI_GRO_CB(skb)->flush))
net/core/gro.c
494
NAPI_GRO_CB(skb)->flush = skb_has_frag_list(skb);
net/core/gro.c
501
NAPI_GRO_CB(skb)->flush = 1;
net/core/gro.c
538
if (NAPI_GRO_CB(skb)->flush)
net/core/netpoll.c
765
goto flush;
net/core/netpoll.c
776
flush:
net/ethernet/eth.c
398
int flush = 1;
net/ethernet/eth.c
406
flush = 0;
net/ethernet/eth.c
423
flush = 1;
net/ethernet/eth.c
435
skb_gro_flush_final(skb, pp, flush);
net/ipv4/af_inet.c
1482
int flush = 1;
net/ipv4/af_inet.c
1507
flush = (u16)((ntohl(*(__be32 *)iph) ^ skb_gro_len(skb)) | (ntohl(*(__be32 *)&iph->id) & ~IP_DF));
net/ipv4/af_inet.c
1529
NAPI_GRO_CB(skb)->flush |= flush;
net/ipv4/af_inet.c
1542
skb_gro_flush_final(skb, pp, flush);
net/ipv4/af_inet.c
1551
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/esp4_offload.c
99
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/fou_core.c
336
int flush = 1;
net/ipv4/fou_core.c
460
flush = 0;
net/ipv4/fou_core.c
463
skb_gro_flush_final_remcsum(skb, pp, flush, &grc);
net/ipv4/gre_offload.c
130
int flush = 1;
net/ipv4/gre_offload.c
184
if ((greh->flags & GRE_CSUM) && !NAPI_GRO_CB(skb)->flush) {
net/ipv4/gre_offload.c
228
flush = 0;
net/ipv4/gre_offload.c
231
skb_gro_flush_final(skb, pp, flush);
net/ipv4/tcp_offload.c
296
int flush = 1;
net/ipv4/tcp_offload.c
307
flush = (__force int)((flags ^ tcp_flag_word(th2)) &
net/ipv4/tcp_offload.c
309
flush |= (__force int)(th->ack_seq ^ th2->ack_seq);
net/ipv4/tcp_offload.c
311
flush |= *(u32 *)((u8 *)th + i) ^
net/ipv4/tcp_offload.c
314
flush |= gro_receive_network_flush(th, th2, p);
net/ipv4/tcp_offload.c
323
flush |= (mss != skb_shinfo(skb)->gso_size);
net/ipv4/tcp_offload.c
325
flush |= (len - 1) >= mss;
net/ipv4/tcp_offload.c
327
flush |= (ntohl(th2->seq) + skb_gro_len(p)) ^ ntohl(th->seq);
net/ipv4/tcp_offload.c
328
flush |= skb_cmp_decrypted(p, skb);
net/ipv4/tcp_offload.c
331
flush |= (__force int)(flags ^ tcp_flag_word(th2));
net/ipv4/tcp_offload.c
332
flush |= skb->ip_summed != p->ip_summed;
net/ipv4/tcp_offload.c
333
flush |= skb->csum_level != p->csum_level;
net/ipv4/tcp_offload.c
334
flush |= NAPI_GRO_CB(p)->count >= 64;
net/ipv4/tcp_offload.c
337
if (flush || skb_gro_receive_list(p, skb))
net/ipv4/tcp_offload.c
343
if (flush || skb_gro_receive(p, skb)) {
net/ipv4/tcp_offload.c
353
flush = len != NAPI_GRO_CB(skb)->count * skb_shinfo(skb)->gso_size;
net/ipv4/tcp_offload.c
355
flush = len < mss;
net/ipv4/tcp_offload.c
357
flush |= (__force int)(flags & (TCP_FLAG_URG | TCP_FLAG_PSH |
net/ipv4/tcp_offload.c
361
if (p && (!NAPI_GRO_CB(skb)->same_flow || flush))
net/ipv4/tcp_offload.c
364
NAPI_GRO_CB(skb)->flush |= (flush != 0);
net/ipv4/tcp_offload.c
424
if (!NAPI_GRO_CB(skb)->flush &&
net/ipv4/tcp_offload.c
427
goto flush;
net/ipv4/tcp_offload.c
431
goto flush;
net/ipv4/tcp_offload.c
437
flush:
net/ipv4/tcp_offload.c
438
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/udp_offload.c
153
NAPI_GRO_CB(skb)->flush |= 1;
net/ipv4/udp_offload.c
27
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/udp_offload.c
711
int flush;
net/ipv4/udp_offload.c
715
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/udp_offload.c
722
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/udp_offload.c
741
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/udp_offload.c
745
flush = gro_receive_network_flush(uh, uh2, p);
net/ipv4/udp_offload.c
753
if (ulen > ntohs(uh2->len) || flush) {
net/ipv4/udp_offload.c
758
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/udp_offload.c
763
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/udp_offload.c
794
int flush = 1;
net/ipv4/udp_offload.c
829
flush = 0;
net/ipv4/udp_offload.c
852
skb_gro_flush_final(skb, pp, flush);
net/ipv4/udp_offload.c
884
goto flush;
net/ipv4/udp_offload.c
887
if (NAPI_GRO_CB(skb)->flush)
net/ipv4/udp_offload.c
892
goto flush;
net/ipv4/udp_offload.c
903
flush:
net/ipv4/udp_offload.c
904
NAPI_GRO_CB(skb)->flush = 1;
net/ipv4/xfrm4_input.c
221
NAPI_GRO_CB(skb)->flush = 1;
net/ipv6/esp6_offload.c
128
NAPI_GRO_CB(skb)->flush = 1;
net/ipv6/ip6_offload.c
213
u16 flush = 1;
net/ipv6/ip6_offload.c
224
flush += ntohs(iph->payload_len) != skb->len - hlen;
net/ipv6/ip6_offload.c
244
flush--;
net/ipv6/ip6_offload.c
278
NAPI_GRO_CB(skb)->flush |= flush;
net/ipv6/ip6_offload.c
283
flush = 1;
net/ipv6/ip6_offload.c
296
skb_gro_flush_final(skb, pp, flush);
net/ipv6/ip6_offload.c
307
NAPI_GRO_CB(skb)->flush = 1;
net/ipv6/ip6_offload.c
322
NAPI_GRO_CB(skb)->flush = 1;
net/ipv6/tcpv6_offload.c
51
if (!NAPI_GRO_CB(skb)->flush &&
net/ipv6/tcpv6_offload.c
54
goto flush;
net/ipv6/tcpv6_offload.c
58
goto flush;
net/ipv6/tcpv6_offload.c
65
flush:
net/ipv6/tcpv6_offload.c
66
NAPI_GRO_CB(skb)->flush = 1;
net/ipv6/udp_offload.c
142
goto flush;
net/ipv6/udp_offload.c
145
if (NAPI_GRO_CB(skb)->flush)
net/ipv6/udp_offload.c
150
goto flush;
net/ipv6/udp_offload.c
162
flush:
net/ipv6/udp_offload.c
163
NAPI_GRO_CB(skb)->flush = 1;
net/ipv6/xfrm6_input.c
221
NAPI_GRO_CB(skb)->flush = 1;
net/mac80211/driver-ops.h
733
if (local->ops->flush)
net/mac80211/driver-ops.h
734
local->ops->flush(&local->hw, vif, queues, drop);
net/mac80211/mesh_plink.c
873
bool flush = false;
net/mac80211/mesh_plink.c
962
flush = true;
net/mac80211/mesh_plink.c
994
if (flush)
net/mac80211/scan.c
1082
if (local->ops->flush) {
net/mac80211/util.c
689
if (!local->ops->flush && !drop)
net/mac80211/util.c
716
if (local->ops->flush)
net/netfilter/ipset/ip_set_bitmap_gen.h
309
.flush = mtype_flush,
net/netfilter/ipset/ip_set_core.c
1302
set->variant->flush(set);
net/netfilter/ipset/ip_set_hash_gen.h
1445
.flush = mtype_flush,
net/netfilter/ipset/ip_set_list_set.c
564
.flush = list_set_flush,
net/netfilter/nf_conntrack_sip.c
1358
goto flush;
net/netfilter/nf_conntrack_sip.c
1400
flush:
net/netfilter/nf_tables_api.c
7843
set->ops->flush(ctx->net, set, elem_priv);
net/netfilter/nft_set_bitmap.c
315
.flush = nft_bitmap_flush,
net/netfilter/nft_set_hash.c
857
.flush = nft_rhash_flush,
net/netfilter/nft_set_hash.c
878
.flush = nft_hash_flush,
net/netfilter/nft_set_hash.c
897
.flush = nft_hash_flush,
net/netfilter/nft_set_pipapo.c
2404
.flush = nft_pipapo_flush,
net/netfilter/nft_set_pipapo.c
2429
.flush = nft_pipapo_flush,
net/netfilter/nft_set_rbtree.c
1193
.flush = nft_rbtree_flush,
net/netfilter/nft_set_rbtree.c
600
static int nft_array_may_resize(const struct nft_set *set, bool flush)
net/netfilter/nft_set_rbtree.c
628
if (flush) {
net/sunrpc/cache.c
1566
if (cd->flush)
net/sunrpc/cache.c
1567
cd->flush();
security/tomoyo/common.c
2060
goto flush;
security/tomoyo/common.c
2068
goto flush;
security/tomoyo/common.c
2079
goto flush;
security/tomoyo/common.c
2085
goto flush;
security/tomoyo/common.c
2103
goto flush;
security/tomoyo/common.c
2118
flush:
sound/core/seq/seq_ump_convert.c
806
bool flush)
sound/core/seq/seq_ump_convert.c
811
if (!flush && (!cc->cc_data_lsb_set || !cc->cc_data_msb_set))
sound/core/ump_convert.c
301
bool flush)
sound/core/ump_convert.c
306
if (!flush && (!cc->cc_data_lsb_set || !cc->cc_data_msb_set))
sound/soc/samsung/i2s.c
479
static inline void i2s_fifo(struct i2s_dai *i2s, u32 flush)
sound/soc/samsung/i2s.c
493
writel(readl(fic) | flush, fic);
sound/soc/samsung/i2s.c
500
writel(readl(fic) & ~flush, fic);
tools/lib/bpf/netlink.c
645
const bool flush);
tools/lib/bpf/netlink.c
810
const bool flush)
tools/lib/bpf/netlink.c
835
if (!flush) {
tools/lib/bpf/netlink.c
850
if (!flush) {
tools/lib/bpf/netlink.c
860
if (!flush) {
tools/lib/perf/include/internal/mmap.h
34
u64 flush;
tools/lib/perf/mmap.c
157
if ((md->end - md->start) < md->flush)
tools/perf/builtin-record.c
1662
u64 flush = 0;
tools/perf/builtin-record.c
1668
flush = map->core.flush;
tools/perf/builtin-record.c
1669
map->core.flush = 1;
tools/perf/builtin-record.c
1674
map->core.flush = flush;
tools/perf/builtin-record.c
1682
map->core.flush = flush;
tools/perf/builtin-record.c
1688
map->core.flush = flush;
tools/perf/builtin-trace.c
4192
u64 flush = trace->oe.last - NSEC_PER_SEC;
tools/perf/builtin-trace.c
4195
if (first && first < flush)
tools/perf/builtin-trace.c
4196
return ordered_events__flush_time(&trace->oe.data, flush);
tools/perf/util/evlist.c
1013
bool auxtrace_overwrite, int nr_cblocks, int affinity, int flush,
tools/perf/util/evlist.c
1024
.flush = flush,
tools/perf/util/evlist.h
200
int affinity, int flush, int comp_level);
tools/perf/util/mmap.c
295
map->core.flush = mp->flush;
tools/perf/util/mmap.h
48
int nr_cblocks, affinity, flush, comp_level;
tools/testing/nvdimm/test/nfit.c
1888
struct acpi_nfit_flush_address *flush;
tools/testing/nvdimm/test/nfit.c
1895
+ struct_size(flush, hint_address, NUM_HINTS) * NUM_DCR
tools/testing/nvdimm/test/nfit.c
192
void **flush;
tools/testing/nvdimm/test/nfit.c
1926
t->flush[i] = test_alloc(t, max(PAGE_SIZE,
tools/testing/nvdimm/test/nfit.c
1929
if (!t->flush[i])
tools/testing/nvdimm/test/nfit.c
2004
struct acpi_nfit_flush_address *flush;
tools/testing/nvdimm/test/nfit.c
2506
flush = nfit_buf + offset;
tools/testing/nvdimm/test/nfit.c
2507
flush->header.type = ACPI_NFIT_TYPE_FLUSH_ADDRESS;
tools/testing/nvdimm/test/nfit.c
2508
flush->header.length = flush_hint_size;
tools/testing/nvdimm/test/nfit.c
2509
flush->device_handle = handle[0];
tools/testing/nvdimm/test/nfit.c
2510
flush->hint_count = NUM_HINTS;
tools/testing/nvdimm/test/nfit.c
2512
flush->hint_address[i] = t->flush_dma[0] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2513
offset += flush->header.length;
tools/testing/nvdimm/test/nfit.c
2516
flush = nfit_buf + offset;
tools/testing/nvdimm/test/nfit.c
2517
flush->header.type = ACPI_NFIT_TYPE_FLUSH_ADDRESS;
tools/testing/nvdimm/test/nfit.c
2518
flush->header.length = flush_hint_size;
tools/testing/nvdimm/test/nfit.c
2519
flush->device_handle = handle[1];
tools/testing/nvdimm/test/nfit.c
2520
flush->hint_count = NUM_HINTS;
tools/testing/nvdimm/test/nfit.c
2522
flush->hint_address[i] = t->flush_dma[1] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2523
offset += flush->header.length;
tools/testing/nvdimm/test/nfit.c
2526
flush = nfit_buf + offset;
tools/testing/nvdimm/test/nfit.c
2527
flush->header.type = ACPI_NFIT_TYPE_FLUSH_ADDRESS;
tools/testing/nvdimm/test/nfit.c
2528
flush->header.length = flush_hint_size;
tools/testing/nvdimm/test/nfit.c
2529
flush->device_handle = handle[2];
tools/testing/nvdimm/test/nfit.c
2530
flush->hint_count = NUM_HINTS;
tools/testing/nvdimm/test/nfit.c
2532
flush->hint_address[i] = t->flush_dma[2] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2533
offset += flush->header.length;
tools/testing/nvdimm/test/nfit.c
2536
flush = nfit_buf + offset;
tools/testing/nvdimm/test/nfit.c
2537
flush->header.type = ACPI_NFIT_TYPE_FLUSH_ADDRESS;
tools/testing/nvdimm/test/nfit.c
2538
flush->header.length = flush_hint_size;
tools/testing/nvdimm/test/nfit.c
2539
flush->device_handle = handle[3];
tools/testing/nvdimm/test/nfit.c
2540
flush->hint_count = NUM_HINTS;
tools/testing/nvdimm/test/nfit.c
2542
flush->hint_address[i] = t->flush_dma[3] + i * sizeof(u64);
tools/testing/nvdimm/test/nfit.c
2543
offset += flush->header.length;
tools/testing/nvdimm/test/nfit.c
2678
flush = nfit_buf + offset;
tools/testing/nvdimm/test/nfit.c
2679
flush->header.type = ACPI_NFIT_TYPE_FLUSH_ADDRESS;
tools/testing/nvdimm/test/nfit.c
2680
flush->header.length = flush_hint_size;
tools/testing/nvdimm/test/nfit.c
2681
flush->device_handle = handle[4];
tools/testing/nvdimm/test/nfit.c
2682
flush->hint_count = NUM_HINTS;
tools/testing/nvdimm/test/nfit.c
2684
flush->hint_address[i] = t->flush_dma[4]
tools/testing/nvdimm/test/nfit.c
2686
offset += flush->header.length;
tools/testing/nvdimm/test/nfit.c
3156
nfit_test->flush = devm_kcalloc(dev, num, sizeof(void *),
tools/testing/nvdimm/test/nfit.c
3177
&& nfit_test->dcr_dma && nfit_test->flush
tools/testing/selftests/arm64/fp/fp-stress.c
200
bool flush)
tools/testing/selftests/arm64/fp/fp-stress.c
213
flush = true;
tools/testing/selftests/arm64/fp/fp-stress.c
216
if (flush && child->output) {
tools/testing/selftests/arm64/fp/fp-stress.c
427
static void drain_output(bool flush)
tools/testing/selftests/arm64/fp/fp-stress.c
442
child_output(evs[i].data.ptr, evs[i].events, flush);
tools/testing/selftests/arm64/gcs/gcs-stress.c
204
bool flush)
tools/testing/selftests/arm64/gcs/gcs-stress.c
217
flush = true;
tools/testing/selftests/arm64/gcs/gcs-stress.c
220
if (flush && child->output) {
tools/testing/selftests/arm64/gcs/gcs-stress.c
328
static void drain_output(bool flush)
tools/testing/selftests/arm64/gcs/gcs-stress.c
343
child_output(evs[i].data.ptr, evs[i].events, flush);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
202
struct hv_tlb_flush *flush = (struct hv_tlb_flush *)data->hcall_gva;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
217
flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
218
flush->processor_mask = BIT(WORKER_VCPU_ID_1);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
229
flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
230
flush->processor_mask = BIT(WORKER_VCPU_ID_1);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
231
flush->gva_list[0] = (u64)data->test_pages;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
243
flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES |
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
245
flush->processor_mask = 0;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
256
flush->flags = HV_FLUSH_ALL_VIRTUAL_ADDRESS_SPACES |
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
258
flush->gva_list[0] = (u64)data->test_pages;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
373
flush->processor_mask = BIT(WORKER_VCPU_ID_1);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
374
hyperv_write_xmm_input(&flush->processor_mask, 1);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
386
flush->processor_mask = BIT(WORKER_VCPU_ID_1);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
387
flush->gva_list[0] = (u64)data->test_pages;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
388
hyperv_write_xmm_input(&flush->processor_mask, 1);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
401
hyperv_write_xmm_input(&flush->processor_mask, 1);
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
415
flush->gva_list[0] = (u64)data->test_pages;
tools/testing/selftests/kvm/x86/hyperv_tlb_flush.c
416
hyperv_write_xmm_input(&flush->processor_mask, 1);
tools/testing/selftests/powerpc/nx-gzip/gzfht_test.c
152
uint64_t flush;
tools/testing/selftests/powerpc/nx-gzip/gzfht_test.c
161
flush = ((0x1ULL & final) << shift) | *buf;
tools/testing/selftests/powerpc/nx-gzip/gzfht_test.c
164
flush |= (0xFFFF0000ULL) << shift; /* Zero length block */
tools/testing/selftests/powerpc/nx-gzip/gzfht_test.c
167
*buf++ = (unsigned char) (flush & 0xffULL);
tools/testing/selftests/powerpc/nx-gzip/gzfht_test.c
168
flush = flush >> 8;
virt/kvm/guest_memfd.c
164
bool flush = false, found_memslot = false;
virt/kvm/guest_memfd.c
187
flush |= kvm_mmu_unmap_gfn_range(kvm, &gfn_range);
virt/kvm/guest_memfd.c
190
if (flush)
virt/kvm/kvm_main.c
2236
bool flush;
virt/kvm/kvm_main.c
2257
flush = false;
virt/kvm/kvm_main.c
2280
flush = true;
virt/kvm/kvm_main.c
2291
if (flush)
virt/kvm/kvm_main.c
2348
bool flush;
virt/kvm/kvm_main.c
2378
flush = false;
virt/kvm/kvm_main.c
2401
flush = true;
virt/kvm/kvm_main.c
2408
if (flush)