overwrite
if (overwrite && nr_buf > 1)
bb->snapshot = overwrite;
int nr_pages, bool overwrite)
.overwrite = intel_security_overwrite,
int nr_pages, bool overwrite)
buf->snapshot = overwrite;
int nr_pages, bool overwrite)
nr_pages, overwrite);
int nr_pages, bool overwrite)
buf->snapshot = overwrite;
int nr_pages, bool overwrite)
buf->snapshot = overwrite;
int nr_pages, bool overwrite)
if (overwrite) {
unsigned int *slot, bool *overwrite)
*overwrite = false;
*overwrite = true;
bool overwrite;
ret = ad4130_find_slot(st, setup_info, &slot, &overwrite);
if (overwrite) {
unsigned int *setup_num, bool *overwrite)
*overwrite = false;
*overwrite = true;
bool overwrite;
ret = ad4170_find_setup(st, &chan_info->setup, &setup_num, &overwrite);
if (overwrite) {
overwrite(mg, mg_update_metadata_after_copy);
hw->blk[i].overwrite = blk_sizes[i].overwrite;
u8 overwrite; /* overwrite existing entries allowed */
u8 overwrite; /* set to true to allow overwrite of table entries */
|| nvdimm->sec.ops->overwrite)
|| !nvdimm->sec.ops->overwrite)
if (!nvdimm->sec.ops || !nvdimm->sec.ops->overwrite
rc = nvdimm->sec.ops->overwrite(nvdimm, data);
bool overwrite, unaligned_io, unwritten;
overwrite = ext4_overwrite_io(inode, offset, count, &unwritten);
((!IS_NOSEC(inode) || *extend || !overwrite ||
if (unaligned_io && (!overwrite || unwritten))
bool overwrite;
if (!ovlrd->overwrite && !ovl_can_move(new))
if (ovlrd->overwrite && new_is_dir && !ovl_pure_upper(new)) {
if (ovlrd->overwrite) {
if (!ovlrd->overwrite) {
if (!ovlrd->overwrite && ovl_type_origin(new)) {
if (!ovlrd->overwrite && ovl_type_merge_or_lower(new))
else if (!ovlrd->overwrite && new_is_dir && !new_opaque &&
if (ovlrd->overwrite && d_inode(new)) {
(!ovlrd->overwrite && ovl_type_origin(new)));
.overwrite = !(flags & RENAME_EXCHANGE),
rename_info->overwrite = cpu_to_le32(1);
__le32 overwrite; /* 1 = overwrite dest */
int nr_pages, bool overwrite);
int (*overwrite)(struct nvdimm *nvdimm,
int nr_pages, bool overwrite);
int overwrite; /* can overwrite itself */
if (!rb->overwrite) {
rb->overwrite = 0;
rb->overwrite = 1;
bool overwrite = !(flags & RING_BUFFER_WRITABLE);
if (!overwrite) {
overwrite) {
overwrite);
rb->aux_overwrite = overwrite;
unsigned long overwrite;
overwrite = local_read(&(cpu_buffer->overrun));
if (overwrite != cpu_buffer->last_overrun) {
cpu_buffer->lost_events = overwrite - cpu_buffer->last_overrun;
cpu_buffer->last_overrun = overwrite;
int i, overwrite = 76, max = 240, count = 20000000;
mtree_store_range(mt, overwrite, overwrite + 15,
xa_mk_value(overwrite), GFP_KERNEL);
overwrite += 5;
if (overwrite >= 135)
overwrite = 76;
int conf_write_autoconf(int overwrite)
if (!overwrite && is_present(autoconf_name))
int conf_write_autoconf(int overwrite);
int ifindex, bool overwrite)
if (!overwrite)
bool overwrite = false;
overwrite = true;
err = do_attach_detach_xdp(progfd, attach_type, ifindex, overwrite);
static struct perf_mmap* perf_evlist__alloc_mmap(struct perf_evlist *evlist, bool overwrite)
perf_mmap__init(&map[i], prev, overwrite, NULL);
perf_evlist__mmap_cb_get(struct perf_evlist *evlist, bool overwrite, int idx)
maps = overwrite ? evlist->mmap_ovw : evlist->mmap;
maps = perf_evlist__alloc_mmap(evlist, overwrite);
if (overwrite)
bool overwrite)
if (overwrite)
bool overwrite = evsel->attr.write_backward;
map = ops->get(evlist, overwrite, idx);
if (overwrite) {
perf_evlist__set_mmap_first(evlist, map, overwrite);
revent = !overwrite ? POLLIN : 0;
bool overwrite)
return overwrite ? evlist->mmap_ovw_first : evlist->mmap_first;
bool overwrite;
bool overwrite, libperf_unmap_cb_t unmap_cb);
bool overwrite);
#define perf_evlist__for_each_mmap(evlist, pos, overwrite) \
for ((pos) = perf_evlist__next_mmap((evlist), NULL, overwrite); \
(pos) = perf_evlist__next_mmap((evlist), (pos), overwrite))
if (!map->overwrite) {
md->start = md->overwrite ? head : old;
md->end = md->overwrite ? old : head;
if (!md->overwrite) {
bool overwrite, libperf_unmap_cb_t unmap_cb)
map->overwrite = overwrite;
if (!map->overwrite)
if (!map->overwrite)
bool overwrite)
if (overwrite) {
bool overwrite, bool synch)
maps = overwrite ? thread->overwrite_maps : thread->maps;
if (overwrite && evlist->bkw_mmap_state != BKW_MMAP_DATA_PENDING)
if (overwrite)
OPT_BOOLEAN(0, "overwrite", &record.opts.overwrite, "use overwrite mode"),
if (record.opts.overwrite)
if (!opts->overwrite)
opts->overwrite = false;
if (opts->overwrite || (hits == top->samples))
.overwrite = 0,
OPT_BOOLEAN(0, "overwrite", &top.record_opts.overwrite,
if (!top->record_opts.overwrite &&
!top->record_opts.overwrite);
md = opts->overwrite ? &evlist->overwrite_mmap[idx] : &evlist->mmap[idx];
bool overwrite = top->record_opts.overwrite;
if (overwrite)
if (overwrite) {
int set, overwrite = -1;
set = term->val.overwrite ? 1 : 0;
if ((overwrite < 0) && (set < 0))
if ((overwrite >= 0) && (set >= 0) && (overwrite != set))
if ((overwrite >= 0) && (set < 0))
if ((overwrite < 0) && (set >= 0)) {
overwrite = set;
if ((overwrite >= 0) && (opts->overwrite != overwrite))
opts->overwrite = overwrite;
bool overwrite)
perf_mmap__init(&map[i].core, prev, overwrite, perf_mmap__unmap_cb);
perf_evlist__mmap_cb_get(struct perf_evlist *_evlist, bool overwrite, int idx)
maps = overwrite ? evlist->overwrite_mmap : evlist->mmap;
maps = evlist__alloc_mmap(evlist, overwrite);
if (overwrite) {
attr->write_backward = term->val.overwrite ? 1 : 0;
attr->write_backward = opts->overwrite ? 1 : 0;
bool overwrite;
t->val.overwrite = val;
int pages = 128, overwrite = false;
&pages, &overwrite))
bool overwrite;
if (args.overwrite && !args.bench_producer) {
if (args.overwrite) {
bool overwrite;
.overwrite = false,
args.overwrite = true;
int overwrite(void *ctx)
options->overwrite = 1;
if (!options->overwrite && !access(path, F_OK)) {
int overwrite;