Symbol: is_atomic
block/blk-merge.c
211
bool is_atomic)
block/blk-merge.c
217
if (is_atomic && lim->atomic_write_boundary_sectors)
block/blk-merge.c
236
bool is_atomic = bio->bi_opf & REQ_ATOMIC;
block/blk-merge.c
237
unsigned boundary_sectors = blk_boundary_sectors(lim, is_atomic);
block/blk-merge.c
246
else if (is_atomic)
block/blk-merge.c
528
bool is_atomic = rq->cmd_flags & REQ_ATOMIC;
block/blk-merge.c
533
boundary_sectors = blk_boundary_sectors(lim, is_atomic);
drivers/firmware/arm_scmi/common.h
324
bool is_atomic;
drivers/firmware/arm_scmi/driver.c
3044
dbg->is_atomic = info->desc->atomic_enabled &&
drivers/firmware/arm_scmi/driver.c
3055
debugfs_create_bool("is_atomic", 0400, trans, &dbg->is_atomic);
drivers/gpu/drm/xe/xe_vm.c
4342
int xe_vma_need_vram_for_atomic(struct xe_device *xe, struct xe_vma *vma, bool is_atomic)
drivers/gpu/drm/xe/xe_vm.c
4347
if (!IS_DGFX(xe) || !is_atomic)
drivers/gpu/drm/xe/xe_vm.h
179
int xe_vma_need_vram_for_atomic(struct xe_device *xe, struct xe_vma *vma, bool is_atomic);
drivers/infiniband/hw/bng_re/bng_sp.c
126
attr->is_atomic = bng_re_is_atomic_cap(rcfw);
drivers/infiniband/hw/bng_re/bng_sp.h
40
bool is_atomic;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
230
if (dev_attr->is_atomic) {
drivers/infiniband/hw/bnxt_re/qplib_sp.c
195
attr->is_atomic = bnxt_qplib_is_atomic_cap(rcfw);
drivers/infiniband/hw/bnxt_re/qplib_sp.h
75
bool is_atomic;
drivers/net/ethernet/qlogic/qed/qed_dev_api.h
210
bool is_atomic);
drivers/net/ethernet/qlogic/qed/qed_fcoe.c
697
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_fcoe.c
703
p_ptt = qed_ptt_acquire_context(p_hwfn, is_atomic);
drivers/net/ethernet/qlogic/qed/qed_fcoe.c
979
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_fcoe.c
981
return qed_fcoe_get_stats(QED_AFFIN_HWFN(cdev), stats, is_atomic);
drivers/net/ethernet/qlogic/qed/qed_fcoe.c
991
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_fcoe.c
997
if (qed_fcoe_stats_context(cdev, &proto_stats, is_atomic)) {
drivers/net/ethernet/qlogic/qed/qed_fcoe.h
44
bool is_atomic);
drivers/net/ethernet/qlogic/qed/qed_fcoe.h
56
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_hw.c
111
if (is_atomic)
drivers/net/ethernet/qlogic/qed/qed_hw.c
83
struct qed_ptt *qed_ptt_acquire_context(struct qed_hwfn *p_hwfn, bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_hw.c
88
if (is_atomic)
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1003
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1009
p_ptt = qed_ptt_acquire_context(p_hwfn, is_atomic);
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1342
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1344
return qed_iscsi_get_stats(QED_AFFIN_HWFN(cdev), stats, is_atomic);
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1370
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_iscsi.c
1376
if (qed_iscsi_stats_context(cdev, &proto_stats, is_atomic)) {
drivers/net/ethernet/qlogic/qed/qed_iscsi.h
49
bool is_atomic);
drivers/net/ethernet/qlogic/qed/qed_iscsi.h
63
bool is_atomic) {}
drivers/net/ethernet/qlogic/qed/qed_l2.c
1867
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_l2.c
1879
p_ptt = IS_PF(cdev) ? qed_ptt_acquire_context(p_hwfn, is_atomic)
drivers/net/ethernet/qlogic/qed/qed_l2.c
1911
bool is_atomic)
drivers/net/ethernet/qlogic/qed/qed_l2.c
1920
_qed_get_vport_stats(cdev, stats, is_atomic);
drivers/net/ethernet/qlogic/qed/qed_l2.h
276
bool is_atomic);
include/trace/events/percpu.h
14
bool reserved, bool is_atomic, size_t size,
include/trace/events/percpu.h
18
TP_ARGS(call_site, reserved, is_atomic, size, align, base_addr, off,
include/trace/events/percpu.h
24
__field( bool, is_atomic )
include/trace/events/percpu.h
36
__entry->is_atomic = is_atomic;
include/trace/events/percpu.h
48
__entry->reserved, __entry->is_atomic,
include/trace/events/percpu.h
78
TP_PROTO(bool reserved, bool is_atomic, size_t size, size_t align),
include/trace/events/percpu.h
80
TP_ARGS(reserved, is_atomic, size, align),
include/trace/events/percpu.h
84
__field( bool, is_atomic )
include/trace/events/percpu.h
91
__entry->is_atomic = is_atomic;
include/trace/events/percpu.h
97
__entry->reserved, __entry->is_atomic,
kernel/kcsan/core.c
1055
const bool is_atomic = size <= sizeof(long long) && \
kernel/kcsan/core.c
1057
if (IS_ENABLED(CONFIG_KCSAN_IGNORE_ATOMICS) && is_atomic) \
kernel/kcsan/core.c
1059
check_access(ptr, size, is_atomic ? KCSAN_ACCESS_ATOMIC : 0, \
kernel/kcsan/core.c
1069
const bool is_atomic = size <= sizeof(long long) && \
kernel/kcsan/core.c
1071
if (IS_ENABLED(CONFIG_KCSAN_IGNORE_ATOMICS) && is_atomic) \
kernel/kcsan/core.c
1075
(is_atomic ? KCSAN_ACCESS_ATOMIC : 0), \
kernel/kcsan/core.c
277
if (is_atomic(ctx, ptr, size, type))
kernel/kcsan/kcsan_test.c
220
const bool is_atomic = (ty & KCSAN_ACCESS_ATOMIC);
kernel/kcsan/kcsan_test.c
223
(is_atomic && is_scoped) ? " (marked, reordered)"
kernel/kcsan/kcsan_test.c
224
: (is_atomic ? " (marked)"
mm/percpu.c
1740
bool is_atomic;
mm/percpu.c
1754
is_atomic = !gfpflags_allow_blocking(gfp);
mm/percpu.c
1780
if (!is_atomic) {
mm/percpu.c
1800
off = pcpu_find_block_fit(chunk, bits, bit_align, is_atomic);
mm/percpu.c
1820
is_atomic);
mm/percpu.c
1837
if (is_atomic) {
mm/percpu.c
1867
if (!is_atomic) {
mm/percpu.c
1898
trace_percpu_alloc_percpu(_RET_IP_, reserved, is_atomic, size, align,
mm/percpu.c
1911
trace_percpu_alloc_percpu_fail(reserved, is_atomic, size, align);
mm/percpu.c
1918
size, align, is_atomic, err);
mm/percpu.c
1919
if (!is_atomic)
mm/percpu.c
1926
if (is_atomic) {