is_atomic
bool is_atomic)
if (is_atomic && lim->atomic_write_boundary_sectors)
bool is_atomic = bio->bi_opf & REQ_ATOMIC;
unsigned boundary_sectors = blk_boundary_sectors(lim, is_atomic);
else if (is_atomic)
bool is_atomic = rq->cmd_flags & REQ_ATOMIC;
boundary_sectors = blk_boundary_sectors(lim, is_atomic);
bool is_atomic;
dbg->is_atomic = info->desc->atomic_enabled &&
debugfs_create_bool("is_atomic", 0400, trans, &dbg->is_atomic);
int xe_vma_need_vram_for_atomic(struct xe_device *xe, struct xe_vma *vma, bool is_atomic)
if (!IS_DGFX(xe) || !is_atomic)
int xe_vma_need_vram_for_atomic(struct xe_device *xe, struct xe_vma *vma, bool is_atomic);
attr->is_atomic = bng_re_is_atomic_cap(rcfw);
bool is_atomic;
if (dev_attr->is_atomic) {
attr->is_atomic = bnxt_qplib_is_atomic_cap(rcfw);
bool is_atomic;
bool is_atomic);
bool is_atomic)
p_ptt = qed_ptt_acquire_context(p_hwfn, is_atomic);
bool is_atomic)
return qed_fcoe_get_stats(QED_AFFIN_HWFN(cdev), stats, is_atomic);
bool is_atomic)
if (qed_fcoe_stats_context(cdev, &proto_stats, is_atomic)) {
bool is_atomic);
bool is_atomic)
if (is_atomic)
struct qed_ptt *qed_ptt_acquire_context(struct qed_hwfn *p_hwfn, bool is_atomic)
if (is_atomic)
bool is_atomic)
p_ptt = qed_ptt_acquire_context(p_hwfn, is_atomic);
bool is_atomic)
return qed_iscsi_get_stats(QED_AFFIN_HWFN(cdev), stats, is_atomic);
bool is_atomic)
if (qed_iscsi_stats_context(cdev, &proto_stats, is_atomic)) {
bool is_atomic);
bool is_atomic) {}
bool is_atomic)
p_ptt = IS_PF(cdev) ? qed_ptt_acquire_context(p_hwfn, is_atomic)
bool is_atomic)
_qed_get_vport_stats(cdev, stats, is_atomic);
bool is_atomic);
bool reserved, bool is_atomic, size_t size,
TP_ARGS(call_site, reserved, is_atomic, size, align, base_addr, off,
__field( bool, is_atomic )
__entry->is_atomic = is_atomic;
__entry->reserved, __entry->is_atomic,
TP_PROTO(bool reserved, bool is_atomic, size_t size, size_t align),
TP_ARGS(reserved, is_atomic, size, align),
__field( bool, is_atomic )
__entry->is_atomic = is_atomic;
__entry->reserved, __entry->is_atomic,
const bool is_atomic = size <= sizeof(long long) && \
if (IS_ENABLED(CONFIG_KCSAN_IGNORE_ATOMICS) && is_atomic) \
check_access(ptr, size, is_atomic ? KCSAN_ACCESS_ATOMIC : 0, \
const bool is_atomic = size <= sizeof(long long) && \
if (IS_ENABLED(CONFIG_KCSAN_IGNORE_ATOMICS) && is_atomic) \
(is_atomic ? KCSAN_ACCESS_ATOMIC : 0), \
if (is_atomic(ctx, ptr, size, type))
const bool is_atomic = (ty & KCSAN_ACCESS_ATOMIC);
(is_atomic && is_scoped) ? " (marked, reordered)"
: (is_atomic ? " (marked)"
bool is_atomic;
is_atomic = !gfpflags_allow_blocking(gfp);
if (!is_atomic) {
off = pcpu_find_block_fit(chunk, bits, bit_align, is_atomic);
is_atomic);
if (is_atomic) {
if (!is_atomic) {
trace_percpu_alloc_percpu(_RET_IP_, reserved, is_atomic, size, align,
trace_percpu_alloc_percpu_fail(reserved, is_atomic, size, align);
size, align, is_atomic, err);
if (!is_atomic)
if (is_atomic) {