Symbol: rq_flags
block/bfq-iosched.c
5239
rq->rq_flags |= RQF_STARTED;
block/bfq-iosched.c
6661
if (rq->rq_flags & RQF_STARTED)
block/bfq-iosched.c
6668
if (likely(rq->rq_flags & RQF_STARTED)) {
block/blk-flush.c
120
rq->rq_flags &= ~RQF_FLUSH_SEQ;
block/blk-flush.c
320
flush_rq->rq_flags |= RQF_FLUSH_SEQ;
block/blk-flush.c
374
rq->rq_flags |= RQF_FLUSH_SEQ;
block/blk-merge.c
678
if (rq->rq_flags & RQF_MIXED_MERGE)
block/blk-merge.c
691
rq->rq_flags |= RQF_MIXED_MERGE;
block/blk-merge.c
710
if (req->rq_flags & RQF_MIXED_MERGE) {
block/blk-merge.c
723
if (req->rq_flags & RQF_IO_STAT) {
block/blk-merge.c
822
if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
block/blk-merge.c
937
if (req->rq_flags & RQF_IO_STAT) {
block/blk-merge.c
960
if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING)
block/blk-merge.c
985
if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING)
block/blk-mq-debugfs.c
281
blk_flags_show(m, (__force unsigned int)rq->rq_flags, rqf_name,
block/blk-mq-dma.c
145
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) {
block/blk-mq-sched.h
105
if (rq->rq_flags & RQF_USE_SCHED) {
block/blk-mq-sched.h
84
if (rq->rq_flags & RQF_USE_SCHED) {
block/blk-mq-sched.h
95
if (rq->rq_flags & RQF_USE_SCHED) {
block/blk-mq.c
1046
if (req->rq_flags & RQF_MIXED_MERGE) {
block/blk-mq.c
1051
if (!(req->rq_flags & RQF_SPECIAL_PAYLOAD)) {
block/blk-mq.c
1078
if ((req->rq_flags & (RQF_IO_STAT|RQF_FLUSH_SEQ)) == RQF_IO_STAT) {
block/blk-mq.c
1130
req->rq_flags |= RQF_IO_STAT;
block/blk-mq.c
1152
if (rq->rq_flags & RQF_STATS)
block/blk-mq.c
1378
rq->rq_flags |= RQF_STATS;
block/blk-mq.c
1428
if (!plug->has_elevator && (rq->rq_flags & RQF_SCHED_TAGS))
block/blk-mq.c
1545
rq->rq_flags &= ~RQF_TIMED_OUT;
block/blk-mq.c
1590
if (rq->rq_flags & RQF_DONTPREP)
block/blk-mq.c
1621
return (rq->rq_flags & RQF_FLUSH_SEQ) && !is_flush_rq(rq);
block/blk-mq.c
1659
req->rq_flags |= RQF_TIMED_OUT;
block/blk-mq.c
1684
if (rq->rq_flags & RQF_TIMED_OUT)
block/blk-mq.c
2779
if ((rq->rq_flags & RQF_USE_SCHED) || !blk_mq_get_budget_and_tag(rq)) {
block/blk-mq.c
3055
.rq_flags = 0,
block/blk-mq.c
3248
if ((rq->rq_flags & RQF_USE_SCHED) ||
block/blk-mq.c
3398
if (rq_src->rq_flags & RQF_SPECIAL_PAYLOAD) {
block/blk-mq.c
3399
rq->rq_flags |= RQF_SPECIAL_PAYLOAD;
block/blk-mq.c
424
data->rq_flags |= RQF_PM;
block/blk-mq.c
425
rq->rq_flags = data->rq_flags;
block/blk-mq.c
427
if (data->rq_flags & RQF_SCHED_TAGS) {
block/blk-mq.c
450
if (rq->rq_flags & RQF_USE_SCHED) {
block/blk-mq.c
492
if (!(data->rq_flags & RQF_SCHED_TAGS))
block/blk-mq.c
515
data->rq_flags |= RQF_SCHED_TAGS;
block/blk-mq.c
526
data->rq_flags |= RQF_USE_SCHED;
block/blk-mq.c
557
data->rq_flags |= RQF_RESV;
block/blk-mq.c
590
if (!(data->rq_flags & RQF_SCHED_TAGS))
block/blk-mq.c
607
.rq_flags = 0,
block/blk-mq.c
673
.rq_flags = 0,
block/blk-mq.c
708
.rq_flags = 0,
block/blk-mq.c
755
data.rq_flags |= RQF_SCHED_TAGS;
block/blk-mq.c
760
data.rq_flags |= RQF_RESV;
block/blk-mq.c
766
if (!(data.rq_flags & RQF_SCHED_TAGS))
block/blk-mq.c
788
if (rq->rq_flags & RQF_USE_SCHED) {
block/blk-mq.c
795
rq->rq_flags &= ~RQF_USE_SCHED;
block/blk-mq.c
858
if (req->rq_flags & RQF_IO_STAT) {
block/blk-mq.c
887
const bool is_flush = (req->rq_flags & RQF_FLUSH_SEQ) != 0;
block/blk-mq.c
957
bool is_flush = req->rq_flags & RQF_FLUSH_SEQ;
block/blk-mq.c
958
bool quiet = req->rq_flags & RQF_QUIET;
block/blk-mq.c
97
if (rq->rq_flags & RQF_IO_STAT &&
block/blk-mq.h
166
req_flags_t rq_flags;
block/blk-mq.h
234
if (data->rq_flags & RQF_SCHED_TAGS)
block/blk-pm.h
21
if (rq->q->dev && !(rq->rq_flags & RQF_PM))
block/blk-timeout.c
140
req->rq_flags &= ~RQF_TIMED_OUT;
block/blk-zoned.c
1306
req->rq_flags |= RQF_ZONE_WRITE_PLUGGING;
block/blk-zoned.c
1716
req->rq_flags &= ~RQF_ZONE_WRITE_PLUGGING;
block/blk.h
177
if (rq->rq_flags & RQF_NOMERGE_FLAGS)
block/blk.h
335
#define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED)
block/blk.h
529
if (rq->rq_flags & RQF_ZONE_WRITE_PLUGGING)
block/elevator.c
168
rq->rq_flags &= ~RQF_HASHED;
block/elevator.c
184
rq->rq_flags |= RQF_HASHED;
block/mq-deadline.c
317
rq->rq_flags |= RQF_STARTED;
drivers/ata/libata-eh.c
2001
if (qc->scsicmd && scsi_cmd_to_rq(qc->scsicmd)->rq_flags & RQF_QUIET)
drivers/ata/libata-scsi.c
777
if (scsi_cmd_to_rq(cmd)->rq_flags & RQF_QUIET)
drivers/block/virtio_blk.c
201
req->rq_flags |= RQF_SPECIAL_PAYLOAD;
drivers/block/virtio_blk.c
234
if (req->rq_flags & RQF_SPECIAL_PAYLOAD)
drivers/md/dm-rq.c
273
if (rq->rq_flags & RQF_FAILED)
drivers/md/dm-rq.c
299
rq->rq_flags |= RQF_FAILED;
drivers/mmc/core/block.c
2091
req->rq_flags |= RQF_QUIET;
drivers/mmc/core/queue.c
243
req->rq_flags |= RQF_QUIET;
drivers/mmc/core/queue.c
291
if (!(req->rq_flags & RQF_DONTPREP)) {
drivers/mmc/core/queue.c
293
req->rq_flags |= RQF_DONTPREP;
drivers/net/ethernet/fungible/funcore/fun_queue.c
459
funq->rq_flags = req->rq_flags | FUN_ADMIN_EPSQ_CREATE_FLAG_RQ;
drivers/net/ethernet/fungible/funcore/fun_queue.c
491
rc = fun_sq_create(fdev, funq->rq_flags, funq->rqid, funq->cqid, 0,
drivers/net/ethernet/fungible/funcore/fun_queue.h
120
u16 rq_flags;
drivers/net/ethernet/fungible/funcore/fun_queue.h
69
u16 rq_flags;
drivers/nvme/host/core.c
1069
if (req->rq_flags & RQF_SPECIAL_PAYLOAD) {
drivers/nvme/host/core.c
1076
req->rq_flags &= ~RQF_SPECIAL_PAYLOAD;
drivers/nvme/host/core.c
1086
if (!(req->rq_flags & RQF_DONTPREP))
drivers/nvme/host/core.c
437
if (unlikely(nvme_req(req)->status && !(req->rq_flags & RQF_QUIET))) {
drivers/nvme/host/core.c
714
req->rq_flags |= RQF_DONTPREP;
drivers/nvme/host/core.c
734
req->rq_flags |= RQF_QUIET;
drivers/nvme/host/core.c
768
if (!(rq->rq_flags & RQF_DONTPREP))
drivers/nvme/host/core.c
889
req->rq_flags |= RQF_SPECIAL_PAYLOAD;
drivers/nvme/host/ioctl.c
102
struct nvme_command *cmd, blk_opf_t rq_flags,
drivers/nvme/host/ioctl.c
107
req = blk_mq_alloc_request(q, nvme_req_op(cmd) | rq_flags, blk_flags);
drivers/nvme/host/ioctl.c
458
blk_opf_t rq_flags = 0;
drivers/nvme/host/ioctl.c
509
rq_flags |= REQ_NOWAIT;
drivers/nvme/host/ioctl.c
513
rq_flags |= REQ_POLLED;
drivers/nvme/host/ioctl.c
515
req = nvme_alloc_user_request(q, &c, rq_flags, blk_flags);
drivers/nvme/host/tcp.c
348
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) {
drivers/scsi/device_handler/scsi_dh_alua.c
1221
req->rq_flags |= RQF_QUIET;
drivers/scsi/device_handler/scsi_dh_emc.c
340
req->rq_flags |= RQF_QUIET;
drivers/scsi/device_handler/scsi_dh_hp_sw.c
191
req->rq_flags |= RQF_QUIET;
drivers/scsi/device_handler/scsi_dh_rdac.c
665
req->rq_flags |= RQF_QUIET;
drivers/scsi/scsi_error.c
2158
req->rq_flags |= RQF_QUIET;
drivers/scsi/scsi_error.c
634
req->rq_flags |= RQF_QUIET;
drivers/scsi/scsi_error.c
746
req->rq_flags |= RQF_QUIET;
drivers/scsi/scsi_lib.c
1022
else if (req->rq_flags & RQF_QUIET)
drivers/scsi/scsi_lib.c
118
if (rq->rq_flags & RQF_DONTPREP) {
drivers/scsi/scsi_lib.c
119
rq->rq_flags &= ~RQF_DONTPREP;
drivers/scsi/scsi_lib.c
1269
if (rq->rq_flags & RQF_DONTPREP) {
drivers/scsi/scsi_lib.c
1271
rq->rq_flags &= ~RQF_DONTPREP;
drivers/scsi/scsi_lib.c
1350
if (req && WARN_ON_ONCE(!(req->rq_flags & RQF_PM)))
drivers/scsi/scsi_lib.c
1358
if (req && !(req->rq_flags & RQF_PM))
drivers/scsi/scsi_lib.c
1876
if (!(req->rq_flags & RQF_DONTPREP)) {
drivers/scsi/scsi_lib.c
1880
req->rq_flags |= RQF_DONTPREP;
drivers/scsi/scsi_lib.c
1930
if (req->rq_flags & RQF_DONTPREP)
drivers/scsi/scsi_lib.c
1943
if (req->rq_flags & RQF_DONTPREP)
drivers/scsi/scsi_lib.c
327
req->rq_flags |= RQF_QUIET;
drivers/scsi/scsi_lib.c
747
if (!(rq->rq_flags & RQF_MIXED_MERGE))
drivers/scsi/scsi_lib.c
931
if (!(req->rq_flags & RQF_QUIET)) {
drivers/scsi/sd.c
1105
rq->rq_flags |= RQF_QUIET;
drivers/scsi/sd.c
1517
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
drivers/scsi/sd.c
2404
req->rq_flags |= RQF_QUIET;
drivers/scsi/sd.c
940
rq->rq_flags |= RQF_SPECIAL_PAYLOAD;
drivers/scsi/sd_zbc.c
378
rq->rq_flags |= RQF_QUIET;
drivers/scsi/st.c
568
req->rq_flags |= RQF_QUIET;
fs/nfsd/nfs4idmap.c
669
WARN_ON_ONCE(test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags));
fs/nfsd/nfs4idmap.c
706
WARN_ON_ONCE(test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags));
fs/nfsd/nfs4proc.c
3066
clear_bit(RQ_USEDEFERRAL, &rqstp->rq_flags);
fs/nfsd/nfs4xdr.c
6370
clear_bit(RQ_USEDEFERRAL, &rqstp->rq_flags);
fs/nfsd/nfscache.c
523
if (!test_bit(RQ_SECURE, &rqstp->rq_flags) && rp->c_secure)
fs/nfsd/nfscache.c
619
rp->c_secure = test_bit(RQ_SECURE, &rqstp->rq_flags);
fs/nfsd/nfsctl.c
1428
nla_put_u32(skb, NFSD_A_RPC_STATUS_FLAGS, genl_rqstp->rq_flags) ||
fs/nfsd/nfsctl.c
1531
genl_rqstp.rq_flags = rqstp->rq_flags;
fs/nfsd/nfsd.h
66
unsigned long rq_flags;
fs/nfsd/nfsfh.c
99
return test_bit(RQ_SECURE, &rqstp->rq_flags);
fs/nfsd/nfsproc.c
232
set_bit(RQ_DROPME, &rqstp->rq_flags);
fs/nfsd/nfsproc.c
265
set_bit(RQ_DROPME, &rqstp->rq_flags);
fs/nfsd/nfssvc.c
1013
if (test_bit(RQ_DROPME, &rqstp->rq_flags))
fs/nfsd/nfssvc.c
914
set_bit(RQ_VICTIM, &rqstp->rq_flags);
fs/nfsd/vfs.c
1449
if (test_bit(RQ_LOCAL, &rqstp->rq_flags) &&
include/linux/blk-mq.h
1154
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
include/linux/blk-mq.h
1165
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
include/linux/blk-mq.h
1204
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
include/linux/blk-mq.h
875
return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS | RQF_USE_SCHED));
include/linux/blk-mq.h
880
return rq->rq_flags & RQF_RESV;
include/linux/blk-mq.h
908
if (req->rq_flags & RQF_SCHED_TAGS)
include/linux/sched.h
641
struct rq_flags;
include/linux/sunrpc/svc.h
219
unsigned long rq_flags; /* flags field */
include/linux/sunrpc/svc.h
319
set_bit(RQ_VICTIM, &rqstp->rq_flags);
include/linux/sunrpc/svc.h
321
return test_bit(RQ_VICTIM, &rqstp->rq_flags);
include/trace/events/sunrpc.h
1816
__entry->flags = rqst->rq_flags;
include/trace/events/sunrpc.h
1852
__entry->flags = rqst->rq_flags;
kernel/sched/core.c
1851
struct rq_flags rf;
kernel/sched/core.c
2248
struct rq_flags rf;
kernel/sched/core.c
2448
static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf,
kernel/sched/core.c
2495
static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf,
kernel/sched/core.c
2520
struct rq_flags rf;
kernel/sched/core.c
2861
static int affine_move_task(struct rq *rq, struct task_struct *p, struct rq_flags *rf,
kernel/sched/core.c
3017
struct rq_flags *rf)
kernel/sched/core.c
3101
struct rq_flags rf;
kernel/sched/core.c
3145
struct rq_flags rf;
kernel/sched/core.c
3301
struct rq_flags srf, drf;
kernel/sched/core.c
3648
struct rq_flags *rf)
kernel/sched/core.c
3710
struct rq_flags rf;
kernel/sched/core.c
3739
struct rq_flags rf;
kernel/sched/core.c
3911
struct rq_flags rf;
kernel/sched/core.c
4300
struct rq_flags rf;
kernel/sched/core.c
4350
struct rq_flags rf;
kernel/sched/core.c
4766
struct rq_flags rf;
kernel/sched/core.c
4981
void __balance_callbacks(struct rq *rq, struct rq_flags *rf)
kernel/sched/core.c
5002
prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf)
kernel/sched/core.c
5240
struct task_struct *next, struct rq_flags *rf)
kernel/sched/core.c
5460
struct rq_flags rf;
kernel/sched/core.c
5552
struct rq_flags rf;
kernel/sched/core.c
5886
struct rq_flags *rf)
kernel/sched/core.c
5909
__pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
kernel/sched/core.c
5989
static inline struct task_struct *pick_task(struct rq *rq, struct rq_flags *rf)
kernel/sched/core.c
6010
pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
kernel/sched/core.c
6465
pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
kernel/sched/core.c
6595
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
kernel/sched/core.c
6701
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
kernel/sched/core.c
6775
struct rq_flags rf;
kernel/sched/core.c
717
struct rq *___task_rq_lock(struct task_struct *p, struct rq_flags *rf)
kernel/sched/core.c
7314
struct rq_flags rf;
kernel/sched/core.c
740
struct rq *_task_rq_lock(struct task_struct *p, struct rq_flags *rf)
kernel/sched/core.c
8088
struct rq_flags rf;
kernel/sched/core.c
8187
struct rq_flags rf;
kernel/sched/core.c
8263
struct rq_flags rf;
kernel/sched/core.c
8275
struct rq_flags rf;
kernel/sched/core.c
8519
struct rq_flags rf;
kernel/sched/core.c
888
struct rq_flags rf;
kernel/sched/core.c
914
struct rq_flags rf;
kernel/sched/core_sched.c
59
struct rq_flags rf;
kernel/sched/cputime.c
312
struct rq_flags rf;
kernel/sched/deadline.c
1130
static void __push_dl_task(struct rq *rq, struct rq_flags *rf)
kernel/sched/deadline.c
1156
struct rq_flags *rf = &scope.rf;
kernel/sched/deadline.c
1226
struct rq_flags rf;
kernel/sched/deadline.c
1966
struct rq_flags rf;
kernel/sched/deadline.c
2457
struct rq_flags rf;
kernel/sched/deadline.c
2508
static int balance_dl(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
kernel/sched/deadline.c
2602
static struct task_struct *__pick_task_dl(struct rq *rq, struct rq_flags *rf)
kernel/sched/deadline.c
2629
static struct task_struct *pick_task_dl(struct rq *rq, struct rq_flags *rf)
kernel/sched/deadline.c
3187
struct rq_flags rf;
kernel/sched/ext.c
2518
do_pick_task_scx(struct rq *rq, struct rq_flags *rf, bool force_scx)
kernel/sched/ext.c
2592
static struct task_struct *pick_task_scx(struct rq *rq, struct rq_flags *rf)
kernel/sched/ext.c
2604
ext_server_pick_task(struct sched_dl_entity *dl_se, struct rq_flags *rf)
kernel/sched/ext.c
2801
struct rq_flags rf;
kernel/sched/ext.c
2974
struct rq_flags rf;
kernel/sched/ext.c
3120
struct rq_flags rf;
kernel/sched/ext.c
3140
struct rq_flags rf;
kernel/sched/ext.c
3199
struct rq_flags rf;
kernel/sched/ext.c
4747
struct rq_flags rf;
kernel/sched/ext.c
514
struct rq_flags rf;
kernel/sched/ext.c
7109
struct rq_flags rf;
kernel/sched/ext_idle.c
896
struct rq_flags rf;
kernel/sched/fair.c
11873
struct rq_flags rf;
kernel/sched/fair.c
12203
struct rq_flags rf;
kernel/sched/fair.c
12798
struct rq_flags rf;
kernel/sched/fair.c
12922
static int sched_balance_newidle(struct rq *this_rq, struct rq_flags *rf)
kernel/sched/fair.c
13733
struct rq_flags rf;
kernel/sched/fair.c
13836
struct rq_flags rf;
kernel/sched/fair.c
13890
struct rq_flags rf;
kernel/sched/fair.c
4845
static int sched_balance_newidle(struct rq *this_rq, struct rq_flags *rf)
kernel/sched/fair.c
6099
struct rq_flags rf;
kernel/sched/fair.c
6171
struct rq_flags rf;
kernel/sched/fair.c
8685
struct rq_flags rf;
kernel/sched/fair.c
8941
static struct task_struct *pick_task_fair(struct rq *rq, struct rq_flags *rf)
kernel/sched/fair.c
8978
pick_next_task_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
kernel/sched/fair.c
9060
fair_server_pick_task(struct sched_dl_entity *dl_se, struct rq_flags *rf)
kernel/sched/fair.c
9801
struct rq_flags rf;
kernel/sched/fair.c
9817
struct rq_flags rf;
kernel/sched/idle.c
465
balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
kernel/sched/idle.c
501
struct task_struct *pick_task_idle(struct rq *rq, struct rq_flags *rf)
kernel/sched/psi.c
1058
struct rq_flags rf;
kernel/sched/psi.c
1089
struct rq_flags rf;
kernel/sched/psi.c
1158
struct rq_flags rf;
kernel/sched/rt.c
1594
static int balance_rt(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
kernel/sched/rt.c
1704
static struct task_struct *pick_task_rt(struct rq *rq, struct rq_flags *rf)
kernel/sched/rt.c
801
struct rq_flags rf;
kernel/sched/sched.h
1869
static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1878
static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1887
static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1898
extern struct rq *___task_rq_lock(struct task_struct *p, struct rq_flags *rf) __acquires_ret;
kernel/sched/sched.h
1901
extern struct rq *_task_rq_lock(struct task_struct *p, struct rq_flags *rf)
kernel/sched/sched.h
1905
__task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
kernel/sched/sched.h
1913
task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
kernel/sched/sched.h
1923
struct rq *rq; struct rq_flags rf)
kernel/sched/sched.h
1930
struct rq *rq; struct rq_flags rf)
kernel/sched/sched.h
1932
static inline void rq_lock_irqsave(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1939
static inline void rq_lock_irq(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1946
static inline void rq_lock(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1953
static inline void rq_unlock_irqrestore(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1960
static inline void rq_unlock_irq(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1967
static inline void rq_unlock(struct rq *rq, struct rq_flags *rf)
kernel/sched/sched.h
1977
struct rq_flags rf)
kernel/sched/sched.h
1985
struct rq_flags rf)
kernel/sched/sched.h
1993
struct rq_flags rf)
kernel/sched/sched.h
1999
static inline struct rq *_this_rq_lock_irq(struct rq_flags *rf) __acquires_ret
kernel/sched/sched.h
2545
int (*balance)(struct rq *rq, struct task_struct *prev, struct rq_flags *rf);
kernel/sched/sched.h
2550
struct task_struct *(*pick_task)(struct rq *rq, struct rq_flags *rf);
kernel/sched/sched.h
2561
struct rq_flags *rf);
kernel/sched/sched.h
2786
struct rq_flags *rf);
kernel/sched/sched.h
2787
extern struct task_struct *pick_task_idle(struct rq *rq, struct rq_flags *rf);
kernel/sched/sched.h
4072
extern void __balance_callbacks(struct rq *rq, struct rq_flags *rf);
kernel/sched/stats.h
209
struct rq_flags rf;
kernel/sched/stop_task.c
19
balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
kernel/sched/stop_task.c
35
static struct task_struct *pick_task_stop(struct rq *rq, struct rq_flags *rf)
kernel/sched/syscalls.c
1331
struct rq_flags rf;
kernel/sched/syscalls.c
501
struct rq_flags rf;
kernel/sched/topology.c
475
struct rq_flags rf;
net/sunrpc/svc.c
1385
set_bit(RQ_USEDEFERRAL, &rqstp->rq_flags);
net/sunrpc/svc.c
1386
clear_bit(RQ_DROPME, &rqstp->rq_flags);
net/sunrpc/svc_xprt.c
1222
if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags))
net/sunrpc/svc_xprt.c
1253
set_bit(RQ_DROPME, &rqstp->rq_flags);
net/sunrpc/svc_xprt.c
419
if (!test_bit(RQ_DATA, &rqstp->rq_flags)) {
net/sunrpc/svc_xprt.c
423
set_bit(RQ_DATA, &rqstp->rq_flags);
net/sunrpc/svc_xprt.c
431
if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) {
net/sunrpc/svcsock.c
1184
set_bit(RQ_LOCAL, &rqstp->rq_flags);
net/sunrpc/svcsock.c
1186
clear_bit(RQ_LOCAL, &rqstp->rq_flags);
net/sunrpc/svcsock.c
395
set_bit(RQ_SECURE, &rqstp->rq_flags);
net/sunrpc/svcsock.c
397
clear_bit(RQ_SECURE, &rqstp->rq_flags);
net/sunrpc/xprtrdma/svc_rdma_recvfrom.c
989
set_bit(RQ_SECURE, &rqstp->rq_flags);