rq_flags
rq->rq_flags |= RQF_STARTED;
if (rq->rq_flags & RQF_STARTED)
if (likely(rq->rq_flags & RQF_STARTED)) {
rq->rq_flags &= ~RQF_FLUSH_SEQ;
flush_rq->rq_flags |= RQF_FLUSH_SEQ;
rq->rq_flags |= RQF_FLUSH_SEQ;
if (rq->rq_flags & RQF_MIXED_MERGE)
rq->rq_flags |= RQF_MIXED_MERGE;
if (req->rq_flags & RQF_MIXED_MERGE) {
if (req->rq_flags & RQF_IO_STAT) {
if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) ||
if (req->rq_flags & RQF_IO_STAT) {
if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING)
if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING)
blk_flags_show(m, (__force unsigned int)rq->rq_flags, rqf_name,
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) {
if (rq->rq_flags & RQF_USE_SCHED) {
if (rq->rq_flags & RQF_USE_SCHED) {
if (rq->rq_flags & RQF_USE_SCHED) {
if (req->rq_flags & RQF_MIXED_MERGE) {
if (!(req->rq_flags & RQF_SPECIAL_PAYLOAD)) {
if ((req->rq_flags & (RQF_IO_STAT|RQF_FLUSH_SEQ)) == RQF_IO_STAT) {
req->rq_flags |= RQF_IO_STAT;
if (rq->rq_flags & RQF_STATS)
rq->rq_flags |= RQF_STATS;
if (!plug->has_elevator && (rq->rq_flags & RQF_SCHED_TAGS))
rq->rq_flags &= ~RQF_TIMED_OUT;
if (rq->rq_flags & RQF_DONTPREP)
return (rq->rq_flags & RQF_FLUSH_SEQ) && !is_flush_rq(rq);
req->rq_flags |= RQF_TIMED_OUT;
if (rq->rq_flags & RQF_TIMED_OUT)
if ((rq->rq_flags & RQF_USE_SCHED) || !blk_mq_get_budget_and_tag(rq)) {
.rq_flags = 0,
if ((rq->rq_flags & RQF_USE_SCHED) ||
if (rq_src->rq_flags & RQF_SPECIAL_PAYLOAD) {
rq->rq_flags |= RQF_SPECIAL_PAYLOAD;
data->rq_flags |= RQF_PM;
rq->rq_flags = data->rq_flags;
if (data->rq_flags & RQF_SCHED_TAGS) {
if (rq->rq_flags & RQF_USE_SCHED) {
if (!(data->rq_flags & RQF_SCHED_TAGS))
data->rq_flags |= RQF_SCHED_TAGS;
data->rq_flags |= RQF_USE_SCHED;
data->rq_flags |= RQF_RESV;
if (!(data->rq_flags & RQF_SCHED_TAGS))
.rq_flags = 0,
.rq_flags = 0,
.rq_flags = 0,
data.rq_flags |= RQF_SCHED_TAGS;
data.rq_flags |= RQF_RESV;
if (!(data.rq_flags & RQF_SCHED_TAGS))
if (rq->rq_flags & RQF_USE_SCHED) {
rq->rq_flags &= ~RQF_USE_SCHED;
if (req->rq_flags & RQF_IO_STAT) {
const bool is_flush = (req->rq_flags & RQF_FLUSH_SEQ) != 0;
bool is_flush = req->rq_flags & RQF_FLUSH_SEQ;
bool quiet = req->rq_flags & RQF_QUIET;
if (rq->rq_flags & RQF_IO_STAT &&
req_flags_t rq_flags;
if (data->rq_flags & RQF_SCHED_TAGS)
if (rq->q->dev && !(rq->rq_flags & RQF_PM))
req->rq_flags &= ~RQF_TIMED_OUT;
req->rq_flags |= RQF_ZONE_WRITE_PLUGGING;
req->rq_flags &= ~RQF_ZONE_WRITE_PLUGGING;
if (rq->rq_flags & RQF_NOMERGE_FLAGS)
#define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED)
if (rq->rq_flags & RQF_ZONE_WRITE_PLUGGING)
rq->rq_flags &= ~RQF_HASHED;
rq->rq_flags |= RQF_HASHED;
rq->rq_flags |= RQF_STARTED;
if (qc->scsicmd && scsi_cmd_to_rq(qc->scsicmd)->rq_flags & RQF_QUIET)
if (scsi_cmd_to_rq(cmd)->rq_flags & RQF_QUIET)
req->rq_flags |= RQF_SPECIAL_PAYLOAD;
if (req->rq_flags & RQF_SPECIAL_PAYLOAD)
if (rq->rq_flags & RQF_FAILED)
rq->rq_flags |= RQF_FAILED;
req->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
if (!(req->rq_flags & RQF_DONTPREP)) {
req->rq_flags |= RQF_DONTPREP;
funq->rq_flags = req->rq_flags | FUN_ADMIN_EPSQ_CREATE_FLAG_RQ;
rc = fun_sq_create(fdev, funq->rq_flags, funq->rqid, funq->cqid, 0,
u16 rq_flags;
u16 rq_flags;
if (req->rq_flags & RQF_SPECIAL_PAYLOAD) {
req->rq_flags &= ~RQF_SPECIAL_PAYLOAD;
if (!(req->rq_flags & RQF_DONTPREP))
if (unlikely(nvme_req(req)->status && !(req->rq_flags & RQF_QUIET))) {
req->rq_flags |= RQF_DONTPREP;
req->rq_flags |= RQF_QUIET;
if (!(rq->rq_flags & RQF_DONTPREP))
req->rq_flags |= RQF_SPECIAL_PAYLOAD;
struct nvme_command *cmd, blk_opf_t rq_flags,
req = blk_mq_alloc_request(q, nvme_req_op(cmd) | rq_flags, blk_flags);
blk_opf_t rq_flags = 0;
rq_flags |= REQ_NOWAIT;
rq_flags |= REQ_POLLED;
req = nvme_alloc_user_request(q, &c, rq_flags, blk_flags);
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) {
req->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
else if (req->rq_flags & RQF_QUIET)
if (rq->rq_flags & RQF_DONTPREP) {
rq->rq_flags &= ~RQF_DONTPREP;
if (rq->rq_flags & RQF_DONTPREP) {
rq->rq_flags &= ~RQF_DONTPREP;
if (req && WARN_ON_ONCE(!(req->rq_flags & RQF_PM)))
if (req && !(req->rq_flags & RQF_PM))
if (!(req->rq_flags & RQF_DONTPREP)) {
req->rq_flags |= RQF_DONTPREP;
if (req->rq_flags & RQF_DONTPREP)
if (req->rq_flags & RQF_DONTPREP)
req->rq_flags |= RQF_QUIET;
if (!(rq->rq_flags & RQF_MIXED_MERGE))
if (!(req->rq_flags & RQF_QUIET)) {
rq->rq_flags |= RQF_QUIET;
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
req->rq_flags |= RQF_QUIET;
rq->rq_flags |= RQF_SPECIAL_PAYLOAD;
rq->rq_flags |= RQF_QUIET;
req->rq_flags |= RQF_QUIET;
WARN_ON_ONCE(test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags));
WARN_ON_ONCE(test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags));
clear_bit(RQ_USEDEFERRAL, &rqstp->rq_flags);
clear_bit(RQ_USEDEFERRAL, &rqstp->rq_flags);
if (!test_bit(RQ_SECURE, &rqstp->rq_flags) && rp->c_secure)
rp->c_secure = test_bit(RQ_SECURE, &rqstp->rq_flags);
nla_put_u32(skb, NFSD_A_RPC_STATUS_FLAGS, genl_rqstp->rq_flags) ||
genl_rqstp.rq_flags = rqstp->rq_flags;
unsigned long rq_flags;
return test_bit(RQ_SECURE, &rqstp->rq_flags);
set_bit(RQ_DROPME, &rqstp->rq_flags);
set_bit(RQ_DROPME, &rqstp->rq_flags);
if (test_bit(RQ_DROPME, &rqstp->rq_flags))
set_bit(RQ_VICTIM, &rqstp->rq_flags);
if (test_bit(RQ_LOCAL, &rqstp->rq_flags) &&
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
if (rq->rq_flags & RQF_SPECIAL_PAYLOAD)
return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS | RQF_USE_SCHED));
return rq->rq_flags & RQF_RESV;
if (req->rq_flags & RQF_SCHED_TAGS)
struct rq_flags;
unsigned long rq_flags; /* flags field */
set_bit(RQ_VICTIM, &rqstp->rq_flags);
return test_bit(RQ_VICTIM, &rqstp->rq_flags);
__entry->flags = rqst->rq_flags;
__entry->flags = rqst->rq_flags;
struct rq_flags rf;
struct rq_flags rf;
static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf,
static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf,
struct rq_flags rf;
static int affine_move_task(struct rq *rq, struct task_struct *p, struct rq_flags *rf,
struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags srf, drf;
struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
void __balance_callbacks(struct rq *rq, struct rq_flags *rf)
prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf)
struct task_struct *next, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags *rf)
__pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
static inline struct task_struct *pick_task(struct rq *rq, struct rq_flags *rf)
pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
struct rq_flags rf;
struct rq *___task_rq_lock(struct task_struct *p, struct rq_flags *rf)
struct rq_flags rf;
struct rq *_task_rq_lock(struct task_struct *p, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
static void __push_dl_task(struct rq *rq, struct rq_flags *rf)
struct rq_flags *rf = &scope.rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
static int balance_dl(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
static struct task_struct *__pick_task_dl(struct rq *rq, struct rq_flags *rf)
static struct task_struct *pick_task_dl(struct rq *rq, struct rq_flags *rf)
struct rq_flags rf;
do_pick_task_scx(struct rq *rq, struct rq_flags *rf, bool force_scx)
static struct task_struct *pick_task_scx(struct rq *rq, struct rq_flags *rf)
ext_server_pick_task(struct sched_dl_entity *dl_se, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
static int sched_balance_newidle(struct rq *this_rq, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
static int sched_balance_newidle(struct rq *this_rq, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
static struct task_struct *pick_task_fair(struct rq *rq, struct rq_flags *rf)
pick_next_task_fair(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
fair_server_pick_task(struct sched_dl_entity *dl_se, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
balance_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
struct task_struct *pick_task_idle(struct rq *rq, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
static int balance_rt(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
static struct task_struct *pick_task_rt(struct rq *rq, struct rq_flags *rf)
struct rq_flags rf;
static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf)
static inline void rq_unpin_lock(struct rq *rq, struct rq_flags *rf)
static inline void rq_repin_lock(struct rq *rq, struct rq_flags *rf)
extern struct rq *___task_rq_lock(struct task_struct *p, struct rq_flags *rf) __acquires_ret;
extern struct rq *_task_rq_lock(struct task_struct *p, struct rq_flags *rf)
__task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf)
struct rq *rq; struct rq_flags rf)
struct rq *rq; struct rq_flags rf)
static inline void rq_lock_irqsave(struct rq *rq, struct rq_flags *rf)
static inline void rq_lock_irq(struct rq *rq, struct rq_flags *rf)
static inline void rq_lock(struct rq *rq, struct rq_flags *rf)
static inline void rq_unlock_irqrestore(struct rq *rq, struct rq_flags *rf)
static inline void rq_unlock_irq(struct rq *rq, struct rq_flags *rf)
static inline void rq_unlock(struct rq *rq, struct rq_flags *rf)
struct rq_flags rf)
struct rq_flags rf)
struct rq_flags rf)
static inline struct rq *_this_rq_lock_irq(struct rq_flags *rf) __acquires_ret
int (*balance)(struct rq *rq, struct task_struct *prev, struct rq_flags *rf);
struct task_struct *(*pick_task)(struct rq *rq, struct rq_flags *rf);
struct rq_flags *rf);
struct rq_flags *rf);
extern struct task_struct *pick_task_idle(struct rq *rq, struct rq_flags *rf);
extern void __balance_callbacks(struct rq *rq, struct rq_flags *rf);
struct rq_flags rf;
balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
static struct task_struct *pick_task_stop(struct rq *rq, struct rq_flags *rf)
struct rq_flags rf;
struct rq_flags rf;
struct rq_flags rf;
set_bit(RQ_USEDEFERRAL, &rqstp->rq_flags);
clear_bit(RQ_DROPME, &rqstp->rq_flags);
if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags))
set_bit(RQ_DROPME, &rqstp->rq_flags);
if (!test_bit(RQ_DATA, &rqstp->rq_flags)) {
set_bit(RQ_DATA, &rqstp->rq_flags);
if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) {
set_bit(RQ_LOCAL, &rqstp->rq_flags);
clear_bit(RQ_LOCAL, &rqstp->rq_flags);
set_bit(RQ_SECURE, &rqstp->rq_flags);
clear_bit(RQ_SECURE, &rqstp->rq_flags);
set_bit(RQ_SECURE, &rqstp->rq_flags);