Home
last modified time | relevance | path

Searched refs:rq_flags (Results 1 – 25 of 65) sorted by relevance

123

/linux/block/
A Dblk-mq-sched.h40 if (rq->rq_flags & RQF_USE_SCHED) { in blk_mq_sched_allow_merge()
51 if (rq->rq_flags & RQF_USE_SCHED) { in blk_mq_sched_completed_request()
61 if (rq->rq_flags & RQF_USE_SCHED) { in blk_mq_sched_requeue_request()
A Dblk-mq.c361 data->rq_flags |= RQF_PM; in blk_mq_rq_ctx_init()
363 data->rq_flags |= RQF_IO_STAT; in blk_mq_rq_ctx_init()
364 rq->rq_flags = data->rq_flags; in blk_mq_rq_ctx_init()
479 data->rq_flags |= RQF_RESV; in __blk_mq_alloc_requests()
662 data.rq_flags |= RQF_SCHED_TAGS; in blk_mq_alloc_request_hctx()
667 data.rq_flags |= RQF_RESV; in blk_mq_alloc_request_hctx()
701 rq->rq_flags &= ~RQF_USE_SCHED; in blk_mq_finish_request()
1025 if (rq->rq_flags & RQF_STATS) in __blk_mq_end_request_acct()
1251 rq->rq_flags |= RQF_STATS; in blk_mq_start_request()
1418 rq->rq_flags &= ~RQF_TIMED_OUT; in __blk_mq_requeue_request()
[all …]
A Dblk-merge.c581 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in __blk_rq_map_sg()
755 if (rq->rq_flags & RQF_MIXED_MERGE) in blk_rq_set_mixed_merge()
768 rq->rq_flags |= RQF_MIXED_MERGE; in blk_rq_set_mixed_merge()
787 if (req->rq_flags & RQF_MIXED_MERGE) { in blk_update_mixed_merge()
885 if (((req->rq_flags | next->rq_flags) & RQF_MIXED_MERGE) || in attempt_merge()
1032 if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) in bio_attempt_back_merge()
1055 if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) in bio_attempt_front_merge()
A Dblk-flush.c120 rq->rq_flags &= ~RQF_FLUSH_SEQ; in blk_flush_restore_request()
319 flush_rq->rq_flags |= RQF_FLUSH_SEQ; in blk_kick_flush()
372 rq->rq_flags |= RQF_FLUSH_SEQ; in blk_rq_init_flush()
A Dblk-pm.h21 if (rq->q->dev && !(rq->rq_flags & RQF_PM)) in blk_pm_mark_last_busy()
A Dblk.h156 if (rq->rq_flags & RQF_NOMERGE_FLAGS) in rq_mergeable()
310 #define ELV_ON_HASH(rq) ((rq)->rq_flags & RQF_HASHED)
416 return (rq->rq_flags & RQF_IO_STAT) && !blk_rq_is_passthrough(rq); in blk_do_io_stat()
501 if (rq->rq_flags & RQF_ZONE_WRITE_PLUGGING) in blk_zone_finish_request()
A Dblk-timeout.c140 req->rq_flags &= ~RQF_TIMED_OUT; in blk_add_timer()
A Dblk-mq.h154 req_flags_t rq_flags; member
226 if (data->rq_flags & RQF_SCHED_TAGS) in blk_mq_tags_from_data()
/linux/include/linux/
A Dblk-mq.h108 req_flags_t rq_flags; member
866 return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS | RQF_USE_SCHED)); in blk_mq_need_time_stamp()
871 return rq->rq_flags & RQF_RESV; in blk_mq_is_reserved_rq()
886 if (!iob || (req->rq_flags & RQF_SCHED_TAGS) || ioerror || in blk_mq_add_to_batch()
1104 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_payload_bytes()
1115 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in req_bvec()
1154 if (rq->rq_flags & RQF_SPECIAL_PAYLOAD) in blk_rq_nr_phys_segments()
/linux/kernel/sched/
A Dsched.h1697 struct rq_flags { struct
1722 static inline void rq_pin_lock(struct rq *rq, struct rq_flags *rf) in rq_pin_lock()
1762 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf)
1786 struct rq *rq; struct rq_flags rf)
1795 static inline void rq_lock_irq(struct rq *rq, struct rq_flags *rf) in rq_lock_irq()
1802 static inline void rq_lock(struct rq *rq, struct rq_flags *rf) in rq_lock()
1823 static inline void rq_unlock(struct rq *rq, struct rq_flags *rf) in rq_unlock()
1833 struct rq_flags rf)
1838 struct rq_flags rf)
1843 struct rq_flags rf)
[all …]
A Dcore.c829 struct rq_flags rf; in hrtick()
857 struct rq_flags rf; in __hrtick_start()
1771 struct rq_flags rf; in uclamp_update_active()
2193 struct rq_flags rf; in wait_task_inactive()
2498 struct rq_flags rf; in migration_cpu_stop()
3102 struct rq_flags rf; in __set_cpus_allowed_ptr()
3146 struct rq_flags rf; in restrict_cpus_allowed_ptr()
3737 struct rq_flags rf; in ttwu_runnable()
3767 struct rq_flags rf; in sched_ttwu_pending()
3946 struct rq_flags rf; in ttwu_queue()
[all …]
A Dstop_task.c19 balance_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in balance_stop()
A Dcore_sched.c57 struct rq_flags rf; in sched_core_update_cookie()
A Dpsi.c1049 struct rq_flags rf; in psi_memstall_enter()
1080 struct rq_flags rf; in psi_memstall_leave()
1149 struct rq_flags rf; in cgroup_move_task()
1225 struct rq_flags rf; in psi_cgroup_restart()
A Ddeadline.c1176 static void __push_dl_task(struct rq *rq, struct rq_flags *rf) in __push_dl_task()
1204 struct rq_flags *rf = &scope.rf; in dl_server_timer()
1270 struct rq_flags rf; in dl_task_timer()
1764 struct rq_flags rf; in inactive_task_timer()
2264 struct rq_flags rf; in migrate_task_rq_dl()
2316 static int balance_dl(struct rq *rq, struct task_struct *p, struct rq_flags *rf) in balance_dl()
2949 struct rq_flags rf; in dl_add_task_root_domain()
/linux/drivers/net/ethernet/fungible/funcore/
A Dfun_queue.h69 u16 rq_flags; member
120 u16 rq_flags; member
A Dfun_queue.c459 funq->rq_flags = req->rq_flags | FUN_ADMIN_EPSQ_CREATE_FLAG_RQ; in fun_alloc_queue()
528 rc = fun_sq_create(fdev, funq->rq_flags, funq->rqid, funq->cqid, 0, in fun_create_rq()
/linux/include/linux/sunrpc/
A Dsvc.h210 unsigned long rq_flags; /* flags field */ member
310 set_bit(RQ_VICTIM, &rqstp->rq_flags); in svc_thread_should_stop()
312 return test_bit(RQ_VICTIM, &rqstp->rq_flags); in svc_thread_should_stop()
/linux/drivers/scsi/
A Dscsi_lib.c118 if (rq->rq_flags & RQF_DONTPREP) { in scsi_mq_requeue_cmd()
119 rq->rq_flags &= ~RQF_DONTPREP; in scsi_mq_requeue_cmd()
320 req->rq_flags |= RQF_QUIET; in scsi_execute_cmd()
731 if (!(rq->rq_flags & RQF_MIXED_MERGE)) in scsi_rq_err_bytes()
910 if (!(req->rq_flags & RQF_QUIET)) { in scsi_io_completion_action()
1001 else if (req->rq_flags & RQF_QUIET) in scsi_io_completion_nz_result()
1235 if (rq->rq_flags & RQF_DONTPREP) { in scsi_cleanup_rq()
1237 rq->rq_flags &= ~RQF_DONTPREP; in scsi_cleanup_rq()
1828 req->rq_flags |= RQF_DONTPREP; in scsi_queue_rq()
1870 if (req->rq_flags & RQF_DONTPREP) in scsi_queue_rq()
[all …]
/linux/drivers/nvme/host/
A Dioctl.c102 struct nvme_command *cmd, blk_opf_t rq_flags, in nvme_alloc_user_request() argument
107 req = blk_mq_alloc_request(q, nvme_req_op(cmd) | rq_flags, blk_flags); in nvme_alloc_user_request()
461 blk_opf_t rq_flags = REQ_ALLOC_CACHE; in nvme_uring_cmd_io() local
496 rq_flags |= REQ_NOWAIT; in nvme_uring_cmd_io()
500 rq_flags |= REQ_POLLED; in nvme_uring_cmd_io()
502 req = nvme_alloc_user_request(q, &c, rq_flags, blk_flags); in nvme_uring_cmd_io()
/linux/drivers/mmc/core/
A Dqueue.c243 req->rq_flags |= RQF_QUIET; in mmc_mq_queue_rq()
291 if (!(req->rq_flags & RQF_DONTPREP)) { in mmc_mq_queue_rq()
293 req->rq_flags |= RQF_DONTPREP; in mmc_mq_queue_rq()
/linux/net/sunrpc/
A Dsvc_xprt.c419 if (!test_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_reserve_slot()
423 set_bit(RQ_DATA, &rqstp->rq_flags); in svc_xprt_reserve_slot()
431 if (test_and_clear_bit(RQ_DATA, &rqstp->rq_flags)) { in svc_xprt_release_slot()
1185 if (rqstp->rq_arg.page_len || !test_bit(RQ_USEDEFERRAL, &rqstp->rq_flags)) in svc_defer()
1216 set_bit(RQ_DROPME, &rqstp->rq_flags); in svc_defer()
/linux/drivers/md/
A Ddm-rq.c266 if (rq->rq_flags & RQF_FAILED) in dm_softirq_done()
293 rq->rq_flags |= RQF_FAILED; in dm_kill_unmapped_request()
/linux/drivers/scsi/device_handler/
A Dscsi_dh_hp_sw.c191 req->rq_flags |= RQF_QUIET; in hp_sw_prep_fn()
/linux/fs/nfsd/
A Dnfscache.c536 if (!test_bit(RQ_SECURE, &rqstp->rq_flags) && rp->c_secure) in nfsd_cache_lookup()
632 rp->c_secure = test_bit(RQ_SECURE, &rqstp->rq_flags); in nfsd_cache_update()

Completed in 88 milliseconds

123