Home
last modified time | relevance | path

Searched refs:hctx (Results 1 – 25 of 73) sorted by relevance

123

/linux/block/
A Dblk-mq-sched.c58 struct blk_mq_hw_ctx *hctx = in blk_mq_dispatch_hctx_list() local
65 if (rq->mq_hctx != hctx) { in blk_mq_dispatch_hctx_list()
97 if (hctx->dispatch_busy) in __blk_mq_do_dispatch_sched()
141 if (rq->mq_hctx != hctx) in __blk_mq_do_dispatch_sched()
204 return hctx->ctxs[idx]; in blk_mq_next_ctx()
280 spin_lock(&hctx->lock); in __blk_mq_sched_dispatch_requests()
353 type = hctx->type; in blk_mq_sched_bio_merge()
392 if (!hctx->sched_tags) in blk_mq_sched_alloc_map_and_rqs()
410 if (hctx->sched_tags) { in blk_mq_sched_tags_teardown()
517 if (hctx->sched_tags) in blk_mq_sched_free_rqs()
[all …]
A Dblk-mq.h162 struct blk_mq_hw_ctx *hctx; member
191 struct blk_mq_hw_ctx *hctx) in bt_wait_ptr() argument
193 if (!hctx) in bt_wait_ptr()
204 __blk_mq_tag_busy(hctx); in blk_mq_tag_busy()
210 __blk_mq_tag_idle(hctx); in blk_mq_tag_idle()
227 return data->hctx->sched_tags; in blk_mq_tags_from_data()
228 return data->hctx->tags; in blk_mq_tags_from_data()
238 return hctx->nr_ctx && hctx->tags; in blk_mq_hw_queue_mapped()
282 atomic_add(val, &hctx->nr_active); in __blk_mq_add_active_requests()
339 blk_mq_dec_active_requests(hctx); in __blk_mq_put_driver_tag()
[all …]
A Dblk-mq-sysfs.c38 sbitmap_free(&hctx->ctx_map); in blk_mq_hw_sysfs_release()
40 kfree(hctx->ctxs); in blk_mq_hw_sysfs_release()
41 kfree(hctx); in blk_mq_hw_sysfs_release()
53 struct blk_mq_hw_ctx *hctx; in blk_mq_hw_sysfs_show() local
59 q = hctx->queue; in blk_mq_hw_sysfs_show()
149 if (!hctx->nr_ctx) in blk_mq_unregister_hctx()
155 kobject_del(&hctx->kobj); in blk_mq_unregister_hctx()
164 if (!hctx->nr_ctx) in blk_mq_register_hctx()
167 ret = kobject_add(&hctx->kobj, q->mq_kobj, "%u", hctx->queue_num); in blk_mq_register_hctx()
183 kobject_del(&hctx->kobj); in blk_mq_register_hctx()
[all …]
A Dblk-mq.c351 struct blk_mq_hw_ctx *hctx = data->hctx; in blk_mq_rq_ctx_init() local
1677 struct blk_mq_hw_ctx *hctx = flush_data->hctx; in flush_busy_ctx() local
1695 .hctx = hctx, in blk_mq_flush_busy_ctxs()
1712 struct blk_mq_hw_ctx *hctx = dispatch_data->hctx; in dispatch_rq_from_ctx() local
1733 .hctx = hctx, in blk_mq_dequeue_from_ctx()
1985 hctx->queue->mq_ops->commit_rqs(hctx); in blk_mq_commit_rqs()
3499 hctx->sched_tags : hctx->tags; in blk_mq_hctx_has_requests()
3501 .hctx = hctx, in blk_mq_hctx_has_requests()
3947 ctx->index_hw[hctx->type] = hctx->nr_ctx; in blk_mq_map_swqueue()
3948 hctx->ctxs[hctx->nr_ctx++] = ctx; in blk_mq_map_swqueue()
[all …]
A Dblk-mq-debugfs.c315 __acquires(&hctx->lock) in hctx_dispatch_start()
319 spin_lock(&hctx->lock); in hctx_dispatch_start()
331 __releases(&hctx->lock) in hctx_dispatch_stop()
368 struct show_busy_params params = { .m = m, .hctx = hctx }; in hctx_busy_show()
425 if (hctx->tags) in hctx_tags_show()
442 if (hctx->tags) in hctx_tags_bitmap_show()
459 if (hctx->sched_tags) in hctx_sched_tags_show()
476 if (hctx->sched_tags) in hctx_sched_tags_bitmap_show()
696 debugfs_create_files(hctx->debugfs_dir, hctx, blk_mq_debugfs_hctx_attrs); in blk_mq_debugfs_register_hctx()
811 if (!hctx->debugfs_dir) in blk_mq_debugfs_register_sched_hctx()
[all …]
A Dblk-mq-tag.c42 struct blk_mq_tags *tags = hctx->tags; in __blk_mq_tag_busy()
49 struct request_queue *q = hctx->queue; in __blk_mq_tag_busy()
110 !hctx_may_queue(data->hctx, bt)) in __blk_mq_get_tag()
162 ws = bt_wait_ptr(bt, data->hctx); in blk_mq_get_tag()
209 ws = bt_wait_ptr(bt, data->hctx); in blk_mq_get_tag()
246 struct blk_mq_hw_ctx *hctx; member
270 struct blk_mq_hw_ctx *hctx = iter_data->hctx; in bt_iter() local
280 tags = hctx->tags; in bt_iter()
292 if (rq->q == q && (!hctx || rq->mq_hctx == hctx)) in bt_iter()
318 .hctx = hctx, in bt_for_each()
[all …]
A Dblk-mq-sched.h16 void blk_mq_sched_mark_restart_hctx(struct blk_mq_hw_ctx *hctx);
17 void __blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx);
19 void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx);
25 static inline void blk_mq_sched_restart(struct blk_mq_hw_ctx *hctx) in blk_mq_sched_restart() argument
27 if (test_bit(BLK_MQ_S_SCHED_RESTART, &hctx->state)) in blk_mq_sched_restart()
28 __blk_mq_sched_restart(hctx); in blk_mq_sched_restart()
70 static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx *hctx) in blk_mq_sched_has_work() argument
72 struct elevator_queue *e = hctx->queue->elevator; in blk_mq_sched_has_work()
75 return e->type->ops.has_work(hctx); in blk_mq_sched_has_work()
80 static inline bool blk_mq_sched_needs_restart(struct blk_mq_hw_ctx *hctx) in blk_mq_sched_needs_restart() argument
[all …]
A Dkyber-iosched.c475 GFP_KERNEL, hctx->numa_node); in kyber_init_hctx()
479 for (i = 0; i < hctx->nr_ctx; i++) in kyber_init_hctx()
507 hctx->sched_data = khd; in kyber_init_hctx()
508 kyber_depth_updated(hctx); in kyber_init_hctx()
527 kfree(hctx->sched_data); in kyber_exit_hctx()
609 rq->mq_ctx->index_hw[hctx->type]); in kyber_insert_requests()
702 blk_mq_run_hw_queue(hctx, true); in kyber_domain_wake()
708 struct blk_mq_hw_ctx *hctx) in kyber_get_domain_token() argument
756 struct blk_mq_hw_ctx *hctx) in kyber_dispatch_cur_domain() argument
971 struct blk_mq_hw_ctx *hctx = data; in kyber_cur_domain_show() local
[all …]
A Dblk-mq-debugfs.h25 struct blk_mq_hw_ctx *hctx);
26 void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx);
33 struct blk_mq_hw_ctx *hctx);
34 void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx);
44 struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_register_hctx() argument
48 static inline void blk_mq_debugfs_unregister_hctx(struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_unregister_hctx() argument
69 struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_register_sched_hctx() argument
73 static inline void blk_mq_debugfs_unregister_sched_hctx(struct blk_mq_hw_ctx *hctx) in blk_mq_debugfs_unregister_sched_hctx() argument
A Dmq-deadline.c498 struct sbitmap_queue *bt = &hctx->sched_tags->bitmap_tags; in dd_to_word_depth()
499 const unsigned int nrr = hctx->queue->nr_requests; in dd_to_word_depth()
524 static void dd_depth_updated(struct blk_mq_hw_ctx *hctx) in dd_depth_updated() argument
526 struct request_queue *q = hctx->queue; in dd_depth_updated()
528 struct blk_mq_tags *tags = hctx->sched_tags; in dd_depth_updated()
538 dd_depth_updated(hctx); in dd_init_hctx()
676 struct request_queue *q = hctx->queue; in dd_insert_request()
724 static void dd_insert_requests(struct blk_mq_hw_ctx *hctx, in dd_insert_requests() argument
728 struct request_queue *q = hctx->queue; in dd_insert_requests()
738 dd_insert_request(hctx, rq, flags, &free); in dd_insert_requests()
[all …]
/linux/samples/hid/
A Dhid_mouse.bpf.c8 static int hid_y_event(struct hid_bpf_ctx *hctx) in hid_y_event() argument
11 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 9 /* size */); in hid_y_event()
16 bpf_printk("event: size: %d", hctx->size); in hid_y_event()
53 static int hid_x_event(struct hid_bpf_ctx *hctx) in hid_x_event() argument
56 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 9 /* size */); in hid_x_event()
71 int BPF_PROG(hid_event, struct hid_bpf_ctx *hctx, enum hid_report_type type) in BPF_PROG() argument
73 int ret = hid_y_event(hctx); in BPF_PROG()
78 return hid_x_event(hctx); in BPF_PROG()
83 int BPF_PROG(hid_rdesc_fixup, struct hid_bpf_ctx *hctx) in BPF_PROG() argument
85 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in BPF_PROG()
A Dhid_surface_dial.bpf.c14 int BPF_PROG(hid_event, struct hid_bpf_ctx *hctx) in BPF_PROG() argument
16 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 9 /* size */); in BPF_PROG()
105 int BPF_PROG(hid_rdesc_fixup, struct hid_bpf_ctx *hctx) in BPF_PROG() argument
107 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in BPF_PROG()
/linux/drivers/hid/bpf/progs/
A DXPPen__ArtistPro16Gen2.bpf.c88 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in SEC()
98 if (hctx->hid->product == PID_ARTIST_PRO14_GEN2) { in SEC()
108 static int xppen_16_fix_eraser(struct hid_bpf_ctx *hctx) in xppen_16_fix_eraser() argument
110 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in xppen_16_fix_eraser()
209 static int xppen_16_fix_angle_offset(struct hid_bpf_ctx *hctx) in xppen_16_fix_angle_offset() argument
211 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in xppen_16_fix_angle_offset()
244 if (hctx->hid->product == PID_ARTIST_PRO14_GEN2) { in xppen_16_fix_angle_offset()
247 } else if (hctx->hid->product == PID_ARTIST_PRO16_GEN2) { in xppen_16_fix_angle_offset()
256 int BPF_PROG(xppen_artist_pro_16_device_event, struct hid_bpf_ctx *hctx) in SEC()
258 int ret = xppen_16_fix_angle_offset(hctx); in SEC()
[all …]
A DHuion__Kamvas-Pro-19.bpf.c195 int BPF_PROG(hid_fix_rdesc_huion_kamvas_pro_19, struct hid_bpf_ctx *hctx) in SEC()
197 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC()
219 int BPF_PROG(kamvas_pro_19_fix_3rd_button, struct hid_bpf_ctx *hctx) in SEC()
221 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in SEC()
274 struct hid_bpf_ctx *hctx = hid_bpf_allocate_context(ctx->hid); in probe() local
276 if (!hctx) { in probe()
281 const char *name = hctx->hid->name; in probe()
290 hid_bpf_release_context(hctx); in probe()
A DXPPen__DecoMini4.bpf.c153 int BPF_PROG(hid_rdesc_fixup_xppen_deco_mini_4, struct hid_bpf_ctx *hctx) in SEC()
155 __u8 *data = hid_bpf_get_data(hctx, 0, HID_MAX_DESCRIPTOR_SIZE); in SEC()
160 if (hctx->size == RDESC_SIZE_PAD) { in SEC()
163 } else if (hctx->size == RDESC_SIZE_PEN) { in SEC()
172 int BPF_PROG(hid_device_event_xppen_deco_mini_4, struct hid_bpf_ctx *hctx) in SEC()
174 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 8 /* size */); in SEC()
A DFR-TEC__Raptor-Mach-2.bpf.c137 int BPF_PROG(hid_fix_rdesc_raptor_mach_2, struct hid_bpf_ctx *hctx) in SEC()
139 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC()
156 int BPF_PROG(raptor_mach_2_fix_hat_switch, struct hid_bpf_ctx *hctx) in SEC()
158 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 64 /* size */); in SEC()
A DXPPen__Artist24.bpf.c93 int BPF_PROG(hid_fix_rdesc_xppen_artist24, struct hid_bpf_ctx *hctx) in SEC()
95 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in SEC()
154 int BPF_PROG(xppen_24_fix_eraser, struct hid_bpf_ctx *hctx) in SEC()
156 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in SEC()
A DThrustmaster__TCA-Yoke-Boeing.bpf.c100 int BPF_PROG(hid_fix_rdesc_tca_yoke, struct hid_bpf_ctx *hctx) in SEC()
104 if (hctx->size != expected_length) in SEC()
107 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC()
A DHuion__Inspiroy-2-S.bpf.c359 int BPF_PROG(hid_fix_rdesc, struct hid_bpf_ctx *hctx) in SEC()
361 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC()
362 __s32 rdesc_size = hctx->size; in SEC()
405 int BPF_PROG(inspiroy_2_fix_events, struct hid_bpf_ctx *hctx) in SEC()
407 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 10 /* size */); in SEC()
A DHuion__Dial-2.bpf.c420 int BPF_PROG(dial_2_fix_rdesc, struct hid_bpf_ctx *hctx) in SEC()
422 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, HID_MAX_DESCRIPTOR_SIZE /* size */); in SEC()
423 __s32 rdesc_size = hctx->size; in SEC()
466 int BPF_PROG(dial_2_fix_events, struct hid_bpf_ctx *hctx) in SEC()
468 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 16 /* size */); in SEC()
A DHP__Elite-Presenter.bpf.c34 int BPF_PROG(hid_fix_rdesc, struct hid_bpf_ctx *hctx) in SEC()
36 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 4096 /* size */); in SEC()
/linux/drivers/crypto/marvell/octeontx2/
A Dcn10k_cpt.c115 void cn10k_cpt_hw_ctx_set(union cn10k_cpt_hw_ctx *hctx, u16 ctx_sz) in cn10k_cpt_hw_ctx_set() argument
117 hctx->w0.aop_valid = 1; in cn10k_cpt_hw_ctx_set()
118 hctx->w0.ctx_hdr_sz = 0; in cn10k_cpt_hw_ctx_set()
119 hctx->w0.ctx_sz = ctx_sz; in cn10k_cpt_hw_ctx_set()
120 hctx->w0.ctx_push_sz = 1; in cn10k_cpt_hw_ctx_set()
127 union cn10k_cpt_hw_ctx *hctx; in cn10k_cpt_hw_ctx_init() local
136 hctx = kmalloc(CN10K_CPT_HW_CTX_SIZE, GFP_KERNEL); in cn10k_cpt_hw_ctx_init()
137 if (unlikely(!hctx)) in cn10k_cpt_hw_ctx_init()
142 kfree(hctx); in cn10k_cpt_hw_ctx_init()
146 cn10k_cpt_hw_ctx_set(hctx, 1); in cn10k_cpt_hw_ctx_init()
[all …]
/linux/tools/testing/selftests/hid/progs/
A Dhid.c335 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 3 /* size */); in BPF_PROG()
345 ret = hid_bpf_hw_request(hctx, data, 2, rtype, reqtype); in BPF_PROG()
366 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 3 /* size */); in BPF_PROG()
377 ret = hid_bpf_hw_request(hctx, data, 2, rtype, reqtype); in BPF_PROG()
404 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 3 /* size */); in BPF_PROG()
412 return hid_bpf_hw_output_report(hctx, data, 2); in BPF_PROG()
425 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 3 /* size */); in BPF_PROG()
435 ret = hid_bpf_hw_output_report(hctx, data, 2); in BPF_PROG()
566 int BPF_PROG(hid_test_infinite_loop_input_report, struct hid_bpf_ctx *hctx, in BPF_PROG() argument
569 __u8 *data = hid_bpf_get_data(hctx, 0 /* offset */, 6 /* size */); in BPF_PROG()
[all …]
/linux/net/dccp/ccids/
A Dccid3.h104 struct ccid3_hc_tx_sock *hctx = ccid_priv(dccp_sk(sk)->dccps_hc_tx_ccid); in ccid3_hc_tx_sk() local
105 BUG_ON(hctx == NULL); in ccid3_hc_tx_sk()
106 return hctx; in ccid3_hc_tx_sk()
/linux/include/linux/
A Dblk-mq.h904 void blk_mq_stop_hw_queue(struct blk_mq_hw_ctx *hctx);
905 void blk_mq_start_hw_queue(struct blk_mq_hw_ctx *hctx);
908 void blk_mq_start_stopped_hw_queue(struct blk_mq_hw_ctx *hctx, bool async);
915 void blk_mq_delay_run_hw_queue(struct blk_mq_hw_ctx *hctx, unsigned long msecs);
916 void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async);
973 #define queue_for_each_hw_ctx(q, hctx, i) \ argument
974 xa_for_each(&(q)->hctx_table, (i), (hctx))
976 #define hctx_for_each_ctx(hctx, ctx, i) \ argument
977 for ((i) = 0; (i) < (hctx)->nr_ctx && \
978 ({ ctx = (hctx)->ctxs[(i)]; 1; }); (i)++)
[all …]

Completed in 845 milliseconds

123