Home
last modified time | relevance | path

Searched refs:q (Results 1 – 25 of 58) sorted by relevance

123

/include/linux/
A Dlwq.h33 static inline void lwq_init(struct lwq *q) in lwq_init() argument
35 spin_lock_init(&q->lock); in lwq_init()
36 q->ready = NULL; in lwq_init()
37 init_llist_head(&q->new); in lwq_init()
48 static inline bool lwq_empty(struct lwq *q) in lwq_empty() argument
51 return smp_load_acquire(&q->ready) == NULL && llist_empty(&q->new); in lwq_empty()
54 struct llist_node *__lwq_dequeue(struct lwq *q);
65 #define lwq_dequeue(q, type, member) \ argument
103 return llist_add(&n->node, &q->new) && in lwq_enqueue()
104 smp_load_acquire(&q->ready) == NULL; in lwq_enqueue()
[all …]
A Dblktrace_api.h49 #define blk_add_cgroup_trace_msg(q, css, fmt, ...) \ argument
54 bt = rcu_dereference((q)->blk_trace); \
59 #define blk_add_trace_msg(q, fmt, ...) \ argument
60 blk_add_cgroup_trace_msg(q, NULL, fmt, ##__VA_ARGS__)
69 bt = rcu_dereference(q->blk_trace); in blk_trace_note_message_enabled()
80 extern int blk_trace_remove(struct request_queue *q);
84 # define blk_trace_shutdown(q) do { } while (0) argument
86 # define blk_trace_setup(q, name, dev, bdev, arg) (-ENOTTY) argument
87 # define blk_trace_startstop(q, start) (-ENOTTY) argument
88 # define blk_add_trace_msg(q, fmt, ...) do { } while (0) argument
[all …]
A Dblk-pm.h13 extern void blk_pm_runtime_init(struct request_queue *q, struct device *dev);
14 extern int blk_pre_runtime_suspend(struct request_queue *q);
15 extern void blk_post_runtime_suspend(struct request_queue *q, int err);
16 extern void blk_pre_runtime_resume(struct request_queue *q);
17 extern void blk_post_runtime_resume(struct request_queue *q);
19 static inline void blk_pm_runtime_init(struct request_queue *q, in blk_pm_runtime_init() argument
A Dblkdev.h667 #define blk_queue_dying(q) test_bit(QUEUE_FLAG_DYING, &(q)->queue_flags) argument
672 #define blk_queue_nonrot(q) (!((q)->limits.features & BLK_FEAT_ROTATIONAL)) argument
673 #define blk_queue_io_stat(q) ((q)->limits.features & BLK_FEAT_IO_STAT) argument
676 #define blk_queue_dax(q) ((q)->limits.features & BLK_FEAT_DAX) argument
677 #define blk_queue_pci_p2pdma(q) ((q)->limits.features & BLK_FEAT_PCI_P2PDMA) argument
689 #define blk_queue_pm_only(q) atomic_read(&(q)->pm_only) argument
710 return q->mq_ops; in queue_is_mq()
716 return q->rpm_status; in queue_rpm_status()
750 if (q->queue_depth) in blk_queue_depth()
1067 return q->limits; in queue_limits_start_update()
[all …]
A Dfortify-string.h286 return __real_strscpy(p, q, size); in sized_strscpy()
297 len = __compiletime_strlen(q); in sized_strscpy()
309 len = strnlen(q, size); in sized_strscpy()
329 return __real_strscpy(p, q, len); in sized_strscpy()
371 copy_len = strlen(q); in strlcat()
453 copy_len = strnlen(q, count); in strncat()
690 #define memcpy(p, q, s) __fortify_memcpy_chk(p, q, s, \ argument
694 #define memmove(p, q, s) __fortify_memcpy_chk(p, q, s, \ argument
795 return __underlying_strcpy(p, q); in strcpy()
796 size = strlen(q) + 1; in strcpy()
[all …]
A Dt10-pi.h42 unsigned int shift = ilog2(queue_logical_block_size(rq->q)); in t10_pi_ref_tag()
45 rq->q->limits.integrity.interval_exp) in t10_pi_ref_tag()
46 shift = rq->q->limits.integrity.interval_exp; in t10_pi_ref_tag()
67 unsigned int shift = ilog2(queue_logical_block_size(rq->q)); in ext_pi_ref_tag()
70 rq->q->limits.integrity.interval_exp) in ext_pi_ref_tag()
71 shift = rq->q->limits.integrity.interval_exp; in ext_pi_ref_tag()
A Dblk-integrity.h36 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity() argument
38 return q->limits.integrity.metadata_size; in blk_integrity_queue_supports_integrity()
55 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments() argument
57 return q->limits.max_integrity_segments; in queue_max_integrity_segments()
102 static inline int blk_rq_count_integrity_sg(struct request_queue *q, in blk_rq_count_integrity_sg() argument
107 static inline int blk_rq_map_integrity_sg(struct request *q, in blk_rq_map_integrity_sg() argument
127 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity() argument
132 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments() argument
A Dswait.h69 extern void __init_swait_queue_head(struct swait_queue_head *q, const char *name,
72 #define init_swait_queue_head(q) \ argument
75 __init_swait_queue_head((q), #q, &__key); \
147 extern void swake_up_one(struct swait_queue_head *q);
148 extern void swake_up_all(struct swait_queue_head *q);
149 extern void swake_up_locked(struct swait_queue_head *q, int wake_flags);
151 extern void prepare_to_swait_exclusive(struct swait_queue_head *q, struct swait_queue *wait, int st…
152 extern long prepare_to_swait_event(struct swait_queue_head *q, struct swait_queue *wait, int state);
154 extern void __finish_swait(struct swait_queue_head *q, struct swait_queue *wait);
155 extern void finish_swait(struct swait_queue_head *q, struct swait_queue *wait);
A Duacce.h41 struct uacce_queue *q);
42 void (*put_queue)(struct uacce_queue *q);
43 int (*start_queue)(struct uacce_queue *q);
44 void (*stop_queue)(struct uacce_queue *q);
45 int (*is_q_updated)(struct uacce_queue *q);
46 int (*mmap)(struct uacce_queue *q, struct vm_area_struct *vma,
48 long (*ioctl)(struct uacce_queue *q, unsigned int cmd,
A Dblk-mq.h104 struct request_queue *q; member
720 struct request_queue *q);
915 void blk_mq_quiesce_queue(struct request_queue *q);
930 blk_mq_freeze_queue(struct request_queue *q) in blk_mq_freeze_queue() argument
934 blk_mq_freeze_queue_nomemsave(q); in blk_mq_freeze_queue()
940 blk_mq_unfreeze_queue_nomemrestore(q); in blk_mq_unfreeze_queue()
966 return __blk_should_fake_timeout(q); in blk_should_fake_timeout()
998 #define queue_for_each_hw_ctx(q, hctx, i) \ argument
999 xa_for_each(&(q)->hctx_table, (i), (hctx))
1007 if (rq->q->mq_ops->cleanup_rq) in blk_mq_cleanup_rq()
[all …]
/include/media/
A Dvideobuf2-core.h700 return q->allow_cache_hints && q->memory == VB2_MEMORY_MMAP; in vb2_queue_allows_cache_hints()
1163 return q->streaming; in vb2_is_streaming()
1181 return q->fileio; in vb2_fileio_is_active()
1190 if (q->bufs_bitmap) in vb2_get_num_buffers()
1191 return bitmap_weight(q->bufs_bitmap, q->max_num_buffers); in vb2_get_num_buffers()
1204 return !!q->is_busy; in vb2_is_busy()
1213 return q->drv_priv; in vb2_get_drv_priv()
1280 q->last_buffer_dequeued = false; in vb2_clear_last_buffer_dequeued()
1296 if (!q->bufs) in vb2_get_buffer()
1299 if (index >= q->max_num_buffers) in vb2_get_buffer()
[all …]
A Dvideobuf2-v4l2.h75 int vb2_querybuf(struct vb2_queue *q, struct v4l2_buffer *b);
85 int vb2_reqbufs(struct vb2_queue *q, struct v4l2_requestbuffers *req);
119 int vb2_prepare_buf(struct vb2_queue *q, struct media_device *mdev,
145 int vb2_qbuf(struct vb2_queue *q, struct media_device *mdev,
157 int vb2_expbuf(struct vb2_queue *q, struct v4l2_exportbuffer *eb);
201 int vb2_streamon(struct vb2_queue *q, enum v4l2_buf_type type);
220 int vb2_streamoff(struct vb2_queue *q, enum v4l2_buf_type type);
233 int __must_check vb2_queue_init(struct vb2_queue *q);
256 void vb2_queue_release(struct vb2_queue *q);
272 int vb2_queue_change_type(struct vb2_queue *q, unsigned int type);
[all …]
/include/trace/events/
A Dvb2.h12 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb),
13 TP_ARGS(q, vb),
26 __entry->owner = q->owner;
27 __entry->queued_count = q->queued_count;
29 atomic_read(&q->owned_by_drv_count);
47 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb),
48 TP_ARGS(q, vb)
52 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb),
53 TP_ARGS(q, vb)
58 TP_ARGS(q, vb)
[all …]
A Dqdisc.h79 TP_PROTO(struct Qdisc *q),
81 TP_ARGS(q),
84 __string( dev, qdisc_dev(q) ? qdisc_dev(q)->name : "(null)" )
85 __string( kind, q->ops->id )
93 __entry->parent = q->parent;
94 __entry->handle = q->handle;
104 TP_PROTO(struct Qdisc *q),
106 TP_ARGS(q),
110 __string( kind, q->ops->id )
118 __entry->parent = q->parent;
[all …]
A Dblock.h99 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0;
135 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0;
208 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0;
311 TP_ARGS(q, bio),
428 TP_PROTO(struct request_queue *q),
430 TP_ARGS(q),
447 TP_ARGS(q, depth, explicit),
475 TP_ARGS(q, depth, explicit)
642 TP_ARGS(q, zno, sector, nr_sectors),
669 TP_ARGS(q, zno, sector, nr_sectors)
[all …]
/include/net/
A Dipv6_frag.h24 struct inet_frag_queue q; member
34 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init()
37 q->key.v6 = *key; in ip6frag_init()
73 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue()
75 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue()
80 fq->q.flags |= INET_FRAG_DROP; in ip6frag_expire_frag_queue()
81 inet_frag_kill(&fq->q, &refs); in ip6frag_expire_frag_queue()
98 head = inet_frag_pull_head(&fq->q); in ip6frag_expire_frag_queue()
103 spin_unlock(&fq->q.lock); in ip6frag_expire_frag_queue()
110 spin_unlock(&fq->q.lock); in ip6frag_expire_frag_queue()
[all …]
A Dinet_frag.h110 void (*constructor)(struct inet_frag_queue *q,
140 void inet_frag_kill(struct inet_frag_queue *q, int *refs);
141 void inet_frag_destroy(struct inet_frag_queue *q);
148 static inline void inet_frag_putn(struct inet_frag_queue *q, int refs) in inet_frag_putn() argument
150 if (refs && refcount_sub_and_test(refs, &q->refcnt)) in inet_frag_putn()
151 inet_frag_destroy(q); in inet_frag_putn()
186 int inet_frag_queue_insert(struct inet_frag_queue *q, struct sk_buff *skb,
188 void *inet_frag_reasm_prepare(struct inet_frag_queue *q, struct sk_buff *skb,
190 void inet_frag_reasm_finish(struct inet_frag_queue *q, struct sk_buff *head,
192 struct sk_buff *inet_frag_pull_head(struct inet_frag_queue *q);
A Dpkt_sched.h23 #define qdisc_priv(q) \ argument
24 _Generic(q, \
25 const struct Qdisc * : (const void *)&q->privdata, \
26 struct Qdisc * : (void *)&q->privdata)
109 void qdisc_hash_del(struct Qdisc *q);
121 void __qdisc_run(struct Qdisc *q);
123 static inline void qdisc_run(struct Qdisc *q) in qdisc_run() argument
125 if (qdisc_run_begin(q)) { in qdisc_run()
126 __qdisc_run(q); in qdisc_run()
127 qdisc_run_end(q); in qdisc_run()
[all …]
A Dsch_generic.h117 struct qdisc_skb_head q; member
473 struct Qdisc *q; member
519 return q->q.qlen; in qdisc_qlen()
531 qlen += q->q.qlen; in qdisc_qlen_sum()
544 return &qdisc->q.lock; in qdisc_lock()
551 return q; in qdisc_root()
1040 sch->q.qlen--; in qdisc_dequeue_internal()
1148 sch->q.qlen++; in qdisc_peek_dequeued()
1165 sch->q.qlen--; in qdisc_update_stats_at_dequeue()
1177 sch->q.qlen++; in qdisc_update_stats_at_enqueue()
[all …]
A Dpkt_cls.h89 return block->q; in tcf_block_q()
173 cl = q->ops->cl_ops->bind_tcf(q, base, r->classid); in __tcf_bind_filter()
176 q->ops->cl_ops->unbind_tcf(q, cl); in __tcf_bind_filter()
182 struct Qdisc *q = tp->chain->block->q; in tcf_bind_filter() local
187 if (!q) in tcf_bind_filter()
189 sch_tree_lock(q); in tcf_bind_filter()
191 sch_tree_unlock(q); in tcf_bind_filter()
200 q->ops->cl_ops->unbind_tcf(q, cl); in __tcf_unbind_filter()
206 struct Qdisc *q = tp->chain->block->q; in tcf_unbind_filter() local
208 if (!q) in tcf_unbind_filter()
[all …]
/include/crypto/
A Db128ops.h60 static inline void be128_xor(be128 *r, const be128 *p, const be128 *q) in be128_xor() argument
62 r->a = p->a ^ q->a; in be128_xor()
63 r->b = p->b ^ q->b; in be128_xor()
66 static inline void le128_xor(le128 *r, const le128 *p, const le128 *q) in le128_xor() argument
68 r->a = p->a ^ q->a; in le128_xor()
69 r->b = p->b ^ q->b; in le128_xor()
A Daria.h436 int q = 4 - (n / 32); in aria_gsrk() local
440 ((y[q % 4]) >> r) ^ in aria_gsrk()
441 ((y[(q + 3) % 4]) << (32 - r)); in aria_gsrk()
443 ((y[(q + 1) % 4]) >> r) ^ in aria_gsrk()
444 ((y[q % 4]) << (32 - r)); in aria_gsrk()
446 ((y[(q + 2) % 4]) >> r) ^ in aria_gsrk()
447 ((y[(q + 1) % 4]) << (32 - r)); in aria_gsrk()
449 ((y[(q + 3) % 4]) >> r) ^ in aria_gsrk()
450 ((y[(q + 2) % 4]) << (32 - r)); in aria_gsrk()
/include/net/sctp/
A Dstream_sched.h33 void (*enqueue)(struct sctp_outq *q, struct sctp_datamsg *msg);
35 struct sctp_chunk *(*dequeue)(struct sctp_outq *q);
37 void (*dequeue_done)(struct sctp_outq *q, struct sctp_chunk *chunk);
51 void sctp_sched_dequeue_done(struct sctp_outq *q, struct sctp_chunk *ch);
53 void sctp_sched_dequeue_common(struct sctp_outq *q, struct sctp_chunk *ch);
/include/linux/dsa/
A Dbrcm.h12 #define BRCM_TAG_SET_PORT_QUEUE(p, q) ((p) << 8 | q) argument
/include/linux/sunrpc/
A Dsched.h214 #define RPC_IS_PRIORITY(q) ((q)->maxpriority > 0) argument
287 static inline const char * rpc_qname(const struct rpc_wait_queue *q) in rpc_qname() argument
289 return ((q && q->name) ? q->name : "unknown"); in rpc_qname()
292 static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q, in rpc_assign_waitqueue_name() argument
295 q->name = name; in rpc_assign_waitqueue_name()
298 static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q, in rpc_assign_waitqueue_name() argument

Completed in 57 milliseconds

123