| /include/linux/ |
| A D | lwq.h | 33 static inline void lwq_init(struct lwq *q) in lwq_init() argument 35 spin_lock_init(&q->lock); in lwq_init() 36 q->ready = NULL; in lwq_init() 37 init_llist_head(&q->new); in lwq_init() 48 static inline bool lwq_empty(struct lwq *q) in lwq_empty() argument 51 return smp_load_acquire(&q->ready) == NULL && llist_empty(&q->new); in lwq_empty() 54 struct llist_node *__lwq_dequeue(struct lwq *q); 65 #define lwq_dequeue(q, type, member) \ argument 103 return llist_add(&n->node, &q->new) && in lwq_enqueue() 104 smp_load_acquire(&q->ready) == NULL; in lwq_enqueue() [all …]
|
| A D | blktrace_api.h | 49 #define blk_add_cgroup_trace_msg(q, css, fmt, ...) \ argument 54 bt = rcu_dereference((q)->blk_trace); \ 59 #define blk_add_trace_msg(q, fmt, ...) \ argument 60 blk_add_cgroup_trace_msg(q, NULL, fmt, ##__VA_ARGS__) 69 bt = rcu_dereference(q->blk_trace); in blk_trace_note_message_enabled() 80 extern int blk_trace_remove(struct request_queue *q); 84 # define blk_trace_shutdown(q) do { } while (0) argument 86 # define blk_trace_setup(q, name, dev, bdev, arg) (-ENOTTY) argument 87 # define blk_trace_startstop(q, start) (-ENOTTY) argument 88 # define blk_add_trace_msg(q, fmt, ...) do { } while (0) argument [all …]
|
| A D | blk-pm.h | 13 extern void blk_pm_runtime_init(struct request_queue *q, struct device *dev); 14 extern int blk_pre_runtime_suspend(struct request_queue *q); 15 extern void blk_post_runtime_suspend(struct request_queue *q, int err); 16 extern void blk_pre_runtime_resume(struct request_queue *q); 17 extern void blk_post_runtime_resume(struct request_queue *q); 19 static inline void blk_pm_runtime_init(struct request_queue *q, in blk_pm_runtime_init() argument
|
| A D | blkdev.h | 667 #define blk_queue_dying(q) test_bit(QUEUE_FLAG_DYING, &(q)->queue_flags) argument 672 #define blk_queue_nonrot(q) (!((q)->limits.features & BLK_FEAT_ROTATIONAL)) argument 673 #define blk_queue_io_stat(q) ((q)->limits.features & BLK_FEAT_IO_STAT) argument 676 #define blk_queue_dax(q) ((q)->limits.features & BLK_FEAT_DAX) argument 677 #define blk_queue_pci_p2pdma(q) ((q)->limits.features & BLK_FEAT_PCI_P2PDMA) argument 689 #define blk_queue_pm_only(q) atomic_read(&(q)->pm_only) argument 710 return q->mq_ops; in queue_is_mq() 716 return q->rpm_status; in queue_rpm_status() 750 if (q->queue_depth) in blk_queue_depth() 1067 return q->limits; in queue_limits_start_update() [all …]
|
| A D | fortify-string.h | 286 return __real_strscpy(p, q, size); in sized_strscpy() 297 len = __compiletime_strlen(q); in sized_strscpy() 309 len = strnlen(q, size); in sized_strscpy() 329 return __real_strscpy(p, q, len); in sized_strscpy() 371 copy_len = strlen(q); in strlcat() 453 copy_len = strnlen(q, count); in strncat() 690 #define memcpy(p, q, s) __fortify_memcpy_chk(p, q, s, \ argument 694 #define memmove(p, q, s) __fortify_memcpy_chk(p, q, s, \ argument 795 return __underlying_strcpy(p, q); in strcpy() 796 size = strlen(q) + 1; in strcpy() [all …]
|
| A D | t10-pi.h | 42 unsigned int shift = ilog2(queue_logical_block_size(rq->q)); in t10_pi_ref_tag() 45 rq->q->limits.integrity.interval_exp) in t10_pi_ref_tag() 46 shift = rq->q->limits.integrity.interval_exp; in t10_pi_ref_tag() 67 unsigned int shift = ilog2(queue_logical_block_size(rq->q)); in ext_pi_ref_tag() 70 rq->q->limits.integrity.interval_exp) in ext_pi_ref_tag() 71 shift = rq->q->limits.integrity.interval_exp; in ext_pi_ref_tag()
|
| A D | blk-integrity.h | 36 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity() argument 38 return q->limits.integrity.metadata_size; in blk_integrity_queue_supports_integrity() 55 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments() argument 57 return q->limits.max_integrity_segments; in queue_max_integrity_segments() 102 static inline int blk_rq_count_integrity_sg(struct request_queue *q, in blk_rq_count_integrity_sg() argument 107 static inline int blk_rq_map_integrity_sg(struct request *q, in blk_rq_map_integrity_sg() argument 127 blk_integrity_queue_supports_integrity(struct request_queue *q) in blk_integrity_queue_supports_integrity() argument 132 queue_max_integrity_segments(const struct request_queue *q) in queue_max_integrity_segments() argument
|
| A D | swait.h | 69 extern void __init_swait_queue_head(struct swait_queue_head *q, const char *name, 72 #define init_swait_queue_head(q) \ argument 75 __init_swait_queue_head((q), #q, &__key); \ 147 extern void swake_up_one(struct swait_queue_head *q); 148 extern void swake_up_all(struct swait_queue_head *q); 149 extern void swake_up_locked(struct swait_queue_head *q, int wake_flags); 151 extern void prepare_to_swait_exclusive(struct swait_queue_head *q, struct swait_queue *wait, int st… 152 extern long prepare_to_swait_event(struct swait_queue_head *q, struct swait_queue *wait, int state); 154 extern void __finish_swait(struct swait_queue_head *q, struct swait_queue *wait); 155 extern void finish_swait(struct swait_queue_head *q, struct swait_queue *wait);
|
| A D | uacce.h | 41 struct uacce_queue *q); 42 void (*put_queue)(struct uacce_queue *q); 43 int (*start_queue)(struct uacce_queue *q); 44 void (*stop_queue)(struct uacce_queue *q); 45 int (*is_q_updated)(struct uacce_queue *q); 46 int (*mmap)(struct uacce_queue *q, struct vm_area_struct *vma, 48 long (*ioctl)(struct uacce_queue *q, unsigned int cmd,
|
| A D | blk-mq.h | 104 struct request_queue *q; member 720 struct request_queue *q); 915 void blk_mq_quiesce_queue(struct request_queue *q); 930 blk_mq_freeze_queue(struct request_queue *q) in blk_mq_freeze_queue() argument 934 blk_mq_freeze_queue_nomemsave(q); in blk_mq_freeze_queue() 940 blk_mq_unfreeze_queue_nomemrestore(q); in blk_mq_unfreeze_queue() 966 return __blk_should_fake_timeout(q); in blk_should_fake_timeout() 998 #define queue_for_each_hw_ctx(q, hctx, i) \ argument 999 xa_for_each(&(q)->hctx_table, (i), (hctx)) 1007 if (rq->q->mq_ops->cleanup_rq) in blk_mq_cleanup_rq() [all …]
|
| /include/media/ |
| A D | videobuf2-core.h | 700 return q->allow_cache_hints && q->memory == VB2_MEMORY_MMAP; in vb2_queue_allows_cache_hints() 1163 return q->streaming; in vb2_is_streaming() 1181 return q->fileio; in vb2_fileio_is_active() 1190 if (q->bufs_bitmap) in vb2_get_num_buffers() 1191 return bitmap_weight(q->bufs_bitmap, q->max_num_buffers); in vb2_get_num_buffers() 1204 return !!q->is_busy; in vb2_is_busy() 1213 return q->drv_priv; in vb2_get_drv_priv() 1280 q->last_buffer_dequeued = false; in vb2_clear_last_buffer_dequeued() 1296 if (!q->bufs) in vb2_get_buffer() 1299 if (index >= q->max_num_buffers) in vb2_get_buffer() [all …]
|
| A D | videobuf2-v4l2.h | 75 int vb2_querybuf(struct vb2_queue *q, struct v4l2_buffer *b); 85 int vb2_reqbufs(struct vb2_queue *q, struct v4l2_requestbuffers *req); 119 int vb2_prepare_buf(struct vb2_queue *q, struct media_device *mdev, 145 int vb2_qbuf(struct vb2_queue *q, struct media_device *mdev, 157 int vb2_expbuf(struct vb2_queue *q, struct v4l2_exportbuffer *eb); 201 int vb2_streamon(struct vb2_queue *q, enum v4l2_buf_type type); 220 int vb2_streamoff(struct vb2_queue *q, enum v4l2_buf_type type); 233 int __must_check vb2_queue_init(struct vb2_queue *q); 256 void vb2_queue_release(struct vb2_queue *q); 272 int vb2_queue_change_type(struct vb2_queue *q, unsigned int type); [all …]
|
| /include/trace/events/ |
| A D | vb2.h | 12 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb), 13 TP_ARGS(q, vb), 26 __entry->owner = q->owner; 27 __entry->queued_count = q->queued_count; 29 atomic_read(&q->owned_by_drv_count); 47 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb), 48 TP_ARGS(q, vb) 52 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb), 53 TP_ARGS(q, vb) 58 TP_ARGS(q, vb) [all …]
|
| A D | qdisc.h | 79 TP_PROTO(struct Qdisc *q), 81 TP_ARGS(q), 84 __string( dev, qdisc_dev(q) ? qdisc_dev(q)->name : "(null)" ) 85 __string( kind, q->ops->id ) 93 __entry->parent = q->parent; 94 __entry->handle = q->handle; 104 TP_PROTO(struct Qdisc *q), 106 TP_ARGS(q), 110 __string( kind, q->ops->id ) 118 __entry->parent = q->parent; [all …]
|
| A D | block.h | 99 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0; 135 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0; 208 __entry->dev = rq->q->disk ? disk_devt(rq->q->disk) : 0; 311 TP_ARGS(q, bio), 428 TP_PROTO(struct request_queue *q), 430 TP_ARGS(q), 447 TP_ARGS(q, depth, explicit), 475 TP_ARGS(q, depth, explicit) 642 TP_ARGS(q, zno, sector, nr_sectors), 669 TP_ARGS(q, zno, sector, nr_sectors) [all …]
|
| /include/net/ |
| A D | ipv6_frag.h | 24 struct inet_frag_queue q; member 34 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() 37 q->key.v6 = *key; in ip6frag_init() 73 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue() 75 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue() 80 fq->q.flags |= INET_FRAG_DROP; in ip6frag_expire_frag_queue() 81 inet_frag_kill(&fq->q, &refs); in ip6frag_expire_frag_queue() 98 head = inet_frag_pull_head(&fq->q); in ip6frag_expire_frag_queue() 103 spin_unlock(&fq->q.lock); in ip6frag_expire_frag_queue() 110 spin_unlock(&fq->q.lock); in ip6frag_expire_frag_queue() [all …]
|
| A D | inet_frag.h | 110 void (*constructor)(struct inet_frag_queue *q, 140 void inet_frag_kill(struct inet_frag_queue *q, int *refs); 141 void inet_frag_destroy(struct inet_frag_queue *q); 148 static inline void inet_frag_putn(struct inet_frag_queue *q, int refs) in inet_frag_putn() argument 150 if (refs && refcount_sub_and_test(refs, &q->refcnt)) in inet_frag_putn() 151 inet_frag_destroy(q); in inet_frag_putn() 186 int inet_frag_queue_insert(struct inet_frag_queue *q, struct sk_buff *skb, 188 void *inet_frag_reasm_prepare(struct inet_frag_queue *q, struct sk_buff *skb, 190 void inet_frag_reasm_finish(struct inet_frag_queue *q, struct sk_buff *head, 192 struct sk_buff *inet_frag_pull_head(struct inet_frag_queue *q);
|
| A D | pkt_sched.h | 23 #define qdisc_priv(q) \ argument 24 _Generic(q, \ 25 const struct Qdisc * : (const void *)&q->privdata, \ 26 struct Qdisc * : (void *)&q->privdata) 109 void qdisc_hash_del(struct Qdisc *q); 121 void __qdisc_run(struct Qdisc *q); 123 static inline void qdisc_run(struct Qdisc *q) in qdisc_run() argument 125 if (qdisc_run_begin(q)) { in qdisc_run() 126 __qdisc_run(q); in qdisc_run() 127 qdisc_run_end(q); in qdisc_run() [all …]
|
| A D | sch_generic.h | 117 struct qdisc_skb_head q; member 473 struct Qdisc *q; member 519 return q->q.qlen; in qdisc_qlen() 531 qlen += q->q.qlen; in qdisc_qlen_sum() 544 return &qdisc->q.lock; in qdisc_lock() 551 return q; in qdisc_root() 1040 sch->q.qlen--; in qdisc_dequeue_internal() 1148 sch->q.qlen++; in qdisc_peek_dequeued() 1165 sch->q.qlen--; in qdisc_update_stats_at_dequeue() 1177 sch->q.qlen++; in qdisc_update_stats_at_enqueue() [all …]
|
| A D | pkt_cls.h | 89 return block->q; in tcf_block_q() 173 cl = q->ops->cl_ops->bind_tcf(q, base, r->classid); in __tcf_bind_filter() 176 q->ops->cl_ops->unbind_tcf(q, cl); in __tcf_bind_filter() 182 struct Qdisc *q = tp->chain->block->q; in tcf_bind_filter() local 187 if (!q) in tcf_bind_filter() 189 sch_tree_lock(q); in tcf_bind_filter() 191 sch_tree_unlock(q); in tcf_bind_filter() 200 q->ops->cl_ops->unbind_tcf(q, cl); in __tcf_unbind_filter() 206 struct Qdisc *q = tp->chain->block->q; in tcf_unbind_filter() local 208 if (!q) in tcf_unbind_filter() [all …]
|
| /include/crypto/ |
| A D | b128ops.h | 60 static inline void be128_xor(be128 *r, const be128 *p, const be128 *q) in be128_xor() argument 62 r->a = p->a ^ q->a; in be128_xor() 63 r->b = p->b ^ q->b; in be128_xor() 66 static inline void le128_xor(le128 *r, const le128 *p, const le128 *q) in le128_xor() argument 68 r->a = p->a ^ q->a; in le128_xor() 69 r->b = p->b ^ q->b; in le128_xor()
|
| A D | aria.h | 436 int q = 4 - (n / 32); in aria_gsrk() local 440 ((y[q % 4]) >> r) ^ in aria_gsrk() 441 ((y[(q + 3) % 4]) << (32 - r)); in aria_gsrk() 443 ((y[(q + 1) % 4]) >> r) ^ in aria_gsrk() 444 ((y[q % 4]) << (32 - r)); in aria_gsrk() 446 ((y[(q + 2) % 4]) >> r) ^ in aria_gsrk() 447 ((y[(q + 1) % 4]) << (32 - r)); in aria_gsrk() 449 ((y[(q + 3) % 4]) >> r) ^ in aria_gsrk() 450 ((y[(q + 2) % 4]) << (32 - r)); in aria_gsrk()
|
| /include/net/sctp/ |
| A D | stream_sched.h | 33 void (*enqueue)(struct sctp_outq *q, struct sctp_datamsg *msg); 35 struct sctp_chunk *(*dequeue)(struct sctp_outq *q); 37 void (*dequeue_done)(struct sctp_outq *q, struct sctp_chunk *chunk); 51 void sctp_sched_dequeue_done(struct sctp_outq *q, struct sctp_chunk *ch); 53 void sctp_sched_dequeue_common(struct sctp_outq *q, struct sctp_chunk *ch);
|
| /include/linux/dsa/ |
| A D | brcm.h | 12 #define BRCM_TAG_SET_PORT_QUEUE(p, q) ((p) << 8 | q) argument
|
| /include/linux/sunrpc/ |
| A D | sched.h | 214 #define RPC_IS_PRIORITY(q) ((q)->maxpriority > 0) argument 287 static inline const char * rpc_qname(const struct rpc_wait_queue *q) in rpc_qname() argument 289 return ((q && q->name) ? q->name : "unknown"); in rpc_qname() 292 static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q, in rpc_assign_waitqueue_name() argument 295 q->name = name; in rpc_assign_waitqueue_name() 298 static inline void rpc_assign_waitqueue_name(struct rpc_wait_queue *q, in rpc_assign_waitqueue_name() argument
|