| /linux/drivers/staging/media/atomisp/pci/hive_isp_css_common/host/ |
| A D | debug_private.h | 56 if (remote_tail > debug_data_ptr->tail) { in debug_synch_queue() 60 debug_data_ptr->tail * sizeof(uint32_t), in debug_synch_queue() 62 } else if (remote_tail < debug_data_ptr->tail) { in debug_synch_queue() 66 debug_data_ptr->tail * sizeof(uint32_t), in debug_synch_queue() 72 debug_data_ptr->tail = remote_tail; in debug_synch_queue() 80 if (remote_tail > debug_data_ptr->tail) { in debug_synch_queue_isp() 96 debug_data_ptr->tail = remote_tail; in debug_synch_queue_isp() 106 if (remote_tail > debug_data_ptr->tail) { in debug_synch_queue_ddr() 110 debug_data_ptr->tail * sizeof(uint32_t), in debug_synch_queue_ddr() 116 debug_data_ptr->tail * sizeof(uint32_t), in debug_synch_queue_ddr() [all …]
|
| /linux/tools/lib/ |
| A D | list_sort.c | 23 *tail = a; in merge() 24 tail = &a->next; in merge() 27 *tail = b; in merge() 31 *tail = b; in merge() 35 *tail = a; in merge() 61 tail = a; in merge_final() 68 tail = b; in merge_final() 78 tail->next = b; in merge_final() 81 tail = b; in merge_final() 209 tail = &(*tail)->prev; in list_sort() [all …]
|
| /linux/lib/ |
| A D | list_sort.c | 24 *tail = a; in merge() 25 tail = &a->next; in merge() 28 *tail = b; in merge() 32 *tail = b; in merge() 36 *tail = a; in merge() 63 tail = a; in merge_final() 70 tail = b; in merge_final() 80 tail->next = b; in merge_final() 92 tail = b; in merge_final() 220 tail = &(*tail)->prev; in list_sort() [all …]
|
| /linux/Documentation/translations/zh_CN/core-api/ |
| A D | circular-buffers.rst | 55 (2) 'tail'索引 - 消费者在缓冲区中找到下一个元素的位置。 57 通常,当tail指针等于head指针时,表明缓冲区是空的;而当head指针比tail指针少一个时, 60 添加元素时,递增head索引;删除元素时,递增tail索引。tail索引不应该跳过head索引, 111 产者控制着head索引,但消费者可能仍然在另一个CPU上耗尽缓冲区并移动tail索引。 116 者控制着tail索引,但生产者可能仍然在另一个CPU上填充缓冲区并移动head索引。 146 unsigned long tail = READ_ONCE(buffer->tail); 181 unsigned long tail = buffer->tail; 186 struct item *item = buffer[tail]; 190 /* 在递增tail之前完成对描述符的读取。 */ 191 smp_store_release(buffer->tail, [all …]
|
| /linux/drivers/rpmsg/ |
| A D | qcom_glink_smem.c | 53 __le32 *tail; member 70 u32 tail; in glink_smem_rx_avail() local 86 tail = le32_to_cpu(*pipe->tail); in glink_smem_rx_avail() 99 u32 tail; in glink_smem_rx_peek() local 101 tail = le32_to_cpu(*pipe->tail); in glink_smem_rx_peek() 118 u32 tail; in glink_smem_rx_advance() local 120 tail = le32_to_cpu(*pipe->tail); in glink_smem_rx_advance() 122 tail += count; in glink_smem_rx_advance() 126 *pipe->tail = cpu_to_le32(tail); in glink_smem_rx_advance() 133 u32 tail; in glink_smem_tx_avail() local [all …]
|
| A D | qcom_glink_rpm.c | 51 void __iomem *tail; member 73 unsigned int tail; in glink_rpm_rx_avail() local 76 tail = readl(pipe->tail); in glink_rpm_rx_avail() 78 if (head < tail) in glink_rpm_rx_avail() 91 tail = readl(pipe->tail); in glink_rpm_rx_peek() 92 tail += offset; in glink_rpm_rx_peek() 114 tail = readl(pipe->tail); in glink_rpm_rx_advance() 116 tail += count; in glink_rpm_rx_advance() 120 writel(tail, pipe->tail); in glink_rpm_rx_advance() 130 tail = readl(pipe->tail); in glink_rpm_tx_avail() [all …]
|
| /linux/fs/netfs/ |
| A D | misc.c | 23 prev = tail; in netfs_buffer_make_space() 25 if (!folioq_full(tail)) in netfs_buffer_make_space() 26 return tail; in netfs_buffer_make_space() 30 tail = kmalloc(sizeof(*tail), GFP_NOFS); in netfs_buffer_make_space() 31 if (!tail) in netfs_buffer_make_space() 34 folioq_init(tail); in netfs_buffer_make_space() 35 tail->prev = prev; in netfs_buffer_make_space() 44 rreq->buffer = tail; in netfs_buffer_make_space() 57 return tail; in netfs_buffer_make_space() 70 if (IS_ERR(tail)) in netfs_buffer_append_folio() [all …]
|
| /linux/drivers/md/dm-vdo/ |
| A D | murmurhash3.c | 85 k2 ^= ((u64)tail[14]) << 48; in murmurhash3_128() 100 k2 ^= ((u64)tail[9]) << 8; in murmurhash3_128() 103 k2 ^= ((u64)tail[8]) << 0; in murmurhash3_128() 111 k1 ^= ((u64)tail[7]) << 56; in murmurhash3_128() 114 k1 ^= ((u64)tail[6]) << 48; in murmurhash3_128() 117 k1 ^= ((u64)tail[5]) << 40; in murmurhash3_128() 120 k1 ^= ((u64)tail[4]) << 32; in murmurhash3_128() 123 k1 ^= ((u64)tail[3]) << 24; in murmurhash3_128() 126 k1 ^= ((u64)tail[2]) << 16; in murmurhash3_128() 129 k1 ^= ((u64)tail[1]) << 8; in murmurhash3_128() [all …]
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| A D | cq.c | 35 u32 tail; in rvt_cq_enter() local 43 tail = RDMA_READ_UAPI_ATOMIC(u_wc->tail); in rvt_cq_enter() 48 tail = k_wc->tail; in rvt_cq_enter() 388 tail = RDMA_READ_UAPI_ATOMIC(old_u_wc->tail); in rvt_resize_cq() 392 tail = old_k_wc->tail; in rvt_resize_cq() 413 tail = 0; in rvt_resize_cq() 415 tail++; in rvt_resize_cq() 484 u32 tail; in rvt_poll_cq() local 493 tail = wc->tail; in rvt_poll_cq() 505 tail++; in rvt_poll_cq() [all …]
|
| /linux/drivers/net/ethernet/intel/fm10k/ |
| A D | fm10k_mbx.c | 17 fifo->tail = 0; in fm10k_fifo_init() 137 if (len > tail) in fm10k_mbx_index_len() 156 return (tail > mbx->tail) ? --tail : ++tail; in fm10k_mbx_tail_add() 172 return (tail < mbx->tail) ? ++tail : --tail; in fm10k_mbx_tail_sub() 346 tail++; in fm10k_mbx_write_copy() 360 if (!tail) in fm10k_mbx_write_copy() 361 tail++; in fm10k_mbx_write_copy() 439 tail += end; in fm10k_mbx_read_copy() 979 if (!tail || (tail == FM10K_MSG_HDR_MASK(TAIL))) in fm10k_mbx_validate_msg_hdr() 995 if (tail) in fm10k_mbx_validate_msg_hdr() [all …]
|
| /linux/include/linux/ |
| A D | circ_buf.h | 12 int tail; member 16 #define CIRC_CNT(head,tail,size) (((head) - (tail)) & ((size)-1)) argument 21 #define CIRC_SPACE(head,tail,size) CIRC_CNT((tail),((head)+1),(size)) argument 26 #define CIRC_CNT_TO_END(head,tail,size) \ argument 27 ({int end = (size) - (tail); \ 32 #define CIRC_SPACE_TO_END(head,tail,size) \ argument 34 int n = (end + (tail)) & ((size)-1); \
|
| A D | objpool.h | 55 uint32_t tail; member 195 uint32_t head, tail; in __objpool_try_add_slot() local 198 tail = READ_ONCE(slot->tail); in __objpool_try_add_slot() 203 WARN_ON_ONCE(tail - head > pool->nr_objs); in __objpool_try_add_slot() 204 } while (!try_cmpxchg_acquire(&slot->tail, &tail, tail + 1)); in __objpool_try_add_slot() 207 WRITE_ONCE(slot->entries[tail & slot->mask], obj); in __objpool_try_add_slot() 209 smp_store_release(&slot->last, tail + 1); in __objpool_try_add_slot()
|
| /linux/drivers/net/wireguard/ |
| A D | queueing.c | 56 queue->head = queue->tail = STUB(queue); in wg_prev_queue_init() 82 struct sk_buff *tail = queue->tail, *next = smp_load_acquire(&NEXT(tail)); in wg_prev_queue_dequeue() local 84 if (tail == STUB(queue)) { in wg_prev_queue_dequeue() 87 queue->tail = next; in wg_prev_queue_dequeue() 88 tail = next; in wg_prev_queue_dequeue() 92 queue->tail = next; in wg_prev_queue_dequeue() 94 return tail; in wg_prev_queue_dequeue() 96 if (tail != READ_ONCE(queue->head)) in wg_prev_queue_dequeue() 99 next = smp_load_acquire(&NEXT(tail)); in wg_prev_queue_dequeue() 101 queue->tail = next; in wg_prev_queue_dequeue() [all …]
|
| /linux/drivers/gpu/drm/i915/gt/ |
| A D | intel_ring.h | 26 void intel_ring_reset(struct intel_ring *ring, u32 tail); 91 assert_ring_tail_valid(const struct intel_ring *ring, unsigned int tail) in assert_ring_tail_valid() argument 95 GEM_BUG_ON(!intel_ring_offset_valid(ring, tail)); in assert_ring_tail_valid() 112 GEM_BUG_ON(cacheline(tail) == cacheline(head) && tail < head); in assert_ring_tail_valid() 117 intel_ring_set_tail(struct intel_ring *ring, unsigned int tail) in intel_ring_set_tail() argument 125 assert_ring_tail_valid(ring, tail); in intel_ring_set_tail() 126 ring->tail = tail; in intel_ring_set_tail() 127 return tail; in intel_ring_set_tail() 131 __intel_ring_space(unsigned int head, unsigned int tail, unsigned int size) in __intel_ring_space() argument 139 return (head - tail - CACHELINE_BYTES) & (size - 1); in __intel_ring_space()
|
| /linux/arch/arm/kernel/ |
| A D | perf_callchain.c | 34 user_backtrace(struct frame_tail __user *tail, in user_backtrace() argument 40 if (!access_ok(tail, sizeof(buftail))) in user_backtrace() 44 err = __copy_from_user_inatomic(&buftail, tail, sizeof(buftail)); in user_backtrace() 56 if (tail + 1 >= buftail.fp) in user_backtrace() 65 struct frame_tail __user *tail; in perf_callchain_user() local 72 tail = (struct frame_tail __user *)regs->ARM_fp - 1; in perf_callchain_user() 75 tail && !((unsigned long)tail & 0x3)) in perf_callchain_user() 76 tail = user_backtrace(tail, entry); in perf_callchain_user()
|
| /linux/arch/arm64/kernel/ |
| A D | stacktrace.c | 351 if (!access_ok(tail, sizeof(buftail))) in unwind_user_frame() 370 if (tail >= buftail.fp) in unwind_user_frame() 399 if (!access_ok(tail, sizeof(buftail))) in unwind_compat_user_frame() 416 if (tail + 1 >= (struct compat_frame_tail __user *) in unwind_compat_user_frame() 433 struct frame_tail __user *tail; in arch_stack_walk_user() local 435 tail = (struct frame_tail __user *)regs->regs[29]; in arch_stack_walk_user() 436 while (tail && !((unsigned long)tail & 0x7)) in arch_stack_walk_user() 437 tail = unwind_user_frame(tail, cookie, consume_entry); in arch_stack_walk_user() 441 struct compat_frame_tail __user *tail; in arch_stack_walk_user() local 444 while (tail && !((unsigned long)tail & 0x3)) in arch_stack_walk_user() [all …]
|
| /linux/kernel/locking/ |
| A D | qspinlock.c | 118 u32 tail; in encode_tail() local 120 tail = (cpu + 1) << _Q_TAIL_CPU_OFFSET; in encode_tail() 123 return tail; in encode_tail() 128 int cpu = (tail >> _Q_TAIL_CPU_OFFSET) - 1; in decode_tail() 183 return (u32)xchg_relaxed(&lock->tail, in xchg_tail() 184 tail >> _Q_TAIL_OFFSET) << _Q_TAIL_OFFSET; in xchg_tail() 227 new = (old & _Q_LOCKED_PENDING_MASK) | tail; in xchg_tail() 316 u32 old, tail; in queued_spin_lock_slowpath() local 400 tail = encode_tail(smp_processor_id(), idx); in queued_spin_lock_slowpath() 460 old = xchg_tail(lock, tail); in queued_spin_lock_slowpath() [all …]
|
| /linux/include/drm/ |
| A D | spsc_queue.h | 43 atomic_long_t tail; member 51 atomic_long_set(&queue->tail, (long)&queue->head); in spsc_queue_init() 67 struct spsc_node **tail; in spsc_queue_push() local 73 tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next); in spsc_queue_push() 74 WRITE_ONCE(*tail, node); in spsc_queue_push() 85 return tail == &queue->head; in spsc_queue_push() 107 if (atomic_long_cmpxchg(&queue->tail, in spsc_queue_pop()
|
| /linux/drivers/media/test-drivers/vidtv/ |
| A D | vidtv_channel.c | 182 tail = vidtv_psi_eit_event_init(tail, event_id); in vidtv_channel_eit_event_cat_into_new() 183 if (!tail) { in vidtv_channel_eit_event_cat_into_new() 192 head = tail; in vidtv_channel_eit_event_cat_into_new() 229 tail = vidtv_psi_sdt_service_init(tail, in vidtv_channel_sdt_serv_cat_into_new() 233 if (!tail) in vidtv_channel_sdt_serv_cat_into_new() 242 head = tail; in vidtv_channel_sdt_serv_cat_into_new() 284 tail = vidtv_psi_pat_program_init(tail, in vidtv_channel_pat_prog_cat_into_new() 287 if (!tail) { in vidtv_channel_pat_prog_cat_into_new() 293 head = tail; in vidtv_channel_pat_prog_cat_into_new() 341 tail = vidtv_psi_pmt_stream_init(tail, in vidtv_channel_pmt_match_sections() [all …]
|
| /linux/Documentation/trace/ |
| A D | ring-buffer-design.rst | 21 tail 227 tail page 323 tail page 337 and possibly the tail page. That is when the commit (and tail) page has been 378 tail page 390 tail page 624 tail page 635 tail page 681 tail page 743 on the new tail page. [all …]
|
| /linux/net/sunrpc/ |
| A D | xdr.c | 179 const struct kvec *tail = xdr->tail; in xdr_buf_to_bvec() local 231 struct kvec *tail = xdr->tail; in xdr_inline_pages() local 561 struct kvec *tail = buf->tail; in xdr_buf_try_expand() local 588 const struct kvec *tail = buf->tail; in xdr_buf_tail_copy_right() local 602 const struct kvec *tail = buf->tail; in xdr_buf_pages_copy_right() local 636 const struct kvec *tail = buf->tail; in xdr_buf_head_copy_right() local 682 const struct kvec *tail = buf->tail; in xdr_buf_tail_shift_right() local 727 const struct kvec *tail = buf->tail; in xdr_buf_tail_copy_left() local 1189 struct kvec *tail = buf->tail; in xdr_truncate_encode() local 1203 xdr->p = tail->iov_base + tail->iov_len; in xdr_truncate_encode() [all …]
|
| /linux/fs/ |
| A D | splice.c | 201 unsigned int tail = pipe->tail; in splice_to_pipe() local 249 unsigned int tail = pipe->tail; in add_to_pipe() local 482 pipe->tail = tail; in splice_from_pipe_feed() 503 pipe->tail = tail+1; in eat_empty_buffer() 713 tail = pipe->tail; in iter_file_splice_write() 753 tail = pipe->tail; in iter_file_splice_write() 761 pipe->tail = tail; in iter_file_splice_write() 848 tail = pipe->tail; in splice_to_socket() 895 tail = pipe->tail; in splice_to_socket() 910 if (tail != pipe->tail) { in splice_to_socket() [all …]
|
| A D | pipe.c | 214 unsigned int tail = READ_ONCE(pipe->tail); in pipe_readable() local 237 pipe->tail = ++tail; in pipe_update_tail() 246 pipe->tail = ++tail; in pipe_update_tail() 278 unsigned int tail = pipe->tail; in pipe_read() local 344 tail = pipe_update_tail(pipe, buf, tail); in pipe_read() 420 unsigned int tail = READ_ONCE(pipe->tail); in pipe_writable() local 625 tail = pipe->tail; in pipe_ioctl() 630 tail++; in pipe_ioctl() 683 tail = READ_ONCE(pipe->tail); in pipe_poll() 1279 tail = pipe->tail; in pipe_resize_ring() [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| A D | xe_trace_guc.h | 62 TP_PROTO(struct xe_device *xe, u8 gt_id, u32 action, u32 len, u32 _head, u32 tail), 63 TP_ARGS(xe, gt_id, action, len, _head, tail), 70 __field(u32, tail) 79 __entry->tail = tail; 85 __entry->tail, __entry->_head) 89 TP_PROTO(struct xe_device *xe, u8 gt_id, u32 action, u32 len, u32 _head, u32 tail), 90 TP_ARGS(xe, gt_id, action, len, _head, tail) 94 TP_PROTO(struct xe_device *xe, u8 gt_id, u32 action, u32 len, u32 _head, u32 tail), 95 TP_ARGS(xe, gt_id, action, len, _head, tail), 99 __entry->tail, __entry->_head)
|
| /linux/net/sched/ |
| A D | sch_choke.c | 67 unsigned int tail; member 97 if (q->head == q->tail) in choke_zap_head_holes() 106 q->tail = (q->tail - 1) & q->tab_mask; in choke_zap_tail_holes() 107 if (q->head == q->tail) in choke_zap_tail_holes() 123 if (idx == q->tail) in choke_drop_by_idx() 205 if (q->head == q->tail) in choke_match_random() 268 q->tab[q->tail] = skb; in choke_enqueue() 269 q->tail = (q->tail + 1) & q->tab_mask; in choke_enqueue() 319 q->head = q->tail = 0; in choke_reset() 389 if (tail < mask) { in choke_change() [all …]
|