| /linux/drivers/virtio/ |
| A D | virtio_ring.c | 422 vq->vq.num_free = num; in virtqueue_init() 603 vq->notify(&vq->vq); in virtqueue_add_split() 780 vq->vq.num_free++; in detach_buf_split() 788 vq->vq.num_free++; in detach_buf_split() 1006 vdev = vq->vq.vdev; in virtqueue_vring_init_split() 1193 vring_free(&vq->vq); in virtqueue_resize_split() 2078 vq->vq.vdev = vdev; in vring_create_virtqueue_packed() 2079 vq->vq.name = name; in vring_create_virtqueue_packed() 2115 return &vq->vq; in vring_create_virtqueue_packed() 2595 vq->vq.callback(&vq->vq); in vring_interrupt() [all …]
|
| A D | virtio_pci_modern.c | 70 struct virtqueue *vq; in virtqueue_exec_admin_cmd() local 74 vq = admin_vq->info->vq; in virtqueue_exec_admin_cmd() 75 if (!vq) in virtqueue_exec_admin_cmd() 238 vq = vp_dev->admin_vq.info->vq; in vp_modern_avq_cleanup() 239 if (!vq) in vp_modern_avq_cleanup() 505 vq->reset = true; in vp_modern_disable_vq_and_reset() 518 if (!vq->reset) in vp_modern_enable_vq_after_reset() 603 if (!vq) in setup_vq() 613 if (!vq->priv) { in setup_vq() 618 return vq; in setup_vq() [all …]
|
| A D | virtio_vdpa.c | 40 struct virtqueue *vq; member 155 struct virtqueue *vq; in virtio_vdpa_setup_vq() local 211 if (!vq) { in virtio_vdpa_setup_vq() 216 vq->num_max = max_num; in virtio_vdpa_setup_vq() 251 vq->priv = info; in virtio_vdpa_setup_vq() 252 info->vq = vq; in virtio_vdpa_setup_vq() 258 return vq; in virtio_vdpa_setup_vq() 261 vring_del_virtqueue(vq); in virtio_vdpa_setup_vq() 286 vring_del_virtqueue(vq); in virtio_vdpa_del_vq() 293 struct virtqueue *vq, *n; in virtio_vdpa_del_vqs() local [all …]
|
| A D | virtio_pci_common.c | 55 iowrite16(vq->index, (void __iomem *)vq->priv); in vp_notify() 224 if (IS_ERR(vq)) in vp_setup_vq() 227 info->vq = vq; in vp_setup_vq() 240 return vq; in vp_setup_vq() 244 return vq; in vp_setup_vq() 256 if (!vq->reset) { in vp_del_vq() 354 if (IS_ERR(vq)) in vp_find_one_vq_msix() 355 return vq; in vp_find_one_vq_msix() 360 return vq; in vp_find_one_vq_msix() 373 return vq; in vp_find_one_vq_msix() [all …]
|
| /linux/drivers/vhost/ |
| A D | vhost.c | 49 #define vhost_used_event(vq) ((__virtio16 __user *)&vq->avail->ring[vq->num]) argument 50 #define vhost_avail_event(vq) ((__virtio16 __user *)&vq->used->ring[vq->num]) argument 111 vq->is_le = vhost_has_feature(vq, VIRTIO_F_VERSION_1) || !vq->user_be; in vhost_init_is_le() 198 poll->vq = vq; in vhost_poll_init() 355 return vq->avail && vq->desc && vq->used && vhost_vq_access_ok(vq); in vhost_vq_is_setup() 475 if (!vq->indirect || !vq->log || !vq->heads) in vhost_dev_alloc_iovecs() 1285 return vhost_put_user(vq, cpu_to_vhost16(vq, vq->avail_idx), in vhost_put_avail_event() 1300 return vhost_put_user(vq, cpu_to_vhost16(vq, vq->used_flags), in vhost_put_used_flags() 1790 return vq_access_ok(vq, vq->num, vq->desc, vq->avail, vq->used); in vhost_vq_access_ok() 1912 if (!vq_access_ok(vq, vq->num, in vhost_vring_set_addr() [all …]
|
| A D | net.c | 244 ubufs->vq = vq; in vhost_net_ubuf_alloc() 388 struct vhost_virtqueue *vq = ubufs->vq; in vhost_zerocopy_complete() local 454 struct vhost_virtqueue *vq = &nvq->vq; in vhost_net_signal_used() local 605 struct vhost_virtqueue *vq = &nvq->vq; in vhost_exceeds_maxpend() local 629 struct vhost_virtqueue *vq = &nvq->vq; in get_tx_bufs() local 665 struct vhost_virtqueue *vq = &nvq->vq; in vhost_net_build_xdp() local 744 struct vhost_virtqueue *vq = &nvq->vq; in handle_tx_copy() local 834 struct vhost_virtqueue *vq = &nvq->vq; in handle_tx_zerocopy() local 943 struct vhost_virtqueue *vq = &nvq->vq; in handle_tx() local 1090 struct vhost_virtqueue *vq = &nvq->vq; in handle_rx() local [all …]
|
| A D | test.c | 51 mutex_lock(&vq->mutex); in handle_vq() 54 mutex_unlock(&vq->mutex); in handle_vq() 61 head = vhost_get_vq_desc(vq, vq->iov, in handle_vq() 62 ARRAY_SIZE(vq->iov), in handle_vq() 69 if (head == vq->num) { in handle_vq() 135 mutex_lock(&vq->mutex); in vhost_test_stop_vq() 189 vq = n->vqs + index; in vhost_test_run() 252 mutex_lock(&vq->mutex); in vhost_test_set_features() 276 vq = &n->vqs[index]; in vhost_test_set_backend() 277 mutex_lock(&vq->mutex); in vhost_test_set_backend() [all …]
|
| A D | vhost.h | 50 struct vhost_virtqueue *vq; member 55 struct vhost_virtqueue *vq); 158 struct vhost_virtqueue *vq; member 211 bool vhost_vq_has_work(struct vhost_virtqueue *vq); 212 bool vhost_vq_is_setup(struct vhost_virtqueue *vq); 229 int vq_meta_prefetch(struct vhost_virtqueue *vq); 253 eventfd_signal((vq)->error_ctx);\ 276 vq->private_data = private_data; in vhost_vq_set_backend() 289 return vq->private_data; in vhost_vq_get_backend() 294 return vq->acked_features & (1ULL << bit); in vhost_has_feature() [all …]
|
| A D | vsock.c | 98 mutex_lock(&vq->mutex); in vhost_transport_do_send_pkt() 103 if (!vq_meta_prefetch(vq)) in vhost_transport_do_send_pkt() 127 head = vhost_get_vq_desc(vq, vq->iov, ARRAY_SIZE(vq->iov), in vhost_transport_do_send_pkt() 134 if (head == vq->num) { in vhost_transport_do_send_pkt() 398 return val < vq->num; in vhost_vsock_more_replies() 489 mutex_lock(&vq->mutex); in vhost_vsock_handle_tx_kick() 509 head = vhost_get_vq_desc(vq, vq->iov, ARRAY_SIZE(vq->iov), in vhost_vsock_handle_tx_kick() 514 if (head == vq->num) { in vhost_vsock_handle_tx_kick() 578 vq = &vsock->vqs[i]; in vhost_vsock_start() 610 vq = &vsock->vqs[i]; in vhost_vsock_start() [all …]
|
| A D | scsi.c | 261 vq = &vs->vqs[i].vq; in vhost_scsi_init_inflight() 477 head = vhost_get_vq_desc(vq, vq->iov, in vhost_scsi_do_evt_work() 957 vc->head = vhost_get_vq_desc(vq, vq->iov, in vhost_scsi_get_desc() 1309 struct vhost_virtqueue *vq = &tmf->svq->vq; in vhost_scsi_tmf_flush_work() local 1759 vq = &vs->vqs[i].vq; in vhost_scsi_set_endpoint() 1763 ret = vhost_scsi_setup_vq_cmds(vq, vq->num); in vhost_scsi_set_endpoint() 1769 vq = &vs->vqs[i].vq; in vhost_scsi_set_endpoint() 1864 vq = &vs->vqs[i].vq; in vhost_scsi_clear_endpoint() 1873 vq = &vs->vqs[i].vq; in vhost_scsi_clear_endpoint() 1932 vq = &vs->vqs[i].vq; in vhost_scsi_set_features() [all …]
|
| /linux/arch/arm64/include/uapi/asm/ |
| A D | ptrace.h | 200 #define SVE_PT_SVE_ZREG_SIZE(vq) __SVE_ZREG_SIZE(vq) argument 201 #define SVE_PT_SVE_PREG_SIZE(vq) __SVE_PREG_SIZE(vq) argument 202 #define SVE_PT_SVE_FFR_SIZE(vq) __SVE_FFR_SIZE(vq) argument 212 #define SVE_PT_SVE_ZREGS_SIZE(vq) \ argument 219 #define SVE_PT_SVE_PREGS_SIZE(vq) \ argument 221 SVE_PT_SVE_PREGS_OFFSET(vq)) 224 #define SVE_PT_SVE_FFR_OFFSET(vq) \ argument 228 ((SVE_PT_SVE_FFR_OFFSET(vq) + SVE_PT_SVE_FFR_SIZE(vq) + \ 319 #define ZA_PT_ZAV_OFFSET(vq, n) \ argument 322 #define ZA_PT_ZA_SIZE(vq) ((vq * __SVE_VQ_BYTES) * (vq * __SVE_VQ_BYTES)) argument [all …]
|
| A D | sve_context.h | 42 #define __sve_vl_from_vq(vq) ((vq) * __SVE_VQ_BYTES) argument 44 #define __SVE_ZREG_SIZE(vq) ((__u32)(vq) * __SVE_VQ_BYTES) argument 45 #define __SVE_PREG_SIZE(vq) ((__u32)(vq) * (__SVE_VQ_BYTES / 8)) argument 46 #define __SVE_FFR_SIZE(vq) __SVE_PREG_SIZE(vq) argument 51 #define __SVE_ZREGS_SIZE(vq) \ argument 54 #define __SVE_PREGS_OFFSET(vq) \ argument 57 (__SVE_PREGS_OFFSET(vq) + __SVE_PREG_SIZE(vq) * (n)) 58 #define __SVE_PREGS_SIZE(vq) \ argument 59 (__SVE_PREG_OFFSET(vq, __SVE_NUM_PREGS) - __SVE_PREGS_OFFSET(vq)) 61 #define __SVE_FFR_OFFSET(vq) \ argument [all …]
|
| A D | sigcontext.h | 211 #define sve_vl_from_vq(vq) __sve_vl_from_vq(vq) argument 276 #define SVE_SIG_ZREG_SIZE(vq) __SVE_ZREG_SIZE(vq) argument 277 #define SVE_SIG_PREG_SIZE(vq) __SVE_PREG_SIZE(vq) argument 278 #define SVE_SIG_FFR_SIZE(vq) __SVE_FFR_SIZE(vq) argument 288 #define SVE_SIG_ZREGS_SIZE(vq) __SVE_ZREGS_SIZE(vq) argument 294 #define SVE_SIG_PREGS_SIZE(vq) __SVE_PREGS_SIZE(vq) argument 296 #define SVE_SIG_FFR_OFFSET(vq) \ argument 299 #define SVE_SIG_REGS_SIZE(vq) \ argument 300 (__SVE_FFR_OFFSET(vq) + __SVE_FFR_SIZE(vq)) 330 #define ZA_SIG_REGS_SIZE(vq) (((vq) * __SVE_VQ_BYTES) * ((vq) * __SVE_VQ_BYTES)) argument [all …]
|
| /linux/tools/virtio/ |
| A D | vhost_net_test.c | 287 virtqueue_disable_cb(vq->vq); in run_tx_test() 290 (vq->started - vq->completed) < 1) { in run_tx_test() 292 r = virtqueue_add_outbuf(vq->vq, &sl, 1, in run_tx_test() 300 if (unlikely(!virtqueue_kick(vq->vq))) { in run_tx_test() 310 while (virtqueue_get_buf(vq->vq, &len)) { in run_tx_test() 331 if (virtqueue_enable_cb_delayed(vq->vq)) in run_tx_test() 334 if (virtqueue_enable_cb(vq->vq)) in run_tx_test() 339 spurious, vq->started, vq->completed); in run_tx_test() 359 r = virtqueue_add_inbuf(vq->vq, &sl, 1, in run_rx_test() 369 if (unlikely(!virtqueue_kick(vq->vq))) { in run_rx_test() [all …]
|
| A D | virtio_test.c | 35 struct virtqueue *vq; member 100 if (info->vq) in vq_reset() 107 assert(info->vq); in vq_reset() 108 info->vq->priv = info; in vq_reset() 187 virtqueue_disable_cb(vq->vq); in run_test() 198 r = virtqueue_add_outbuf(vq->vq, &sl, 1, in run_test() 212 if (unlikely(!virtqueue_kick(vq->vq))) { in run_test() 228 while (virtqueue_get_buf(vq->vq, &len)) { in run_test() 236 vq_reset(vq, vq->vring.num, &dev->vdev); in run_test() 263 if (virtqueue_enable_cb_delayed(vq->vq)) in run_test() [all …]
|
| /linux/include/linux/ |
| A D | virtio.h | 33 void (*callback)(struct virtqueue *vq); 43 int virtqueue_add_outbuf(struct virtqueue *vq, 48 int virtqueue_add_inbuf(struct virtqueue *vq, 59 int virtqueue_add_sgs(struct virtqueue *vq, 68 bool virtqueue_kick(struct virtqueue *vq); 72 bool virtqueue_notify(struct virtqueue *vq); 79 void virtqueue_disable_cb(struct virtqueue *vq); 81 bool virtqueue_enable_cb(struct virtqueue *vq); 104 int virtqueue_reset(struct virtqueue *vq, 184 #define virtio_device_for_each_vq(vdev, vq) \ argument [all …]
|
| /linux/tools/virtio/linux/ |
| A D | virtio.h | 22 void (*callback)(struct virtqueue *vq); 33 int virtqueue_add_sgs(struct virtqueue *vq, 40 int virtqueue_add_outbuf(struct virtqueue *vq, 45 int virtqueue_add_inbuf(struct virtqueue *vq, 50 bool virtqueue_kick(struct virtqueue *vq); 54 void virtqueue_disable_cb(struct virtqueue *vq); 56 bool virtqueue_enable_cb(struct virtqueue *vq); 57 bool virtqueue_enable_cb_delayed(struct virtqueue *vq); 67 bool (*notify)(struct virtqueue *vq), 68 void (*callback)(struct virtqueue *vq), [all …]
|
| /linux/drivers/gpu/drm/virtio/ |
| A D | virtgpu_trace.h | 12 TP_PROTO(struct virtqueue *vq, struct virtio_gpu_ctrl_hdr *hdr, u32 seqno), 13 TP_ARGS(vq, hdr, seqno), 16 __field(unsigned int, vq) 17 __string(name, vq->name) 26 __entry->dev = vq->vdev->index; 27 __entry->vq = vq->index; 33 __entry->num_free = vq->num_free; 37 __entry->dev, __entry->vq, __get_str(name), 43 TP_PROTO(struct virtqueue *vq, struct virtio_gpu_ctrl_hdr *hdr, u32 seqno), 44 TP_ARGS(vq, hdr, seqno) [all …]
|
| /linux/drivers/vdpa/vdpa_sim/ |
| A D | vdpa_sim_blk.c | 123 ret = vringh_getdesc_iotlb(&vq->vring, &vq->out_iov, &vq->in_iov, in vdpasim_blk_handle_req() 128 if (vq->out_iov.used < 1 || vq->in_iov.used < 1) { in vdpasim_blk_handle_req() 130 vq->out_iov.used, vq->in_iov.used); in vdpasim_blk_handle_req() 134 if (vq->in_iov.iov[vq->in_iov.used - 1].iov_len < 1) { in vdpasim_blk_handle_req() 146 bytes = vringh_iov_pull_iotlb(&vq->vring, &vq->out_iov, &hdr, in vdpasim_blk_handle_req() 179 bytes = vringh_iov_push_iotlb(&vq->vring, &vq->in_iov, in vdpasim_blk_handle_req() 202 bytes = vringh_iov_pull_iotlb(&vq->vring, &vq->out_iov, in vdpasim_blk_handle_req() 215 bytes = vringh_iov_push_iotlb(&vq->vring, &vq->in_iov, in vdpasim_blk_handle_req() 245 bytes = vringh_iov_pull_iotlb(&vq->vring, &vq->out_iov, &range, in vdpasim_blk_handle_req() 319 vringh_complete_iotlb(&vq->vring, vq->head, pushed); in vdpasim_blk_handle_req() [all …]
|
| A D | vdpa_sim.c | 87 if (!vq->cb) in vdpasim_vq_notify() 90 vq->cb(vq->private); in vdpasim_vq_notify() 105 vringh_init_iotlb_va(&vq->vring, vdpasim->features, vq->num, in vdpasim_queue_ready() 108 vringh_init_iotlb(&vq->vring, vdpasim->features, vq->num, in vdpasim_queue_ready() 130 vq->ready = false; in vdpasim_vq_reset() 131 vq->desc_addr = 0; in vdpasim_vq_reset() 134 vq->cb = NULL; in vdpasim_vq_reset() 311 vq->num = num; in vdpasim_set_vq_num() 320 return vq->num; in vdpasim_get_vq_size() 336 if (vq->ready) in vdpasim_kick_vq() [all …]
|
| /linux/drivers/vdpa/vdpa_user/ |
| A D | vduse_dev.c | 456 vq->num = 0; in vduse_dev_reset() 457 memset(&vq->state, 0, sizeof(vq->state)); in vduse_dev_reset() 495 if (!vq->ready) in vduse_vq_kick() 498 if (vq->kickfd) in vduse_vq_kick() 544 vq->num = num; in vduse_vdpa_set_vq_num() 552 if (vq->num) in vduse_vdpa_get_vq_size() 950 if (vq->ready && vq->kicked && vq->kickfd) { in vduse_kickfd_setup() 986 if (vq->ready && vq->cb.callback) in vduse_vq_irq_inject() 987 vq->cb.callback(vq->cb.private); in vduse_vq_irq_inject() 999 if (vq->ready && vq->cb.trigger) { in vduse_vq_signal_irqfd() [all …]
|
| /linux/net/vmw_vsock/ |
| A D | virtio_transport.c | 159 struct virtqueue *vq; in virtio_transport_send_pkt_work() local 202 virtqueue_kick(vq); in virtio_transport_send_pkt_work() 226 virtqueue_kick(vq); in virtio_transport_send_skb_fast_path() 312 struct virtqueue *vq; in virtio_vsock_rx_fill() local 333 } while (vq->num_free); in virtio_vsock_rx_fill() 336 virtqueue_kick(vq); in virtio_vsock_rx_fill() 343 struct virtqueue *vq; in virtio_transport_tx_work() local 356 virtqueue_disable_cb(vq); in virtio_transport_tx_work() 387 struct virtqueue *vq; in virtio_vsock_event_fill_one() local 449 struct virtqueue *vq; in virtio_transport_event_work() local [all …]
|
| /linux/tools/testing/selftests/arm64/fp/ |
| A D | sve-ptrace.c | 335 unsigned int vq = sve_vq_from_vl(vl); in ptrace_set_sve_get_sve_data() local 357 SVE_PT_SVE_ZREG_SIZE(vq)); in ptrace_set_sve_get_sve_data() 361 SVE_PT_SVE_PREG_SIZE(vq)); in ptrace_set_sve_get_sve_data() 394 SVE_PT_SVE_ZREG_SIZE(vq)) != 0) { in ptrace_set_sve_get_sve_data() 403 SVE_PT_SVE_PREG_SIZE(vq)) != 0) { in ptrace_set_sve_get_sve_data() 430 unsigned int vq = sve_vq_from_vl(vl); in ptrace_set_sve_get_fpsimd_data() local 458 SVE_PT_SVE_ZREG_SIZE(vq)); in ptrace_set_sve_get_fpsimd_data() 514 unsigned int vq = sve_vq_from_vl(vl); in ptrace_set_fpsimd_get_sve_data() local 621 unsigned int vq, vl; in do_parent() local 702 for (vq = SVE_VQ_MIN; vq <= TEST_VQ_MAX; vq++) { in do_parent() [all …]
|
| A D | fp-ptrace.c | 359 int ret, vq; in check_ptrace_values_sve() local 436 int ret, vq; in check_ptrace_values_ssve() local 511 int ret, vq; in check_ptrace_values_za() local 882 int vq, sme_vq; in check_memory_values() local 917 __SVE_PREG_SIZE(vq))) in check_memory_values() 1031 int ret, vl, vq, regset; in sve_write() local 1034 vq = __sve_vq_from_vl(vl); in sve_write() 1121 int ret, vq; in za_write() local 1296 unsigned int vq; in probe_vls() local 1301 for (vq = ARCH_VQ_MAX; vq > 0; vq /= 2) { in probe_vls() [all …]
|
| /linux/drivers/staging/media/sunxi/cedrus/ |
| A D | cedrus_video.c | 292 struct vb2_queue *vq; in cedrus_s_fmt_vid_cap() local 296 if (vb2_is_busy(vq)) in cedrus_s_fmt_vid_cap() 317 struct vb2_queue *vq; in cedrus_s_fmt_vid_out_p() local 331 vq->subsystem_flags |= in cedrus_s_fmt_vid_out_p() 335 vq->subsystem_flags &= in cedrus_s_fmt_vid_out_p() 375 struct vb2_queue *vq; in cedrus_s_fmt_vid_out() local 384 if (vb2_is_streaming(vq) || (vb2_is_busy(vq) && in cedrus_s_fmt_vid_out() 438 if (V4L2_TYPE_IS_OUTPUT(vq->type)) in cedrus_queue_setup() 460 if (V4L2_TYPE_IS_OUTPUT(vq->type)) in cedrus_queue_cleanup() 488 if (V4L2_TYPE_IS_OUTPUT(vq->type)) in cedrus_buf_prepare() [all …]
|