Home
last modified time | relevance | path

Searched refs:ucmd (Results 1 – 25 of 36) sorted by relevance

12

/linux-6.3-rc2/drivers/iommu/iommufd/
A Dioas.c39 struct iommu_ioas_alloc *cmd = ucmd->cmd; in iommufd_ioas_alloc_ioctl()
46 ioas = iommufd_ioas_alloc(ucmd->ictx); in iommufd_ioas_alloc_ioctl()
51 rc = iommufd_ucmd_respond(ucmd, sizeof(*cmd)); in iommufd_ioas_alloc_ioctl()
65 struct iommu_ioas_iova_ranges *cmd = ucmd->cmd; in iommufd_ioas_iova_ranges()
74 ioas = iommufd_get_ioas(ucmd, cmd->ioas_id); in iommufd_ioas_iova_ranges()
154 ioas = iommufd_get_ioas(ucmd, cmd->ioas_id); in iommufd_ioas_allow_iovas()
202 struct iommu_ioas_map *cmd = ucmd->cmd; in iommufd_ioas_map()
216 ioas = iommufd_get_ioas(ucmd, cmd->ioas_id); in iommufd_ioas_map()
237 struct iommu_ioas_copy *cmd = ucmd->cmd; in iommufd_ioas_copy()
290 struct iommu_ioas_unmap *cmd = ucmd->cmd; in iommufd_ioas_unmap()
[all …]
A Diommufd_private.h105 if (copy_to_user(ucmd->ubuffer, ucmd->cmd, in iommufd_ucmd_respond()
106 min_t(size_t, ucmd->user_size, cmd_len))) in iommufd_ucmd_respond()
217 return container_of(iommufd_get_object(ucmd->ictx, id, in iommufd_get_ioas()
223 int iommufd_ioas_alloc_ioctl(struct iommufd_ucmd *ucmd);
225 int iommufd_ioas_iova_ranges(struct iommufd_ucmd *ucmd);
227 int iommufd_ioas_map(struct iommufd_ucmd *ucmd);
228 int iommufd_ioas_copy(struct iommufd_ucmd *ucmd);
229 int iommufd_ioas_unmap(struct iommufd_ucmd *ucmd);
230 int iommufd_ioas_option(struct iommufd_ucmd *ucmd);
234 int iommufd_vfio_ioas(struct iommufd_ucmd *ucmd);
[all …]
A Dmain.c227 struct iommu_option *cmd = ucmd->cmd; in iommufd_option()
238 rc = iommufd_ioas_option(ucmd); in iommufd_option()
310 struct iommufd_ucmd ucmd = {}; in iommufd_fops_ioctl() local
320 ucmd.ictx = ictx; in iommufd_fops_ioctl()
321 ucmd.ubuffer = (void __user *)arg; in iommufd_fops_ioctl()
322 ret = get_user(ucmd.user_size, (u32 __user *)ucmd.ubuffer); in iommufd_fops_ioctl()
329 if (ucmd.user_size < op->min_size) in iommufd_fops_ioctl()
332 ucmd.cmd = &buf; in iommufd_fops_ioctl()
333 ret = copy_struct_from_user(ucmd.cmd, op->size, ucmd.ubuffer, in iommufd_fops_ioctl()
334 ucmd.user_size); in iommufd_fops_ioctl()
[all …]
A Dselftest.c78 ioas = iommufd_get_ioas(ucmd, ioas_id); in iommufd_test_syz_conv_iova_id()
282 ioas = iommufd_get_ioas(ucmd, cmd->id); in iommufd_test_mock_domain()
291 sobj->idev.ictx = ucmd->ictx; in iommufd_test_mock_domain()
325 ioas = iommufd_get_ioas(ucmd, mockpt_id); in iommufd_test_add_reserved()
554 struct iommu_test_cmd *cmd = ucmd->cmd; in iommufd_test_create_access()
574 ucmd->ictx, ioas_id, in iommufd_test_create_access()
628 struct iommu_test_cmd *cmd = ucmd->cmd; in iommufd_test_access_pages()
723 struct iommu_test_cmd *cmd = ucmd->cmd; in iommufd_test_access_rw()
791 struct iommu_test_cmd *cmd = ucmd->cmd; in iommufd_test()
802 ucmd, cmd->id, cmd->check_map.iova, in iommufd_test()
[all …]
A Dvfio_compat.c123 int iommufd_vfio_ioas(struct iommufd_ucmd *ucmd) in iommufd_vfio_ioas() argument
125 struct iommu_vfio_ioas *cmd = ucmd->cmd; in iommufd_vfio_ioas()
132 ioas = get_compat_ioas(ucmd->ictx); in iommufd_vfio_ioas()
137 return iommufd_ucmd_respond(ucmd, sizeof(*cmd)); in iommufd_vfio_ioas()
140 ioas = iommufd_get_ioas(ucmd, cmd->ioas_id); in iommufd_vfio_ioas()
143 xa_lock(&ucmd->ictx->objects); in iommufd_vfio_ioas()
144 ucmd->ictx->vfio_ioas = ioas; in iommufd_vfio_ioas()
145 xa_unlock(&ucmd->ictx->objects); in iommufd_vfio_ioas()
150 xa_lock(&ucmd->ictx->objects); in iommufd_vfio_ioas()
151 ucmd->ictx->vfio_ioas = NULL; in iommufd_vfio_ioas()
[all …]
/linux-6.3-rc2/drivers/infiniband/hw/mana/
A Dwq.c14 struct mana_ib_create_wq ucmd = {}; in mana_ib_create_wq() local
19 if (udata->inlen < sizeof(ucmd)) in mana_ib_create_wq()
22 err = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); in mana_ib_create_wq()
33 ibdev_dbg(&mdev->ib_dev, "ucmd wq_buf_addr 0x%llx\n", ucmd.wq_buf_addr); in mana_ib_create_wq()
35 umem = ib_umem_get(pd->device, ucmd.wq_buf_addr, ucmd.wq_buf_size, in mana_ib_create_wq()
46 wq->wq_buf_size = ucmd.wq_buf_size; in mana_ib_create_wq()
A Dqp.c121 ret = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); in mana_ib_create_qp_rss()
154 ucmd.rx_hash_function); in mana_ib_create_qp_rss()
159 port = ucmd.port; in mana_ib_create_qp_rss()
169 ucmd.rx_hash_function, port); in mana_ib_create_qp_rss()
224 ucmd.rx_hash_key_len, in mana_ib_create_qp_rss()
225 ucmd.rx_hash_key); in mana_ib_create_qp_rss()
268 struct mana_ib_create_qp ucmd = {}; in mana_ib_create_qp_raw() local
283 err = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); in mana_ib_create_qp_raw()
291 port = ucmd.port; in mana_ib_create_qp_raw()
320 ucmd.sq_buf_addr, ucmd.port); in mana_ib_create_qp_raw()
[all …]
A Dcq.c13 struct mana_ib_create_cq ucmd = {}; in mana_ib_create_cq() local
19 if (udata->inlen < sizeof(ucmd)) in mana_ib_create_cq()
22 err = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); in mana_ib_create_cq()
35 cq->umem = ib_umem_get(ibdev, ucmd.buf_addr, cq->cqe * COMP_ENTRY_SIZE, in mana_ib_create_cq()
/linux-6.3-rc2/drivers/infiniband/hw/mthca/
A Dmthca_provider.c405 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mthca_create_srq()
410 ucmd.db_page); in mthca_create_srq()
474 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mthca_create_qp()
479 ucmd.sq_db_index, in mthca_create_qp()
480 ucmd.sq_db_page); in mthca_create_qp()
486 ucmd.rq_db_index, in mthca_create_qp()
595 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mthca_create_cq()
721 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mthca_resize_cq()
725 lkey = ucmd.lkey; in mthca_resize_cq()
847 ucmd.mr_attrs = 0; in mthca_reg_user_mr()
[all …]
/linux-6.3-rc2/drivers/infiniband/hw/hns/
A Dhns_roce_cq.c300 struct hns_roce_ib_create_cq *ucmd) in get_cq_ucmd() argument
305 ret = ib_copy_from_udata(ucmd, udata, min(udata->inlen, sizeof(*ucmd))); in get_cq_ucmd()
315 struct hns_roce_ib_create_cq *ucmd) in set_cq_param() argument
331 struct hns_roce_ib_create_cq *ucmd) in set_cqe_size() argument
341 if (ucmd->cqe_size != HNS_ROCE_V2_CQE_SIZE && in set_cqe_size()
342 ucmd->cqe_size != HNS_ROCE_V3_CQE_SIZE) { in set_cqe_size()
344 "invalid cqe size %u.\n", ucmd->cqe_size); in set_cqe_size()
348 hr_cq->cqe_size = ucmd->cqe_size; in set_cqe_size()
363 struct hns_roce_ib_create_cq ucmd = {}; in hns_roce_create_cq() local
374 ret = get_cq_ucmd(hr_cq, udata, &ucmd); in hns_roce_create_cq()
[all …]
A Dhns_roce_qp.c579 struct hns_roce_ib_create_qp *ucmd) in check_sq_size_with_integrity() argument
585 if (ucmd->log_sq_stride > max_sq_stride || in check_sq_size_with_integrity()
602 struct hns_roce_ib_create_qp *ucmd) in set_user_sq_size() argument
621 hr_qp->sq.wqe_shift = ucmd->log_sq_stride; in set_user_sq_size()
768 struct hns_roce_ib_create_qp *ucmd) in user_qp_has_sdb() argument
826 struct hns_roce_ib_create_qp *ucmd, in alloc_user_qp_db() argument
900 struct hns_roce_ib_create_qp *ucmd, in alloc_qp_db() argument
994 struct hns_roce_ib_create_qp *ucmd) in set_qp_param() argument
1014 ret = ib_copy_from_udata(ucmd, udata, in set_qp_param()
1015 min(udata->inlen, sizeof(*ucmd))); in set_qp_param()
[all …]
A Dhns_roce_srq.c345 struct hns_roce_ib_create_srq ucmd = {}; in alloc_srq_buf() local
349 ret = ib_copy_from_udata(&ucmd, udata, in alloc_srq_buf()
350 min(udata->inlen, sizeof(ucmd))); in alloc_srq_buf()
359 ret = alloc_srq_idx(hr_dev, srq, udata, ucmd.que_addr); in alloc_srq_buf()
363 ret = alloc_srq_wqe_buf(hr_dev, srq, udata, ucmd.buf_addr); in alloc_srq_buf()
/linux-6.3-rc2/drivers/infiniband/hw/mlx5/
A Dqp.c448 if (ucmd) { in set_rq_size()
450 if (ucmd->rq_wqe_shift > BITS_PER_BYTE * sizeof(ucmd->rq_wqe_shift)) in set_rq_size()
646 if (ucmd->sq_wqe_count && !is_power_of_2(ucmd->sq_wqe_count)) { in set_user_buf_size()
1701 void *ucmd; member
1713 struct mlx5_ib_create_qp_rss *ucmd = params->ucmd; in create_rss_raw_qp_tir() local
2066 struct mlx5_ib_create_qp *ucmd = params->ucmd; in create_dci() local
2227 struct mlx5_ib_create_qp *ucmd = params->ucmd; in create_user_qp() local
2723 struct mlx5_ib_create_qp *ucmd = params->ucmd; in create_dct() local
3064 if (inlen > ucmd && !ib_is_udata_cleared(udata, ucmd, inlen - ucmd)) in process_udata_size()
3166 struct mlx5_ib_create_qp *ucmd = params->ucmd; in get_qp_uidx() local
[all …]
A Dcq.c742 if ((ucmd.cqe_size != 64 && ucmd.cqe_size != 128) || in create_cq_user()
743 ucmd.reserved0 || ucmd.reserved1) in create_cq_user()
746 *cqe_size = ucmd.cqe_size; in create_cq_user()
772 ucmd.buf_addr, entries * ucmd.cqe_size, in create_cq_user()
792 *index = ucmd.uar_page_index; in create_cq_user()
800 if (ucmd.cqe_comp_en == 1) { in create_cq_user()
1156 struct mlx5_ib_resize_cq ucmd; in resize_user() local
1160 err = ib_copy_from_udata(&ucmd, udata, sizeof(ucmd)); in resize_user()
1164 if (ucmd.reserved0 || ucmd.reserved1) in resize_user()
1168 if (ucmd.cqe_size && SIZE_MAX / ucmd.cqe_size <= entries - 1) in resize_user()
[all …]
A Dsrq.c48 struct mlx5_ib_create_srq ucmd = {}; in create_srq_user() local
55 ucmdlen = min(udata->inlen, sizeof(ucmd)); in create_srq_user()
57 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { in create_srq_user()
62 if (ucmd.reserved0 || ucmd.reserved1) in create_srq_user()
65 if (udata->inlen > sizeof(ucmd) && in create_srq_user()
66 !ib_is_udata_cleared(udata, sizeof(ucmd), in create_srq_user()
67 udata->inlen - sizeof(ucmd))) in create_srq_user()
71 err = get_srq_user_index(ucontext, &ucmd, udata->inlen, &uidx); in create_srq_user()
76 srq->wq_sig = !!(ucmd.flags & MLX5_SRQ_FLAG_SIGNATURE); in create_srq_user()
78 srq->umem = ib_umem_get(pd->device, ucmd.buf_addr, buf_size, 0); in create_srq_user()
[all …]
A Dcounters.h15 struct mlx5_ib_create_flow *ucmd);
A Dmlx5_ib.h1506 struct mlx5_ib_create_qp *ucmd, in get_qp_user_index() argument
1512 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) && !cqe_version && in get_qp_user_index()
1513 (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_qp_user_index()
1516 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) != !!cqe_version) in get_qp_user_index()
1519 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_qp_user_index()
1523 struct mlx5_ib_create_srq *ucmd, in get_srq_user_index() argument
1529 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) && !cqe_version && in get_srq_user_index()
1530 (ucmd->uidx == MLX5_IB_DEFAULT_UIDX)) in get_srq_user_index()
1533 if ((offsetofend(typeof(*ucmd), uidx) <= inlen) != !!cqe_version) in get_srq_user_index()
1536 return verify_assign_uidx(cqe_version, ucmd->uidx, user_index); in get_srq_user_index()
/linux-6.3-rc2/drivers/dma-buf/
A Ddma-heap.c126 static long dma_heap_ioctl(struct file *file, unsigned int ucmd, in dma_heap_ioctl() argument
133 int nr = _IOC_NR(ucmd); in dma_heap_ioctl()
145 out_size = _IOC_SIZE(ucmd); in dma_heap_ioctl()
147 if ((ucmd & kcmd & IOC_IN) == 0) in dma_heap_ioctl()
149 if ((ucmd & kcmd & IOC_OUT) == 0) in dma_heap_ioctl()
/linux-6.3-rc2/drivers/infiniband/hw/mlx4/
A Dqp.c452 ucmd->log_sq_stride > in set_user_sq_size()
574 ucmd->rx_hash_fields_mask); in set_qp_rss()
723 sizeof(ucmd.reserved1); in _mlx4_ib_create_qp_rss()
729 if (ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen))) { in _mlx4_ib_create_qp_rss()
734 if (memchr_inv(ucmd.reserved, 0, sizeof(ucmd.reserved))) in _mlx4_ib_create_qp_rss()
737 if (ucmd.comp_mask || ucmd.reserved1) in _mlx4_ib_create_qp_rss()
1063 struct mlx4_ib_create_qp ucmd; in create_qp_common() local
4129 sizeof(ucmd.comp_mask); in mlx4_ib_create_wq()
4266 sizeof(ucmd.reserved); in mlx4_ib_modify_wq()
4275 if (ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen))) in mlx4_ib_modify_wq()
[all …]
A Dsrq.c112 struct mlx4_ib_create_srq ucmd; in mlx4_ib_create_srq() local
114 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mlx4_ib_create_srq()
118 ib_umem_get(ib_srq->device, ucmd.buf_addr, buf_size, 0); in mlx4_ib_create_srq()
132 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &srq->db); in mlx4_ib_create_srq()
A Dcq.c205 struct mlx4_ib_create_cq ucmd; in mlx4_ib_create_cq() local
207 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_cq()
212 buf_addr = (void *)(unsigned long)ucmd.buf_addr; in mlx4_ib_create_cq()
214 ucmd.buf_addr, entries); in mlx4_ib_create_cq()
218 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &cq->db); in mlx4_ib_create_cq()
317 struct mlx4_ib_resize_cq ucmd; in mlx4_alloc_resize_umem() local
323 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) in mlx4_alloc_resize_umem()
331 ucmd.buf_addr, entries); in mlx4_alloc_resize_umem()
/linux-6.3-rc2/drivers/infiniband/hw/vmw_pvrdma/
A Dpvrdma_srq.c109 struct pvrdma_create_srq ucmd; in pvrdma_create_srq() local
144 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_srq()
149 srq->umem = ib_umem_get(ibsrq->device, ucmd.buf_addr, ucmd.buf_size, 0); in pvrdma_create_srq()
A Dpvrdma_cq.c116 struct pvrdma_create_cq ucmd; in pvrdma_create_cq() local
136 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_cq()
141 cq->umem = ib_umem_get(ibdev, ucmd.buf_addr, ucmd.buf_size, in pvrdma_create_cq()
/linux-6.3-rc2/arch/mips/cavium-octeon/
A Docteon-platform.c34 u32 ucmd; in octeon2_usb_reset() local
41 ucmd = cvmx_read64_uint32(CVMX_UAHCX_EHCI_USBCMD); in octeon2_usb_reset()
42 ucmd &= ~CMD_RUN; in octeon2_usb_reset()
43 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset()
45 ucmd |= CMD_RESET; in octeon2_usb_reset()
46 cvmx_write64_uint32(CVMX_UAHCX_EHCI_USBCMD, ucmd); in octeon2_usb_reset()
47 ucmd = cvmx_read64_uint32(CVMX_UAHCX_OHCI_USBCMD); in octeon2_usb_reset()
48 ucmd |= CMD_RUN; in octeon2_usb_reset()
49 cvmx_write64_uint32(CVMX_UAHCX_OHCI_USBCMD, ucmd); in octeon2_usb_reset()
/linux-6.3-rc2/drivers/infiniband/sw/rxe/
A Drxe_srq.c146 struct rxe_modify_srq_cmd *ucmd, struct ib_udata *udata) in rxe_srq_from_attr() argument
157 mi = u64_to_user_ptr(ucmd->mmap_info_addr); in rxe_srq_from_attr()

Completed in 59 milliseconds

12