| /drivers/infiniband/core/ |
| A D | ib_core_uverbs.c | 125 xa_lock(&ucontext->mmap_xa); in rdma_user_mmap_entry_get_pgoff() 138 xa_unlock(&ucontext->mmap_xa); in rdma_user_mmap_entry_get_pgoff() 146 xa_unlock(&ucontext->mmap_xa); in rdma_user_mmap_entry_get_pgoff() 183 struct ib_ucontext *ucontext = entry->ucontext; in rdma_user_mmap_entry_free() local 190 xa_lock(&ucontext->mmap_xa); in rdma_user_mmap_entry_free() 193 xa_unlock(&ucontext->mmap_xa); in rdma_user_mmap_entry_free() 235 xa_lock(&entry->ucontext->mmap_xa); in rdma_user_mmap_entry_remove() 277 entry->ucontext = ucontext; in rdma_user_mmap_entry_insert_range() 287 xa_lock(&ucontext->mmap_xa); in rdma_user_mmap_entry_insert_range() 324 xa_unlock(&ucontext->mmap_xa); in rdma_user_mmap_entry_insert_range() [all …]
|
| A D | uverbs_std_types_device.c | 187 if (IS_ERR(ucontext)) in UVERBS_HANDLER() 188 return PTR_ERR(ucontext); in UVERBS_HANDLER() 189 ib_dev = ucontext->device; in UVERBS_HANDLER() 251 if (IS_ERR(ucontext)) in UVERBS_HANDLER() 252 return PTR_ERR(ucontext); in UVERBS_HANDLER() 253 ib_dev = ucontext->device; in UVERBS_HANDLER() 269 return ucontext->device->ops.query_ucontext(ucontext, attrs); in UVERBS_HANDLER() 345 if (IS_ERR(ucontext)) in UVERBS_HANDLER() 346 return PTR_ERR(ucontext); in UVERBS_HANDLER() 347 ib_dev = ucontext->device; in UVERBS_HANDLER() [all …]
|
| A D | rdma_core.c | 267 struct ib_ucontext *ucontext = in alloc_uobj() local 270 if (IS_ERR(ucontext)) in alloc_uobj() 271 return ERR_CAST(ucontext); in alloc_uobj() 272 attrs->context = ucontext; in alloc_uobj() 855 struct ib_ucontext *ucontext = ufile->ucontext; in ufile_destroy_ucontext() local 876 kfree(ucontext); in ufile_destroy_ucontext() 878 ufile->ucontext = NULL; in ufile_destroy_ucontext() 885 struct ib_ucontext *ucontext = ufile->ucontext; in __uverbs_cleanup_ufile() local 939 if (!ufile->ucontext) in uverbs_destroy_ufile_hw() 1036 struct ib_ucontext *ucontext; in rdma_uattrs_has_raw_cap() local [all …]
|
| A D | uverbs_main.c | 105 struct ib_ucontext *ucontext = smp_load_acquire(&ufile->ucontext); in ib_uverbs_get_ucontext_file() local 111 if (!ucontext) in ib_uverbs_get_ucontext_file() 114 return ucontext; in ib_uverbs_get_ucontext_file() 694 struct ib_ucontext *ucontext; in ib_uverbs_mmap() local 699 ucontext = ib_uverbs_get_ucontext_file(file); in ib_uverbs_mmap() 700 if (IS_ERR(ucontext)) { in ib_uverbs_mmap() 701 ret = PTR_ERR(ucontext); in ib_uverbs_mmap() 708 ret = ucontext->device->ops.mmap(ucontext, vma); in ib_uverbs_mmap() 737 if (!ufile->ucontext) in rdma_umap_open() 912 if (ufile->ucontext) in rdma_user_mmap_disassociate()
|
| A D | uverbs_cmd.c | 219 if (!ucontext) in ib_alloc_ucontext() 223 ucontext->ufile = ufile; in ib_alloc_ucontext() 243 if (file->ucontext) { in ib_init_ucontext() 248 ret = ib_rdmacg_try_charge(&ucontext->cg_obj, ucontext->device, in ib_init_ucontext() 269 ret = ucontext->device->ops.alloc_ucontext(ucontext, in ib_init_ucontext() 280 smp_store_release(&file->ucontext, ucontext); in ib_init_ucontext() 287 ib_rdmacg_uncharge(&ucontext->cg_obj, ucontext->device, in ib_init_ucontext() 397 if (IS_ERR(ucontext)) in ib_uverbs_query_device() 405 copy_query_dev_fields(ucontext, &resp, &ucontext->device->attrs); in ib_uverbs_query_device() 420 if (IS_ERR(ucontext)) in ib_uverbs_query_port() [all …]
|
| /drivers/infiniband/hw/cxgb4/ |
| A D | cq.c | 991 ucontext ? &ucontext->uctx : &chp->cq.rdev->uctx, in c4iw_destroy_cq() 1026 ucontext->is_32b_cqe = 1; in c4iw_create_cq() 1066 memsize = hwentries * ((ucontext && ucontext->is_32b_cqe) ? in c4iw_create_cq() 1080 ucontext ? &ucontext->uctx : &rhp->rdev.uctx, in c4iw_create_cq() 1096 if (ucontext) { in c4iw_create_cq() 1111 uresp.key = ucontext->key; in c4iw_create_cq() 1112 ucontext->key += PAGE_SIZE; in c4iw_create_cq() 1114 ucontext->key += PAGE_SIZE; in c4iw_create_cq() 1134 insert_mmap(ucontext, mm); in c4iw_create_cq() 1142 insert_mmap(ucontext, mm2); in c4iw_create_cq() [all …]
|
| A D | qp.c | 2087 ucontext = qhp->ucontext; in c4iw_destroy_qp() 2111 ucontext ? &ucontext->uctx : &rhp->rdev.uctx, !qhp->srq); in c4iw_destroy_qp() 2177 if (ucontext) { in c4iw_create_qp() 2185 ucontext ? &ucontext->uctx : &rhp->rdev.uctx, in c4iw_create_qp() 2223 if (udata && ucontext) { in c4iw_create_qp() 2340 qhp->ucontext = ucontext; in c4iw_create_qp() 2377 ucontext ? &ucontext->uctx : &rhp->rdev.uctx, !attrs->srq); in c4iw_create_qp() 2751 if (ucontext) in c4iw_create_srq() 2754 ret = alloc_srq_queue(srq, ucontext ? &ucontext->uctx : in c4iw_create_srq() 2819 free_srq_queue(srq, ucontext ? &ucontext->uctx : &rhp->rdev.uctx, in c4iw_create_srq() [all …]
|
| A D | provider.c | 64 struct c4iw_ucontext *ucontext = to_c4iw_ucontext(context); in c4iw_dealloc_ucontext() local 69 rhp = to_c4iw_dev(ucontext->ibucontext.device); in c4iw_dealloc_ucontext() 71 list_for_each_entry_safe(mm, tmp, &ucontext->mmaps, entry) in c4iw_dealloc_ucontext() 73 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in c4iw_dealloc_ucontext() 76 static int c4iw_alloc_ucontext(struct ib_ucontext *ucontext, in c4iw_alloc_ucontext() argument 79 struct ib_device *ibdev = ucontext->device; in c4iw_alloc_ucontext() 80 struct c4iw_ucontext *context = to_c4iw_ucontext(ucontext); in c4iw_alloc_ucontext() 135 struct c4iw_ucontext *ucontext; in c4iw_mmap() local 150 ucontext = to_c4iw_ucontext(context); in c4iw_mmap() 152 mm = remove_mmap(ucontext, key, len); in c4iw_mmap()
|
| A D | iw_cxgb4.h | 490 struct c4iw_ucontext *ucontext; member 552 static inline struct c4iw_mm_entry *remove_mmap(struct c4iw_ucontext *ucontext, in remove_mmap() argument 558 spin_lock(&ucontext->mmap_lock); in remove_mmap() 559 list_for_each_safe(pos, nxt, &ucontext->mmaps) { in remove_mmap() 564 spin_unlock(&ucontext->mmap_lock); in remove_mmap() 570 spin_unlock(&ucontext->mmap_lock); in remove_mmap() 600 static inline void insert_mmap(struct c4iw_ucontext *ucontext, in insert_mmap() argument 603 spin_lock(&ucontext->mmap_lock); in insert_mmap() 606 list_add_tail(&mm->entry, &ucontext->mmaps); in insert_mmap() 607 spin_unlock(&ucontext->mmap_lock); in insert_mmap()
|
| A D | restrack.c | 147 if (qhp->ucontext) in c4iw_fill_res_qp_entry()
|
| /drivers/infiniband/hw/irdma/ |
| A D | verbs.c | 127 vma->vm_private_data = ucontext; in irdma_mmap_legacy() 175 struct irdma_ucontext *ucontext; in irdma_mmap() local 179 ucontext = to_ucontext(context); in irdma_mmap() 182 if (ucontext->legacy_mode) in irdma_mmap() 292 ucontext->iwdev = iwdev; in irdma_alloc_ucontext() 296 ucontext->use_raw_attrs = true; in irdma_alloc_ucontext() 303 ucontext->legacy_mode = true; in irdma_alloc_ucontext() 314 ucontext->db_mmap_entry = in irdma_alloc_ucontext() 318 if (!ucontext->db_mmap_entry) in irdma_alloc_ucontext() 1933 if (ucontext->legacy_mode) in irdma_resize_cq() [all …]
|
| /drivers/infiniband/hw/efa/ |
| A D | efa_verbs.c | 550 struct efa_ucontext *ucontext, in qp_mmap_entries_setup() argument 559 efa_user_mmap_entry_insert(&ucontext->ibucontext, in qp_mmap_entries_setup() 573 efa_user_mmap_entry_insert(&ucontext->ibucontext, in qp_mmap_entries_setup() 586 efa_user_mmap_entry_insert(&ucontext->ibucontext, in qp_mmap_entries_setup() 687 struct efa_ucontext *ucontext; in efa_create_qp() local 743 create_qp_params.uarn = ucontext->uarn; in efa_create_qp() 1212 cq->ucontext = ucontext; in efa_create_cq_umem() 1240 params.uarn = cq->ucontext->uarn; in efa_create_cq_umem() 1991 ucontext->uarn = result.uarn; in efa_alloc_ucontext() 2020 efa_dealloc_uar(dev, ucontext->uarn); in efa_dealloc_ucontext() [all …]
|
| A D | efa.h | 101 struct efa_ucontext *ucontext; member
|
| /drivers/infiniband/hw/mthca/ |
| A D | mthca_srq.c | 103 struct mthca_ucontext *ucontext = rdma_udata_to_drv_context( in mthca_tavor_init_srq_context() local 113 context->uar = cpu_to_be32(ucontext->uar.index); in mthca_tavor_init_srq_context() 124 struct mthca_ucontext *ucontext = rdma_udata_to_drv_context( in mthca_arbel_init_srq_context() local 141 context->logstride_usrpage |= cpu_to_be32(ucontext->uar.index); in mthca_arbel_init_srq_context()
|
| /drivers/infiniband/hw/mlx5/ |
| A D | srq.c | 49 struct mlx5_ib_ucontext *ucontext = rdma_udata_to_drv_context( in create_srq_user() local 71 err = get_srq_user_index(ucontext, &ucmd, udata->inlen, &uidx); in create_srq_user() 86 err = mlx5_ib_db_map_user(ucontext, ucmd.db_addr, &srq->db); in create_srq_user()
|
| A D | mlx5_ib.h | 1610 static inline int get_qp_user_index(struct mlx5_ib_ucontext *ucontext, in get_qp_user_index() argument 1615 u8 cqe_version = ucontext->cqe_version; in get_qp_user_index() 1627 static inline int get_srq_user_index(struct mlx5_ib_ucontext *ucontext, in get_srq_user_index() argument 1632 u8 cqe_version = ucontext->cqe_version; in get_srq_user_index()
|
| A D | qp.c | 884 struct mlx5_ib_ucontext *ucontext = rdma_udata_to_drv_context( in create_user_rq() local 921 err = mlx5_ib_db_map_user(ucontext, ucmd->db_addr, &rwq->db); in create_user_rq() 2806 struct mlx5_ib_ucontext *ucontext = rdma_udata_to_drv_context( in check_valid_flow() local 2824 if (attr->qp_type == IB_QPT_RAW_PACKET && !ucontext->cqe_version) { in check_valid_flow() 3193 struct mlx5_ib_ucontext *ucontext = rdma_udata_to_drv_context( in get_qp_uidx() local 4006 struct mlx5_ib_ucontext *ucontext = rdma_udata_to_drv_context( in get_tx_affinity_rr() local 4011 if (ucontext) in get_tx_affinity_rr() 4012 tx_port_affinity = &ucontext->tx_port_affinity; in get_tx_affinity_rr() 4035 struct mlx5_ib_ucontext *ucontext = rdma_udata_to_drv_context( in get_tx_affinity() local 4057 if (ucontext) in get_tx_affinity() [all …]
|
| /drivers/infiniband/hw/mlx4/ |
| A D | srq.c | 77 struct mlx4_ib_ucontext *ucontext = rdma_udata_to_drv_context( in mlx4_ib_create_srq() local 205 mlx4_ib_db_unmap_user(ucontext, &srq->db); in mlx4_ib_create_srq()
|
| A D | qp.c | 2130 struct mlx4_ib_ucontext *ucontext = rdma_udata_to_drv_context( in __mlx4_ib_modify_qp() local 2237 if (ucontext) in __mlx4_ib_modify_qp() 2239 mlx4_to_hw_uar_index(dev->dev, ucontext->uar.index)); in __mlx4_ib_modify_qp() 2373 if (!ucontext) in __mlx4_ib_modify_qp() 2510 if (!ucontext && in __mlx4_ib_modify_qp() 2521 if (!ucontext && in __mlx4_ib_modify_qp() 2585 if (!ucontext) { in __mlx4_ib_modify_qp()
|
| /drivers/infiniband/hw/usnic/ |
| A D | usnic_ib_verbs.c | 465 struct usnic_ib_ucontext *ucontext = rdma_udata_to_drv_context( in usnic_ib_create_qp() local 514 qp_grp->ctx = ucontext; in usnic_ib_create_qp() 515 list_add_tail(&qp_grp->link, &ucontext->qp_grp_list); in usnic_ib_create_qp()
|
| /drivers/infiniband/hw/qedr/ |
| A D | verbs.h | 48 int qedr_mmap(struct ib_ucontext *ucontext, struct vm_area_struct *vma);
|
| /drivers/infiniband/hw/hns/ |
| A D | hns_roce_main.c | 305 hns_roce_user_mmap_entry_insert(struct ib_ucontext *ucontext, u64 address, in hns_roce_user_mmap_entry_insert() argument 323 ucontext, &entry->rdma_entry, length, 0); in hns_roce_user_mmap_entry_insert() 327 ucontext, &entry->rdma_entry, length, 1, in hns_roce_user_mmap_entry_insert()
|
| /drivers/infiniband/hw/ocrdma/ |
| A D | ocrdma_verbs.c | 541 struct ocrdma_ucontext *ucontext = get_ocrdma_ucontext(context); in ocrdma_mmap() local 551 found = ocrdma_search_mmap(ucontext, vma->vm_pgoff << PAGE_SHIFT, len); in ocrdma_mmap() 964 cq->ucontext = uctx; in ocrdma_copy_cq_uresp() 1075 if (cq->ucontext) { in ocrdma_destroy_cq() 1076 pdid = cq->ucontext->cntxt_pd->id; in ocrdma_destroy_cq() 1077 ocrdma_del_mmap(cq->ucontext, (u64) cq->pa, in ocrdma_destroy_cq() 1079 ocrdma_del_mmap(cq->ucontext, in ocrdma_destroy_cq()
|
| A D | ocrdma.h | 333 struct ocrdma_ucontext *ucontext; member
|
| /drivers/infiniband/hw/mana/ |
| A D | main.c | 202 struct mana_ib_ucontext *ucontext = in mana_ib_alloc_ucontext() local 222 ucontext->doorbell = doorbell_page; in mana_ib_alloc_ucontext()
|