Lines Matching refs:srp_dev

235 	iu->dma = ib_dma_map_single(host->srp_dev->dev, iu->buf, size,  in srp_alloc_iu()
237 if (ib_dma_mapping_error(host->srp_dev->dev, iu->dma)) in srp_alloc_iu()
258 ib_dma_unmap_single(host->srp_dev->dev, iu->dma, iu->size, in srp_free_iu()
280 ret = ib_find_cached_pkey(target->srp_host->srp_dev->dev, in srp_init_ib_qp()
308 new_cm_id = ib_create_cm_id(target->srp_host->srp_dev->dev, in srp_new_ib_cm_id()
316 if (rdma_cap_opa_ah(target->srp_host->srp_dev->dev, in srp_new_ib_cm_id()
499 struct srp_device *dev = target->srp_host->srp_dev; in srp_alloc_fr_pool()
526 struct srp_device *dev = target->srp_host->srp_dev; in srp_create_ch_ib()
639 struct srp_device *dev = target->srp_host->srp_dev; in srp_free_ch_ib()
721 target->srp_host->srp_dev->dev, in srp_ib_lookup_path()
784 ret = ib_query_port(host->srp_dev->dev, host->port, &attr); in srp_get_subnet_timeout()
790 dev_name(&host->srp_dev->dev->dev), subnet_timeout); in srp_get_subnet_timeout()
907 memcpy(ipi + 8, &target->srp_host->srp_dev->dev->node_guid, 8); in srp_send_req()
966 struct srp_device *dev = target->srp_host->srp_dev; in srp_exit_cmd_priv()
984 struct srp_device *srp_dev = target->srp_host->srp_dev; in srp_init_cmd_priv() local
985 struct ib_device *ibdev = srp_dev->dev; in srp_init_cmd_priv()
990 if (srp_dev->use_fast_reg) { in srp_init_cmd_priv()
1181 struct srp_device *dev = target->srp_host->srp_dev; in srp_unmap_data()
1426 struct srp_device *dev = target->srp_host->srp_dev; in srp_map_finish_fr()
1555 struct srp_device *dev = target->srp_host->srp_dev; in srp_map_idb()
1593 struct srp_device *dev = ch->target->srp_host->srp_dev; in srp_check_mapping()
1653 dev = target->srp_host->srp_dev; in srp_map_data()
1996 struct ib_device *dev = target->srp_host->srp_dev->dev; in srp_response_common()
2064 struct ib_device *dev = target->srp_host->srp_dev->dev; in srp_recv_done()
2179 dev = target->srp_host->srp_dev->dev; in srp_queuecommand()
2719 struct ib_device *dev = target->srp_host->srp_dev->dev; in srp_send_tsk_mgmt()
2989 dev_name(&target->srp_host->srp_dev->dev->dev)); in local_ib_device_show()
3118 if (scsi_add_host(target->scsi_host, host->srp_dev->dev->dev.parent)) in srp_add_target()
3691 struct srp_device *srp_dev = host->srp_dev; in add_target_store() local
3692 struct ib_device *ibdev = srp_dev->dev; in add_target_store()
3712 target_host->virt_boundary_mask = ~srp_dev->mr_page_mask; in add_target_store()
3720 target->lkey = host->srp_dev->pd->local_dma_lkey; in add_target_store()
3721 target->global_rkey = host->srp_dev->global_rkey; in add_target_store()
3760 if (!srp_dev->has_fr && !target->allow_ext_sg && in add_target_store()
3766 if (srp_dev->use_fast_reg) { in add_target_store()
3770 max_sectors_per_mr = srp_dev->max_pages_per_mr << in add_target_store()
3771 (ilog2(srp_dev->mr_page_size) - 9); in add_target_store()
3792 srp_dev->max_pages_per_mr - 1) / in add_target_store()
3793 srp_dev->max_pages_per_mr; in add_target_store()
3796 target->scsi_host->max_sectors, srp_dev->max_pages_per_mr, srp_dev->mr_page_size, in add_target_store()
3935 return sysfs_emit(buf, "%s\n", dev_name(&host->srp_dev->dev->dev)); in ibdev_show()
3968 host->srp_dev = device; in srp_add_port()
3989 struct srp_device *srp_dev = client_data; in srp_rename_dev() local
3992 list_for_each_entry_safe(host, tmp_host, &srp_dev->dev_list, list) { in srp_rename_dev()
4003 struct srp_device *srp_dev; in srp_add_one() local
4011 srp_dev = kzalloc(sizeof(*srp_dev), GFP_KERNEL); in srp_add_one()
4012 if (!srp_dev) in srp_add_one()
4021 srp_dev->mr_page_size = 1 << mr_page_shift; in srp_add_one()
4022 srp_dev->mr_page_mask = ~((u64) srp_dev->mr_page_size - 1); in srp_add_one()
4024 do_div(max_pages_per_mr, srp_dev->mr_page_size); in srp_add_one()
4026 attr->max_mr_size, srp_dev->mr_page_size, in srp_add_one()
4028 srp_dev->max_pages_per_mr = min_t(u64, SRP_MAX_PAGES_PER_MR, in srp_add_one()
4031 srp_dev->has_fr = (attr->device_cap_flags & in srp_add_one()
4033 if (!never_register && !srp_dev->has_fr) in srp_add_one()
4036 attr->max_mr_size >= 2 * srp_dev->mr_page_size) in srp_add_one()
4037 srp_dev->use_fast_reg = srp_dev->has_fr; in srp_add_one()
4039 if (never_register || !register_always || !srp_dev->has_fr) in srp_add_one()
4042 if (srp_dev->use_fast_reg) { in srp_add_one()
4043 srp_dev->max_pages_per_mr = in srp_add_one()
4044 min_t(u32, srp_dev->max_pages_per_mr, in srp_add_one()
4047 srp_dev->mr_max_size = srp_dev->mr_page_size * in srp_add_one()
4048 srp_dev->max_pages_per_mr; in srp_add_one()
4052 srp_dev->max_pages_per_mr, srp_dev->mr_max_size); in srp_add_one()
4054 INIT_LIST_HEAD(&srp_dev->dev_list); in srp_add_one()
4056 srp_dev->dev = device; in srp_add_one()
4057 srp_dev->pd = ib_alloc_pd(device, flags); in srp_add_one()
4058 if (IS_ERR(srp_dev->pd)) { in srp_add_one()
4059 int ret = PTR_ERR(srp_dev->pd); in srp_add_one()
4061 kfree(srp_dev); in srp_add_one()
4066 srp_dev->global_rkey = srp_dev->pd->unsafe_global_rkey; in srp_add_one()
4067 WARN_ON_ONCE(srp_dev->global_rkey == 0); in srp_add_one()
4071 host = srp_add_port(srp_dev, p); in srp_add_one()
4073 list_add_tail(&host->list, &srp_dev->dev_list); in srp_add_one()
4076 ib_set_client_data(device, &srp_client, srp_dev); in srp_add_one()
4082 struct srp_device *srp_dev; in srp_remove_one() local
4086 srp_dev = client_data; in srp_remove_one()
4088 list_for_each_entry_safe(host, tmp_host, &srp_dev->dev_list, list) { in srp_remove_one()
4114 ib_dealloc_pd(srp_dev->pd); in srp_remove_one()
4116 kfree(srp_dev); in srp_remove_one()