/linux-6.3-rc2/drivers/infiniband/core/ |
A D | iwcm.c | 402 cm_id->device->ops.iw_destroy_listen(cm_id); in destroy_cm_id() 425 cm_id->device->ops.iw_reject(cm_id, NULL, 0); in destroy_cm_id() 439 iwpm_remove_mapinfo(&cm_id->local_addr, &cm_id->m_local_addr); in destroy_cm_id() 519 cm_id->m_local_addr = cm_id->local_addr; in iw_cm_map() 520 cm_id->m_remote_addr = cm_id->remote_addr; in iw_cm_map() 583 ret = cm_id->device->ops.iw_create_listen(cm_id, in iw_cm_listen() 624 ret = cm_id->device->ops.iw_reject(cm_id, private_data, in iw_cm_reject() 660 qp = cm_id->device->ops.iw_get_qp(cm_id->device, iw_param->qpn); in iw_cm_accept() 671 ret = cm_id->device->ops.iw_accept(cm_id, iw_param); in iw_cm_accept() 731 ret = cm_id->device->ops.iw_connect(cm_id, iw_param); in iw_cm_connect() [all …]
|
A D | cm_trace.h | 139 const struct ib_cm_id *cm_id 142 TP_ARGS(cm_id), 153 __entry->cm_id = cm_id; 156 __entry->state = cm_id->state; 173 TP_ARGS(cm_id)) 188 const struct ib_cm_id *cm_id, 192 TP_ARGS(cm_id, reason), 195 __field(const void *, cm_id) 203 __entry->cm_id = cm_id; 223 TP_ARGS(cm_id)) [all …]
|
A D | ucma.c | 93 struct rdma_cm_id *cm_id; member 169 if (!ctx->cm_id->device) { in ucma_get_ctx_dev() 190 ctx->cm_id = NULL; in ucma_close_id() 220 ctx->cm_id = cm_id; in ucma_set_ctx_cm_id() 465 if (IS_ERR(cm_id)) { in ucma_create_id() 466 ret = PTR_ERR(cm_id); in ucma_create_id() 853 if (rdma_cap_ib_sa(ctx->cm_id->device, ctx->cm_id->port_num)) in ucma_query_route() 855 else if (rdma_protocol_roce(ctx->cm_id->device, ctx->cm_id->port_num)) in ucma_query_route() 857 else if (rdma_protocol_iwarp(ctx->cm_id->device, ctx->cm_id->port_num)) in ucma_query_route() 873 if (!cm_id->device) in ucma_query_device_addr() [all …]
|
A D | cma_trace.h | 29 __field(u32, cm_id) 36 __entry->cm_id = id_priv->res.id; 75 __field(u32, cm_id) 82 __entry->cm_id = id_priv->res.id; 104 __field(u32, cm_id) 112 __entry->cm_id = id_priv->res.id; 182 __field(u32, cm_id) 195 __entry->cm_id = id_priv->res.id; 232 __field(u32, cm_id) 264 __field(u32, cm_id) [all …]
|
A D | cm.c | 1036 switch (cm_id->state) { in cm_destroy_id() 1047 cm_id->state = IB_CM_IDLE; in cm_destroy_id() 1053 cm_id->state = IB_CM_IDLE; in cm_destroy_id() 1061 cm_id->state = IB_CM_IDLE; in cm_destroy_id() 1114 cm_id->state = IB_CM_IDLE; in cm_destroy_id() 1150 cm_destroy_id(cm_id, 0); in ib_destroy_cm_id() 2287 trace_icm_send_rep(cm_id); in ib_send_cm_rep() 2362 trace_icm_send_rtu(cm_id); in ib_send_cm_rtu() 3131 cm_id->state = cm_state; in ib_send_cm_mra() 3909 switch (cm_id->state) { in cm_establish() [all …]
|
A D | cma.c | 142 return id_priv->cm_id.iw; in rdma_iw_cm_id() 2409 conn_id->cm_id.ib = cm_id; in cma_ib_req_handler() 2453 if (!cm_id->device) { in rdma_read_gids() 2461 if (rdma_protocol_roce(cm_id->device, cm_id->port_num)) { in rdma_read_gids() 2586 conn_id->cm_id.iw = cm_id; in iw_conn_req_handler() 2621 id_priv->cm_id.ib = id; in cma_ib_listen() 2642 id_priv->cm_id.iw = id; in cma_iw_listen() 4272 id_priv->cm_id.ib = id; in cma_resolve_ib_udp() 4323 id_priv->cm_id.ib = id; in cma_connect_ib() 4377 if (IS_ERR(cm_id)) in cma_connect_iw() [all …]
|
/linux-6.3-rc2/net/rds/ |
A D | rdma_transport.c | 92 ret = rdma_resolve_route(cm_id, in rds_rdma_cm_event_handler_cmn() 105 cm_id->route.path_rec[0].sl = in rds_rdma_cm_event_handler_cmn() 199 struct rdma_cm_id *cm_id; in rds_rdma_listen_init_common() local 204 if (IS_ERR(cm_id)) { in rds_rdma_listen_init_common() 205 ret = PTR_ERR(cm_id); in rds_rdma_listen_init_common() 215 ret = rdma_bind_addr(cm_id, sa); in rds_rdma_listen_init_common() 222 ret = rdma_listen(cm_id, 128); in rds_rdma_listen_init_common() 231 *ret_cm_id = cm_id; in rds_rdma_listen_init_common() 232 cm_id = NULL; in rds_rdma_listen_init_common() 234 if (cm_id) in rds_rdma_listen_init_common() [all …]
|
A D | ib.c | 410 struct rdma_cm_id *cm_id; in rds_ib_laddr_check() local 422 cm_id = rdma_create_id(&init_net, rds_rdma_cm_event_handler, in rds_ib_laddr_check() 424 if (IS_ERR(cm_id)) in rds_ib_laddr_check() 425 return PTR_ERR(cm_id); in rds_ib_laddr_check() 475 ret = rdma_bind_addr(cm_id, sa); in rds_ib_laddr_check() 478 if (ret || !cm_id->device || in rds_ib_laddr_check() 479 cm_id->device->node_type != RDMA_NODE_IB_CA) in rds_ib_laddr_check() 484 cm_id->device ? cm_id->device->node_type : -1); in rds_ib_laddr_check() 487 rdma_destroy_id(cm_id); in rds_ib_laddr_check()
|
A D | rdma_transport.h | 21 int rds_rdma_cm_event_handler(struct rdma_cm_id *cm_id, 23 int rds6_rdma_cm_event_handler(struct rdma_cm_id *cm_id,
|
/linux-6.3-rc2/drivers/infiniband/hw/qedr/ |
A D | qedr_iw_cm.c | 98 ep->cm_id->rem_ref(ep->cm_id); in qedr_iw_free_ep() 167 ep->cm_id->event_handler(ep->cm_id, &event); in qedr_iw_issue_event() 240 ep->cm_id->event_handler(ep->cm_id, &event); in qedr_iw_disconnect_worker() 576 cm_id->add_ref(cm_id); in qedr_iw_connect() 577 ep->cm_id = cm_id; in qedr_iw_connect() 679 cm_id->add_ref(cm_id); in qedr_iw_create_listen() 680 listener->cm_id = cm_id; in qedr_iw_create_listen() 716 cm_id->rem_ref(cm_id); in qedr_iw_create_listen() 731 cm_id->rem_ref(cm_id); in qedr_iw_destroy_listen() 753 cm_id->add_ref(cm_id); in qedr_iw_accept() [all …]
|
A D | qedr_iw_cm.h | 34 int qedr_iw_connect(struct iw_cm_id *cm_id, 37 int qedr_iw_create_listen(struct iw_cm_id *cm_id, int backlog); 39 int qedr_iw_destroy_listen(struct iw_cm_id *cm_id); 41 int qedr_iw_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 43 int qedr_iw_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len);
|
/linux-6.3-rc2/include/rdma/ |
A D | iw_cm.h | 42 typedef int (*iw_cm_handler)(struct iw_cm_id *cm_id, 53 typedef int (*iw_event_handler)(struct iw_cm_id *cm_id, 115 void iw_destroy_cm_id(struct iw_cm_id *cm_id); 128 void iw_cm_unbind_qp(struct iw_cm_id *cm_id, struct ib_qp *qp); 149 int iw_cm_listen(struct iw_cm_id *cm_id, int backlog); 165 int iw_cm_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param); 179 int iw_cm_reject(struct iw_cm_id *cm_id, const void *private_data, 194 int iw_cm_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *iw_param); 206 int iw_cm_disconnect(struct iw_cm_id *cm_id, int abrupt); 217 int iw_cm_init_qp_attr(struct iw_cm_id *cm_id, struct ib_qp_attr *qp_attr,
|
A D | ib_cm.h | 325 void ib_destroy_cm_id(struct ib_cm_id *cm_id); 380 int ib_send_cm_req(struct ib_cm_id *cm_id, 405 int ib_send_cm_rep(struct ib_cm_id *cm_id, 416 int ib_send_cm_rtu(struct ib_cm_id *cm_id, 429 int ib_send_cm_dreq(struct ib_cm_id *cm_id, 444 int ib_send_cm_drep(struct ib_cm_id *cm_id, 476 int ib_send_cm_rej(struct ib_cm_id *cm_id, 496 int ib_send_cm_mra(struct ib_cm_id *cm_id, 517 int ib_cm_init_qp_attr(struct ib_cm_id *cm_id, 538 int ib_send_cm_sidr_req(struct ib_cm_id *cm_id, [all …]
|
/linux-6.3-rc2/drivers/nvme/target/ |
A D | rdma.c | 636 struct rdma_cm_id *cm_id = rsp->queue->cm_id; in nvmet_rdma_rw_ctx_init() local 655 struct rdma_cm_id *cm_id = rsp->queue->cm_id; in nvmet_rdma_rw_ctx_destroy() local 664 rdma_rw_ctx_destroy(&rsp->rw, cm_id->qp, cm_id->port_num, in nvmet_rdma_rw_ctx_destroy() 720 struct rdma_cm_id *cm_id = rsp->queue->cm_id; in nvmet_rdma_queue_response() local 791 struct rdma_cm_id *cm_id = rsp->queue->cm_id; in nvmet_rdma_write_data_done() local 1450 queue->cm_id = cm_id; in nvmet_rdma_alloc_queue() 1742 if (xchg(&port->cm_id, NULL) != cm_id) in nvmet_rdma_device_removal() 1841 struct rdma_cm_id *cm_id = xchg(&port->cm_id, NULL); in nvmet_rdma_disable_port() local 1843 if (cm_id) in nvmet_rdma_disable_port() 1889 port->cm_id = cm_id; in nvmet_rdma_enable_port() [all …]
|
/linux-6.3-rc2/net/9p/ |
A D | trans_rdma.c | 76 struct rdma_cm_id *cm_id; member 280 rdma_disconnect(rdma->cm_id); in p9_cm_event_handler() 348 ib_dma_unmap_single(rdma->cm_id->device, in send_done() 376 if (rdma->cm_id && !IS_ERR(rdma->cm_id)) in rdma_destroy_trans() 377 rdma_destroy_id(rdma->cm_id); in rdma_destroy_trans() 545 rdma_disconnect(rdma->cm_id); in rdma_request() 563 rdma_disconnect(rdma->cm_id); in rdma_close() 659 if (IS_ERR(rdma->cm_id)) in rdma_create_trans() 679 err = rdma_resolve_addr(rdma->cm_id, NULL, in rdma_create_trans() 723 rdma->qp = rdma->cm_id->qp; in rdma_create_trans() [all …]
|
/linux-6.3-rc2/drivers/infiniband/hw/irdma/ |
A D | cm.c | 3466 cm_id = iwqp->cm_id; in irdma_cm_disconn_true() 3680 iwqp->cm_id = cm_id; in irdma_accept() 3681 cm_node->cm_id = cm_id; in irdma_accept() 3689 cm_id->add_ref(cm_id); in irdma_accept() 3828 cm_info.cm_id = cm_id; in irdma_connect() 3870 iwqp->cm_id = cm_id; in irdma_connect() 3872 cm_id->add_ref(cm_id); in irdma_connect() 3962 cm_info.cm_id = cm_id; in irdma_create_listen() 4004 cm_id->add_ref(cm_id); in irdma_create_listen() 4037 cm_id->rem_ref(cm_id); in irdma_destroy_listen() [all …]
|
A D | cm.h | 245 struct iw_cm_id *cm_id; member 296 struct iw_cm_id *cm_id; member 330 struct iw_cm_id *cm_id; member 394 int irdma_accept(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 395 int irdma_reject(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len); 396 int irdma_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param); 397 int irdma_create_listen(struct iw_cm_id *cm_id, int backlog); 398 int irdma_destroy_listen(struct iw_cm_id *cm_id);
|
A D | trace_cm.h | 201 TP_PROTO(struct irdma_cm_node *cm_node, struct iw_cm_id *cm_id, 203 TP_ARGS(cm_node, cm_id, type, status, caller), 206 __field(struct iw_cm_id *, cm_id) 222 __entry->cm_id = cm_id; 241 __entry->cm_id, 257 TP_PROTO(struct iw_cm_id *cm_id, enum iw_cm_event_type type, 259 TP_ARGS(cm_id, type, status, caller), 260 TP_STRUCT__entry(__field(struct iw_cm_id *, cm_id) 265 TP_fast_assign(__entry->cm_id = cm_id; 271 __entry->cm_id,
|
/linux-6.3-rc2/fs/ksmbd/ |
A D | transport_rdma.c | 84 struct rdma_cm_id *cm_id; member 103 struct rdma_cm_id *cm_id; member 369 t->cm_id = cm_id; in alloc_transport() 370 cm_id->context = t; in alloc_transport() 452 if (t->cm_id) in free_transport() 1638 t->cm_id->device->ops.get_port_immutable(t->cm_id->device, in smb_direct_accept_client() 1910 t->qp = t->cm_id->qp; in smb_direct_create_qpair() 2078 cm_id); in smb_direct_listen_handler() 2101 if (IS_ERR(cm_id)) { in smb_direct_listen() 2103 return PTR_ERR(cm_id); in smb_direct_listen() [all …]
|
/linux-6.3-rc2/drivers/infiniband/hw/erdma/ |
A D | erdma_cm.c | 246 cm_id = cep->listen_cep->cm_id; in erdma_cm_upcall() 251 cm_id = cep->cm_id; in erdma_cm_upcall() 267 return cm_id->event_handler(cm_id, &event); in erdma_cm_upcall() 283 if (cep->cm_id) { in erdma_qp_cm_drop() 301 cep->cm_id->rem_ref(cep->cm_id); in erdma_qp_cm_drop() 800 if (cep->cm_id) in erdma_cm_work_handler() 881 cep->cm_id->rem_ref(cep->cm_id); in erdma_cm_work_handler() 1045 cep->cm_id = id; in erdma_connect() 1181 cep->cm_id = id; in erdma_accept() 1355 cep->cm_id->rem_ref(cep->cm_id); in erdma_create_listen() [all …]
|
/linux-6.3-rc2/drivers/infiniband/hw/cxgb4/ |
A D | cm.c | 152 epc->cm_id->rem_ref(epc->cm_id); in deref_cm_id() 153 epc->cm_id = NULL; in deref_cm_id() 160 epc->cm_id->add_ref(epc->cm_id); in ref_cm_id() 1278 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in close_complete_upcall() 1294 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in peer_close_upcall() 1310 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in peer_abort_upcall() 1353 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in connect_reply_upcall() 1408 ep->com.cm_id->event_handler(ep->com.cm_id, &event); in established_upcall() 3181 ep->com.cm_id = cm_id; in c4iw_accept_cr() 3335 ep->com.cm_id = cm_id; in c4iw_connect() [all …]
|
/linux-6.3-rc2/drivers/infiniband/ulp/ipoib/ |
A D | ipoib_cm.c | 439 return ib_send_cm_rep(cm_id, &rep); in ipoib_cm_send_rep() 445 struct net_device *dev = cm_id->context; in ipoib_cm_req_handler() 456 p->id = cm_id; in ipoib_cm_req_handler() 457 cm_id->context = p; in ipoib_cm_req_handler() 512 return ipoib_cm_req_handler(cm_id, event); in ipoib_cm_rx_handler() 514 ib_send_cm_drep(cm_id, NULL, 0); in ipoib_cm_rx_handler() 517 p = cm_id->context; in ipoib_cm_rx_handler() 988 struct ipoib_cm_tx *p = cm_id->context; in ipoib_cm_rep_handler() 1049 ret = ib_send_cm_rtu(cm_id, NULL, 0); in ipoib_cm_rep_handler() 1252 struct ipoib_cm_tx *tx = cm_id->context; in ipoib_cm_tx_handler() [all …]
|
/linux-6.3-rc2/drivers/infiniband/sw/siw/ |
A D | siw_cm.c | 326 id = cep->cm_id; in siw_cm_upcall() 399 if (cep->cm_id) { in siw_qp_cm_drop() 419 cep->cm_id->rem_ref(cep->cm_id); in siw_qp_cm_drop() 1066 if (cep->cm_id) in siw_cm_work_handler() 1122 if (cep->cm_id) in siw_cm_work_handler() 1174 cep->cm_id->rem_ref(cep->cm_id); in siw_cm_work_handler() 1409 cep->cm_id = id; in siw_connect() 1647 cep->cm_id = id; in siw_accept() 1841 cep->cm_id = id; in siw_create_listen() 1887 cep->cm_id->rem_ref(cep->cm_id); in siw_create_listen() [all …]
|
/linux-6.3-rc2/drivers/infiniband/ulp/rtrs/ |
A D | rtrs-srv.c | 1524 rdma_disconnect(con->c.cm_id); in rtrs_srv_close_work() 1548 rdma_destroy_id(con->c.cm_id); in rtrs_srv_close_work() 1646 con->c.cm_id = cm_id; in create_con() 1699 cm_id->context = &con->c; in create_con() 1950 c = cm_id->context; in rtrs_srv_rdma_cm_handler() 1983 struct rdma_cm_id *cm_id; in rtrs_srv_cm_init() local 1988 if (IS_ERR(cm_id)) { in rtrs_srv_cm_init() 1989 ret = PTR_ERR(cm_id); in rtrs_srv_cm_init() 1999 ret = rdma_listen(cm_id, 64); in rtrs_srv_cm_init() 2006 return cm_id; in rtrs_srv_cm_init() [all …]
|
A D | rtrs.c | 216 rdma_notify(con->cm_id, IB_EVENT_COMM_EST); in qp_event_handler() 233 struct rdma_cm_id *cm_id = con->cm_id; in create_cq() local 237 cq = ib_alloc_cq(cm_id->device, con, nr_cqe, cq_vector, in create_cq() 240 cq = ib_cq_pool_get(cm_id->device, nr_cqe, cq_vector, poll_ctx); in create_cq() 257 struct rdma_cm_id *cm_id = con->cm_id; in create_qp() local 272 ret = rdma_create_qp(cm_id, pd, &init_attr); in create_qp() 277 con->qp = cm_id->qp; in create_qp() 319 rdma_destroy_qp(con->cm_id); in rtrs_cq_qp_destroy()
|