Home
last modified time | relevance | path

Searched refs:qp (Results 1 – 19 of 19) sorted by relevance

/include/linux/
A Dntb_transport.h65 void (*rx_handler)(struct ntb_transport_qp *qp, void *qp_data,
67 void (*tx_handler)(struct ntb_transport_qp *qp, void *qp_data,
72 unsigned char ntb_transport_qp_num(struct ntb_transport_qp *qp);
73 unsigned int ntb_transport_max_size(struct ntb_transport_qp *qp);
77 void ntb_transport_free_queue(struct ntb_transport_qp *qp);
78 int ntb_transport_rx_enqueue(struct ntb_transport_qp *qp, void *cb, void *data,
82 void *ntb_transport_rx_remove(struct ntb_transport_qp *qp, unsigned int *len);
83 void ntb_transport_link_up(struct ntb_transport_qp *qp);
84 void ntb_transport_link_down(struct ntb_transport_qp *qp);
85 bool ntb_transport_link_query(struct ntb_transport_qp *qp);
[all …]
A Dhisi_acc_qm.h455 void (*req_cb)(struct hisi_qp *qp, void *data);
456 void (*event_cb)(struct hisi_qp *qp);
530 int hisi_qm_start_qp(struct hisi_qp *qp, unsigned long arg);
531 void hisi_qm_stop_qp(struct hisi_qp *qp);
532 int hisi_qp_send(struct hisi_qp *qp, const void *msg);
/include/rdma/
A Drdmavt_qp.h511 struct rvt_qp *qp; member
559 return !!qp->pid; in rvt_is_user_qp()
577 if (qp && atomic_dec_and_test(&qp->refcount)) in rvt_put_qp()
607 struct rvt_qp *qp, in rvt_qp_wqe_reserve() argument
663 return (len + qp->pmtu - 1) >> qp->log_pmtu; in rvt_div_round_up_mtu()
706 qp = rcu_dereference(rvp->qp[qpn]); in rvt_lookup_qpn()
710 for (qp = rcu_dereference(rdi->qp_dev->qp_table[n]); qp; in rvt_lookup_qpn()
711 qp = rcu_dereference(qp->next)) in rvt_lookup_qpn()
715 return qp; in rvt_lookup_qpn()
852 last = rvt_qp_swqe_incr(qp, qp->s_last); in rvt_qp_complete_swqe()
[all …]
A Dmr_pool.h10 struct ib_mr *ib_mr_pool_get(struct ib_qp *qp, struct list_head *list);
11 void ib_mr_pool_put(struct ib_qp *qp, struct list_head *list, struct ib_mr *mr);
13 int ib_mr_pool_init(struct ib_qp *qp, struct list_head *list, int nr,
15 void ib_mr_pool_destroy(struct ib_qp *qp, struct list_head *list);
A Drw.h45 int rdma_rw_ctx_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num,
48 void rdma_rw_ctx_destroy(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
52 int rdma_rw_ctx_signature_init(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
57 void rdma_rw_ctx_destroy_signature(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
62 struct ib_send_wr *rdma_rw_ctx_wrs(struct rdma_rw_ctx *ctx, struct ib_qp *qp,
64 int rdma_rw_ctx_post(struct rdma_rw_ctx *ctx, struct ib_qp *qp, u32 port_num,
70 int rdma_rw_init_mrs(struct ib_qp *qp, struct ib_qp_init_attr *attr);
71 void rdma_rw_cleanup_mrs(struct ib_qp *qp);
A Drdma_vt.h35 struct rvt_qp __rcu *qp[2]; member
213 bool (*schedule_send)(struct rvt_qp *qp);
214 bool (*schedule_send_no_lock)(struct rvt_qp *qp);
230 void (*do_send)(struct rvt_qp *qp);
263 void (*notify_qp_reset)(struct rvt_qp *qp);
275 void (*flush_qp_waiters)(struct rvt_qp *qp);
281 void (*stop_send_queue)(struct rvt_qp *qp);
286 void (*quiesce_qp)(struct rvt_qp *qp);
291 void (*notify_error_qp)(struct rvt_qp *qp);
523 int rvt_invalidate_rkey(struct rvt_qp *qp, u32 rkey);
[all …]
A Duverbs_std_types.h147 struct ib_qp *qp, struct ib_device *device, in ib_set_flow() argument
155 if (qp) { in ib_set_flow()
156 atomic_inc(&qp->usecnt); in ib_set_flow()
157 ibflow->qp = qp; in ib_set_flow()
A Drdma_counter.h53 int rdma_counter_bind_qp_auto(struct ib_qp *qp, u32 port);
54 int rdma_counter_unbind_qp(struct ib_qp *qp, u32 port, bool force);
A Dib_verbs.h775 struct ib_qp *qp; member
1040 struct ib_qp *qp; member
1779 struct ib_qp *qp; member
2136 struct ib_qp *qp; member
3882 int ib_modify_qp(struct ib_qp *qp,
3897 int ib_query_qp(struct ib_qp *qp,
3937 int ib_close_qp(struct ib_qp *qp);
3958 return qp->device->ops.post_send(qp, send_wr, bad_send_wr ? : &dummy); in ib_post_send()
3975 return qp->device->ops.post_recv(qp, recv_wr, bad_recv_wr ? : &dummy); in ib_post_recv()
4593 void ib_drain_rq(struct ib_qp *qp);
[all …]
A Drdma_cm.h115 struct ib_qp *qp; member
A Dib_mad.h572 struct ib_qp *qp; member
/include/net/
A Derspan.h183 } *qp; in erspan_build_header() local
198 qp = (struct qtag_prefix *)(skb->data + 2 * ETH_ALEN); in erspan_build_header()
199 vlan_tci = ntohs(qp->tci); in erspan_build_header()
277 } *qp; in erspan_build_header_v2() local
292 qp = (struct qtag_prefix *)(skb->data + 2 * ETH_ALEN); in erspan_build_header_v2()
293 vlan_tci = ntohs(qp->tci); in erspan_build_header_v2()
/include/linux/mlx4/
A Dqp.h481 int sqd_event, struct mlx4_qp *qp);
483 int mlx4_qp_query(struct mlx4_dev *dev, struct mlx4_qp *qp,
488 struct mlx4_qp *qp, enum mlx4_qp_state *qp_state);
495 void mlx4_qp_remove(struct mlx4_dev *dev, struct mlx4_qp *qp);
507 void mlx4_put_qp(struct mlx4_qp *qp);
A Ddevice.h831 int qp; member
925 } __packed qp; member
1155 int mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp);
1156 void mlx4_qp_free(struct mlx4_dev *dev, struct mlx4_qp *qp);
1167 int mlx4_unicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16],
1169 int mlx4_unicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16],
1171 int mlx4_multicast_attach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16],
1174 int mlx4_multicast_detach(struct mlx4_dev *dev, struct mlx4_qp *qp, u8 gid[16],
/include/linux/qed/
A Dqed_rdma_if.h505 struct qed_rdma_qp *qp; member
533 struct qed_rdma_qp *qp; member
619 int (*rdma_modify_qp)(void *roce_cxt, struct qed_rdma_qp *qp,
622 int (*rdma_query_qp)(void *rdma_cxt, struct qed_rdma_qp *qp,
624 int (*rdma_destroy_qp)(void *rdma_cxt, struct qed_rdma_qp *qp);
/include/linux/mlx5/
A Dqp.h530 int mlx5_debug_qp_add(struct mlx5_core_dev *dev, struct mlx5_core_qp *qp);
531 void mlx5_debug_qp_remove(struct mlx5_core_dev *dev, struct mlx5_core_qp *qp);
/include/trace/events/
A Dib_mad.h57 __entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
146 __entry->qp_num = wr->mad_agent_priv->qp_info->qp->qp_num;
211 __entry->qp_num = qp_info->qp->qp_num;
A Drpcrdma.h1519 DEFINE_ACCEPT_EVENT(qp);
/include/uapi/rdma/
A Dvmw_pvrdma-abi.h292 __aligned_u64 qp; member

Completed in 65 milliseconds