Home
last modified time | relevance | path

Searched refs:tx_ctx (Results 1 – 12 of 12) sorted by relevance

/drivers/net/ethernet/fungible/funeth/
A Dfuneth_ktls.c32 struct fun_ktls_tx_ctx *tx_ctx; in fun_ktls_add() local
64 tx_ctx = tls_driver_ctx(sk, direction); in fun_ktls_add()
65 tx_ctx->tlsid = rsp.tlsid; in fun_ktls_add()
66 tx_ctx->next_seq = start_offload_tcp_sn; in fun_ktls_add()
77 struct fun_ktls_tx_ctx *tx_ctx; in fun_ktls_del() local
82 tx_ctx = __tls_driver_ctx(tls_ctx, direction); in fun_ktls_del()
89 req.tlsid = tx_ctx->tlsid; in fun_ktls_del()
100 struct fun_ktls_tx_ctx *tx_ctx; in fun_ktls_resync() local
106 tx_ctx = tls_driver_ctx(sk, direction); in fun_ktls_resync()
113 req.tlsid = tx_ctx->tlsid; in fun_ktls_resync()
[all …]
/drivers/infiniband/sw/siw/
A Dsiw_qp.c137 qp->tx_ctx.tx_suspend = 1; in siw_qp_llp_close()
559 if (qp->tx_ctx.mpa_crc_enabled) { in siw_send_terminate()
560 siw_crc_init(&qp->tx_ctx.mpa_crc); in siw_send_terminate()
561 siw_crc_update(&qp->tx_ctx.mpa_crc, in siw_send_terminate()
564 siw_crc_update(&qp->tx_ctx.mpa_crc, in siw_send_terminate()
567 siw_crc_final(&qp->tx_ctx.mpa_crc, (u8 *)&crc); in siw_send_terminate()
613 siw_crc_init(&qp->tx_ctx.mpa_crc); in siw_qp_nextstate_from_idle()
614 qp->tx_ctx.mpa_crc_enabled = true; in siw_qp_nextstate_from_idle()
631 qp->tx_ctx.ddp_msn[RDMAP_UNTAGGED_QN_SEND] = 0; in siw_qp_nextstate_from_idle()
898 qp->tx_ctx.orq_fence = 1; in siw_activate_tx_from_sq()
[all …]
A Dsiw_qp_tx.c708 struct siw_iwarp_tx *c_tx = &qp->tx_ctx; in siw_prepare_fpdu()
795 struct siw_iwarp_tx *c_tx = &qp->tx_ctx; in siw_qp_sq_proc_tx()
797 int rv = 0, burst_len = qp->tx_ctx.burst; in siw_qp_sq_proc_tx()
917 qp->tx_ctx.burst = burst_len; in siw_qp_sq_proc_tx()
1036 if (unlikely(qp->tx_ctx.tx_suspend)) { in siw_qp_sq_process()
1094 qp->tx_ctx.ctrl_sent, qp->tx_ctx.ctrl_len, in siw_qp_sq_process()
1095 qp->tx_ctx.bytes_unsent); in siw_qp_sq_process()
1131 if (!qp->tx_ctx.tx_suspend) in siw_qp_sq_process()
1176 !qp->tx_ctx.tx_suspend)) { in siw_sq_resume()
1184 if (!qp->tx_ctx.tx_suspend) in siw_sq_resume()
A Dsiw.h433 struct siw_iwarp_tx tx_ctx; /* Transmit context */ member
471 #define tx_qp(tx) container_of(tx, struct siw_qp, tx_ctx)
472 #define tx_wqe(qp) (&(qp)->tx_ctx.wqe_active)
A Dsiw_verbs.c436 qp->tx_ctx.gso_seg_limit = 1; in siw_create_qp()
437 qp->tx_ctx.zcopy_tx = zcopy_tx; in siw_create_qp()
585 qp->tx_ctx.tx_suspend = 1; in siw_verbs_modify_qp()
974 qp->tx_ctx.in_syscall = 1; in siw_post_send()
976 if (siw_qp_sq_process(qp) != 0 && !(qp->tx_ctx.tx_suspend)) in siw_post_send()
979 qp->tx_ctx.in_syscall = 0; in siw_post_send()
A Dsiw_cm.c392 qp->tx_ctx.tx_suspend = 1; in siw_qp_cm_drop()
767 qp->tx_ctx.gso_seg_limit = 0; in siw_proc_mpareply()
1314 cep->qp->tx_ctx.tx_suspend = 1; in siw_cm_llp_state_change()
1588 qp->tx_ctx.gso_seg_limit = 0; in siw_accept()
A Dsiw_qp_rx.c1144 if (qp->tx_ctx.orq_fence) { in siw_check_tx_fence()
1165 qp->tx_ctx.orq_fence = 0; in siw_check_tx_fence()
1173 qp->tx_ctx.orq_fence = 0; in siw_check_tx_fence()
/drivers/net/wireless/intel/ipw2x00/
A Dlibipw_crypto_wep.c29 struct arc4_ctx tx_ctx; member
132 arc4_setkey(&wep->tx_ctx, key, klen); in libipw_wep_encrypt()
133 arc4_crypt(&wep->tx_ctx, pos, pos, len + 4); in libipw_wep_encrypt()
/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/
A Dchcr_ktls.c646 struct tls_offload_context_tx *tx_ctx; in chcr_ktls_cpl_act_open_rpl() local
685 tx_ctx = tls_offload_ctx_tx(tls_ctx); in chcr_ktls_cpl_act_open_rpl()
1923 struct tls_offload_context_tx *tx_ctx; in chcr_ktls_xmit() local
1943 tx_ctx = tls_offload_ctx_tx(tls_ctx); in chcr_ktls_xmit()
1978 spin_lock_irqsave(&tx_ctx->lock, flags); in chcr_ktls_xmit()
1984 record = tls_get_record(tx_ctx, tcp_seq, in chcr_ktls_xmit()
2014 spin_unlock_irqrestore(&tx_ctx->lock, in chcr_ktls_xmit()
2096 spin_unlock_irqrestore(&tx_ctx->lock, flags); in chcr_ktls_xmit()
2184 struct tls_offload_context_tx *tx_ctx; in ch_ktls_reset_all_conn() local
2188 xa_for_each(&u_ctx->tid_list, index, tx_ctx) { in ch_ktls_reset_all_conn()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
A Dktls_tx.c98 struct tls_offload_context_tx *tx_ctx; member
496 priv_tx->tx_ctx = tls_offload_ctx_tx(tls_ctx); in mlx5e_ktls_add_tx()
625 struct tls_offload_context_tx *tx_ctx = priv_tx->tx_ctx; in tx_sync_info_get() local
632 spin_lock_irqsave(&tx_ctx->lock, flags); in tx_sync_info_get()
633 record = tls_get_record(tx_ctx, tcp_seq, &info->rcd_sn); in tx_sync_info_get()
672 spin_unlock_irqrestore(&tx_ctx->lock, flags); in tx_sync_info_get()
/drivers/net/ethernet/intel/i40e/
A Di40e_virtchnl_pf.c634 struct i40e_hmc_obj_txq tx_ctx; in i40e_config_vsi_tx_queue() local
652 memset(&tx_ctx, 0, sizeof(struct i40e_hmc_obj_txq)); in i40e_config_vsi_tx_queue()
655 tx_ctx.base = info->dma_ring_addr / 128; in i40e_config_vsi_tx_queue()
656 tx_ctx.qlen = info->ring_len; in i40e_config_vsi_tx_queue()
657 tx_ctx.rdylist = le16_to_cpu(vsi->info.qs_handle[0]); in i40e_config_vsi_tx_queue()
658 tx_ctx.rdylist_act = 0; in i40e_config_vsi_tx_queue()
659 tx_ctx.head_wb_ena = info->headwb_enabled; in i40e_config_vsi_tx_queue()
660 tx_ctx.head_wb_addr = info->dma_headwb_addr; in i40e_config_vsi_tx_queue()
673 ret = i40e_set_lan_tx_queue_context(hw, pf_queue_id, &tx_ctx); in i40e_config_vsi_tx_queue()
A Di40e_main.c3421 struct i40e_hmc_obj_txq tx_ctx; in i40e_configure_tx_ring() local
3440 memset(&tx_ctx, 0, sizeof(tx_ctx)); in i40e_configure_tx_ring()
3442 tx_ctx.new_context = 1; in i40e_configure_tx_ring()
3443 tx_ctx.base = (ring->dma / 128); in i40e_configure_tx_ring()
3444 tx_ctx.qlen = ring->count; in i40e_configure_tx_ring()
3447 tx_ctx.fd_ena = 1; in i40e_configure_tx_ring()
3449 tx_ctx.timesync_ena = 1; in i40e_configure_tx_ring()
3452 tx_ctx.head_wb_ena = 1; in i40e_configure_tx_ring()
3453 tx_ctx.head_wb_addr = ring->dma + in i40e_configure_tx_ring()
3468 tx_ctx.rdylist = in i40e_configure_tx_ring()
[all …]

Completed in 68 milliseconds