Lines Matching refs:iwdev

224 	struct irdma_device *iwdev = rf->iwdev;  in irdma_process_aeq()  local
239 ibdev_dbg(&iwdev->ibdev, in irdma_process_aeq()
251 atomic_dec(&iwdev->vsi.qp_suspend_reqs); in irdma_process_aeq()
252 wake_up(&iwdev->suspend_wq); in irdma_process_aeq()
255 ibdev_dbg(&iwdev->ibdev, "AEQ: qp_id %d is already freed\n", in irdma_process_aeq()
322 if (iwqp->iwdev->vsi.tc_change_pending) { in irdma_process_aeq()
324 wake_up(&iwqp->iwdev->suspend_wq); in irdma_process_aeq()
334 ibdev_err(&iwdev->ibdev, in irdma_process_aeq()
379 ibdev_err(&iwdev->ibdev, "abnormal ae_id = 0x%x bool qp=%d qp_id = %d, ae_src=%d\n", in irdma_process_aeq()
381 if (rdma_protocol_roce(&iwdev->ibdev, 1)) { in irdma_process_aeq()
1058 static int irdma_alloc_set_mac(struct irdma_device *iwdev) in irdma_alloc_set_mac() argument
1062 status = irdma_alloc_local_mac_entry(iwdev->rf, in irdma_alloc_set_mac()
1063 &iwdev->mac_ip_table_idx); in irdma_alloc_set_mac()
1065 status = irdma_add_local_mac_entry(iwdev->rf, in irdma_alloc_set_mac()
1066 (const u8 *)iwdev->netdev->dev_addr, in irdma_alloc_set_mac()
1067 (u8)iwdev->mac_ip_table_idx); in irdma_alloc_set_mac()
1069 irdma_del_local_mac_entry(iwdev->rf, in irdma_alloc_set_mac()
1070 (u8)iwdev->mac_ip_table_idx); in irdma_alloc_set_mac()
1105 ibdev_dbg(&rf->iwdev->ibdev, "ERR: ceq irq config fail\n"); in irdma_cfg_ceq_vector()
1133 ibdev_dbg(&rf->iwdev->ibdev, "ERR: aeq irq config fail\n"); in irdma_cfg_aeq_vector()
1224 ibdev_dbg(&rf->iwdev->ibdev, "ERR: create ceq status = %d\n", in irdma_setup_ceq_0()
1278 ibdev_dbg(&rf->iwdev->ibdev, in irdma_setup_ceqs()
1438 static int irdma_initialize_ilq(struct irdma_device *iwdev) in irdma_initialize_ilq() argument
1449 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ilq()
1455 status = irdma_puda_create_rsrc(&iwdev->vsi, &info); in irdma_initialize_ilq()
1457 ibdev_dbg(&iwdev->ibdev, "ERR: ilq create fail\n"); in irdma_initialize_ilq()
1468 static int irdma_initialize_ieq(struct irdma_device *iwdev) in irdma_initialize_ieq() argument
1475 info.qp_id = iwdev->vsi.exception_lan_q; in irdma_initialize_ieq()
1479 info.sq_size = min(iwdev->rf->max_qp / 2, (u32)32768); in irdma_initialize_ieq()
1481 info.buf_size = iwdev->vsi.mtu + IRDMA_IPV4_PAD; in irdma_initialize_ieq()
1483 status = irdma_puda_create_rsrc(&iwdev->vsi, &info); in irdma_initialize_ieq()
1485 ibdev_dbg(&iwdev->ibdev, "ERR: ieq create fail\n"); in irdma_initialize_ieq()
1496 struct irdma_device *iwdev = vsi->back_vsi; in irdma_reinitialize_ieq() local
1497 struct irdma_pci_f *rf = iwdev->rf; in irdma_reinitialize_ieq()
1500 if (irdma_initialize_ieq(iwdev)) { in irdma_reinitialize_ieq()
1501 iwdev->rf->reset = true; in irdma_reinitialize_ieq()
1625 void irdma_rt_deinit_hw(struct irdma_device *iwdev) in irdma_rt_deinit_hw() argument
1627 ibdev_dbg(&iwdev->ibdev, "INIT: state = %d\n", iwdev->init_state); in irdma_rt_deinit_hw()
1629 switch (iwdev->init_state) { in irdma_rt_deinit_hw()
1631 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_deinit_hw()
1632 irdma_del_local_mac_entry(iwdev->rf, in irdma_rt_deinit_hw()
1633 (u8)iwdev->mac_ip_table_idx); in irdma_rt_deinit_hw()
1639 if (!iwdev->roce_mode) in irdma_rt_deinit_hw()
1640 irdma_puda_dele_rsrc(&iwdev->vsi, IRDMA_PUDA_RSRC_TYPE_IEQ, in irdma_rt_deinit_hw()
1641 iwdev->rf->reset); in irdma_rt_deinit_hw()
1644 if (!iwdev->roce_mode) in irdma_rt_deinit_hw()
1645 irdma_puda_dele_rsrc(&iwdev->vsi, in irdma_rt_deinit_hw()
1647 iwdev->rf->reset); in irdma_rt_deinit_hw()
1650 ibdev_warn(&iwdev->ibdev, "bad init_state = %d\n", iwdev->init_state); in irdma_rt_deinit_hw()
1654 irdma_cleanup_cm_core(&iwdev->cm_core); in irdma_rt_deinit_hw()
1655 if (iwdev->vsi.pestat) { in irdma_rt_deinit_hw()
1656 irdma_vsi_stats_free(&iwdev->vsi); in irdma_rt_deinit_hw()
1657 kfree(iwdev->vsi.pestat); in irdma_rt_deinit_hw()
1659 if (iwdev->cleanup_wq) in irdma_rt_deinit_hw()
1660 destroy_workqueue(iwdev->cleanup_wq); in irdma_rt_deinit_hw()
1703 static void irdma_get_used_rsrc(struct irdma_device *iwdev) in irdma_get_used_rsrc() argument
1705 iwdev->rf->used_pds = find_first_zero_bit(iwdev->rf->allocated_pds, in irdma_get_used_rsrc()
1706 iwdev->rf->max_pd); in irdma_get_used_rsrc()
1707 iwdev->rf->used_qps = find_first_zero_bit(iwdev->rf->allocated_qps, in irdma_get_used_rsrc()
1708 iwdev->rf->max_qp); in irdma_get_used_rsrc()
1709 iwdev->rf->used_cqs = find_first_zero_bit(iwdev->rf->allocated_cqs, in irdma_get_used_rsrc()
1710 iwdev->rf->max_cq); in irdma_get_used_rsrc()
1711 iwdev->rf->used_mrs = find_first_zero_bit(iwdev->rf->allocated_mrs, in irdma_get_used_rsrc()
1712 iwdev->rf->max_mr); in irdma_get_used_rsrc()
1746 ibdev_warn(&rf->iwdev->ibdev, "bad init_state = %d\n", rf->init_state); in irdma_ctrl_deinit_hw()
1759 int irdma_rt_init_hw(struct irdma_device *iwdev, in irdma_rt_init_hw() argument
1762 struct irdma_pci_f *rf = iwdev->rf; in irdma_rt_init_hw()
1769 vsi_info.back_vsi = iwdev; in irdma_rt_init_hw()
1771 vsi_info.pf_data_vsi_num = iwdev->vsi_num; in irdma_rt_init_hw()
1775 irdma_sc_vsi_init(&iwdev->vsi, &vsi_info); in irdma_rt_init_hw()
1777 status = irdma_setup_cm_core(iwdev, rf->rdma_ver); in irdma_rt_init_hw()
1783 irdma_cleanup_cm_core(&iwdev->cm_core); in irdma_rt_init_hw()
1787 status = irdma_vsi_stats_init(&iwdev->vsi, &stats_info); in irdma_rt_init_hw()
1789 irdma_cleanup_cm_core(&iwdev->cm_core); in irdma_rt_init_hw()
1795 if (!iwdev->roce_mode) { in irdma_rt_init_hw()
1796 status = irdma_initialize_ilq(iwdev); in irdma_rt_init_hw()
1799 iwdev->init_state = ILQ_CREATED; in irdma_rt_init_hw()
1800 status = irdma_initialize_ieq(iwdev); in irdma_rt_init_hw()
1803 iwdev->init_state = IEQ_CREATED; in irdma_rt_init_hw()
1806 status = irdma_setup_ceqs(rf, &iwdev->vsi); in irdma_rt_init_hw()
1810 iwdev->init_state = CEQS_CREATED; in irdma_rt_init_hw()
1819 iwdev->init_state = PBLE_CHUNK_MEM; in irdma_rt_init_hw()
1827 iwdev->init_state = AEQ_CREATED; in irdma_rt_init_hw()
1831 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_init_hw()
1832 irdma_alloc_set_mac(iwdev); in irdma_rt_init_hw()
1833 irdma_add_ip(iwdev); in irdma_rt_init_hw()
1834 iwdev->init_state = IP_ADDR_REGISTERED; in irdma_rt_init_hw()
1839 iwdev->cleanup_wq = alloc_workqueue("irdma-cleanup-wq", in irdma_rt_init_hw()
1841 if (!iwdev->cleanup_wq) in irdma_rt_init_hw()
1843 irdma_get_used_rsrc(iwdev); in irdma_rt_init_hw()
1844 init_waitqueue_head(&iwdev->suspend_wq); in irdma_rt_init_hw()
1850 status, iwdev->init_state); in irdma_rt_init_hw()
1851 irdma_rt_deinit_hw(iwdev); in irdma_rt_init_hw()
2059 ibdev_err(&rf->iwdev->ibdev, "cqp opcode = 0x%x maj_err_code = 0x%x min_err_code = 0x%x\n", in irdma_cqp_ce_handler()
2148 irdma_hw_modify_qp(iwqp->iwdev, iwqp, &info, 0); in irdma_next_iw_state()
2255 static int irdma_cqp_manage_apbvt_cmd(struct irdma_device *iwdev, in irdma_cqp_manage_apbvt_cmd() argument
2263 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, add_port); in irdma_cqp_manage_apbvt_cmd()
2274 cqp_info->in.u.manage_apbvt_entry.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_cqp_manage_apbvt_cmd()
2276 ibdev_dbg(&iwdev->ibdev, "DEV: %s: port=0x%04x\n", in irdma_cqp_manage_apbvt_cmd()
2279 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_cqp_manage_apbvt_cmd()
2280 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_cqp_manage_apbvt_cmd()
2290 struct irdma_apbvt_entry *irdma_add_apbvt(struct irdma_device *iwdev, u16 port) in irdma_add_apbvt() argument
2292 struct irdma_cm_core *cm_core = &iwdev->cm_core; in irdma_add_apbvt()
2314 if (irdma_cqp_manage_apbvt_cmd(iwdev, port, true)) { in irdma_add_apbvt()
2327 void irdma_del_apbvt(struct irdma_device *iwdev, in irdma_del_apbvt() argument
2330 struct irdma_cm_core *cm_core = &iwdev->cm_core; in irdma_del_apbvt()
2344 irdma_cqp_manage_apbvt_cmd(iwdev, entry->port, false); in irdma_del_apbvt()
2419 int irdma_manage_qhash(struct irdma_device *iwdev, struct irdma_cm_info *cminfo, in irdma_manage_qhash() argument
2425 struct irdma_cqp *iwcqp = &iwdev->rf->cqp; in irdma_manage_qhash()
2438 info->vsi = &iwdev->vsi; in irdma_manage_qhash()
2449 ether_addr_copy(info->mac_addr, iwdev->netdev->dev_addr); in irdma_manage_qhash()
2474 ibdev_dbg(&iwdev->ibdev, in irdma_manage_qhash()
2482 ibdev_dbg(&iwdev->ibdev, in irdma_manage_qhash()
2490 cqp_info->in.u.manage_qhash_table_entry.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_manage_qhash()
2494 status = irdma_handle_cqp_op(iwdev->rf, cqp_request); in irdma_manage_qhash()
2532 ibdev_err(&iwqp->iwdev->ibdev, "Flush QP[%d] failed, SQ has more work", in irdma_hw_flush_wqes_callback()
2621 ibdev_err(&iwqp->iwdev->ibdev, "fatal QP event: SQ in error but not flushed, qp: %d", in irdma_hw_flush_wqes()
2637 ibdev_dbg(&rf->iwdev->ibdev, in irdma_hw_flush_wqes()
2682 struct irdma_pci_f *rf = iwqp->iwdev->rf; in irdma_flush_wqes()
2712 queue_delayed_work(iwqp->iwdev->cleanup_wq, in irdma_flush_wqes()