| /drivers/scsi/snic/ |
| A D | snic_isr.c | 80 for (i = 0; i < ARRAY_SIZE(snic->msix); i++) { in snic_free_intr() 81 if (snic->msix[i].requested) { in snic_free_intr() 83 snic->msix[i].devid); in snic_free_intr() 104 sprintf(snic->msix[SNIC_MSIX_WQ].devname, in snic_request_intr() 108 snic->msix[SNIC_MSIX_WQ].devid = snic; in snic_request_intr() 110 sprintf(snic->msix[SNIC_MSIX_IO_CMPL].devname, in snic_request_intr() 114 snic->msix[SNIC_MSIX_IO_CMPL].devid = snic; in snic_request_intr() 124 snic->msix[i].isr, in snic_request_intr() 126 snic->msix[i].devname, in snic_request_intr() 127 snic->msix[i].devid); in snic_request_intr() [all …]
|
| /drivers/scsi/fnic/ |
| A D | fnic_isr.c | 122 i = irq - fnic->msix[0].irq_num; in fnic_isr_msix_wq_copy() 126 if (fnic->msix[i].irq_num == irq) in fnic_isr_msix_wq_copy() 165 if (fnic->msix[i].requested) in fnic_free_intr() 167 fnic->msix[i].devid); in fnic_free_intr() 197 fnic->msix[FNIC_MSIX_RQ].devid = fnic; in fnic_request_intr() 205 sprintf(fnic->msix[i].devname, in fnic_request_intr() 208 fnic->msix[i].devid = fnic; in fnic_request_intr() 221 fnic->msix[i].isr, 0, in fnic_request_intr() 222 fnic->msix[i].devname, in fnic_request_intr() 223 fnic->msix[i].devid); in fnic_request_intr() [all …]
|
| /drivers/scsi/bfa/ |
| A D | bfa_hw_ct.c | 108 bfa->msix.nvecs = nvecs; in bfa_hwct_msix_init() 115 if (bfa->msix.nvecs == 0) in bfa_hwct_msix_ctrl_install() 118 if (bfa->msix.nvecs == 1) in bfa_hwct_msix_ctrl_install() 119 bfa->msix.handler[BFI_MSIX_LPU_ERR_CT] = bfa_msix_all; in bfa_hwct_msix_ctrl_install() 129 if (bfa->msix.nvecs == 0) in bfa_hwct_msix_queue_install() 132 if (bfa->msix.nvecs == 1) { in bfa_hwct_msix_queue_install() 134 bfa->msix.handler[i] = bfa_msix_all; in bfa_hwct_msix_queue_install() 139 bfa->msix.handler[i] = bfa_msix_reqq; in bfa_hwct_msix_queue_install() 142 bfa->msix.handler[i] = bfa_msix_rspq; in bfa_hwct_msix_queue_install() 151 bfa->msix.handler[i] = bfa_hwct_msix_dummy; in bfa_hwct_msix_uninstall() [all …]
|
| A D | bfa_hw_cb.c | 110 bfa->msix.nvecs = nvecs; in bfa_hwcb_msix_init() 119 if (bfa->msix.nvecs == 0) in bfa_hwcb_msix_ctrl_install() 122 if (bfa->msix.nvecs == 1) { in bfa_hwcb_msix_ctrl_install() 124 bfa->msix.handler[i] = bfa_msix_all; in bfa_hwcb_msix_ctrl_install() 129 bfa->msix.handler[i] = bfa_msix_lpu_err; in bfa_hwcb_msix_ctrl_install() 137 if (bfa->msix.nvecs == 0) in bfa_hwcb_msix_queue_install() 140 if (bfa->msix.nvecs == 1) { in bfa_hwcb_msix_queue_install() 142 bfa->msix.handler[i] = bfa_msix_all; in bfa_hwcb_msix_queue_install() 147 bfa->msix.handler[i] = bfa_msix_reqq; in bfa_hwcb_msix_queue_install() 150 bfa->msix.handler[i] = bfa_msix_rspq; in bfa_hwcb_msix_queue_install() [all …]
|
| A D | bfa.h | 175 void (*hw_isr_mode_set)(struct bfa_s *bfa, bfa_boolean_t msix); 298 ((__bfa)->msix.handler[__vec](__bfa, __vec)) 328 void bfa_hwcb_isr_mode_set(struct bfa_s *bfa, bfa_boolean_t msix); 342 void bfa_hwct_isr_mode_set(struct bfa_s *bfa, bfa_boolean_t msix);
|
| /drivers/gpu/drm/xe/ |
| A D | xe_irq.c | 900 XA_LIMIT(*msix, *msix); in xe_irq_msix_alloc_vector() 906 *msix = id; in xe_irq_msix_alloc_vector() 913 xa_erase(&xe->irq.msix.indexes, msix); in xe_irq_msix_release_vector() 958 irq_buf = xa_load(&xe->irq.msix.indexes, msix); in xe_irq_msix_free_irq() 975 u16 msix; in xe_irq_msix_request_irqs() local 977 msix = GUC2HOST_MSIX; in xe_irq_msix_request_irqs() 983 msix = DEFAULT_MSIX; in xe_irq_msix_request_irqs() 996 unsigned long msix; in xe_irq_msix_free() local 999 xa_for_each(&xe->irq.msix.indexes, msix, dummy) in xe_irq_msix_free() 1007 unsigned long msix; in xe_irq_msix_synchronize_irq() local [all …]
|
| A D | xe_irq.h | 23 const char *name, bool dynamic_msix, u16 *msix); 24 void xe_irq_msix_free_irq(struct xe_device *xe, u16 msix);
|
| /drivers/infiniband/hw/hfi1/ |
| A D | affinity.c | 749 cpumask_clear(&msix->mask); in hfi1_update_sdma_affinity() 752 msix->irq, irq_type_names[msix->type], in hfi1_update_sdma_affinity() 754 irq_set_affinity_hint(msix->irq, &msix->mask); in hfi1_update_sdma_affinity() 803 notify->irq = msix->irq; in hfi1_setup_sdma_notifier() 837 cpumask_clear(&msix->mask); in get_irq_affinity() 841 switch (msix->type) { in get_irq_affinity() 889 msix->irq, irq_type_names[msix->type], in get_irq_affinity() 891 irq_set_affinity_hint(msix->irq, &msix->mask); in get_irq_affinity() 893 if (msix->type == IRQ_SDMA) { in get_irq_affinity() 920 switch (msix->type) { in hfi1_put_irq_affinity() [all …]
|
| A D | affinity.h | 44 struct hfi1_msix_entry *msix); 50 struct hfi1_msix_entry *msix);
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_irq.c | 135 if (!pf->msix.min) in ice_init_interrupt_scheme() 136 pf->msix.min = ICE_MIN_MSIX; in ice_init_interrupt_scheme() 138 if (!pf->msix.max) in ice_init_interrupt_scheme() 139 pf->msix.max = min(total_vectors, in ice_init_interrupt_scheme() 142 pf->msix.total = total_vectors; in ice_init_interrupt_scheme() 143 pf->msix.rest = total_vectors - pf->msix.max; in ice_init_interrupt_scheme() 146 vectors = pf->msix.min; in ice_init_interrupt_scheme() 148 vectors = pf->msix.max; in ice_init_interrupt_scheme() 150 vectors = pci_alloc_irq_vectors(pf->pdev, pf->msix.min, vectors, in ice_init_interrupt_scheme() 155 ice_init_irq_tracker(pf, pf->msix.max, vectors); in ice_init_interrupt_scheme() [all …]
|
| /drivers/net/ethernet/chelsio/cxgb4/ |
| A D | cxgb4_tc_mqprio.c | 155 int ret, msix = 0; in cxgb4_mqprio_alloc_hw_resources() local 187 if (msix >= 0) { in cxgb4_mqprio_alloc_hw_resources() 189 if (msix < 0) { in cxgb4_mqprio_alloc_hw_resources() 190 ret = msix; in cxgb4_mqprio_alloc_hw_resources() 194 eorxq->msix = &adap->msix_info[msix]; in cxgb4_mqprio_alloc_hw_resources() 195 snprintf(eorxq->msix->desc, in cxgb4_mqprio_alloc_hw_resources() 196 sizeof(eorxq->msix->desc), in cxgb4_mqprio_alloc_hw_resources() 209 dev, msix, &eorxq->fl, in cxgb4_mqprio_alloc_hw_resources() 248 eorxq->msix->aff_mask); in cxgb4_mqprio_alloc_hw_resources() 260 if (eorxq->msix) in cxgb4_mqprio_alloc_hw_resources() [all …]
|
| A D | cxgb4_main.c | 984 int msix; in setup_non_data_intr() local 992 if (msix < 0) in setup_non_data_intr() 1006 int msix, err = 0; in setup_fw_sge_queues() local 1014 if (msix < 0) in setup_fw_sge_queues() 1065 if (msix >= 0) { in setup_sge_queues() 1068 err = msix; in setup_sge_queues() 1075 q->msix = &adap->msix_info[msix]; in setup_sge_queues() 1346 if (msix >= 0) { in cxgb4_port_mirror_alloc_queues() 1348 if (msix < 0) { in cxgb4_port_mirror_alloc_queues() 1349 ret = msix; in cxgb4_port_mirror_alloc_queues() [all …]
|
| A D | cxgb4_uld.c | 144 q->msix = &adap->msix_info[msi_idx]; in alloc_uld_rxqs() 167 if (q->msix) in alloc_uld_rxqs() 168 cxgb4_free_msix_idx_in_bmap(adap, q->msix->idx); in alloc_uld_rxqs() 332 minfo = rxq_info->uldrxq[idx].msix; in request_msix_queue_irqs_uld() 347 minfo = rxq_info->uldrxq[idx].msix; in request_msix_queue_irqs_uld() 363 minfo = rxq_info->uldrxq[idx].msix; in free_msix_queue_irqs_uld()
|
| /drivers/vfio/pci/ |
| A D | vfio_pci_intrs.c | 398 if (!msix) { in vfio_msi_enable() 420 unsigned int vector, bool msix) in vfio_msi_alloc_irq() argument 430 if (irq > 0 || !msix || !vdev->has_dyn_msix) in vfio_msi_alloc_irq() 441 unsigned int vector, int fd, bool msix) in vfio_msi_set_vector_signal() argument 468 irq = vfio_msi_alloc_irq(vdev, vector, msix); in vfio_msi_set_vector_signal() 496 if (msix) { in vfio_msi_set_vector_signal() 528 unsigned count, int32_t *fds, bool msix) in vfio_msi_set_block() argument 556 vfio_msi_set_vector_signal(vdev, i, -1, msix); in vfio_msi_disable() 683 vfio_msi_disable(vdev, msix); in vfio_pci_set_msi_trigger() 696 fds, msix); in vfio_pci_set_msi_trigger() [all …]
|
| /drivers/net/ethernet/emulex/benet/ |
| A D | be_roce.c | 65 dev_info.msix.num_vectors = min(num_vec, MAX_MSIX_VECTORS); in _be_roce_dev_add() 70 dev_info.msix.start_vector = adapter->num_evt_qs; in _be_roce_dev_add() 71 for (i = 0; i < dev_info.msix.num_vectors; i++) { in _be_roce_dev_add() 72 dev_info.msix.vector_list[i] = in _be_roce_dev_add() 76 dev_info.msix.num_vectors = 0; in _be_roce_dev_add()
|
| /drivers/rapidio/devices/ |
| A D | tsi721.c | 803 priv->msix[TSI721_VECT_IMB0_RCV + i].vector = in tsi721_enable_msix() 809 priv->msix[TSI721_VECT_IMB0_INT + i].vector = in tsi721_enable_msix() 815 priv->msix[TSI721_VECT_OMB0_DONE + i].vector = in tsi721_enable_msix() 821 priv->msix[TSI721_VECT_OMB0_INT + i].vector = in tsi721_enable_msix() 830 priv->msix[TSI721_VECT_DMA0_DONE + i].vector = in tsi721_enable_msix() 836 priv->msix[TSI721_VECT_DMA0_INT + i].vector = in tsi721_enable_msix() 1997 priv->msix[idx].irq_name, (void *)priv); in tsi721_open_outb_mbox() 2008 priv->msix[idx].irq_name, (void *)priv); in tsi721_open_outb_mbox() 2014 free_irq(priv->msix[idx].vector, (void *)priv); in tsi721_open_outb_mbox() 2314 priv->msix[idx].irq_name, (void *)priv); in tsi721_open_inb_mbox() [all …]
|
| A D | tsi721_dma.c | 152 rc = request_irq(priv->msix[idx].vector, tsi721_bdma_msix, 0, in tsi721_bdma_ch_init() 153 priv->msix[idx].irq_name, (void *)bdma_chan); in tsi721_bdma_ch_init() 164 rc = request_irq(priv->msix[idx].vector, tsi721_bdma_msix, 0, in tsi721_bdma_ch_init() 165 priv->msix[idx].irq_name, (void *)bdma_chan); in tsi721_bdma_ch_init() 172 priv->msix[TSI721_VECT_DMA0_DONE + in tsi721_bdma_ch_init() 226 free_irq(priv->msix[TSI721_VECT_DMA0_DONE + in tsi721_bdma_ch_free() 228 free_irq(priv->msix[TSI721_VECT_DMA0_INT + in tsi721_bdma_ch_free() 772 synchronize_irq(priv->msix[TSI721_VECT_DMA0_DONE + in tsi721_sync_dma_irq() 774 synchronize_irq(priv->msix[TSI721_VECT_DMA0_INT + in tsi721_sync_dma_irq()
|
| /drivers/net/ethernet/cisco/enic/ |
| A D | enic_main.c | 125 enic->msix[i].affinity_mask); in enic_init_affinity_hint() 1536 if (enic->msix[i].requested) in enic_free_intr() 1538 enic->msix[i].devid); in enic_free_intr() 1603 enic->msix[i].requested = 0; in enic_request_intr() 1607 enic->msix[i].isr, 0, in enic_request_intr() 1608 enic->msix[i].devname, in enic_request_intr() 1609 enic->msix[i].devid); in enic_request_intr() 1614 enic->msix[i].requested = 1; in enic_request_intr() 2450 kfree(enic->msix); in enic_free_enic_resources() 2451 enic->msix = NULL; in enic_free_enic_resources() [all …]
|
| /drivers/edac/ |
| A D | thunderx_edac.c | 1071 struct msix_entry *msix = irq_id; in thunderx_ocx_com_isr() local 1073 msix_ent[msix->entry]); in thunderx_ocx_com_isr() 1100 struct msix_entry *msix = irq_id; in thunderx_ocx_com_threaded_isr() local 1102 msix_ent[msix->entry]); in thunderx_ocx_com_threaded_isr() 1164 struct msix_entry *msix = irq_id; in thunderx_ocx_lnk_isr() local 1166 msix_ent[msix->entry]); in thunderx_ocx_lnk_isr() 1171 ctx->link = msix->entry; in thunderx_ocx_lnk_isr() 1183 struct msix_entry *msix = irq_id; in thunderx_ocx_lnk_threaded_isr() local 1185 msix_ent[msix->entry]); in thunderx_ocx_lnk_threaded_isr() 1755 struct msix_entry *msix = irq_id; in thunderx_l2c_tad_isr() local [all …]
|
| /drivers/ntb/hw/amd/ |
| A D | ntb_hw_amd.c | 727 ndev->msix = kcalloc_node(msix_max, sizeof(*ndev->msix), in ndev_init_isr() 729 if (!ndev->msix) in ndev_init_isr() 733 ndev->msix[i].entry = i; in ndev_init_isr() 735 msix_count = pci_enable_msix_range(pdev, ndev->msix, in ndev_init_isr() 751 rc = request_irq(ndev->msix[i].vector, ndev_vec_isr, 0, in ndev_init_isr() 764 free_irq(ndev->msix[i].vector, &ndev->vec[i]); in ndev_init_isr() 767 kfree(ndev->msix); in ndev_init_isr() 771 ndev->msix = NULL; in ndev_init_isr() 822 if (ndev->msix) { in ndev_deinit_isr() 825 free_irq(ndev->msix[i].vector, &ndev->vec[i]); in ndev_deinit_isr() [all …]
|
| /drivers/scsi/qla2xxx/ |
| A D | qla_mid.c | 598 if (rsp->msix && rsp->msix->have_irq) { in qla25xx_free_rsp_que() 599 free_irq(rsp->msix->vector, rsp->msix->handle); in qla25xx_free_rsp_que() 600 rsp->msix->have_irq = 0; in qla25xx_free_rsp_que() 601 rsp->msix->in_use = 0; in qla25xx_free_rsp_que() 602 rsp->msix->handle = NULL; in qla25xx_free_rsp_que() 864 rsp->msix = qpair->msix; in qla25xx_create_rsp_que() 902 ret = qla25xx_request_irq(ha, qpair, qpair->msix, in qla25xx_create_rsp_que()
|
| /drivers/misc/keba/ |
| A D | cp500.c | 81 unsigned int msix; member 753 cp500->devs->uart0_rfb.msix < cp500->msix_num) { in cp500_register_auxiliary_devs() 755 cp500->devs->uart0_rfb.msix); in cp500_register_auxiliary_devs() 762 cp500->devs->uart1_dbg.msix < cp500->msix_num) { in cp500_register_auxiliary_devs() 764 cp500->devs->uart1_dbg.msix); in cp500_register_auxiliary_devs() 771 cp500->devs->uart2_si1.msix < cp500->msix_num) { in cp500_register_auxiliary_devs() 773 cp500->devs->uart2_si1.msix); in cp500_register_auxiliary_devs()
|
| /drivers/irqchip/ |
| A D | irq-gic-its-msi-parent.c | 42 int msi, msix, *count = data; in its_pci_msi_vec_count() local 45 msix = max(pci_msix_vec_count(pdev), 0); in its_pci_msi_vec_count() 46 *count += max(msi, msix); in its_pci_msi_vec_count()
|
| /drivers/crypto/cavium/nitrox/ |
| A D | nitrox_isr.c | 395 vec = ndev->iov.msix.vector; in nitrox_sriov_unregister_interrupts() 419 ndev->iov.msix.entry = NON_RING_MSIX_BASE; in nitrox_sriov_register_interupts() 420 ret = pci_enable_msix_exact(pdev, &ndev->iov.msix, NR_NON_RING_VECTORS); in nitrox_sriov_register_interupts() 439 vec = ndev->iov.msix.vector; in nitrox_sriov_register_interupts()
|
| /drivers/ntb/hw/intel/ |
| A D | ntb_hw_gen1.c | 387 ndev->msix = kcalloc_node(msix_max, sizeof(*ndev->msix), in ndev_init_isr() 389 if (!ndev->msix) in ndev_init_isr() 393 ndev->msix[i].entry = i; in ndev_init_isr() 395 msix_count = pci_enable_msix_range(pdev, ndev->msix, in ndev_init_isr() 403 rc = request_irq(ndev->msix[i].vector, ndev_vec_isr, 0, in ndev_init_isr() 416 free_irq(ndev->msix[i].vector, &ndev->vec[i]); in ndev_init_isr() 419 kfree(ndev->msix); in ndev_init_isr() 423 ndev->msix = NULL; in ndev_init_isr() 477 if (ndev->msix) { in ndev_deinit_isr() 480 free_irq(ndev->msix[i].vector, &ndev->vec[i]); in ndev_deinit_isr() [all …]
|