| /drivers/dma/ |
| A D | dmaengine.c | 330 if (!dma_has_cap(cap, device->cap_mask) || in min_chan() 331 dma_has_cap(DMA_PRIVATE, device->cap_mask)) in min_chan() 376 if (dma_has_cap(DMA_PRIVATE, device->cap_mask)) in dma_channel_rebalance() 563 if (dma_has_cap(DMA_PRIVATE, device->cap_mask)) in dma_issue_pending_all() 671 dma_cap_set(DMA_PRIVATE, device->cap_mask); in find_candidate() 708 dma_cap_set(DMA_PRIVATE, device->cap_mask); in dma_get_slave_channel() 1032 if (!dma_has_cap(DMA_MEMCPY, device->cap_mask)) in device_has_all_tx_types() 1037 if (!dma_has_cap(DMA_XOR, device->cap_mask)) in device_has_all_tx_types() 1047 if (!dma_has_cap(DMA_PQ, device->cap_mask)) in device_has_all_tx_types() 1230 dma_cap_set(DMA_ASYNC_TX, device->cap_mask); in dma_async_device_register() [all …]
|
| A D | mv_xor.c | 1036 int idx, dma_cap_mask_t cap_mask, int irq) in mv_xor_channel_add() argument 1088 dma_dev->cap_mask = cap_mask; in mv_xor_channel_add() 1103 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in mv_xor_channel_add() 1144 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in mv_xor_channel_add() 1393 dma_cap_mask_t cap_mask; in mv_xor_probe() local 1399 dma_cap_zero(cap_mask); in mv_xor_probe() 1400 dma_cap_set(DMA_MEMCPY, cap_mask); in mv_xor_probe() 1401 dma_cap_set(DMA_XOR, cap_mask); in mv_xor_probe() 1402 dma_cap_set(DMA_INTERRUPT, cap_mask); in mv_xor_probe() 1411 cap_mask, irq); in mv_xor_probe() [all …]
|
| A D | at_hdmac.c | 518 dma_cap_mask_t cap_mask; member 1947 dma_cap_set(DMA_MEMCPY, at91sam9rl_config.cap_mask); in at_dma_probe() 1948 dma_cap_set(DMA_INTERLEAVE, at91sam9g45_config.cap_mask); in at_dma_probe() 1949 dma_cap_set(DMA_MEMCPY, at91sam9g45_config.cap_mask); in at_dma_probe() 1950 dma_cap_set(DMA_MEMSET, at91sam9g45_config.cap_mask); in at_dma_probe() 1951 dma_cap_set(DMA_MEMSET_SG, at91sam9g45_config.cap_mask); in at_dma_probe() 1952 dma_cap_set(DMA_PRIVATE, at91sam9g45_config.cap_mask); in at_dma_probe() 1953 dma_cap_set(DMA_SLAVE, at91sam9g45_config.cap_mask); in at_dma_probe() 1975 atdma->dma_device.cap_mask = plat_dat->cap_mask; in at_dma_probe() 2046 if (dma_has_cap(DMA_MEMCPY, atdma->dma_device.cap_mask)) in at_dma_probe() [all …]
|
| A D | bcm-sba-raid.c | 1553 dma_cap_zero(dma_dev->cap_mask); in sba_async_register() 1554 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in sba_async_register() 1555 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in sba_async_register() 1556 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in sba_async_register() 1557 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in sba_async_register() 1573 if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask)) in sba_async_register() 1577 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) in sba_async_register() 1581 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in sba_async_register() 1587 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in sba_async_register() 1607 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "", in sba_async_register() [all …]
|
| A D | mcf-edma-main.c | 213 dma_cap_set(DMA_PRIVATE, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe() 214 dma_cap_set(DMA_SLAVE, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe() 215 dma_cap_set(DMA_CYCLIC, mcf_edma->dma_dev.cap_mask); in mcf_edma_probe()
|
| A D | xgene-dma.c | 1489 dma_cap_zero(dma_dev->cap_mask); in xgene_dma_set_caps() 1504 dma_cap_set(DMA_PQ, dma_dev->cap_mask); in xgene_dma_set_caps() 1505 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps() 1508 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in xgene_dma_set_caps() 1518 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in xgene_dma_set_caps() 1524 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in xgene_dma_set_caps() 1568 dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "XOR " : "", in xgene_dma_async_register() 1569 dma_has_cap(DMA_PQ, dma_dev->cap_mask) ? "PQ " : ""); in xgene_dma_async_register()
|
| A D | mv_xor_v2.c | 814 dma_cap_zero(dma_dev->cap_mask); in mv_xor_v2_probe() 815 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in mv_xor_v2_probe() 816 dma_cap_set(DMA_XOR, dma_dev->cap_mask); in mv_xor_v2_probe() 817 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in mv_xor_v2_probe()
|
| A D | ep93xx_dma.c | 1449 dma_cap_mask_t mask = edma->dma_dev.cap_mask; in ep93xx_m2p_dma_of_xlate() 1480 dma_cap_mask_t mask = edma->dma_dev.cap_mask; in ep93xx_m2m_dma_of_xlate() 1514 dma_cap_zero(dma_dev->cap_mask); in ep93xx_dma_probe() 1515 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in ep93xx_dma_probe() 1516 dma_cap_set(DMA_CYCLIC, dma_dev->cap_mask); in ep93xx_dma_probe() 1532 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ep93xx_dma_probe() 1540 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ep93xx_dma_probe()
|
| A D | dmatest.c | 858 dev->cap_mask) && in dmatest_func() 1037 if (dma_has_cap(DMA_COMPLETION_NO_ORDER, dma_dev->cap_mask) && in dmatest_add_channel() 1043 if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) { in dmatest_add_channel() 1050 if (dma_has_cap(DMA_MEMSET, dma_dev->cap_mask)) { in dmatest_add_channel() 1057 if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) { in dmatest_add_channel() 1061 if (dma_has_cap(DMA_PQ, dma_dev->cap_mask)) { in dmatest_add_channel()
|
| A D | img-mdc-dma.c | 912 dma_cap_zero(mdma->dma_dev.cap_mask); in mdc_dma_probe() 913 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 914 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 915 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); in mdc_dma_probe() 916 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); in mdc_dma_probe()
|
| A D | moxart-dma.c | 580 dma_cap_zero(mdc->dma_slave.cap_mask); in moxart_probe() 581 dma_cap_set(DMA_SLAVE, mdc->dma_slave.cap_mask); in moxart_probe() 582 dma_cap_set(DMA_PRIVATE, mdc->dma_slave.cap_mask); in moxart_probe()
|
| A D | loongson2-apb-dma.c | 633 dma_cap_zero(ddev->cap_mask); in ls2x_dma_probe() 634 dma_cap_set(DMA_SLAVE, ddev->cap_mask); in ls2x_dma_probe() 635 dma_cap_set(DMA_CYCLIC, ddev->cap_mask); in ls2x_dma_probe()
|
| A D | mmp_tdma.c | 613 dma_cap_mask_t mask = tdev->device.cap_mask; in mmp_tdma_xlate() 686 dma_cap_set(DMA_SLAVE, tdev->device.cap_mask); in mmp_tdma_probe() 687 dma_cap_set(DMA_CYCLIC, tdev->device.cap_mask); in mmp_tdma_probe()
|
| A D | uniphier-xdmac.c | 507 dma_cap_zero(ddev->cap_mask); in uniphier_xdmac_probe() 508 dma_cap_set(DMA_MEMCPY, ddev->cap_mask); in uniphier_xdmac_probe() 509 dma_cap_set(DMA_SLAVE, ddev->cap_mask); in uniphier_xdmac_probe()
|
| A D | ste_dma40.c | 1983 dma_cap_mask_t cap = d40c->chan.device->cap_mask; in d40_config_memcpy() 2837 if (dma_has_cap(DMA_SLAVE, dev->cap_mask)) { in d40_ops_init() 2842 if (dma_has_cap(DMA_MEMCPY, dev->cap_mask)) { in d40_ops_init() 2852 if (dma_has_cap(DMA_CYCLIC, dev->cap_mask)) in d40_ops_init() 2875 dma_cap_zero(base->dma_slave.cap_mask); in d40_dmaengine_init() 2876 dma_cap_set(DMA_SLAVE, base->dma_slave.cap_mask); in d40_dmaengine_init() 2877 dma_cap_set(DMA_CYCLIC, base->dma_slave.cap_mask); in d40_dmaengine_init() 2891 dma_cap_zero(base->dma_memcpy.cap_mask); in d40_dmaengine_init() 2907 dma_cap_zero(base->dma_both.cap_mask); in d40_dmaengine_init() 2908 dma_cap_set(DMA_SLAVE, base->dma_both.cap_mask); in d40_dmaengine_init() [all …]
|
| A D | bcm2835-dma.c | 931 dma_cap_set(DMA_SLAVE, od->ddev.cap_mask); in bcm2835_dma_probe() 932 dma_cap_set(DMA_PRIVATE, od->ddev.cap_mask); in bcm2835_dma_probe() 933 dma_cap_set(DMA_CYCLIC, od->ddev.cap_mask); in bcm2835_dma_probe() 934 dma_cap_set(DMA_MEMCPY, od->ddev.cap_mask); in bcm2835_dma_probe()
|
| /drivers/dma/ioat/ |
| A D | sysfs.c | 22 dma_has_cap(DMA_PQ, dma->cap_mask) ? " pq" : "", in cap_show() 23 dma_has_cap(DMA_PQ_VAL, dma->cap_mask) ? " pq_val" : "", in cap_show() 24 dma_has_cap(DMA_XOR, dma->cap_mask) ? " xor" : "", in cap_show() 25 dma_has_cap(DMA_XOR_VAL, dma->cap_mask) ? " xor_val" : "", in cap_show() 26 dma_has_cap(DMA_INTERRUPT, dma->cap_mask) ? " intr" : ""); in cap_show()
|
| A D | init.c | 512 dma_cap_set(DMA_MEMCPY, dma->cap_mask); in ioat_probe() 795 if (!dma_has_cap(DMA_XOR, dma->cap_mask)) in ioat_xor_val_self_test() 903 if (!dma_has_cap(DMA_XOR_VAL, dma_chan->device->cap_mask)) in ioat_xor_val_self_test() 1092 dma_cap_set(DMA_INTERRUPT, dma->cap_mask); in ioat3_dma_probe() 1108 dma_cap_set(DMA_XOR, dma->cap_mask); in ioat3_dma_probe() 1111 dma_cap_set(DMA_XOR_VAL, dma->cap_mask); in ioat3_dma_probe() 1119 dma_cap_set(DMA_PQ, dma->cap_mask); in ioat3_dma_probe() 1120 dma_cap_set(DMA_PQ_VAL, dma->cap_mask); in ioat3_dma_probe() 1130 dma_cap_set(DMA_XOR, dma->cap_mask); in ioat3_dma_probe() 1131 dma_cap_set(DMA_XOR_VAL, dma->cap_mask); in ioat3_dma_probe() [all …]
|
| /drivers/infiniband/hw/hfi1/ |
| A D | driver.c | 73 cap_mask = *cap_mask_ptr, value, diff, in hfi1_caps_set() local 83 diff = value ^ (cap_mask & ~HFI1_CAP_LOCKED_SMASK); in hfi1_caps_set() 95 cap_mask &= ~diff; in hfi1_caps_set() 97 cap_mask |= (value & diff); in hfi1_caps_set() 100 ((cap_mask & HFI1_CAP_MUST_HAVE_KERN) << HFI1_CAP_USER_SHIFT); in hfi1_caps_set() 101 cap_mask &= ~diff; in hfi1_caps_set() 103 *cap_mask_ptr = cap_mask; in hfi1_caps_set() 110 unsigned long cap_mask = *(unsigned long *)kp->arg; in hfi1_caps_get() local 112 cap_mask &= ~HFI1_CAP_LOCKED_SMASK; in hfi1_caps_get() 113 cap_mask |= ((cap_mask & HFI1_CAP_K2U) << HFI1_CAP_USER_SHIFT); in hfi1_caps_get() [all …]
|
| /drivers/dma/idxd/ |
| A D | dma.c | 212 dma_cap_set(DMA_INTERRUPT, dma->cap_mask); in idxd_register_dma_device() 213 dma_cap_set(DMA_PRIVATE, dma->cap_mask); in idxd_register_dma_device() 214 dma_cap_set(DMA_COMPLETION_NO_ORDER, dma->cap_mask); in idxd_register_dma_device() 219 dma_cap_set(DMA_MEMCPY, dma->cap_mask); in idxd_register_dma_device()
|
| /drivers/net/ethernet/qlogic/qlcnic/ |
| A D | qlcnic_minidump.c | 295 hdr->drv_cap_mask = hdr->cap_mask; in qlcnic_82xx_cache_tmpl_hdr_values() 296 fw_dump->cap_mask = hdr->cap_mask; in qlcnic_82xx_cache_tmpl_hdr_values() 349 hdr->drv_cap_mask = hdr->cap_mask; in qlcnic_83xx_cache_tmpl_hdr_values() 350 fw_dump->cap_mask = hdr->cap_mask; in qlcnic_83xx_cache_tmpl_hdr_values() 1278 fw_dump->cap_mask); in qlcnic_fw_cmd_get_minidump_temp() 1322 if (i & fw_dump->cap_mask) in qlcnic_dump_fw() 1353 if (!(entry->hdr.mask & fw_dump->cap_mask)) { in qlcnic_dump_fw() 1450 fw_dump->cap_mask = 0x1f; in qlcnic_83xx_get_minidump_template()
|
| /drivers/dma/ppc4xx/ |
| A D | adma.c | 3772 dma_cap_set(DMA_MEMCPY, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3773 dma_cap_set(DMA_INTERRUPT, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3774 dma_cap_set(DMA_PQ, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3775 dma_cap_set(DMA_PQ_VAL, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3776 dma_cap_set(DMA_XOR_VAL, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3779 dma_cap_set(DMA_XOR, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3780 dma_cap_set(DMA_PQ, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3781 dma_cap_set(DMA_INTERRUPT, adev->common.cap_mask); in ppc440spe_adma_init_capabilities() 3782 adev->common.cap_mask = adev->common.cap_mask; in ppc440spe_adma_init_capabilities() 3799 if (dma_has_cap(DMA_XOR, adev->common.cap_mask)) { in ppc440spe_adma_init_capabilities() [all …]
|
| /drivers/dma/xilinx/ |
| A D | xdma.c | 1257 dma_cap_set(DMA_SLAVE, xdev->dma_dev.cap_mask); in xdma_probe() 1258 dma_cap_set(DMA_PRIVATE, xdev->dma_dev.cap_mask); in xdma_probe() 1259 dma_cap_set(DMA_CYCLIC, xdev->dma_dev.cap_mask); in xdma_probe() 1260 dma_cap_set(DMA_INTERLEAVE, xdev->dma_dev.cap_mask); in xdma_probe() 1261 dma_cap_set(DMA_REPEAT, xdev->dma_dev.cap_mask); in xdma_probe() 1262 dma_cap_set(DMA_LOAD_EOT, xdev->dma_dev.cap_mask); in xdma_probe()
|
| A D | xilinx_dpdma.c | 1768 dma_cap_set(DMA_SLAVE, ddev->cap_mask); in xilinx_dpdma_probe() 1769 dma_cap_set(DMA_PRIVATE, ddev->cap_mask); in xilinx_dpdma_probe() 1770 dma_cap_set(DMA_CYCLIC, ddev->cap_mask); in xilinx_dpdma_probe() 1771 dma_cap_set(DMA_INTERLEAVE, ddev->cap_mask); in xilinx_dpdma_probe() 1772 dma_cap_set(DMA_REPEAT, ddev->cap_mask); in xilinx_dpdma_probe() 1773 dma_cap_set(DMA_LOAD_EOT, ddev->cap_mask); in xilinx_dpdma_probe()
|
| /drivers/dma/dw-edma/ |
| A D | dw-edma-core.c | 775 dma_cap_zero(dma->cap_mask); in dw_edma_channel_setup() 776 dma_cap_set(DMA_SLAVE, dma->cap_mask); in dw_edma_channel_setup() 777 dma_cap_set(DMA_CYCLIC, dma->cap_mask); in dw_edma_channel_setup() 778 dma_cap_set(DMA_PRIVATE, dma->cap_mask); in dw_edma_channel_setup() 779 dma_cap_set(DMA_INTERLEAVE, dma->cap_mask); in dw_edma_channel_setup()
|