| /drivers/gpu/host1x/ |
| A D | job.c | 116 struct host1x_job_gather *gather = &job->cmds[job->num_cmds].gather; in host1x_job_add_gather() local 118 gather->words = words; in host1x_job_add_gather() 119 gather->bo = bo; in host1x_job_add_gather() 120 gather->offset = offset; in host1x_job_add_gather() 223 g = &job->cmds[i].gather; in pin_job() 544 g = &job->cmds[i].gather; in copy_gathers() 568 void *gather; in copy_gathers() local 572 g = &job->cmds[i].gather; in copy_gathers() 575 gather = host1x_bo_mmap(g->bo); in copy_gathers() 578 host1x_bo_munmap(g->bo, gather); in copy_gathers() [all …]
|
| A D | job.h | 32 struct host1x_job_gather gather; member
|
| /drivers/net/ethernet/cavium/liquidio/ |
| A D | octeon_nic.h | 95 u32 gather:1; member 150 if (!setup->s.gather) { in octnet_prepare_pci_cmd_o2() 153 ih2->gather = 1; in octnet_prepare_pci_cmd_o2() 195 if (!setup->s.gather) { in octnet_prepare_pci_cmd_o3() 198 ih3->gather = 1; in octnet_prepare_pci_cmd_o3()
|
| A D | liquidio_common.h | 357 u64 gather:1; member 391 u64 gather:1; 488 u64 gather:1; member 533 u64 gather:1;
|
| /drivers/gpu/drm/tegra/ |
| A D | submit.c | 70 struct gather_bo *gather = container_of(bo, struct gather_bo, base); in gather_bo_pin() local 89 err = dma_get_sgtable(gather->dev, map->sgt, gather->gather_data, gather->gather_data_dma, in gather_bo_pin() 90 gather->gather_data_words * 4); in gather_bo_pin() 99 map->size = gather->gather_data_words * 4; in gather_bo_pin()
|
| /drivers/net/ethernet/marvell/octeon_ep_vf/ |
| A D | octep_vf_tx.h | 57 u8 gather; member 194 u64 gather:1; member
|
| A D | octep_vf_tx.c | 59 if (!tx_buffer->gather) { in octep_vf_iq_process_completions() 118 if (!tx_buffer->gather) { in octep_vf_iq_free_pending()
|
| A D | octep_vf_main.c | 637 tx_buffer->gather = 0; in octep_vf_start_xmit() 651 ih->gather = 1; in octep_vf_start_xmit() 652 tx_buffer->gather = 1; in octep_vf_start_xmit() 725 tx_buffer->gather = 0; in octep_vf_start_xmit()
|
| /drivers/iommu/ |
| A D | io-pgtable-arm-v7s.c | 558 struct iommu_iotlb_gather *gather, in __arm_v7s_unmap() argument 606 } else if (!iommu_iotlb_gather_queued(gather)) { in __arm_v7s_unmap() 607 io_pgtable_tlb_add_page(iop, gather, iova, blk_size); in __arm_v7s_unmap() 619 return __arm_v7s_unmap(data, gather, iova, size, lvl + 1, ptep); in __arm_v7s_unmap() 624 struct iommu_iotlb_gather *gather) in arm_v7s_unmap_pages() argument 633 ret = __arm_v7s_unmap(data, gather, iova, pgsize, 1, data->pgd); in arm_v7s_unmap_pages() 792 static void __init dummy_tlb_add_page(struct iommu_iotlb_gather *gather, in dummy_tlb_add_page() argument
|
| A D | s390-iommu.c | 790 struct iommu_iotlb_gather *gather) in s390_iommu_iotlb_sync() argument 793 size_t size = gather->end - gather->start + 1; in s390_iommu_iotlb_sync() 797 if (!gather->end) in s390_iommu_iotlb_sync() 803 zpci_refresh_trans((u64)zdev->fh << 32, gather->start, in s390_iommu_iotlb_sync() 1009 struct iommu_iotlb_gather *gather) in s390_iommu_unmap_pages() argument 1023 iommu_iotlb_gather_add_range(gather, iova, size); in s390_iommu_unmap_pages()
|
| A D | io-pgtable-arm.c | 335 struct iommu_iotlb_gather *gather, 625 struct iommu_iotlb_gather *gather, in __arm_lpae_unmap() argument 671 if (gather && !iommu_iotlb_gather_queued(gather)) in __arm_lpae_unmap() 673 io_pgtable_tlb_add_page(iop, gather, iova + j * size, size); in __arm_lpae_unmap() 683 return __arm_lpae_unmap(data, gather, iova, size, pgcount, lvl + 1, ptep); in __arm_lpae_unmap() 688 struct iommu_iotlb_gather *gather) in arm_lpae_unmap_pages() argument 703 return __arm_lpae_unmap(data, gather, iova, pgsize, pgcount, in arm_lpae_unmap_pages() 1287 static void __init dummy_tlb_add_page(struct iommu_iotlb_gather *gather, in dummy_tlb_add_page() argument
|
| A D | io-pgtable-dart.c | 286 struct iommu_iotlb_gather *gather) in dart_unmap_pages() argument 316 if (!iommu_iotlb_gather_queued(gather)) in dart_unmap_pages() 317 io_pgtable_tlb_add_page(&data->iop, gather, in dart_unmap_pages()
|
| A D | mtk_iommu.c | 813 struct iommu_iotlb_gather *gather) in mtk_iommu_unmap() argument 817 iommu_iotlb_gather_add_range(gather, iova, pgsize * pgcount); in mtk_iommu_unmap() 818 return dom->iop->unmap_pages(dom->iop, iova, pgsize, pgcount, gather); in mtk_iommu_unmap() 830 struct iommu_iotlb_gather *gather) in mtk_iommu_iotlb_sync() argument 833 size_t length = gather->end - gather->start + 1; in mtk_iommu_iotlb_sync() 835 mtk_iommu_tlb_flush_range_sync(gather->start, length, dom->bank); in mtk_iommu_iotlb_sync()
|
| A D | msm_iommu.c | 176 static void __flush_iotlb_page(struct iommu_iotlb_gather *gather, in __flush_iotlb_page() argument 512 struct iommu_iotlb_gather *gather) in msm_iommu_unmap() argument 519 ret = priv->iop->unmap_pages(priv->iop, iova, pgsize, pgcount, gather); in msm_iommu_unmap()
|
| A D | ipmmu-vmsa.c | 681 struct iommu_iotlb_gather *gather) in ipmmu_unmap() argument 685 return domain->iop->unmap_pages(domain->iop, iova, pgsize, pgcount, gather); in ipmmu_unmap() 697 struct iommu_iotlb_gather *gather) in ipmmu_iotlb_sync() argument
|
| /drivers/net/ethernet/marvell/octeon_ep/ |
| A D | octep_tx.h | 57 u8 gather; member 236 u64 gather:1; member
|
| A D | octep_tx.c | 58 if (!tx_buffer->gather) { in octep_iq_process_completions() 119 if (!tx_buffer->gather) { in octep_iq_free_pending()
|
| /drivers/net/ethernet/cavium/thunder/ |
| A D | nicvf_queues.c | 558 struct sq_gather_subdesc *gather; in nicvf_unmap_sndq_buffers() local 564 gather = (struct sq_gather_subdesc *)GET_SQ_DESC(sq, hdr_sqe); in nicvf_unmap_sndq_buffers() 566 dma_unmap_page_attrs(&nic->pdev->dev, gather->addr, in nicvf_unmap_sndq_buffers() 567 gather->size, DMA_TO_DEVICE, in nicvf_unmap_sndq_buffers() 1420 struct sq_gather_subdesc *gather; in nicvf_sq_add_gather_subdesc() local 1423 gather = (struct sq_gather_subdesc *)GET_SQ_DESC(sq, qentry); in nicvf_sq_add_gather_subdesc() 1425 memset(gather, 0, SND_QUEUE_DESC_SIZE); in nicvf_sq_add_gather_subdesc() 1426 gather->subdesc_type = SQ_DESC_TYPE_GATHER; in nicvf_sq_add_gather_subdesc() 1427 gather->ld_type = NIC_SEND_LD_TYPE_E_LDD; in nicvf_sq_add_gather_subdesc() 1428 gather->size = size; in nicvf_sq_add_gather_subdesc() [all …]
|
| /drivers/message/fusion/ |
| A D | Kconfig | 65 int "Maximum number of scatter gather entries (16 - 128)" 70 gather entries per I/O. The driver default is 128, which matches
|
| /drivers/iommu/arm/arm-smmu/ |
| A D | qcom_iommu.c | 176 static void qcom_iommu_tlb_add_page(struct iommu_iotlb_gather *gather, in qcom_iommu_tlb_add_page() argument 448 struct iommu_iotlb_gather *gather) in qcom_iommu_unmap() argument 465 ret = ops->unmap_pages(ops, iova, pgsize, pgcount, gather); in qcom_iommu_unmap() 486 struct iommu_iotlb_gather *gather) in qcom_iommu_iotlb_sync() argument
|
| /drivers/scsi/mpt3sas/ |
| A D | Kconfig | 59 gather entries per I/O. The driver default is 128, which matches 71 gather entries per I/O. The driver default is 128, which matches
|
| /drivers/iommu/iommufd/ |
| A D | io_pagetable.c | 534 struct iommu_iotlb_gather gather; in iommu_read_and_clear_dirty() local 549 iommu_dirty_bitmap_init(&dirty, iter, &gather); in iommu_read_and_clear_dirty() 558 iommu_iotlb_sync(domain, &gather); in iommu_read_and_clear_dirty() 613 struct iommu_iotlb_gather gather; in iopt_clear_dirty_data() local 620 iommu_dirty_bitmap_init(&dirty, NULL, &gather); in iopt_clear_dirty_data() 634 iommu_iotlb_sync(domain, &gather); in iopt_clear_dirty_data()
|
| /drivers/iommu/riscv/ |
| A D | iommu.c | 1074 struct iommu_iotlb_gather *gather) in riscv_iommu_iotlb_sync() argument 1078 riscv_iommu_iotlb_inval(domain, gather->start, gather->end); in riscv_iommu_iotlb_sync() 1247 struct iommu_iotlb_gather *gather) in riscv_iommu_unmap_pages() argument 1268 iommu_iotlb_gather_add_page(&domain->domain, gather, iova, in riscv_iommu_unmap_pages()
|
| /drivers/block/ |
| A D | ps3disk.c | 81 struct request *req, int gather) in ps3disk_scatter_gather() argument 88 if (gather) in ps3disk_scatter_gather()
|
| /drivers/gpu/host1x/hw/ |
| A D | debug_hw.c | 224 g = &job->cmds[i].gather; in show_channel_gathers()
|