| /linux/include/linux/ |
| A D | iommu.h | 400 *gather = (struct iommu_iotlb_gather) { in iommu_iotlb_gather_init() 543 return gather->end != 0 && in iommu_iotlb_gather_is_disjoint() 544 (end + 1 < gather->start || start > gather->end + 1); in iommu_iotlb_gather_is_disjoint() 563 if (gather->start > iova) in iommu_iotlb_gather_add_range() 564 gather->start = iova; in iommu_iotlb_gather_add_range() 565 if (gather->end < end) in iommu_iotlb_gather_add_range() 566 gather->end = end; in iommu_iotlb_gather_add_range() 589 if ((gather->pgsize && gather->pgsize != size) || in iommu_iotlb_gather_add_page() 591 iommu_iotlb_sync(domain, gather); in iommu_iotlb_gather_add_page() 593 gather->pgsize = size; in iommu_iotlb_gather_add_page() [all …]
|
| A D | io-pgtable.h | 43 void (*tlb_add_page)(struct iommu_iotlb_gather *gather, 162 size_t size, struct iommu_iotlb_gather *gather); 165 struct iommu_iotlb_gather *gather); 232 struct iommu_iotlb_gather * gather, unsigned long iova, in io_pgtable_tlb_add_page() argument 236 iop->cfg.tlb->tlb_add_page(gather, iova, granule, iop->cookie); in io_pgtable_tlb_add_page()
|
| /linux/drivers/gpu/host1x/ |
| A D | job.c | 110 struct host1x_job_gather *gather = &job->cmds[job->num_cmds].gather; in host1x_job_add_gather() local 112 gather->words = words; in host1x_job_add_gather() 113 gather->bo = bo; in host1x_job_add_gather() 114 gather->offset = offset; in host1x_job_add_gather() 238 g = &job->cmds[i].gather; in pin_job() 578 g = &job->cmds[i].gather; in copy_gathers() 602 void *gather; in copy_gathers() local 606 g = &job->cmds[i].gather; in copy_gathers() 609 gather = host1x_bo_mmap(g->bo); in copy_gathers() 612 host1x_bo_munmap(g->bo, gather); in copy_gathers() [all …]
|
| A D | job.h | 32 struct host1x_job_gather gather; member
|
| /linux/drivers/net/ethernet/cavium/liquidio/ |
| A D | octeon_nic.h | 95 u32 gather:1; member 150 if (!setup->s.gather) { in octnet_prepare_pci_cmd_o2() 153 ih2->gather = 1; in octnet_prepare_pci_cmd_o2() 195 if (!setup->s.gather) { in octnet_prepare_pci_cmd_o3() 198 ih3->gather = 1; in octnet_prepare_pci_cmd_o3()
|
| A D | liquidio_common.h | 357 u64 gather:1; member 391 u64 gather:1; 488 u64 gather:1; member 533 u64 gather:1;
|
| /linux/drivers/iommu/ |
| A D | io-pgtable-arm-v7s.c | 607 struct iommu_iotlb_gather *gather, in arm_v7s_split_blk_unmap() argument 644 return __arm_v7s_unmap(data, gather, iova, size, 2, tablep); in arm_v7s_split_blk_unmap() 647 io_pgtable_tlb_add_page(&data->iop, gather, iova, size); in arm_v7s_split_blk_unmap() 652 struct iommu_iotlb_gather *gather, in __arm_v7s_unmap() argument 703 } else if (!iommu_iotlb_gather_queued(gather)) { in __arm_v7s_unmap() 704 io_pgtable_tlb_add_page(iop, gather, iova, blk_size); in __arm_v7s_unmap() 714 return arm_v7s_split_blk_unmap(data, gather, iova, size, pte[0], in __arm_v7s_unmap() 720 return __arm_v7s_unmap(data, gather, iova, size, lvl + 1, ptep); in __arm_v7s_unmap() 725 struct iommu_iotlb_gather *gather) in arm_v7s_unmap_pages() argument 746 size_t size, struct iommu_iotlb_gather *gather) in arm_v7s_unmap() argument [all …]
|
| A D | tegra-gart.c | 210 size_t bytes, struct iommu_iotlb_gather *gather) in gart_iommu_unmap() argument 271 struct iommu_iotlb_gather *gather) in gart_iommu_sync() argument 273 size_t length = gather->end - gather->start + 1; in gart_iommu_sync() 275 gart_iommu_sync_map(domain, gather->start, length); in gart_iommu_sync()
|
| A D | io-pgtable-arm.c | 258 struct iommu_iotlb_gather *gather, 558 struct iommu_iotlb_gather *gather, in arm_lpae_split_blk_unmap() argument 609 io_pgtable_tlb_add_page(&data->iop, gather, iova + i * size, size); in arm_lpae_split_blk_unmap() 618 struct iommu_iotlb_gather *gather, in __arm_lpae_unmap() argument 653 } else if (!iommu_iotlb_gather_queued(gather)) { in __arm_lpae_unmap() 654 io_pgtable_tlb_add_page(iop, gather, iova + i * size, size); in __arm_lpae_unmap() 667 return arm_lpae_split_blk_unmap(data, gather, iova, size, pte, in __arm_lpae_unmap() 678 struct iommu_iotlb_gather *gather) in arm_lpae_unmap_pages() argument 693 return __arm_lpae_unmap(data, gather, iova, pgsize, pgcount, in arm_lpae_unmap_pages() 698 size_t size, struct iommu_iotlb_gather *gather) in arm_lpae_unmap() argument [all …]
|
| A D | mtk_iommu.c | 513 struct iommu_iotlb_gather *gather) in mtk_iommu_unmap() argument 517 iommu_iotlb_gather_add_range(gather, iova, size); in mtk_iommu_unmap() 518 return dom->iop->unmap(dom->iop, iova, size, gather); in mtk_iommu_unmap() 529 struct iommu_iotlb_gather *gather) in mtk_iommu_iotlb_sync() argument 532 size_t length = gather->end - gather->start + 1; in mtk_iommu_iotlb_sync() 534 mtk_iommu_tlb_flush_range_sync(gather->start, length, gather->pgsize, in mtk_iommu_iotlb_sync()
|
| /linux/drivers/gpu/drm/amd/display/dc/ |
| A D | dc_helper.c | 44 bool gather = false; in submit_dmub_read_modify_write() local 51 gather = ctx->dmub_srv->reg_helper_offload.gather_in_progress; in submit_dmub_read_modify_write() 56 ctx->dmub_srv->reg_helper_offload.gather_in_progress = gather; in submit_dmub_read_modify_write() 69 bool gather = false; in submit_dmub_burst_write() local 74 gather = ctx->dmub_srv->reg_helper_offload.gather_in_progress; in submit_dmub_burst_write() 79 ctx->dmub_srv->reg_helper_offload.gather_in_progress = gather; in submit_dmub_burst_write() 91 bool gather = false; in submit_dmub_reg_wait() local 93 gather = ctx->dmub_srv->reg_helper_offload.gather_in_progress; in submit_dmub_reg_wait() 101 ctx->dmub_srv->reg_helper_offload.gather_in_progress = gather; in submit_dmub_reg_wait()
|
| /linux/Documentation/block/ |
| A D | request.rst | 71 unsigned short nr_phys_segments DB Number of physical scatter gather 74 unsigned short nr_hw_segments DB Number of hardware scatter gather
|
| /linux/drivers/net/ethernet/cavium/thunder/ |
| A D | nicvf_queues.c | 557 struct sq_gather_subdesc *gather; in nicvf_unmap_sndq_buffers() local 563 gather = (struct sq_gather_subdesc *)GET_SQ_DESC(sq, hdr_sqe); in nicvf_unmap_sndq_buffers() 565 dma_unmap_page_attrs(&nic->pdev->dev, gather->addr, in nicvf_unmap_sndq_buffers() 566 gather->size, DMA_TO_DEVICE, in nicvf_unmap_sndq_buffers() 1421 struct sq_gather_subdesc *gather; in nicvf_sq_add_gather_subdesc() local 1424 gather = (struct sq_gather_subdesc *)GET_SQ_DESC(sq, qentry); in nicvf_sq_add_gather_subdesc() 1426 memset(gather, 0, SND_QUEUE_DESC_SIZE); in nicvf_sq_add_gather_subdesc() 1427 gather->subdesc_type = SQ_DESC_TYPE_GATHER; in nicvf_sq_add_gather_subdesc() 1428 gather->ld_type = NIC_SEND_LD_TYPE_E_LDD; in nicvf_sq_add_gather_subdesc() 1429 gather->size = size; in nicvf_sq_add_gather_subdesc() [all …]
|
| /linux/drivers/crypto/cavium/zip/ |
| A D | zip_device.h | 57 union zip_zptr_s *gather; member
|
| /linux/Documentation/networking/ |
| A D | driver.rst | 15 for a driver implementing scatter-gather this means:: 51 For a non-scatter-gather supporting card, the three tests simply become::
|
| /linux/Documentation/driver-api/fpga/ |
| A D | fpga-programming.rst | 55 if (image is in a scatter gather table) { 57 info->sgt = [your scatter gather table]
|
| A D | fpga-mgr.rst | 13 The FPGA image to be programmed can be in a scatter gather list, a single 16 gather list instead if possible.
|
| /linux/drivers/message/fusion/ |
| A D | Kconfig | 65 int "Maximum number of scatter gather entries (16 - 128)" 70 gather entries per I/O. The driver default is 128, which matches
|
| /linux/arch/mips/include/asm/octeon/ |
| A D | cvmx-pko.h | 220 uint64_t gather:1; member 250 uint64_t gather:1;
|
| /linux/kernel/dma/ |
| A D | Kconfig | 129 hardware components that do not support I/O map nor scatter-gather. 226 bool "Debug DMA scatter-gather usage" 236 preparing literal scatter-gather descriptors, where there is a risk of
|
| /linux/drivers/scsi/mpt3sas/ |
| A D | Kconfig | 59 gather entries per I/O. The driver default is 128, which matches 71 gather entries per I/O. The driver default is 128, which matches
|
| /linux/drivers/iommu/arm/arm-smmu/ |
| A D | qcom_iommu.c | 186 static void qcom_iommu_tlb_add_page(struct iommu_iotlb_gather *gather, in qcom_iommu_tlb_add_page() argument 435 size_t size, struct iommu_iotlb_gather *gather) in qcom_iommu_unmap() argument 452 ret = ops->unmap(ops, iova, size, gather); in qcom_iommu_unmap() 473 struct iommu_iotlb_gather *gather) in qcom_iommu_iotlb_sync() argument
|
| /linux/Documentation/devicetree/bindings/arm/ |
| A D | arm-dsu-pmu.txt | 5 form a multicore cluster. The PMU enables to gather various statistics on
|
| /linux/Documentation/scsi/ |
| A D | st.rst | 259 scatter/gather segments permitted by the HBA 293 Scatter/gather buffers (buffers that consist of chunks non-contiguous 296 supporting scatter/gather), buffer allocation is using the following 300 those not supporting scatter/gather. The size of this buffer will be 307 2. The scatter/gather segments to fill the specified buffer size are 313 number of scatter/gather segments allowed for the SCSI adapter is not 314 exceeded if it is smaller than the maximum number of scatter/gather 341 max_sg_segs=xxx the maximum number of scatter/gather 357 with 'st.'. For instance, to set the maximum number of scatter/gather 359 number of scatter/gather segments). [all …]
|
| /linux/drivers/iommu/amd/ |
| A D | iommu.c | 2102 struct iommu_iotlb_gather *gather, in amd_iommu_iotlb_gather_add_page() argument 2116 iommu_iotlb_gather_is_disjoint(gather, iova, size)) in amd_iommu_iotlb_gather_add_page() 2117 iommu_iotlb_sync(domain, gather); in amd_iommu_iotlb_gather_add_page() 2119 iommu_iotlb_gather_add_range(gather, iova, size); in amd_iommu_iotlb_gather_add_page() 2124 struct iommu_iotlb_gather *gather) in amd_iommu_unmap() argument 2134 r = (ops->unmap) ? ops->unmap(ops, iova, page_size, gather) : 0; in amd_iommu_unmap() 2136 amd_iommu_iotlb_gather_add_page(dom, gather, iova, page_size); in amd_iommu_unmap() 2239 struct iommu_iotlb_gather *gather) in amd_iommu_iotlb_sync() argument 2245 domain_flush_pages(dom, gather->start, gather->end - gather->start, 1); in amd_iommu_iotlb_sync()
|