| /linux/arch/mips/cavium-octeon/ |
| A D | dma-octeon.c | 40 return paddr; in octeon_hole_phys_to_dma() 53 if (paddr >= 0x410000000ull && paddr < 0x420000000ull) in octeon_gen1_phys_to_dma() 54 paddr -= 0x400000000ull; in octeon_gen1_phys_to_dma() 90 if (paddr >= 0x410000000ull && paddr < 0x420000000ull) in octeon_big_phys_to_dma() 95 paddr = OCTEON_BAR2_PCI_ADDRESS + paddr; in octeon_big_phys_to_dma() 97 return paddr; in octeon_big_phys_to_dma() 118 if (paddr >= 0x410000000ull && paddr < 0x420000000ull) in octeon_small_phys_to_dma() 123 paddr = paddr - octeon_bar1_pci_phys; in octeon_small_phys_to_dma() 125 paddr = OCTEON_BAR2_PCI_ADDRESS + paddr; in octeon_small_phys_to_dma() 127 return paddr; in octeon_small_phys_to_dma() [all …]
|
| /linux/arch/m68k/mm/ |
| A D | memory.c | 31 : : "a" (paddr)); in clear040() 42 : : "a" (paddr)); in cleari040() 54 : : "a" (paddr)); in push040() 64 push040(paddr); in pushcl040() 66 clear040(paddr); in pushcl040() 113 paddr += tmp; in cache_clear() 116 paddr &= PAGE_MASK; in cache_clear() 119 paddr += tmp; in cache_clear() 164 paddr &= PAGE_MASK; in cache_push() 167 push040(paddr); in cache_push() [all …]
|
| /linux/arch/m68k/kernel/ |
| A D | sys_m68k.c | 70 unsigned long paddr, i; in cache_flush_040() local 135 : : "a" (paddr)); in cache_flush_040() 142 : : "a" (paddr)); in cache_flush_040() 150 : : "a" (paddr)); in cache_flush_040() 174 paddr += 16; in cache_flush_040() 192 : : "a" (paddr)); in cache_flush_040() 199 : : "a" (paddr)); in cache_flush_040() 207 : : "a" (paddr)); in cache_flush_040() 222 : "=a" (paddr) \ 230 unsigned long paddr, i; in cache_flush_060() local [all …]
|
| /linux/arch/riscv/mm/ |
| A D | dma-noncoherent.c | 20 void *vaddr = phys_to_virt(paddr); in arch_dma_cache_wback() 24 noncoherent_cache_ops.wback(paddr, size); in arch_dma_cache_wback() 33 void *vaddr = phys_to_virt(paddr); in arch_dma_cache_inv() 37 noncoherent_cache_ops.inv(paddr, size); in arch_dma_cache_inv() 47 void *vaddr = phys_to_virt(paddr); in arch_dma_cache_wback_inv() 51 noncoherent_cache_ops.wback_inv(paddr, size); in arch_dma_cache_wback_inv() 74 arch_dma_cache_wback(paddr, size); in arch_sync_dma_for_device() 79 arch_dma_cache_inv(paddr, size); in arch_sync_dma_for_device() 88 arch_dma_cache_wback(paddr, size); in arch_sync_dma_for_device() 90 arch_dma_cache_wback_inv(paddr, size); in arch_sync_dma_for_device() [all …]
|
| /linux/arch/arc/mm/ |
| A D | cache.c | 210 paddr &= CACHE_LINE_MASK; in __cache_line_loop_v3() 235 paddr += L1_CACHE_BYTES; in __cache_line_loop_v3() 269 paddr &= CACHE_LINE_MASK; in __cache_line_loop_v4() 292 paddr += L1_CACHE_BYTES; in __cache_line_loop_v4() 319 paddr &= CACHE_LINE_MASK; in __cache_line_loop_v4() 338 write_aux_reg(s, paddr); in __cache_line_loop_v4() 514 phys_addr_t paddr, vaddr; member 529 .paddr = paddr, in __ic_line_inv_vaddr() 641 paddr &= SLC_LINE_MASK; in slc_op_line() 647 paddr += l2_line_sz; in slc_op_line() [all …]
|
| A D | ioremap.c | 13 static inline bool arc_uncached_addr_space(phys_addr_t paddr) in arc_uncached_addr_space() argument 16 if (paddr >= ARC_UNCACHED_ADDR_SPACE) in arc_uncached_addr_space() 18 } else if (paddr >= perip_base && paddr <= perip_end) { in arc_uncached_addr_space() 25 void __iomem *ioremap(phys_addr_t paddr, unsigned long size) in ioremap() argument 31 if (arc_uncached_addr_space(paddr)) in ioremap() 32 return (void __iomem *)(u32)paddr; in ioremap() 34 return ioremap_prot(paddr, size, in ioremap() 46 void __iomem *ioremap_prot(phys_addr_t paddr, size_t size, in ioremap_prot() argument 52 return generic_ioremap_prot(paddr, size, pgprot_noncached(prot)); in ioremap_prot()
|
| A D | dma.c | 51 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device() argument 56 dma_cache_wback(paddr, size); in arch_sync_dma_for_device() 60 dma_cache_inv(paddr, size); in arch_sync_dma_for_device() 64 dma_cache_wback_inv(paddr, size); in arch_sync_dma_for_device() 72 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu() argument 82 dma_cache_inv(paddr, size); in arch_sync_dma_for_cpu()
|
| /linux/arch/arm/mm/ |
| A D | dma-mapping-nommu.c | 16 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device() argument 19 dmac_map_area(__va(paddr), size, dir); in arch_sync_dma_for_device() 22 outer_inv_range(paddr, paddr + size); in arch_sync_dma_for_device() 24 outer_clean_range(paddr, paddr + size); in arch_sync_dma_for_device() 27 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu() argument 31 outer_inv_range(paddr, paddr + size); in arch_sync_dma_for_cpu() 32 dmac_unmap_area(__va(paddr), size, dir); in arch_sync_dma_for_cpu()
|
| /linux/arch/microblaze/kernel/ |
| A D | dma.c | 17 static void __dma_sync(phys_addr_t paddr, size_t size, in __dma_sync() argument 23 flush_dcache_range(paddr, paddr + size); in __dma_sync() 26 invalidate_dcache_range(paddr, paddr + size); in __dma_sync() 33 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device() argument 36 __dma_sync(paddr, size, dir); in arch_sync_dma_for_device() 39 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu() argument 42 __dma_sync(paddr, size, dir); in arch_sync_dma_for_cpu()
|
| /linux/arch/x86/mm/pat/ |
| A D | memtype.c | 659 if (x86_platform.is_untracked_pat_range(paddr, paddr + PAGE_SIZE)) in lookup_memtype() 662 if (pat_pagerange_is_ram(paddr, paddr + PAGE_SIZE)) { in lookup_memtype() 864 is_ram = pat_pagerange_is_ram(paddr, paddr + size); in reserve_pfn_range() 890 ret = memtype_reserve(paddr, paddr + size, want_pcm, &pcm); in reserve_pfn_range() 897 memtype_free(paddr, paddr + size); in reserve_pfn_range() 916 memtype_free(paddr, paddr + size); in reserve_pfn_range() 930 is_ram = pat_pagerange_is_ram(paddr, paddr + size); in free_pfn_range() 932 memtype_free(paddr, paddr + size); in free_pfn_range() 996 resource_size_t paddr; in track_pfn_copy() local 1045 paddr += PAGE_SIZE; in track_pfn_remap() [all …]
|
| /linux/include/linux/ |
| A D | dma-direct.h | 27 phys_addr_t paddr) in translate_phys_to_dma() argument 32 u64 offset = paddr - m->cpu_start; in translate_phys_to_dma() 34 if (paddr >= m->cpu_start && offset < m->size) in translate_phys_to_dma() 82 phys_addr_t paddr) in phys_to_dma_unencrypted() argument 85 return translate_phys_to_dma(dev, paddr); in phys_to_dma_unencrypted() 86 return paddr; in phys_to_dma_unencrypted() 97 return __sme_set(phys_to_dma_unencrypted(dev, paddr)); in phys_to_dma() 102 phys_addr_t paddr; in dma_to_phys() local 105 paddr = translate_dma_to_phys(dev, dma_addr); in dma_to_phys() 107 paddr = dma_addr; in dma_to_phys() [all …]
|
| /linux/drivers/soc/apple/ |
| A D | sart.c | 53 phys_addr_t *paddr, size_t *size); 72 phys_addr_t *paddr, size_t *size) in sart2_get_entry() argument 104 phys_addr_t *paddr, size_t *size) in sart3_get_entry() argument 150 phys_addr_t paddr; in apple_sart_probe() local 159 i, flags, &paddr, size); in apple_sart_probe() 208 phys_addr_t paddr, size_t size) in sart_set_entry() argument 215 paddr >>= sart->ops->size_shift; in sart_set_entry() 241 i, &paddr, size); in apple_sart_add_allowed_region() 253 &paddr, size); in apple_sart_add_allowed_region() 266 &paddr, size); in apple_sart_remove_allowed_region() [all …]
|
| /linux/arch/x86/mm/ |
| A D | ioremap.c | 599 u64 paddr; in memremap_is_efi_data() local 606 paddr <<= 32; in memremap_is_efi_data() 612 paddr <<= 32; in memremap_is_efi_data() 640 u64 paddr, paddr_next; in memremap_is_setup_data() local 643 while (paddr) { in memremap_is_setup_data() 684 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in memremap_is_setup_data() 687 paddr = paddr_next; in memremap_is_setup_data() 702 u64 paddr, paddr_next; in early_memremap_is_setup_data() local 705 while (paddr) { in early_memremap_is_setup_data() 747 if ((phys_addr > paddr) && (phys_addr < (paddr + len))) in early_memremap_is_setup_data() [all …]
|
| A D | mem_encrypt_amd.c | 58 unsigned long paddr, bool decrypt) in snp_memcpy() argument 67 early_snp_set_memory_shared((unsigned long)__va(paddr), paddr, npages); in snp_memcpy() 72 early_snp_set_memory_private((unsigned long)__va(paddr), paddr, npages); in snp_memcpy() 114 early_memremap_encrypted_wp(paddr, len); in __sme_early_enc_dec() 116 dst = enc ? early_memremap_encrypted(paddr, len) : in __sme_early_enc_dec() 117 early_memremap_decrypted(paddr, len); in __sme_early_enc_dec() 141 paddr += len; in __sme_early_enc_dec() 148 __sme_early_enc_dec(paddr, size, true); in sme_early_encrypt() 153 __sme_early_enc_dec(paddr, size, false); in sme_early_decrypt() 166 pmd = map ? (paddr & PMD_MASK) + pmd_flags : 0; in __sme_early_map_unmap_mem() [all …]
|
| A D | init_64.c | 462 pte = pte_page + pte_index(paddr); in phys_pte_init() 463 i = pte_index(paddr); in phys_pte_init() 467 if (paddr >= paddr_end) { in phys_pte_init() 516 int i = pmd_index(paddr); in phys_pmd_init() 524 if (paddr >= paddr_end) { in phys_pmd_init() 540 paddr_last = phys_pte_init(pte, paddr, in phys_pmd_init() 609 vaddr = (unsigned long)__va(paddr); in phys_pud_init() 613 if (paddr >= paddr_end) { in phys_pud_init() 687 vaddr = (unsigned long)__va(paddr); in phys_p4d_init() 699 paddr = __pa(vaddr); in phys_p4d_init() [all …]
|
| /linux/arch/xtensa/kernel/ |
| A D | pci-dma.c | 24 static void do_cache_op(phys_addr_t paddr, size_t size, in do_cache_op() argument 27 unsigned long off = paddr & (PAGE_SIZE - 1); in do_cache_op() 28 unsigned long pfn = PFN_DOWN(paddr); in do_cache_op() 32 fn((unsigned long)phys_to_virt(paddr), size); in do_cache_op() 46 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu() argument 52 do_cache_op(paddr, size, __invalidate_dcache_range); in arch_sync_dma_for_cpu() 64 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device() argument 71 do_cache_op(paddr, size, __flush_dcache_range); in arch_sync_dma_for_device()
|
| /linux/drivers/bus/ |
| A D | uniphier-system-bus.c | 39 int bank, u32 addr, u64 paddr, u32 size) in uniphier_system_bus_add_bank() argument 45 bank, addr, paddr, size); in uniphier_system_bus_add_bank() 58 if (paddr > U32_MAX) { in uniphier_system_bus_add_bank() 63 end = paddr + size; in uniphier_system_bus_add_bank() 65 if (addr > paddr) { in uniphier_system_bus_add_bank() 68 addr, paddr); in uniphier_system_bus_add_bank() 71 paddr -= addr; in uniphier_system_bus_add_bank() 73 paddr = round_down(paddr, 0x00020000); in uniphier_system_bus_add_bank() 80 mask = paddr ^ (end - 1); in uniphier_system_bus_add_bank() 83 paddr = round_down(paddr, mask); in uniphier_system_bus_add_bank() [all …]
|
| /linux/drivers/cache/ |
| A D | starfive_starlink_cache.c | 47 static void starlink_cache_dma_cache_wback(phys_addr_t paddr, unsigned long size) in starlink_cache_dma_cache_wback() argument 49 writeq(FIELD_PREP(STARLINK_CACHE_ADDRESS_RANGE_MASK, paddr), in starlink_cache_dma_cache_wback() 51 writeq(FIELD_PREP(STARLINK_CACHE_ADDRESS_RANGE_MASK, paddr + size), in starlink_cache_dma_cache_wback() 62 static void starlink_cache_dma_cache_invalidate(phys_addr_t paddr, unsigned long size) in starlink_cache_dma_cache_invalidate() argument 64 writeq(FIELD_PREP(STARLINK_CACHE_ADDRESS_RANGE_MASK, paddr), in starlink_cache_dma_cache_invalidate() 66 writeq(FIELD_PREP(STARLINK_CACHE_ADDRESS_RANGE_MASK, paddr + size), in starlink_cache_dma_cache_invalidate() 77 static void starlink_cache_dma_cache_wback_inv(phys_addr_t paddr, unsigned long size) in starlink_cache_dma_cache_wback_inv() argument 79 writeq(FIELD_PREP(STARLINK_CACHE_ADDRESS_RANGE_MASK, paddr), in starlink_cache_dma_cache_wback_inv() 81 writeq(FIELD_PREP(STARLINK_CACHE_ADDRESS_RANGE_MASK, paddr + size), in starlink_cache_dma_cache_wback_inv()
|
| /linux/arch/csky/mm/ |
| A D | dma-mapping.c | 14 static inline void cache_op(phys_addr_t paddr, size_t size, in cache_op() argument 17 struct page *page = phys_to_page(paddr); in cache_op() 19 unsigned long offset = offset_in_page(paddr); in cache_op() 58 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device() argument 63 cache_op(paddr, size, dma_wb_range); in arch_sync_dma_for_device() 67 cache_op(paddr, size, dma_wbinv_range); in arch_sync_dma_for_device() 74 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu() argument 82 cache_op(paddr, size, dma_inv_range); in arch_sync_dma_for_cpu()
|
| A D | tcm.c | 29 unsigned long vaddr, paddr; local 32 paddr = CONFIG_ITCM_RAM_BASE; 47 set_pte(tcm_pte, pfn_pte(__phys_to_pfn(paddr), PAGE_KERNEL)); 51 paddr = paddr + PAGE_SIZE; 58 paddr = CONFIG_DTCM_RAM_BASE; 66 set_pte(tcm_pte, pfn_pte(__phys_to_pfn(paddr), PAGE_KERNEL)); 70 paddr = paddr + PAGE_SIZE;
|
| /linux/drivers/xen/ |
| A D | swiotlb-xen.c | 49 unsigned long bfn = pfn_to_bfn(XEN_PFN_DOWN(paddr)); in xen_phys_to_bus() 52 baddr |= paddr & ~XEN_PAGE_MASK; in xen_phys_to_bus() 58 return phys_to_dma(dev, xen_phys_to_bus(dev, paddr)); in xen_phys_to_dma() 65 phys_addr_t paddr = (xen_pfn << XEN_PAGE_SHIFT) | in xen_bus_to_phys() local 68 return paddr; in xen_bus_to_phys() 108 if (pfn_valid(PFN_DOWN(paddr))) in xen_swiotlb_find_pool() 109 return swiotlb_find_pool(dev, paddr); in xen_swiotlb_find_pool() 271 arch_sync_dma_for_cpu(paddr, size, dir); in xen_swiotlb_unmap_page() 287 phys_addr_t paddr = xen_dma_to_phys(dev, dma_addr); in xen_swiotlb_sync_single_for_cpu() local 292 arch_sync_dma_for_cpu(paddr, size, dir); in xen_swiotlb_sync_single_for_cpu() [all …]
|
| /linux/drivers/media/platform/st/sti/hva/ |
| A D | hva-mem.c | 16 dma_addr_t paddr; in hva_mem_alloc() local 25 base = dma_alloc_attrs(dev, size, &paddr, GFP_KERNEL, in hva_mem_alloc() 36 b->paddr = paddr; in hva_mem_alloc() 42 ctx->name, size, b->vaddr, &b->paddr, b->name); in hva_mem_alloc() 56 ctx->name, buf->size, buf->vaddr, &buf->paddr, buf->name); in hva_mem_free() 58 dma_free_attrs(dev, buf->size, buf->vaddr, buf->paddr, in hva_mem_free()
|
| /linux/drivers/net/phy/ |
| A D | bcm87xx.c | 36 const __be32 *paddr; in bcm87xx_of_reg_init() local 43 paddr = of_get_property(phydev->mdio.dev.of_node, in bcm87xx_of_reg_init() 45 if (!paddr) in bcm87xx_of_reg_init() 48 paddr_end = paddr + (len /= sizeof(*paddr)); in bcm87xx_of_reg_init() 52 while (paddr + 3 < paddr_end) { in bcm87xx_of_reg_init() 53 u16 devid = be32_to_cpup(paddr++); in bcm87xx_of_reg_init() 54 u16 reg = be32_to_cpup(paddr++); in bcm87xx_of_reg_init() 55 u16 mask = be32_to_cpup(paddr++); in bcm87xx_of_reg_init() 56 u16 val_bits = be32_to_cpup(paddr++); in bcm87xx_of_reg_init()
|
| /linux/arch/sparc/include/asm/ |
| A D | mxcc.h | 89 static inline void mxcc_set_stream_src(unsigned long *paddr) in mxcc_set_stream_src() argument 91 unsigned long data0 = paddr[0]; in mxcc_set_stream_src() 92 unsigned long data1 = paddr[1]; in mxcc_set_stream_src() 102 static inline void mxcc_set_stream_dst(unsigned long *paddr) in mxcc_set_stream_dst() argument 104 unsigned long data0 = paddr[0]; in mxcc_set_stream_dst() 105 unsigned long data1 = paddr[1]; in mxcc_set_stream_dst()
|
| /linux/arch/x86/include/asm/uv/ |
| A D | uv_hub.h | 422 return uv_gam_range(paddr)->nasid; in uv_soc_phys_ram_to_nasid() 441 paddr |= uv_hub_info->gnode_upper; in uv_soc_phys_ram_to_gpa() 442 paddr = ((paddr << uv_hub_info->m_shift) in uv_soc_phys_ram_to_gpa() 444 ((paddr >> uv_hub_info->m_val) in uv_soc_phys_ram_to_gpa() 447 paddr |= uv_soc_phys_ram_to_nasid(paddr) in uv_soc_phys_ram_to_gpa() 450 return paddr; in uv_soc_phys_ram_to_gpa() 469 unsigned long paddr; in uv_gpa_to_soc_phys_ram() local 478 paddr = gpa & uv_hub_info->gpa_mask; in uv_gpa_to_soc_phys_ram() 479 if (paddr >= remap_base && paddr < remap_base + remap_top) in uv_gpa_to_soc_phys_ram() 480 paddr -= remap_base; in uv_gpa_to_soc_phys_ram() [all …]
|