| /arch/arm/mm/ |
| A D | dma-mapping.c | 827 BUG_ON(addr < mapping->base || bitmap_index > mapping->extensions); in __free_iova() 1568 if (!mapping) in arm_iommu_create_mapping() 1583 mapping->base = base; in arm_iommu_create_mapping() 1595 return mapping; in arm_iommu_create_mapping() 1601 kfree(mapping); in arm_iommu_create_mapping() 1617 kfree(mapping); in release_iommu_mapping() 1624 if (mapping->nr_bitmaps >= mapping->extensions) in extend_iommu_mapping() 1628 mapping->bitmaps[next_bitmap] = kzalloc(mapping->bitmap_size, in extend_iommu_mapping() 1640 if (mapping) in arm_iommu_release_mapping() 1700 if (!mapping) { in arm_iommu_detach_device() [all …]
|
| A D | flush.c | 253 flush_dcache_mmap_lock(mapping); in __flush_dcache_aliases() 281 flush_dcache_mmap_unlock(mapping); in __flush_dcache_aliases() 289 struct address_space *mapping; in __sync_icache_dcache() local 303 mapping = folio_flush_mapping(folio); in __sync_icache_dcache() 305 mapping = NULL; in __sync_icache_dcache() 308 __flush_dcache_folio(mapping, folio); in __sync_icache_dcache() 336 struct address_space *mapping; in flush_dcache_folio() local 351 mapping = folio_flush_mapping(folio); in flush_dcache_folio() 354 mapping && !folio_mapped(folio)) in flush_dcache_folio() 358 if (mapping && cache_is_vivt()) in flush_dcache_folio() [all …]
|
| A D | fault-armv.c | 124 make_coherent(struct address_space *mapping, struct vm_area_struct *vma, in make_coherent() argument 142 flush_dcache_mmap_lock(mapping); in make_coherent() 143 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in make_coherent() 169 flush_dcache_mmap_unlock(mapping); in make_coherent() 191 struct address_space *mapping; in update_mmu_cache_range() local 205 mapping = folio_flush_mapping(folio); in update_mmu_cache_range() 207 __flush_dcache_folio(mapping, folio); in update_mmu_cache_range() 208 if (mapping) { in update_mmu_cache_range() 210 make_coherent(mapping, vma, addr, ptep, pfn); in update_mmu_cache_range()
|
| /arch/arm64/kvm/ |
| A D | pkvm.c | 301 struct pkvm_mapping *mapping; in __pkvm_pgtable_stage2_unmap() local 309 mapping->nr_pages); in __pkvm_pgtable_stage2_unmap() 313 kfree(mapping); in __pkvm_pgtable_stage2_unmap() 346 if (mapping) { in pkvm_pgtable_stage2_map() 354 mapping = NULL; in pkvm_pgtable_stage2_map() 361 swap(mapping, cache->mapping); in pkvm_pgtable_stage2_map() 362 mapping->gfn = gfn; in pkvm_pgtable_stage2_map() 363 mapping->pfn = pfn; in pkvm_pgtable_stage2_map() 381 struct pkvm_mapping *mapping; in pkvm_pgtable_stage2_wrprotect() local 387 mapping->nr_pages); in pkvm_pgtable_stage2_wrprotect() [all …]
|
| /arch/nios2/include/asm/ |
| A D | cacheflush.h | 54 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) argument 55 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) argument 56 #define flush_dcache_mmap_lock_irqsave(mapping, flags) \ argument 57 xa_lock_irqsave(&mapping->i_pages, flags) 58 #define flush_dcache_mmap_unlock_irqrestore(mapping, flags) \ argument 59 xa_unlock_irqrestore(&mapping->i_pages, flags)
|
| /arch/parisc/include/asm/ |
| A D | cacheflush.h | 54 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) argument 55 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) argument 56 #define flush_dcache_mmap_lock_irqsave(mapping, flags) \ argument 57 xa_lock_irqsave(&mapping->i_pages, flags) 58 #define flush_dcache_mmap_unlock_irqrestore(mapping, flags) \ argument 59 xa_unlock_irqrestore(&mapping->i_pages, flags)
|
| /arch/nios2/mm/ |
| A D | cacheflush.c | 84 flush_dcache_mmap_lock_irqsave(mapping, flags); in flush_aliases() 96 flush_dcache_mmap_unlock_irqrestore(mapping, flags); in flush_aliases() 177 struct address_space *mapping; in flush_dcache_folio() local 186 mapping = folio_flush_mapping(folio); in flush_dcache_folio() 189 if (mapping && !mapping_mapped(mapping)) { in flush_dcache_folio() 193 if (mapping) { in flush_dcache_folio() 195 flush_aliases(mapping, folio); in flush_dcache_folio() 215 struct address_space *mapping; in update_mmu_cache_range() local 233 mapping = folio_flush_mapping(folio); in update_mmu_cache_range() 234 if (mapping) { in update_mmu_cache_range() [all …]
|
| /arch/csky/abiv1/ |
| A D | cacheflush.c | 20 struct address_space *mapping; in flush_dcache_folio() local 25 mapping = folio_flush_mapping(folio); in flush_dcache_folio() 27 if (mapping && !folio_mapped(folio)) in flush_dcache_folio() 31 if (mapping) in flush_dcache_folio()
|
| /arch/powerpc/boot/dts/fsl/ |
| A D | t4240si-pre.dtsi | 95 fsl,portid-mapping = <0x80000000>; 102 fsl,portid-mapping = <0x80000000>; 109 fsl,portid-mapping = <0x80000000>; 116 fsl,portid-mapping = <0x80000000>; 123 fsl,portid-mapping = <0x40000000>; 130 fsl,portid-mapping = <0x40000000>; 137 fsl,portid-mapping = <0x40000000>; 144 fsl,portid-mapping = <0x40000000>; 151 fsl,portid-mapping = <0x20000000>; 158 fsl,portid-mapping = <0x20000000>; [all …]
|
| A D | p4080si-pre.dtsi | 99 fsl,portid-mapping = <0x80000000>; 109 fsl,portid-mapping = <0x40000000>; 119 fsl,portid-mapping = <0x20000000>; 129 fsl,portid-mapping = <0x10000000>; 139 fsl,portid-mapping = <0x08000000>; 149 fsl,portid-mapping = <0x04000000>; 159 fsl,portid-mapping = <0x02000000>; 169 fsl,portid-mapping = <0x01000000>;
|
| A D | b4860si-pre.dtsi | 80 fsl,portid-mapping = <0x80000000>; 87 fsl,portid-mapping = <0x80000000>; 94 fsl,portid-mapping = <0x80000000>; 101 fsl,portid-mapping = <0x80000000>;
|
| A D | t208xsi-pre.dtsi | 86 fsl,portid-mapping = <0x80000000>; 93 fsl,portid-mapping = <0x80000000>; 100 fsl,portid-mapping = <0x80000000>; 107 fsl,portid-mapping = <0x80000000>;
|
| A D | p2041si-pre.dtsi | 94 fsl,portid-mapping = <0x80000000>; 104 fsl,portid-mapping = <0x40000000>; 114 fsl,portid-mapping = <0x20000000>; 124 fsl,portid-mapping = <0x10000000>;
|
| /arch/csky/abiv1/inc/abi/ |
| A D | cacheflush.h | 19 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) argument 20 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) argument
|
| /arch/x86/kernel/ |
| A D | jailhouse.c | 203 void *mapping; in jailhouse_init_platform() local 224 mapping = early_memremap(pa_data, sizeof(header)); in jailhouse_init_platform() 225 memcpy(&header, mapping, sizeof(header)); in jailhouse_init_platform() 226 early_memunmap(mapping, sizeof(header)); in jailhouse_init_platform() 244 mapping = early_memremap(pa_data, setup_data_len); in jailhouse_init_platform() 245 memcpy(&setup_data, mapping, setup_data_len); in jailhouse_init_platform() 246 early_memunmap(mapping, setup_data_len); in jailhouse_init_platform()
|
| /arch/arm/include/asm/ |
| A D | device.h | 10 struct dma_iommu_mapping *mapping; member 24 #define to_dma_iommu_mapping(dev) ((dev)->archdata.mapping)
|
| A D | dma-iommu.h | 29 void arm_iommu_release_mapping(struct dma_iommu_mapping *mapping); 32 struct dma_iommu_mapping *mapping);
|
| A D | cacheflush.h | 321 #define flush_dcache_mmap_lock(mapping) xa_lock_irq(&mapping->i_pages) argument 322 #define flush_dcache_mmap_unlock(mapping) xa_unlock_irq(&mapping->i_pages) argument
|
| /arch/xtensa/mm/ |
| A D | cache.c | 128 struct address_space *mapping = folio_flush_mapping(folio); in flush_dcache_folio() local 136 if (mapping && !mapping_mapped(mapping)) { in flush_dcache_folio() 155 if (!alias && !mapping) in flush_dcache_folio() 168 if (mapping) in flush_dcache_folio()
|
| /arch/mips/pci/ |
| A D | pci-rc32434.c | 155 rc32434_pci->pcilba[0].mapping = (unsigned int) (PCI_ADDR_START); in rc32434_pcibridge_init() 163 rc32434_pci->pcilba[1].mapping = 0x60000000; in rc32434_pcibridge_init() 170 rc32434_pci->pcilba[2].mapping = 0x18FFFFFF; in rc32434_pcibridge_init() 179 rc32434_pci->pcilba[3].mapping = 0x18800000; in rc32434_pcibridge_init()
|
| /arch/sh/mm/ |
| A D | cache-sh7705.c | 138 struct address_space *mapping = folio_flush_mapping(folio); in sh7705_flush_dcache_folio() local 140 if (mapping && !mapping_mapped(mapping)) in sh7705_flush_dcache_folio()
|
| /arch/loongarch/include/asm/ |
| A D | cacheflush.h | 48 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 49 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
| /arch/csky/abiv2/inc/abi/ |
| A D | cacheflush.h | 34 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 35 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
| /arch/arc/include/asm/ |
| A D | cacheflush.h | 38 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 39 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
| /arch/sparc/include/asm/ |
| A D | cacheflush_32.h | 47 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 48 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|