/arch/arm/mm/ |
D | flush.c | 202 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 236 if (mapping && cache_is_vipt_aliasing()) in __flush_dcache_page() 241 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) in __flush_dcache_aliases() argument 255 flush_dcache_mmap_lock(mapping); in __flush_dcache_aliases() 256 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in __flush_dcache_aliases() 269 flush_dcache_mmap_unlock(mapping); in __flush_dcache_aliases() 277 struct address_space *mapping; in __sync_icache_dcache() local 288 mapping = page_mapping(page); in __sync_icache_dcache() 290 mapping = NULL; in __sync_icache_dcache() 293 __flush_dcache_page(mapping, page); in __sync_icache_dcache() [all …]
|
D | dma-mapping.c | 1048 static int extend_iommu_mapping(struct dma_iommu_mapping *mapping); 1050 static inline dma_addr_t __alloc_iova(struct dma_iommu_mapping *mapping, in __alloc_iova() argument 1056 size_t mapping_size = mapping->bits << PAGE_SHIFT; in __alloc_iova() 1067 spin_lock_irqsave(&mapping->lock, flags); in __alloc_iova() 1068 for (i = 0; i < mapping->nr_bitmaps; i++) { in __alloc_iova() 1069 start = bitmap_find_next_zero_area(mapping->bitmaps[i], in __alloc_iova() 1070 mapping->bits, 0, count, align); in __alloc_iova() 1072 if (start > mapping->bits) in __alloc_iova() 1075 bitmap_set(mapping->bitmaps[i], start, count); in __alloc_iova() 1084 if (i == mapping->nr_bitmaps) { in __alloc_iova() [all …]
|
D | fault-armv.c | 132 make_coherent(struct address_space *mapping, struct vm_area_struct *vma, in make_coherent() argument 148 flush_dcache_mmap_lock(mapping); in make_coherent() 149 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in make_coherent() 162 flush_dcache_mmap_unlock(mapping); in make_coherent() 184 struct address_space *mapping; in update_mmu_cache() local 198 mapping = page_mapping(page); in update_mmu_cache() 200 __flush_dcache_page(mapping, page); in update_mmu_cache() 201 if (mapping) { in update_mmu_cache() 203 make_coherent(mapping, vma, addr, ptep, pfn); in update_mmu_cache()
|
/arch/unicore32/mm/ |
D | flush.c | 61 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 77 struct address_space *mapping; in flush_dcache_page() local 86 mapping = page_mapping(page); in flush_dcache_page() 88 if (mapping && !mapping_mapped(mapping)) in flush_dcache_page() 91 __flush_dcache_page(mapping, page); in flush_dcache_page() 92 if (mapping) in flush_dcache_page()
|
/arch/nios2/mm/ |
D | cacheflush.c | 73 static void flush_aliases(struct address_space *mapping, struct page *page) in flush_aliases() argument 81 flush_dcache_mmap_lock(mapping); in flush_aliases() 82 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in flush_aliases() 94 flush_dcache_mmap_unlock(mapping); in flush_aliases() 160 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 174 struct address_space *mapping; in flush_dcache_page() local 183 mapping = page_mapping(page); in flush_dcache_page() 186 if (mapping && !mapping_mapped(mapping)) { in flush_dcache_page() 189 __flush_dcache_page(mapping, page); in flush_dcache_page() 190 if (mapping) { in flush_dcache_page() [all …]
|
/arch/m32r/include/asm/ |
D | cacheflush.h | 17 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 18 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument 39 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 40 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument 53 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 54 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
/arch/arm64/include/asm/ |
D | cacheflush.h | 132 #define flush_dcache_mmap_lock(mapping) \ argument 133 spin_lock_irq(&(mapping)->tree_lock) 134 #define flush_dcache_mmap_unlock(mapping) \ argument 135 spin_unlock_irq(&(mapping)->tree_lock)
|
/arch/powerpc/boot/dts/fsl/ |
D | t4240si-pre.dtsi | 93 fsl,portid-mapping = <0x80000000>; 100 fsl,portid-mapping = <0x80000000>; 107 fsl,portid-mapping = <0x80000000>; 114 fsl,portid-mapping = <0x80000000>; 121 fsl,portid-mapping = <0x40000000>; 128 fsl,portid-mapping = <0x40000000>; 135 fsl,portid-mapping = <0x40000000>; 142 fsl,portid-mapping = <0x40000000>; 149 fsl,portid-mapping = <0x20000000>; 156 fsl,portid-mapping = <0x20000000>; [all …]
|
D | p4080si-pre.dtsi | 99 fsl,portid-mapping = <0x80000000>; 109 fsl,portid-mapping = <0x40000000>; 119 fsl,portid-mapping = <0x20000000>; 129 fsl,portid-mapping = <0x10000000>; 139 fsl,portid-mapping = <0x08000000>; 149 fsl,portid-mapping = <0x04000000>; 159 fsl,portid-mapping = <0x02000000>; 169 fsl,portid-mapping = <0x01000000>;
|
/arch/c6x/platforms/ |
D | megamod-pic.c | 179 int *mapping, int size) in parse_priority_map() argument 195 mapping[i] = val; in parse_priority_map() 205 int mapping[NR_MUX_OUTPUTS]; in init_megamod_pic() local 233 for (i = 0; i < ARRAY_SIZE(mapping); i++) in init_megamod_pic() 234 mapping[i] = IRQ_UNMAPPED; in init_megamod_pic() 236 parse_priority_map(pic, mapping, ARRAY_SIZE(mapping)); in init_megamod_pic() 274 mapping[hwirq - 4] = i; in init_megamod_pic() 292 if (mapping[i] != IRQ_UNMAPPED) { in init_megamod_pic() 294 np->full_name, mapping[i], i + 4); in init_megamod_pic() 295 set_megamod_mux(pic, mapping[i], i); in init_megamod_pic()
|
/arch/unicore32/include/asm/ |
D | cacheflush.h | 182 #define flush_dcache_mmap_lock(mapping) \ argument 183 spin_lock_irq(&(mapping)->tree_lock) 184 #define flush_dcache_mmap_unlock(mapping) \ argument 185 spin_unlock_irq(&(mapping)->tree_lock)
|
/arch/arm/include/asm/ |
D | device.h | 18 struct dma_iommu_mapping *mapping; member 32 #define to_dma_iommu_mapping(dev) ((dev)->archdata.mapping)
|
D | dma-iommu.h | 30 void arm_iommu_release_mapping(struct dma_iommu_mapping *mapping); 33 struct dma_iommu_mapping *mapping);
|
/arch/parisc/include/asm/ |
D | cacheflush.h | 77 #define flush_dcache_mmap_lock(mapping) \ argument 78 spin_lock_irq(&(mapping)->tree_lock) 79 #define flush_dcache_mmap_unlock(mapping) \ argument 80 spin_unlock_irq(&(mapping)->tree_lock)
|
/arch/cris/include/asm/ |
D | cacheflush.h | 17 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 18 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
/arch/metag/include/asm/ |
D | cacheflush.h | 67 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 68 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument 95 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 96 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
/arch/nios2/include/asm/ |
D | cacheflush.h | 49 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 50 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
/arch/xtensa/mm/ |
D | cache.c | 137 struct address_space *mapping = page_mapping(page); in flush_dcache_page() local 145 if (mapping && !mapping_mapped(mapping)) { in flush_dcache_page() 164 if (!alias && !mapping) in flush_dcache_page() 176 if (mapping) in flush_dcache_page()
|
/arch/hexagon/include/asm/ |
D | cacheflush.h | 48 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 49 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
/arch/c6x/include/asm/ |
D | cacheflush.h | 34 #define flush_dcache_mmap_lock(mapping) do {} while (0) argument 35 #define flush_dcache_mmap_unlock(mapping) do {} while (0) argument
|
/arch/ia64/include/asm/ |
D | cacheflush.h | 34 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 35 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|
/arch/mips/pci/ |
D | pci-rc32434.c | 155 rc32434_pci->pcilba[0].mapping = (unsigned int) (PCI_ADDR_START); in rc32434_pcibridge_init() 163 rc32434_pci->pcilba[1].mapping = 0x60000000; in rc32434_pcibridge_init() 170 rc32434_pci->pcilba[2].mapping = 0x18FFFFFF; in rc32434_pcibridge_init() 179 rc32434_pci->pcilba[3].mapping = 0x18800000; in rc32434_pcibridge_init()
|
/arch/sh/mm/ |
D | cache-sh7705.c | 139 struct address_space *mapping = page_mapping(page); in sh7705_flush_dcache_page() local 141 if (mapping && !mapping_mapped(mapping)) in sh7705_flush_dcache_page()
|
/arch/score/include/asm/ |
D | cacheflush.h | 23 #define flush_dcache_mmap_lock(mapping) do {} while (0) argument 24 #define flush_dcache_mmap_unlock(mapping) do {} while (0) argument
|
/arch/alpha/include/asm/ |
D | cacheflush.h | 14 #define flush_dcache_mmap_lock(mapping) do { } while (0) argument 15 #define flush_dcache_mmap_unlock(mapping) do { } while (0) argument
|