• Home
  • Raw
  • Download

Lines Matching refs:dev

36 static inline int cpu_is_noncoherent_r10000(struct device *dev)  in cpu_is_noncoherent_r10000()  argument
38 return !plat_device_is_coherent(dev) && in cpu_is_noncoherent_r10000()
43 static gfp_t massage_gfp_flags(const struct device *dev, gfp_t gfp) in massage_gfp_flags() argument
49 if (dev == NULL) in massage_gfp_flags()
51 else if (dev->coherent_dma_mask < DMA_BIT_MASK(24)) in massage_gfp_flags()
56 if (dev->coherent_dma_mask < DMA_BIT_MASK(32)) in massage_gfp_flags()
68 void *dma_alloc_noncoherent(struct device *dev, size_t size, in dma_alloc_noncoherent() argument
73 gfp = massage_gfp_flags(dev, gfp); in dma_alloc_noncoherent()
79 *dma_handle = plat_map_dma_mem(dev, ret, size); in dma_alloc_noncoherent()
87 void *dma_alloc_coherent(struct device *dev, size_t size, in dma_alloc_coherent() argument
92 gfp = massage_gfp_flags(dev, gfp); in dma_alloc_coherent()
98 *dma_handle = plat_map_dma_mem(dev, ret, size); in dma_alloc_coherent()
100 if (!plat_device_is_coherent(dev)) { in dma_alloc_coherent()
111 void dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, in dma_free_noncoherent() argument
114 plat_unmap_dma_mem(dev, dma_handle); in dma_free_noncoherent()
120 void dma_free_coherent(struct device *dev, size_t size, void *vaddr, in dma_free_coherent() argument
125 plat_unmap_dma_mem(dev, dma_handle); in dma_free_coherent()
127 if (!plat_device_is_coherent(dev)) in dma_free_coherent()
156 dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size, in dma_map_single() argument
161 if (!plat_device_is_coherent(dev)) in dma_map_single()
164 return plat_map_dma_mem(dev, ptr, size); in dma_map_single()
169 void dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, in dma_unmap_single() argument
172 if (cpu_is_noncoherent_r10000(dev)) in dma_unmap_single()
176 plat_unmap_dma_mem(dev, dma_addr); in dma_unmap_single()
181 int dma_map_sg(struct device *dev, struct scatterlist *sg, int nents, in dma_map_sg() argument
192 if (!plat_device_is_coherent(dev) && addr) in dma_map_sg()
194 sg->dma_address = plat_map_dma_mem(dev, in dma_map_sg()
203 dma_addr_t dma_map_page(struct device *dev, struct page *page, in dma_map_page() argument
208 if (!plat_device_is_coherent(dev)) { in dma_map_page()
215 return plat_map_dma_mem_page(dev, page) + offset; in dma_map_page()
220 void dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, in dma_unmap_page() argument
225 if (!plat_device_is_coherent(dev) && direction != DMA_TO_DEVICE) { in dma_unmap_page()
232 plat_unmap_dma_mem(dev, dma_address); in dma_unmap_page()
237 void dma_unmap_sg(struct device *dev, struct scatterlist *sg, int nhwentries, in dma_unmap_sg() argument
246 if (!plat_device_is_coherent(dev) && in dma_unmap_sg()
252 plat_unmap_dma_mem(dev, sg->dma_address); in dma_unmap_sg()
258 void dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, in dma_sync_single_for_cpu() argument
263 if (cpu_is_noncoherent_r10000(dev)) { in dma_sync_single_for_cpu()
273 void dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, in dma_sync_single_for_device() argument
278 plat_extra_sync_for_device(dev); in dma_sync_single_for_device()
279 if (!plat_device_is_coherent(dev)) { in dma_sync_single_for_device()
289 void dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, in dma_sync_single_range_for_cpu() argument
294 if (cpu_is_noncoherent_r10000(dev)) { in dma_sync_single_range_for_cpu()
304 void dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, in dma_sync_single_range_for_device() argument
309 plat_extra_sync_for_device(dev); in dma_sync_single_range_for_device()
310 if (!plat_device_is_coherent(dev)) { in dma_sync_single_range_for_device()
320 void dma_sync_sg_for_cpu(struct device *dev, struct scatterlist *sg, int nelems, in dma_sync_sg_for_cpu() argument
329 if (cpu_is_noncoherent_r10000(dev)) in dma_sync_sg_for_cpu()
337 void dma_sync_sg_for_device(struct device *dev, struct scatterlist *sg, int nelems, in dma_sync_sg_for_device() argument
346 if (!plat_device_is_coherent(dev)) in dma_sync_sg_for_device()
354 int dma_mapping_error(struct device *dev, dma_addr_t dma_addr) in dma_mapping_error() argument
356 return plat_dma_mapping_error(dev, dma_addr); in dma_mapping_error()
361 int dma_supported(struct device *dev, u64 mask) in dma_supported() argument
363 return plat_dma_supported(dev, mask); in dma_supported()
368 int dma_is_consistent(struct device *dev, dma_addr_t dma_addr) in dma_is_consistent() argument
370 return plat_device_is_coherent(dev); in dma_is_consistent()
375 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
380 plat_extra_sync_for_device(dev); in dma_cache_sync()
381 if (!plat_device_is_coherent(dev)) in dma_cache_sync()