Lines Matching refs:size
207 dma_contiguous_early_fixup(phys_addr_t base, unsigned long size) in dma_contiguous_early_fixup() argument
228 int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, in dma_contiguous_reserve_area() argument
234 ret = cma_declare_contiguous(base, size, limit, 0, 0, fixed, in dma_contiguous_reserve_area()
283 static struct page *cma_alloc_aligned(struct cma *cma, size_t size, gfp_t gfp) in cma_alloc_aligned() argument
285 unsigned int align = min(get_order(size), CONFIG_CMA_ALIGNMENT); in cma_alloc_aligned()
287 return cma_alloc(cma, size >> PAGE_SHIFT, align, gfp & __GFP_NOWARN); in cma_alloc_aligned()
305 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp) in dma_alloc_contiguous() argument
315 return cma_alloc_aligned(dev->cma_area, size, gfp); in dma_alloc_contiguous()
316 if (size <= PAGE_SIZE) in dma_alloc_contiguous()
325 page = cma_alloc_aligned(cma, size, gfp); in dma_alloc_contiguous()
334 return cma_alloc_aligned(dma_contiguous_default_area, size, gfp); in dma_alloc_contiguous()
348 void dma_free_contiguous(struct device *dev, struct page *page, size_t size) in dma_free_contiguous() argument
350 unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT; in dma_free_contiguous()
370 __free_pages(page, get_order(size)); in dma_free_contiguous()
418 if (!IS_ALIGNED(rmem->base | rmem->size, CMA_MIN_ALIGNMENT_BYTES)) { in rmem_cma_setup()
423 err = cma_init_reserved_mem(rmem->base, rmem->size, 0, rmem->name, &cma); in rmem_cma_setup()
429 dma_contiguous_early_fixup(rmem->base, rmem->size); in rmem_cma_setup()
438 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_cma_setup()