/arch/x86/lib/ |
D | usercopy_64.c | 17 unsigned long __clear_user(void __user *addr, unsigned long size) in __clear_user() argument 44 : [size8] "=&c"(size), [dst] "=&D" (__d0) in __clear_user() 45 : [size1] "r"(size & 7), "[size8]" (size / 8), "[dst]"(addr)); in __clear_user() 47 return size; in __clear_user() 69 static void clean_cache_range(void *addr, size_t size) in clean_cache_range() argument 73 void *vend = addr + size; in clean_cache_range() 81 void arch_wb_cache_pmem(void *addr, size_t size) in arch_wb_cache_pmem() argument 83 clean_cache_range(addr, size); in arch_wb_cache_pmem() 87 long __copy_user_flushcache(void *dst, const void __user *src, unsigned size) in __copy_user_flushcache() argument 90 long rc = __copy_user_nocache(dst, src, size, 0); in __copy_user_flushcache() [all …]
|
/arch/arm/boot/dts/ |
D | aspeed-bmc-facebook-cmm.dts | 334 #size-cells = <0>; 341 #size-cells = <0>; 347 #size-cells = <0>; 353 #size-cells = <0>; 358 #size-cells = <0>; 363 #size-cells = <0>; 368 #size-cells = <0>; 373 #size-cells = <0>; 378 #size-cells = <0>; 383 #size-cells = <0>; [all …]
|
D | bcm2837.dtsi | 40 #size-cells = <0>; 43 /* Source for d/i-cache-line-size and d/i-cache-sets 47 * Source for d/i-cache-size 56 d-cache-size = <0x8000>; 57 d-cache-line-size = <64>; 58 d-cache-sets = <128>; // 32KiB(size)/64(line-size)=512ways/4-way set 59 i-cache-size = <0x8000>; 60 i-cache-line-size = <64>; 61 i-cache-sets = <256>; // 32KiB(size)/64(line-size)=512ways/2-way set 71 d-cache-size = <0x8000>; [all …]
|
/arch/s390/lib/ |
D | find.c | 17 unsigned long find_first_bit_inv(const unsigned long *addr, unsigned long size) in find_first_bit_inv() argument 23 while (size & ~(BITS_PER_LONG - 1)) { in find_first_bit_inv() 27 size -= BITS_PER_LONG; in find_first_bit_inv() 29 if (!size) in find_first_bit_inv() 31 tmp = (*p) & (~0UL << (BITS_PER_LONG - size)); in find_first_bit_inv() 33 return result + size; /* Nope. */ in find_first_bit_inv() 39 unsigned long find_next_bit_inv(const unsigned long *addr, unsigned long size, in find_next_bit_inv() argument 46 if (offset >= size) in find_next_bit_inv() 47 return size; in find_next_bit_inv() 48 size -= result; in find_next_bit_inv() [all …]
|
D | uaccess.c | 105 unsigned long size) in copy_from_user_mvcos() argument 129 : "+a" (size), "+a" (ptr), "+a" (x), "+a" (tmp1), "=a" (tmp2) in copy_from_user_mvcos() 131 return size; in copy_from_user_mvcos() 135 unsigned long size) in copy_from_user_mvcp() argument 165 : "+a" (size), "+a" (ptr), "+a" (x), "+a" (tmp1), "=a" (tmp2) in copy_from_user_mvcp() 168 return size; in copy_from_user_mvcp() 180 unsigned long size) in copy_to_user_mvcos() argument 204 : "+a" (size), "+a" (ptr), "+a" (x), "+a" (tmp1), "=a" (tmp2) in copy_to_user_mvcos() 206 return size; in copy_to_user_mvcos() 210 unsigned long size) in copy_to_user_mvcs() argument [all …]
|
/arch/arm64/boot/dts/amazon/ |
D | alpine-v3.dtsi | 17 #size-cells = <2>; 21 #size-cells = <0>; 28 d-cache-size = <0x8000>; 29 d-cache-line-size = <64>; 31 i-cache-size = <0xc000>; 32 i-cache-line-size = <64>; 42 d-cache-size = <0x8000>; 43 d-cache-line-size = <64>; 45 i-cache-size = <0xc000>; 46 i-cache-line-size = <64>; [all …]
|
/arch/x86/kernel/cpu/mtrr/ |
D | if.c | 38 mtrr_file_add(unsigned long base, unsigned long size, in mtrr_file_add() argument 52 if ((base & (PAGE_SIZE - 1)) || (size & (PAGE_SIZE - 1))) in mtrr_file_add() 55 size >>= PAGE_SHIFT; in mtrr_file_add() 57 reg = mtrr_add_page(base, size, type, true); in mtrr_file_add() 64 mtrr_file_del(unsigned long base, unsigned long size, in mtrr_file_del() argument 71 if ((base & (PAGE_SIZE - 1)) || (size & (PAGE_SIZE - 1))) in mtrr_file_del() 74 size >>= PAGE_SHIFT; in mtrr_file_del() 76 reg = mtrr_del_page(-1, base, size); in mtrr_file_del() 98 unsigned long long base, size; in mtrr_write() local 133 size = simple_strtoull(ptr + 5, &ptr, 0); in mtrr_write() [all …]
|
/arch/openrisc/include/asm/ |
D | uaccess.h | 51 static inline int __range_ok(unsigned long addr, unsigned long size) in __range_ok() argument 55 return size <= fs && addr <= (fs - size); in __range_ok() 58 #define access_ok(addr, size) \ argument 61 __range_ok((unsigned long)(addr), (size)); \ 94 #define __put_user_nocheck(x, ptr, size) \ argument 97 __put_user_size((x), (ptr), (size), __pu_err); \ 101 #define __put_user_check(x, ptr, size) \ argument 105 if (access_ok(__pu_addr, size)) \ 106 __put_user_size((x), __pu_addr, (size), __pu_err); \ 110 #define __put_user_size(x, ptr, size, retval) \ argument [all …]
|
/arch/mips/mm/ |
D | c-r3k.c | 30 unsigned long flags, status, dummy, size; in r3k_cache_size() local 45 size = 0; in r3k_cache_size() 47 for (size = 128; size <= 0x40000; size <<= 1) in r3k_cache_size() 48 *(p + size) = 0; in r3k_cache_size() 50 for (size = 128; in r3k_cache_size() 51 (size <= 0x40000) && (*(p + size) == 0); in r3k_cache_size() 52 size <<= 1) in r3k_cache_size() 54 if (size > 0x40000) in r3k_cache_size() 55 size = 0; in r3k_cache_size() 60 return size * sizeof(*p); in r3k_cache_size() [all …]
|
D | dma-noncoherent.c | 47 void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent() argument 49 dma_cache_wback_inv((unsigned long)page_address(page), size); in arch_dma_prep_coherent() 52 void *arch_dma_set_uncached(void *addr, size_t size) in arch_dma_set_uncached() argument 57 static inline void dma_sync_virt_for_device(void *addr, size_t size, in dma_sync_virt_for_device() argument 62 dma_cache_wback((unsigned long)addr, size); in dma_sync_virt_for_device() 65 dma_cache_inv((unsigned long)addr, size); in dma_sync_virt_for_device() 68 dma_cache_wback_inv((unsigned long)addr, size); in dma_sync_virt_for_device() 75 static inline void dma_sync_virt_for_cpu(void *addr, size_t size, in dma_sync_virt_for_cpu() argument 83 dma_cache_inv((unsigned long)addr, size); in dma_sync_virt_for_cpu() 95 static inline void dma_sync_phys(phys_addr_t paddr, size_t size, in dma_sync_phys() argument [all …]
|
/arch/powerpc/boot/ |
D | simple_alloc.c | 24 unsigned long size; member 37 static void *simple_malloc(unsigned long size) in simple_malloc() argument 42 if (size == 0) in simple_malloc() 45 size = _ALIGN_UP(size, alloc_min); in simple_malloc() 49 if (size <= space_left) { in simple_malloc() 51 p->size = size; in simple_malloc() 53 next_base += size; in simple_malloc() 54 space_left -= size; in simple_malloc() 60 else if (!(p->flags & ENTRY_IN_USE) && (size <= p->size)) { in simple_malloc() 97 static void *simple_realloc(void *ptr, unsigned long size) in simple_realloc() argument [all …]
|
/arch/powerpc/lib/ |
D | rheap.c | 139 blk->size = 0; in get_slot() 157 int size; in attach_free_block() local 162 size = blkn->size; in attach_free_block() 164 e = s + size; in attach_free_block() 176 be = bs + blk->size; in attach_free_block() 193 if (before && s != (before->start + before->size)) in attach_free_block() 215 before->size += size; in attach_free_block() 221 after->start -= size; in attach_free_block() 222 after->size += size; in attach_free_block() 227 before->size += size + after->size; in attach_free_block() [all …]
|
/arch/parisc/kernel/ |
D | pci-dma.c | 78 unsigned long size, unsigned long *paddr_ptr) in map_pte_uncached() argument 84 end = vaddr + size; in map_pte_uncached() 105 unsigned long size, unsigned long *paddr_ptr) in map_pmd_uncached() argument 111 end = vaddr + size; in map_pmd_uncached() 127 static inline int map_uncached_pages(unsigned long vaddr, unsigned long size, in map_uncached_pages() argument 131 unsigned long end = vaddr + size; in map_uncached_pages() 154 unsigned long size) in unmap_uncached_pte() argument 169 end = vaddr + size; in unmap_uncached_pte() 190 unsigned long size) in unmap_uncached_pmd() argument 205 end = vaddr + size; in unmap_uncached_pmd() [all …]
|
/arch/arm/mm/ |
D | ioremap.c | 49 size_t size, unsigned int mtype) in find_static_vm_paddr() argument 62 paddr + size - 1 > vm->phys_addr + vm->size - 1) in find_static_vm_paddr() 83 if (vm->addr <= vaddr && vm->addr + vm->size > vaddr) in find_static_vm_vaddr() 142 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument 144 unsigned long addr = virt, end = virt + (size & ~(SZ_1M - 1)); in unmap_area_sections() 184 size_t size, const struct mem_type *type) in remap_area_sections() argument 186 unsigned long addr = virt, end = virt + size; in remap_area_sections() 193 unmap_area_sections(virt, size); in remap_area_sections() 211 size_t size, const struct mem_type *type) in remap_area_supersections() argument 213 unsigned long addr = virt, end = virt + size; in remap_area_supersections() [all …]
|
D | dma-mapping.c | 43 size_t size; member 53 size_t size; member 126 unsigned long offset, size_t size, enum dma_data_direction dir, in arm_dma_map_page() argument 130 __dma_page_cpu_to_dev(page, offset, size, dir); in arm_dma_map_page() 135 unsigned long offset, size_t size, enum dma_data_direction dir, in arm_coherent_dma_map_page() argument 156 size_t size, enum dma_data_direction dir, unsigned long attrs) in arm_dma_unmap_page() argument 160 handle & ~PAGE_MASK, size, dir); in arm_dma_unmap_page() 164 dma_addr_t handle, size_t size, enum dma_data_direction dir) in arm_dma_sync_single_for_cpu() argument 168 __dma_page_dev_to_cpu(page, offset, size, dir); in arm_dma_sync_single_for_cpu() 172 dma_addr_t handle, size_t size, enum dma_data_direction dir) in arm_dma_sync_single_for_device() argument [all …]
|
/arch/um/kernel/ |
D | initrd.c | 15 static int load_initrd(char *filename, void *buf, int size); 20 long long size; in read_initrd() local 26 err = os_file_size(initrd, &size); in read_initrd() 34 if (size == 0) { in read_initrd() 39 area = memblock_alloc(size, SMP_CACHE_BYTES); in read_initrd() 41 panic("%s: Failed to allocate %llu bytes\n", __func__, size); in read_initrd() 43 if (load_initrd(initrd, area, size) == -1) in read_initrd() 47 initrd_end = initrd_start + size; in read_initrd() 63 static int load_initrd(char *filename, void *buf, int size) in load_initrd() argument 73 n = os_read_file(fd, buf, size); in load_initrd() [all …]
|
/arch/um/include/asm/ |
D | uaccess.h | 12 #define __under_task_size(addr, size) \ argument 14 (((unsigned long) (addr) + (size)) < TASK_SIZE)) 16 #define __access_ok_vsyscall(addr, size) \ argument 18 ((unsigned long) (addr) + (size) <= FIXADDR_USER_END) && \ 19 ((unsigned long) (addr) + (size) >= (unsigned long)(addr))) 21 #define __addr_range_nowrap(addr, size) \ argument 22 ((unsigned long) (addr) <= ((unsigned long) (addr) + (size))) 29 static inline int __access_ok(unsigned long addr, unsigned long size); 41 static inline int __access_ok(unsigned long addr, unsigned long size) in __access_ok() argument 43 return __addr_range_nowrap(addr, size) && in __access_ok() [all …]
|
/arch/xtensa/include/asm/ |
D | uaccess.h | 41 #define __user_ok(addr, size) \ argument 42 (((size) <= TASK_SIZE)&&((addr) <= TASK_SIZE-(size))) 43 #define __access_ok(addr, size) (__kernel_ok || __user_ok((addr), (size))) argument 44 #define access_ok(addr, size) __access_ok((unsigned long)(addr), (size)) argument 77 #define __put_user_nocheck(x, ptr, size) \ argument 80 __put_user_size((x), (ptr), (size), __pu_err); \ 84 #define __put_user_check(x, ptr, size) \ argument 88 if (access_ok(__pu_addr, size)) \ 89 __put_user_size((x), __pu_addr, (size), __pu_err); \ 93 #define __put_user_size(x, ptr, size, retval) \ argument [all …]
|
/arch/parisc/include/asm/ |
D | floppy.h | 41 #define fd_dma_mem_alloc(size) SW._dma_mem_alloc(size) argument 42 #define fd_dma_setup(addr, size, mode, io) SW._dma_setup(addr, size, mode, io) argument 154 static unsigned long dma_mem_alloc(unsigned long size) in dma_mem_alloc() argument 156 return __get_dma_pages(GFP_KERNEL, get_order(size)); in dma_mem_alloc() 160 static unsigned long vdma_mem_alloc(unsigned long size) in vdma_mem_alloc() argument 162 return (unsigned long) vmalloc(size); in vdma_mem_alloc() 166 #define nodma_mem_alloc(size) vdma_mem_alloc(size) argument 168 static void _fd_dma_mem_free(unsigned long addr, unsigned long size) in _fd_dma_mem_free() argument 173 free_pages(addr, get_order(size)); in _fd_dma_mem_free() 176 #define fd_dma_mem_free(addr, size) _fd_dma_mem_free(addr, size) argument [all …]
|
/arch/arm64/boot/dts/marvell/ |
D | armada-ap807-quad.dtsi | 16 #size-cells = <0>; 25 i-cache-size = <0xc000>; 26 i-cache-line-size = <64>; 28 d-cache-size = <0x8000>; 29 d-cache-line-size = <64>; 40 i-cache-size = <0xc000>; 41 i-cache-line-size = <64>; 43 d-cache-size = <0x8000>; 44 d-cache-line-size = <64>; 55 i-cache-size = <0xc000>; [all …]
|
D | armada-ap806-quad.dtsi | 16 #size-cells = <0>; 25 i-cache-size = <0xc000>; 26 i-cache-line-size = <64>; 28 d-cache-size = <0x8000>; 29 d-cache-line-size = <64>; 40 i-cache-size = <0xc000>; 41 i-cache-line-size = <64>; 43 d-cache-size = <0x8000>; 44 d-cache-line-size = <64>; 55 i-cache-size = <0xc000>; [all …]
|
/arch/c6x/mm/ |
D | dma-coherent.c | 74 void *arch_dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, in arch_dma_alloc() argument 81 if (!dma_size || !size) in arch_dma_alloc() 84 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in arch_dma_alloc() 102 void arch_dma_free(struct device *dev, size_t size, void *vaddr, in arch_dma_free() argument 107 if (!dma_size || !size) in arch_dma_free() 110 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in arch_dma_free() 118 void __init coherent_mem_init(phys_addr_t start, u32 size) in coherent_mem_init() argument 120 if (!size) in coherent_mem_init() 125 start, size); in coherent_mem_init() 128 dma_size = size; in coherent_mem_init() [all …]
|
/arch/powerpc/mm/ |
D | ioremap.c | 12 void __iomem *ioremap(phys_addr_t addr, unsigned long size) in ioremap() argument 18 return iowa_ioremap(addr, size, prot, caller); in ioremap() 19 return __ioremap_caller(addr, size, prot, caller); in ioremap() 23 void __iomem *ioremap_wc(phys_addr_t addr, unsigned long size) in ioremap_wc() argument 29 return iowa_ioremap(addr, size, prot, caller); in ioremap_wc() 30 return __ioremap_caller(addr, size, prot, caller); in ioremap_wc() 34 void __iomem *ioremap_coherent(phys_addr_t addr, unsigned long size) in ioremap_coherent() argument 40 return iowa_ioremap(addr, size, prot, caller); in ioremap_coherent() 41 return __ioremap_caller(addr, size, prot, caller); in ioremap_coherent() 44 void __iomem *ioremap_prot(phys_addr_t addr, unsigned long size, unsigned long flags) in ioremap_prot() argument [all …]
|
/arch/x86/include/asm/ |
D | uaccess_64.h | 50 raw_copy_from_user(void *dst, const void __user *src, unsigned long size) in raw_copy_from_user() argument 52 return copy_user_generic(dst, (__force void *)src, size); in raw_copy_from_user() 56 raw_copy_to_user(void __user *dst, const void *src, unsigned long size) in raw_copy_to_user() argument 58 return copy_user_generic((__force void *)dst, src, size); in raw_copy_to_user() 62 unsigned long raw_copy_in_user(void __user *dst, const void __user *src, unsigned long size) in raw_copy_in_user() argument 65 (__force void *)src, size); in raw_copy_in_user() 69 unsigned size, int zerorest); 71 extern long __copy_user_flushcache(void *dst, const void __user *src, unsigned size); 77 unsigned size) in __copy_from_user_inatomic_nocache() argument 79 kasan_check_write(dst, size); in __copy_from_user_inatomic_nocache() [all …]
|
/arch/sh/mm/ |
D | pmb.c | 40 unsigned long size; member 55 unsigned long size; member 58 { .size = SZ_512M, .flag = PMB_SZ_512M, }, 59 { .size = SZ_128M, .flag = PMB_SZ_128M, }, 60 { .size = SZ_64M, .flag = PMB_SZ_64M, }, 61 { .size = SZ_16M, .flag = PMB_SZ_16M, }, 132 return (b->vpn == (a->vpn + a->size)) && in pmb_can_merge() 133 (b->ppn == (a->ppn + a->size)) && in pmb_can_merge() 138 unsigned long size) in pmb_mapping_exists() argument 156 if ((vaddr < pmbe->vpn) || (vaddr >= (pmbe->vpn + pmbe->size))) in pmb_mapping_exists() [all …]
|