/arch/x86/lib/ |
D | usercopy_64.c | 28 static void clean_cache_range(void *addr, size_t size) in clean_cache_range() argument 32 void *vend = addr + size; in clean_cache_range() 40 void arch_wb_cache_pmem(void *addr, size_t size) in arch_wb_cache_pmem() argument 42 clean_cache_range(addr, size); in arch_wb_cache_pmem() 46 long __copy_user_flushcache(void *dst, const void __user *src, unsigned size) in __copy_user_flushcache() argument 52 rc = __copy_user_nocache(dst, src, size); in __copy_user_flushcache() 63 if (size < 8) { in __copy_user_flushcache() 64 if (!IS_ALIGNED(dest, 4) || size != 4) in __copy_user_flushcache() 65 clean_cache_range(dst, size); in __copy_user_flushcache() 73 if (size > flushed && !IS_ALIGNED(size - flushed, 8)) in __copy_user_flushcache() [all …]
|
/arch/arm/boot/dts/aspeed/ |
D | aspeed-bmc-facebook-cmm.dts | 334 #size-cells = <0>; 341 #size-cells = <0>; 347 #size-cells = <0>; 353 #size-cells = <0>; 358 #size-cells = <0>; 363 #size-cells = <0>; 368 #size-cells = <0>; 373 #size-cells = <0>; 378 #size-cells = <0>; 383 #size-cells = <0>; [all …]
|
D | aspeed-bmc-facebook-fuji.dts | 219 data-size = <16>; 239 #size-cells = <0>; 245 #size-cells = <0>; 252 #size-cells = <0>; 259 #size-cells = <0>; 265 #size-cells = <0>; 271 #size-cells = <0>; 277 #size-cells = <0>; 283 #size-cells = <0>; 289 #size-cells = <0>; [all …]
|
D | aspeed-bmc-facebook-minipack.dts | 249 #size-cells = <1>; 299 #size-cells = <1>; 350 #size-cells = <0>; 356 #size-cells = <0>; 362 #size-cells = <0>; 368 #size-cells = <0>; 374 #size-cells = <0>; 380 #size-cells = <0>; 386 #size-cells = <0>; 392 #size-cells = <0>; [all …]
|
D | ibm-power10-quad.dtsi | 11 #size-cells = <0>; 16 #size-cells = <1>; 28 #size-cells = <0>; 39 #size-cells = <0>; 44 #size-cells = <1>; 56 #size-cells = <0>; 67 #size-cells = <0>; 72 #size-cells = <1>; 84 #size-cells = <0>; 95 #size-cells = <0>; [all …]
|
D | aspeed-bmc-facebook-cloudripper.dts | 88 data-size = <16>; 138 #size-cells = <0>; 144 #size-cells = <0>; 150 #size-cells = <0>; 156 #size-cells = <0>; 162 #size-cells = <0>; 168 #size-cells = <0>; 174 #size-cells = <0>; 180 #size-cells = <0>; 186 #size-cells = <0>; [all …]
|
D | ibm-power10-dual.dtsi | 8 #size-cells = <0>; 15 #size-cells = <1>; 27 #size-cells = <0>; 31 #size-cells = <0>; 37 #size-cells = <0>; 43 #size-cells = <0>; 49 #size-cells = <0>; 55 #size-cells = <0>; 61 #size-cells = <0>; 67 #size-cells = <0>; [all …]
|
/arch/s390/lib/ |
D | find.c | 17 unsigned long find_first_bit_inv(const unsigned long *addr, unsigned long size) in find_first_bit_inv() argument 23 while (size & ~(BITS_PER_LONG - 1)) { in find_first_bit_inv() 27 size -= BITS_PER_LONG; in find_first_bit_inv() 29 if (!size) in find_first_bit_inv() 31 tmp = (*p) & (~0UL << (BITS_PER_LONG - size)); in find_first_bit_inv() 33 return result + size; /* Nope. */ in find_first_bit_inv() 39 unsigned long find_next_bit_inv(const unsigned long *addr, unsigned long size, in find_next_bit_inv() argument 46 if (offset >= size) in find_next_bit_inv() 47 return size; in find_next_bit_inv() 48 size -= result; in find_next_bit_inv() [all …]
|
/arch/arm64/boot/dts/amd/ |
D | amd-seattle-cpus.dtsi | 6 #size-cells = <0x0>; 49 i-cache-size = <0xC000>; 50 i-cache-line-size = <64>; 52 d-cache-size = <0x8000>; 53 d-cache-line-size = <64>; 65 i-cache-size = <0xC000>; 66 i-cache-line-size = <64>; 68 d-cache-size = <0x8000>; 69 d-cache-line-size = <64>; 80 i-cache-size = <0xC000>; [all …]
|
/arch/arm/boot/dts/broadcom/ |
D | bcm2836.dtsi | 40 #size-cells = <0>; 43 /* Source for d/i-cache-line-size and d/i-cache-sets 47 * Source for d/i-cache-size 56 d-cache-size = <0x8000>; 57 d-cache-line-size = <64>; 58 d-cache-sets = <128>; // 32KiB(size)/64(line-size)=512ways/4-way set 59 i-cache-size = <0x8000>; 60 i-cache-line-size = <32>; 61 i-cache-sets = <512>; // 32KiB(size)/32(line-size)=1024ways/2-way set 70 d-cache-size = <0x8000>; [all …]
|
D | bcm2837.dtsi | 39 #size-cells = <0>; 42 /* Source for d/i-cache-line-size and d/i-cache-sets 46 * Source for d/i-cache-size 55 d-cache-size = <0x8000>; 56 d-cache-line-size = <64>; 57 d-cache-sets = <128>; // 32KiB(size)/64(line-size)=512ways/4-way set 58 i-cache-size = <0x8000>; 59 i-cache-line-size = <64>; 60 i-cache-sets = <256>; // 32KiB(size)/64(line-size)=512ways/2-way set 70 d-cache-size = <0x8000>; [all …]
|
/arch/arm64/boot/dts/amazon/ |
D | alpine-v3.dtsi | 17 #size-cells = <2>; 21 #size-cells = <0>; 28 d-cache-size = <0x8000>; 29 d-cache-line-size = <64>; 31 i-cache-size = <0xc000>; 32 i-cache-line-size = <64>; 42 d-cache-size = <0x8000>; 43 d-cache-line-size = <64>; 45 i-cache-size = <0xc000>; 46 i-cache-line-size = <64>; [all …]
|
/arch/x86/kernel/cpu/mtrr/ |
D | if.c | 38 mtrr_file_add(unsigned long base, unsigned long size, in mtrr_file_add() argument 52 if ((base & (PAGE_SIZE - 1)) || (size & (PAGE_SIZE - 1))) in mtrr_file_add() 55 size >>= PAGE_SHIFT; in mtrr_file_add() 57 reg = mtrr_add_page(base, size, type, true); in mtrr_file_add() 64 mtrr_file_del(unsigned long base, unsigned long size, in mtrr_file_del() argument 71 if ((base & (PAGE_SIZE - 1)) || (size & (PAGE_SIZE - 1))) in mtrr_file_del() 74 size >>= PAGE_SHIFT; in mtrr_file_del() 76 reg = mtrr_del_page(-1, base, size); in mtrr_file_del() 98 unsigned long long base, size; in mtrr_write() local 133 size = simple_strtoull(ptr + 5, &ptr, 0); in mtrr_write() [all …]
|
/arch/arm64/kernel/ |
D | patch-scs.c | 91 static int __always_inline skip_xleb128(const u8 **opcode, int size) in skip_xleb128() argument 97 size--; in skip_xleb128() 100 return size; in skip_xleb128() 107 u32 size; member 136 int size = frame->size - offsetof(struct eh_frame, opcodes) + 4; in scs_handle_fde_frame() local 149 size -= l + 1; in scs_handle_fde_frame() 156 while (size-- > 0) { in scs_handle_fde_frame() 165 size--; in scs_handle_fde_frame() 171 size -= 2; in scs_handle_fde_frame() 176 size = skip_xleb128(&opcode, size); in scs_handle_fde_frame() [all …]
|
/arch/riscv/mm/ |
D | dma-noncoherent.c | 18 static inline void arch_dma_cache_wback(phys_addr_t paddr, size_t size) in arch_dma_cache_wback() argument 24 noncoherent_cache_ops.wback(paddr, size); in arch_dma_cache_wback() 28 ALT_CMO_OP(clean, vaddr, size, riscv_cbom_block_size); in arch_dma_cache_wback() 31 static inline void arch_dma_cache_inv(phys_addr_t paddr, size_t size) in arch_dma_cache_inv() argument 37 noncoherent_cache_ops.inv(paddr, size); in arch_dma_cache_inv() 42 ALT_CMO_OP(inval, vaddr, size, riscv_cbom_block_size); in arch_dma_cache_inv() 45 static inline void arch_dma_cache_wback_inv(phys_addr_t paddr, size_t size) in arch_dma_cache_wback_inv() argument 51 noncoherent_cache_ops.wback_inv(paddr, size); in arch_dma_cache_wback_inv() 56 ALT_CMO_OP(flush, vaddr, size, riscv_cbom_block_size); in arch_dma_cache_wback_inv() 69 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device() argument [all …]
|
/arch/mips/mm/ |
D | c-r3k.c | 30 unsigned long flags, status, dummy, size; in r3k_cache_size() local 45 size = 0; in r3k_cache_size() 47 for (size = 128; size <= 0x40000; size <<= 1) in r3k_cache_size() 48 *(p + size) = 0; in r3k_cache_size() 50 for (size = 128; in r3k_cache_size() 51 (size <= 0x40000) && (*(p + size) == 0); in r3k_cache_size() 52 size <<= 1) in r3k_cache_size() 54 if (size > 0x40000) in r3k_cache_size() 55 size = 0; in r3k_cache_size() 60 return size * sizeof(*p); in r3k_cache_size() [all …]
|
/arch/powerpc/lib/ |
D | rheap.c | 139 blk->size = 0; in get_slot() 157 int size; in attach_free_block() local 162 size = blkn->size; in attach_free_block() 164 e = s + size; in attach_free_block() 176 be = bs + blk->size; in attach_free_block() 193 if (before && s != (before->start + before->size)) in attach_free_block() 215 before->size += size; in attach_free_block() 221 after->start -= size; in attach_free_block() 222 after->size += size; in attach_free_block() 227 before->size += size + after->size; in attach_free_block() [all …]
|
/arch/parisc/kernel/ |
D | pci-dma.c | 78 unsigned long size, unsigned long *paddr_ptr) in map_pte_uncached() argument 84 end = vaddr + size; in map_pte_uncached() 105 unsigned long size, unsigned long *paddr_ptr) in map_pmd_uncached() argument 111 end = vaddr + size; in map_pmd_uncached() 127 static inline int map_uncached_pages(unsigned long vaddr, unsigned long size, in map_uncached_pages() argument 131 unsigned long end = vaddr + size; in map_uncached_pages() 154 unsigned long size) in unmap_uncached_pte() argument 169 end = vaddr + size; in unmap_uncached_pte() 190 unsigned long size) in unmap_uncached_pmd() argument 205 end = vaddr + size; in unmap_uncached_pmd() [all …]
|
/arch/arm/mm/ |
D | ioremap.c | 50 size_t size, unsigned int mtype) in find_static_vm_paddr() argument 63 paddr + size - 1 > vm->phys_addr + vm->size - 1) in find_static_vm_paddr() 84 if (vm->addr <= vaddr && vm->addr + vm->size > vaddr) in find_static_vm_vaddr() 148 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument 150 unsigned long addr = virt, end = virt + (size & ~(SZ_1M - 1)); in unmap_area_sections() 189 size_t size, const struct mem_type *type) in remap_area_sections() argument 191 unsigned long addr = virt, end = virt + size; in remap_area_sections() 198 unmap_area_sections(virt, size); in remap_area_sections() 216 size_t size, const struct mem_type *type) in remap_area_supersections() argument 218 unsigned long addr = virt, end = virt + size; in remap_area_supersections() [all …]
|
D | dma-mapping.c | 43 size_t size; member 53 size_t size; member 107 static void __dma_clear_buffer(struct page *page, size_t size, int coherent_flag) in __dma_clear_buffer() argument 115 phys_addr_t end = base + size; in __dma_clear_buffer() 116 while (size > 0) { in __dma_clear_buffer() 123 size -= PAGE_SIZE; in __dma_clear_buffer() 129 memset(ptr, 0, size); in __dma_clear_buffer() 131 dmac_flush_range(ptr, ptr + size); in __dma_clear_buffer() 132 outer_flush_range(__pa(ptr), __pa(ptr) + size); in __dma_clear_buffer() 141 static struct page *__dma_alloc_buffer(struct device *dev, size_t size, in __dma_alloc_buffer() argument [all …]
|
/arch/openrisc/include/asm/ |
D | uaccess.h | 57 #define __put_user_nocheck(x, ptr, size) \ argument 60 __put_user_size((x), (ptr), (size), __pu_err); \ 64 #define __put_user_check(x, ptr, size) \ argument 68 if (access_ok(__pu_addr, size)) \ 69 __put_user_size((x), __pu_addr, (size), __pu_err); \ 73 #define __put_user_size(x, ptr, size, retval) \ argument 76 switch (size) { \ 129 #define __get_user_nocheck(x, ptr, size) \ argument 132 __get_user_size((x), (ptr), (size), __gu_err); \ 136 #define __get_user_check(x, ptr, size) \ argument [all …]
|
/arch/powerpc/boot/ |
D | simple_alloc.c | 22 unsigned long size; member 35 static void *simple_malloc(unsigned long size) in simple_malloc() argument 40 if (size == 0) in simple_malloc() 43 size = _ALIGN_UP(size, alloc_min); in simple_malloc() 47 if (size <= space_left) { in simple_malloc() 49 p->size = size; in simple_malloc() 51 next_base += size; in simple_malloc() 52 space_left -= size; in simple_malloc() 58 else if (!(p->flags & ENTRY_IN_USE) && (size <= p->size)) { in simple_malloc() 95 static void *simple_realloc(void *ptr, unsigned long size) in simple_realloc() argument [all …]
|
/arch/loongarch/kernel/ |
D | unaligned.c | 254 unsigned int res, size = 0; in emulate_load_store_insn() local 264 size = 2; in emulate_load_store_insn() 269 size = 2; in emulate_load_store_insn() 274 size = 2; in emulate_load_store_insn() 279 size = 4; in emulate_load_store_insn() 284 size = 4; in emulate_load_store_insn() 289 size = 4; in emulate_load_store_insn() 294 size = 8; in emulate_load_store_insn() 299 size = 8; in emulate_load_store_insn() 304 size = 4; in emulate_load_store_insn() [all …]
|
/arch/um/kernel/ |
D | load_file.c | 10 static int __init __uml_load_file(const char *filename, void *buf, int size) in __uml_load_file() argument 20 n = os_read_file(fd, buf, size); in __uml_load_file() 21 if (n != size) { in __uml_load_file() 23 "err = %d\n", size, in __uml_load_file() 32 void *uml_load_file(const char *filename, unsigned long long *size) in uml_load_file() argument 37 *size = 0; in uml_load_file() 42 err = os_file_size(filename, size); in uml_load_file() 46 if (*size == 0) { in uml_load_file() 51 area = memblock_alloc(*size, SMP_CACHE_BYTES); in uml_load_file() 53 panic("%s: Failed to allocate %llu bytes\n", __func__, *size); in uml_load_file() [all …]
|
/arch/x86/include/asm/ |
D | uaccess_64.h | 85 static inline bool __access_ok(const void __user *ptr, unsigned long size) in __access_ok() argument 87 if (__builtin_constant_p(size <= PAGE_SIZE) && size <= PAGE_SIZE) { in __access_ok() 90 unsigned long sum = size + (unsigned long)ptr; in __access_ok() 125 raw_copy_from_user(void *dst, const void __user *src, unsigned long size) in raw_copy_from_user() argument 127 return copy_user_generic(dst, (__force void *)src, size); in raw_copy_from_user() 131 raw_copy_to_user(void __user *dst, const void *src, unsigned long size) in raw_copy_to_user() argument 133 return copy_user_generic((__force void *)dst, src, size); in raw_copy_to_user() 136 extern long __copy_user_nocache(void *dst, const void __user *src, unsigned size); 137 extern long __copy_user_flushcache(void *dst, const void __user *src, unsigned size); 141 unsigned size) in __copy_from_user_inatomic_nocache() argument [all …]
|