/kernel/dma/ |
D | coherent.c | 32 struct dma_coherent_mem * mem) in dma_get_device_base() argument 34 if (mem->use_dev_dma_pfn_offset) in dma_get_device_base() 35 return (mem->pfn_base - dev->dma_pfn_offset) << PAGE_SHIFT; in dma_get_device_base() 37 return mem->device_base; in dma_get_device_base() 42 struct dma_coherent_mem **mem) in dma_init_coherent_memory() argument 77 *mem = dma_mem; in dma_init_coherent_memory() 87 static void dma_release_coherent_memory(struct dma_coherent_mem *mem) in dma_release_coherent_memory() argument 89 if (!mem) in dma_release_coherent_memory() 92 memunmap(mem->virt_base); in dma_release_coherent_memory() 93 kfree(mem->bitmap); in dma_release_coherent_memory() [all …]
|
/kernel/ |
D | kexec_file.c | 434 i, ksegment->buf, ksegment->bufsz, ksegment->mem, in SYSCALL_DEFINE5() 492 kbuf->mem = temp_start; in locate_mem_hole_top_down() 526 kbuf->mem = temp_start; in locate_mem_hole_bottom_up() 640 if (kbuf->mem != KEXEC_BUF_MEM_UNKNOWN) in kexec_locate_mem_hole() 699 ksegment->mem = kbuf->mem; in kexec_add_buffer() 788 sha_regions[j].start = ksegment->mem; in kexec_calculate_store_digests() 912 bss_addr = kbuf->mem + kbuf->bufsz; in kexec_purgatory_setup_sechdrs() 936 kbuf->image->start += kbuf->mem + offset; in kexec_purgatory_setup_sechdrs() 943 sechdrs[i].sh_addr = kbuf->mem + offset; in kexec_purgatory_setup_sechdrs() 1169 int crash_exclude_mem_range(struct crash_mem *mem, in crash_exclude_mem_range() argument [all …]
|
D | kexec_core.c | 171 mstart = image->segment[i].mem; in sanity_check_segment_list() 190 mstart = image->segment[i].mem; in sanity_check_segment_list() 195 pstart = image->segment[j].mem; in sanity_check_segment_list() 242 mstart = image->segment[i].mem; in sanity_check_segment_list() 290 mstart = image->segment[i].mem; in kimage_is_destination_range() 458 mstart = image->segment[i].mem; in kimage_alloc_crash_control_pages() 791 maddr = segment->mem; in kimage_load_normal_segment() 864 maddr = segment->mem; in kimage_load_crash_segment()
|
D | kexec.c | 292 out.mem = in.mem; in COMPAT_SYSCALL_DEFINE4()
|
D | relay.c | 125 void *mem; in relay_alloc_buf() local 141 mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL); in relay_alloc_buf() 142 if (!mem) in relay_alloc_buf() 145 memset(mem, 0, *size); in relay_alloc_buf() 147 return mem; in relay_alloc_buf()
|
D | kexec_elf.c | 416 kbuf->mem = KEXEC_BUF_MEM_UNKNOWN; in kexec_elf_load() 420 load_addr = kbuf->mem; in kexec_elf_load()
|
/kernel/trace/ |
D | trace_seq.c | 262 void trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len) in trace_seq_putmem() argument 274 seq_buf_putmem(&s->seq, mem, len); in trace_seq_putmem() 288 void trace_seq_putmem_hex(struct trace_seq *s, const void *mem, in trace_seq_putmem_hex() argument 305 seq_buf_putmem_hex(&s->seq, mem, len); in trace_seq_putmem_hex()
|
/kernel/bpf/ |
D | queue_stack_maps.c | 70 struct bpf_map_memory mem = {0}; in queue_stack_map_alloc() local 77 ret = bpf_map_charge_init(&mem, cost); in queue_stack_map_alloc() 83 bpf_map_charge_finish(&mem); in queue_stack_map_alloc() 91 bpf_map_charge_move(&qs->map.memory, &mem); in queue_stack_map_alloc()
|
D | reuseport_array.c | 154 struct bpf_map_memory mem; in reuseport_array_alloc() local 163 err = bpf_map_charge_init(&mem, array_size); in reuseport_array_alloc() 170 bpf_map_charge_finish(&mem); in reuseport_array_alloc() 176 bpf_map_charge_move(&array->map.memory, &mem); in reuseport_array_alloc()
|
D | stackmap.c | 89 struct bpf_map_memory mem; in stack_map_alloc() local 118 err = bpf_map_charge_init(&mem, cost); in stack_map_alloc() 124 bpf_map_charge_finish(&mem); in stack_map_alloc() 140 bpf_map_charge_move(&smap->map.memory, &mem); in stack_map_alloc() 147 bpf_map_charge_finish(&mem); in stack_map_alloc()
|
D | local_storage.c | 275 struct bpf_map_memory mem; in cgroup_storage_map_alloc() local 295 ret = bpf_map_charge_init(&mem, sizeof(struct bpf_cgroup_storage_map)); in cgroup_storage_map_alloc() 302 bpf_map_charge_finish(&mem); in cgroup_storage_map_alloc() 306 bpf_map_charge_move(&map->map.memory, &mem); in cgroup_storage_map_alloc()
|
D | syscall.c | 203 int bpf_map_charge_init(struct bpf_map_memory *mem, u64 size) in bpf_map_charge_init() argument 219 mem->pages = pages; in bpf_map_charge_init() 220 mem->user = user; in bpf_map_charge_init() 225 void bpf_map_charge_finish(struct bpf_map_memory *mem) in bpf_map_charge_finish() argument 227 bpf_uncharge_memlock(mem->user, mem->pages); in bpf_map_charge_finish() 228 free_uid(mem->user); in bpf_map_charge_finish() 305 struct bpf_map_memory mem; in bpf_map_free_deferred() local 307 bpf_map_charge_move(&mem, &map->memory); in bpf_map_free_deferred() 311 bpf_map_charge_finish(&mem); in bpf_map_free_deferred() 553 struct bpf_map_memory mem; in map_create() local [all …]
|
D | arraymap.c | 78 struct bpf_map_memory mem; in array_map_alloc() local 115 ret = bpf_map_charge_init(&mem, cost); in array_map_alloc() 122 bpf_map_charge_finish(&mem); in array_map_alloc() 130 bpf_map_charge_move(&array->map.memory, &mem); in array_map_alloc()
|
/kernel/debug/ |
D | gdbstub.c | 239 char *kgdb_mem2hex(char *mem, char *buf, int count) in kgdb_mem2hex() argument 250 err = probe_kernel_read(tmp, mem, count); in kgdb_mem2hex() 268 int kgdb_hex2mem(char *buf, char *mem, int count) in kgdb_hex2mem() argument 286 return probe_kernel_write(mem, tmp_raw, count); in kgdb_hex2mem() 326 static int kgdb_ebin2mem(char *buf, char *mem, int count) in kgdb_ebin2mem() argument 338 return probe_kernel_write(mem, c, size); in kgdb_ebin2mem()
|
/kernel/power/ |
D | Kconfig | 189 Enable this with a kernel parameter like "test_suspend=mem".
|