/kernel/dma/ |
D | coherent.c | 33 struct dma_coherent_mem * mem) in dma_get_device_base() argument 35 if (mem->use_dev_dma_pfn_offset) in dma_get_device_base() 36 return phys_to_dma(dev, PFN_PHYS(mem->pfn_base)); in dma_get_device_base() 37 return mem->device_base; in dma_get_device_base() 42 struct dma_coherent_mem **mem) in dma_init_coherent_memory() argument 77 *mem = dma_mem; in dma_init_coherent_memory() 87 static void _dma_release_coherent_memory(struct dma_coherent_mem *mem) in _dma_release_coherent_memory() argument 89 if (!mem) in _dma_release_coherent_memory() 92 memunmap(mem->virt_base); in _dma_release_coherent_memory() 93 kfree(mem->bitmap); in _dma_release_coherent_memory() [all …]
|
/kernel/ |
D | kexec_file.c | 393 i, ksegment->buf, ksegment->bufsz, ksegment->mem, in SYSCALL_DEFINE5() 455 kbuf->mem = temp_start; in locate_mem_hole_top_down() 489 kbuf->mem = temp_start; in locate_mem_hole_bottom_up() 608 if (kbuf->mem != KEXEC_BUF_MEM_UNKNOWN) in kexec_locate_mem_hole() 679 ksegment->mem = kbuf->mem; in kexec_add_buffer() 770 sha_regions[j].start = ksegment->mem; in kexec_calculate_store_digests() 894 bss_addr = kbuf->mem + kbuf->bufsz; in kexec_purgatory_setup_sechdrs() 930 kbuf->image->start += kbuf->mem + offset; in kexec_purgatory_setup_sechdrs() 937 sechdrs[i].sh_addr = kbuf->mem + offset; in kexec_purgatory_setup_sechdrs() 1163 int crash_exclude_mem_range(struct crash_mem *mem, in crash_exclude_mem_range() argument [all …]
|
D | kexec_core.c | 171 mstart = image->segment[i].mem; in sanity_check_segment_list() 190 mstart = image->segment[i].mem; in sanity_check_segment_list() 195 pstart = image->segment[j].mem; in sanity_check_segment_list() 242 mstart = image->segment[i].mem; in sanity_check_segment_list() 290 mstart = image->segment[i].mem; in kimage_is_destination_range() 458 mstart = image->segment[i].mem; in kimage_alloc_crash_control_pages() 797 maddr = segment->mem; in kimage_load_normal_segment() 870 maddr = segment->mem; in kimage_load_crash_segment()
|
D | kexec.c | 294 out.mem = in.mem; in COMPAT_SYSCALL_DEFINE4()
|
D | relay.c | 125 void *mem; in relay_alloc_buf() local 141 mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL); in relay_alloc_buf() 142 if (!mem) in relay_alloc_buf() 145 memset(mem, 0, *size); in relay_alloc_buf() 147 return mem; in relay_alloc_buf()
|
D | kexec_elf.c | 416 kbuf->mem = KEXEC_BUF_MEM_UNKNOWN; in kexec_elf_load() 420 load_addr = kbuf->mem; in kexec_elf_load()
|
/kernel/trace/ |
D | trace_seq.c | 259 void trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len) in trace_seq_putmem() argument 271 seq_buf_putmem(&s->seq, mem, len); in trace_seq_putmem() 285 void trace_seq_putmem_hex(struct trace_seq *s, const void *mem, in trace_seq_putmem_hex() argument 302 seq_buf_putmem_hex(&s->seq, mem, len); in trace_seq_putmem_hex()
|
/kernel/bpf/ |
D | queue_stack_maps.c | 70 struct bpf_map_memory mem = {0}; in queue_stack_map_alloc() local 77 ret = bpf_map_charge_init(&mem, cost); in queue_stack_map_alloc() 83 bpf_map_charge_finish(&mem); in queue_stack_map_alloc() 91 bpf_map_charge_move(&qs->map.memory, &mem); in queue_stack_map_alloc()
|
D | reuseport_array.c | 155 struct bpf_map_memory mem; in reuseport_array_alloc() local 164 err = bpf_map_charge_init(&mem, array_size); in reuseport_array_alloc() 171 bpf_map_charge_finish(&mem); in reuseport_array_alloc() 177 bpf_map_charge_move(&array->map.memory, &mem); in reuseport_array_alloc()
|
D | bpf_struct_ops.c | 562 struct bpf_map_memory mem; in bpf_struct_ops_map_alloc() local 589 err = bpf_map_charge_init(&mem, map_total_size); in bpf_struct_ops_map_alloc() 595 bpf_map_charge_finish(&mem); in bpf_struct_ops_map_alloc() 608 bpf_map_charge_finish(&mem); in bpf_struct_ops_map_alloc() 615 bpf_map_charge_move(&map->memory, &mem); in bpf_struct_ops_map_alloc()
|
D | local_storage.c | 291 struct bpf_map_memory mem; in cgroup_storage_map_alloc() local 312 ret = bpf_map_charge_init(&mem, sizeof(struct bpf_cgroup_storage_map)); in cgroup_storage_map_alloc() 319 bpf_map_charge_finish(&mem); in cgroup_storage_map_alloc() 323 bpf_map_charge_move(&map->map.memory, &mem); in cgroup_storage_map_alloc()
|
D | stackmap.c | 94 struct bpf_map_memory mem; in stack_map_alloc() local 124 err = bpf_map_charge_init(&mem, cost + attr->max_entries * in stack_map_alloc() 131 bpf_map_charge_finish(&mem); in stack_map_alloc() 147 bpf_map_charge_move(&smap->map.memory, &mem); in stack_map_alloc() 154 bpf_map_charge_finish(&mem); in stack_map_alloc()
|
D | arraymap.c | 88 struct bpf_map_memory mem; in array_map_alloc() local 134 ret = bpf_map_charge_init(&mem, cost); in array_map_alloc() 145 bpf_map_charge_finish(&mem); in array_map_alloc() 154 bpf_map_charge_finish(&mem); in array_map_alloc() 162 bpf_map_charge_move(&array->map.memory, &mem); in array_map_alloc()
|
D | syscall.c | 379 int bpf_map_charge_init(struct bpf_map_memory *mem, u64 size) in bpf_map_charge_init() argument 395 mem->pages = pages; in bpf_map_charge_init() 396 mem->user = user; in bpf_map_charge_init() 401 void bpf_map_charge_finish(struct bpf_map_memory *mem) in bpf_map_charge_finish() argument 403 bpf_uncharge_memlock(mem->user, mem->pages); in bpf_map_charge_finish() 404 free_uid(mem->user); in bpf_map_charge_finish() 481 struct bpf_map_memory mem; in bpf_map_free_deferred() local 483 bpf_map_charge_move(&mem, &map->memory); in bpf_map_free_deferred() 487 bpf_map_charge_finish(&mem); in bpf_map_free_deferred() 812 struct bpf_map_memory mem; in map_create() local [all …]
|
D | cpumap.c | 250 rxq.mem = xdpf->mem; in cpu_map_bpf_prog_run_xdp()
|
/kernel/debug/ |
D | gdbstub.c | 239 char *kgdb_mem2hex(char *mem, char *buf, int count) in kgdb_mem2hex() argument 250 err = copy_from_kernel_nofault(tmp, mem, count); in kgdb_mem2hex() 268 int kgdb_hex2mem(char *buf, char *mem, int count) in kgdb_hex2mem() argument 286 return copy_to_kernel_nofault(mem, tmp_raw, count); in kgdb_hex2mem() 326 static int kgdb_ebin2mem(char *buf, char *mem, int count) in kgdb_ebin2mem() argument 338 return copy_to_kernel_nofault(mem, c, size); in kgdb_ebin2mem()
|
/kernel/power/ |
D | Kconfig | 201 Enable this with a kernel parameter like "test_suspend=mem".
|