/kernel/ |
D | resource.c | 153 static void free_resource(struct resource *res) in free_resource() argument 161 if (res && PageSlab(virt_to_head_page(res))) in free_resource() 162 kfree(res); in free_resource() 327 struct resource *res) in find_next_iomem_res() argument 331 if (!res) in find_next_iomem_res() 361 *res = (struct resource) { in find_next_iomem_res() 379 struct resource res; in __walk_iomem_res_desc() local 383 !find_next_iomem_res(start, end, flags, desc, &res)) { in __walk_iomem_res_desc() 384 ret = (*func)(&res, arg); in __walk_iomem_res_desc() 388 start = res.end + 1; in __walk_iomem_res_desc() [all …]
|
D | regset.c | 12 int res; in __regset_get() local 23 res = regset->regset_get(target, regset, in __regset_get() 25 if (res < 0) { in __regset_get() 27 return res; in __regset_get() 30 return size - res; in __regset_get()
|
D | kallsyms.c | 168 char *res; in cleanup_symbol_name() local 180 res = strstr(s, ".llvm."); in cleanup_symbol_name() 181 if (res) { in cleanup_symbol_name() 182 *res = '\0'; in cleanup_symbol_name() 460 int res; in lookup_symbol_name() local 475 res = lookup_module_symbol_name(addr, symname); in lookup_symbol_name() 476 if (res) in lookup_symbol_name() 477 return res; in lookup_symbol_name() 487 int res; in lookup_symbol_attrs() local 503 res = lookup_module_symbol_attrs(addr, size, offset, modname, name); in lookup_symbol_attrs() [all …]
|
D | acct.c | 165 struct bsd_acct_struct *res; in acct_get() local 169 res = to_acct(READ_ONCE(ns->bacct)); in acct_get() 170 if (!res) { in acct_get() 174 if (!atomic_long_inc_not_zero(&res->count)) { in acct_get() 180 mutex_lock(&res->lock); in acct_get() 181 if (res != to_acct(READ_ONCE(ns->bacct))) { in acct_get() 182 mutex_unlock(&res->lock); in acct_get() 183 acct_put(res); in acct_get() 186 return res; in acct_get()
|
D | iomem.c | 129 static void devm_memremap_release(struct device *dev, void *res) in devm_memremap_release() argument 131 memunmap(*(void **)res); in devm_memremap_release() 134 static int devm_memremap_match(struct device *dev, void *res, void *match_data) in devm_memremap_match() argument 136 return *(void **)res == match_data; in devm_memremap_match()
|
D | kexec_file.c | 491 static int locate_mem_hole_callback(struct resource *res, void *arg) in locate_mem_hole_callback() argument 494 u64 start = res->start, end = res->end; in locate_mem_hole_callback() 500 if (res->flags & IORESOURCE_SYSRAM_DRIVER_MANAGED) in locate_mem_hole_callback() 525 struct resource res = { }; in kexec_walk_memblock() local 543 res.start = mstart; in kexec_walk_memblock() 544 res.end = mend - 1; in kexec_walk_memblock() 545 ret = func(&res, kbuf); in kexec_walk_memblock() 557 res.start = mstart; in kexec_walk_memblock() 558 res.end = mend - 1; in kexec_walk_memblock() 559 ret = func(&res, kbuf); in kexec_walk_memblock()
|
D | kcov.c | 478 int res = 0; in kcov_mmap() local 488 res = -EINVAL; in kcov_mmap() 495 res = vm_insert_page(vma, vma->vm_start + off, page); in kcov_mmap() 496 if (res) { in kcov_mmap() 498 return res; in kcov_mmap() 504 return res; in kcov_mmap() 683 int res; in kcov_ioctl() local 740 res = kcov_ioctl_locked(kcov, cmd, arg); in kcov_ioctl() 743 return res; in kcov_ioctl()
|
/kernel/power/ |
D | user.c | 140 ssize_t res; in snapshot_read() local 146 res = -ENODATA; in snapshot_read() 150 res = snapshot_read_next(&data->handle); in snapshot_read() 151 if (res <= 0) in snapshot_read() 154 res = PAGE_SIZE - pg_offp; in snapshot_read() 157 res = simple_read_from_buffer(buf, count, &pg_offp, in snapshot_read() 158 data_of(data->handle), res); in snapshot_read() 159 if (res > 0) in snapshot_read() 160 *offp += res; in snapshot_read() 165 return res; in snapshot_read() [all …]
|
D | swap.c | 351 int res; in swsusp_swap_check() local 354 res = swap_type_of(swsusp_resume_device, swsusp_resume_block); in swsusp_swap_check() 356 res = find_first_swap(&swsusp_resume_device); in swsusp_swap_check() 357 if (res < 0) in swsusp_swap_check() 358 return res; in swsusp_swap_check() 359 root_swap = res; in swsusp_swap_check() 366 res = set_blocksize(hib_resume_bdev, PAGE_SIZE); in swsusp_swap_check() 367 if (res < 0) in swsusp_swap_check() 370 return res; in swsusp_swap_check()
|
/kernel/cgroup/ |
D | misc.c | 80 return atomic_long_read(&root_cg.res[type].usage); in misc_cg_res_total_usage() 119 WARN_ONCE(atomic_long_add_negative(-amount, &cg->res[type].usage), in misc_cg_cancel_charge() 145 struct misc_res *res; in misc_cg_try_charge() local 155 res = &i->res[type]; in misc_cg_try_charge() 157 new_usage = atomic_long_add_return(amount, &res->usage); in misc_cg_try_charge() 158 if (new_usage > READ_ONCE(res->max) || in misc_cg_try_charge() 168 atomic_long_inc(&j->res[type].events); in misc_cg_try_charge() 216 max = READ_ONCE(cg->res[i].max); in misc_cg_max_show() 282 WRITE_ONCE(cg->res[type].max, max); in misc_cg_max_write() 304 usage = atomic_long_read(&cg->res[i].usage); in misc_cg_current_show() [all …]
|
/kernel/irq/ |
D | devres.c | 18 static void devm_irq_release(struct device *dev, void *res) in devm_irq_release() argument 20 struct irq_devres *this = res; in devm_irq_release() 25 static int devm_irq_match(struct device *dev, void *res, void *data) in devm_irq_match() argument 27 struct irq_devres *this = res, *match = data; in devm_irq_match() 154 static void devm_irq_desc_release(struct device *dev, void *res) in devm_irq_desc_release() argument 156 struct irq_desc_devres *this = res; in devm_irq_desc_release() 240 static void devm_irq_remove_generic_chip(struct device *dev, void *res) in devm_irq_remove_generic_chip() argument 242 struct irq_generic_chip_devres *this = res; in devm_irq_remove_generic_chip()
|
D | handle.c | 148 irqreturn_t res; in __handle_irq_event_percpu() local 158 res = action->handler(irq, action->dev_id); in __handle_irq_event_percpu() 159 trace_irq_handler_exit(irq, action, res); in __handle_irq_event_percpu() 165 switch (res) { in __handle_irq_event_percpu() 183 retval |= res; in __handle_irq_event_percpu()
|
D | chip.c | 759 irqreturn_t res; in handle_fasteoi_nmi() local 767 res = action->handler(irq, action->dev_id); in handle_fasteoi_nmi() 768 trace_irq_handler_exit(irq, action, res); in handle_fasteoi_nmi() 932 irqreturn_t res; in handle_percpu_devid_irq() local 945 res = action->handler(irq, raw_cpu_ptr(action->percpu_dev_id)); in handle_percpu_devid_irq() 946 trace_irq_handler_exit(irq, action, res); in handle_percpu_devid_irq() 975 irqreturn_t res; in handle_percpu_devid_fasteoi_nmi() local 980 res = action->handler(irq, raw_cpu_ptr(action->percpu_dev_id)); in handle_percpu_devid_fasteoi_nmi() 981 trace_irq_handler_exit(irq, action, res); in handle_percpu_devid_fasteoi_nmi()
|
/kernel/time/ |
D | sched_clock.c | 86 u64 cyc, res; in sched_clock() local 95 res = rd->epoch_ns + cyc_to_ns(cyc, rd->mult, rd->shift); in sched_clock() 98 return res; in sched_clock() 156 u64 res, wrap, new_mask, new_epoch, cyc, ns; in sched_clock_register() local 213 res = cyc_to_ns(1ULL, new_mult, new_shift); in sched_clock_register() 216 bits, r, r_unit, res, wrap); in sched_clock_register()
|
D | clockevents.c | 29 int res; member 403 int res; in __clockevents_unbind() local 406 res = __clockevents_try_unbind(cu->ce, smp_processor_id()); in __clockevents_unbind() 407 if (res == -EAGAIN) in __clockevents_unbind() 408 res = clockevents_replace(cu->ce); in __clockevents_unbind() 409 cu->res = res; in __clockevents_unbind() 419 struct ce_unbind cu = { .ce = ced, .res = -ENODEV }; in clockevents_unbind() 422 return cu.res; in clockevents_unbind()
|
D | time.c | 769 struct timespec64 res; in timespec64_add_safe() local 771 set_normalized_timespec64(&res, (timeu64_t) lhs.tv_sec + rhs.tv_sec, in timespec64_add_safe() 774 if (unlikely(res.tv_sec < lhs.tv_sec || res.tv_sec < rhs.tv_sec)) { in timespec64_add_safe() 775 res.tv_sec = TIME64_MAX; in timespec64_add_safe() 776 res.tv_nsec = 0; in timespec64_add_safe() 779 return res; in timespec64_add_safe()
|
D | ntp.c | 631 int res = -EAGAIN; in sync_hw_clock() local 651 res = update_persistent_clock64(to_set); in sync_hw_clock() 652 if (res != -ENODEV) in sync_hw_clock() 656 res = update_rtc(&to_set, &offset_nsec); in sync_hw_clock() 657 if (res == -ENODEV) in sync_hw_clock() 660 sched_sync_hw_clock(offset_nsec, res != 0); in sync_hw_clock()
|
/kernel/debug/kdb/ |
D | kdb_support.c | 266 int kdb_getarea_size(void *res, unsigned long addr, size_t size) in kdb_getarea_size() argument 268 int ret = copy_from_kernel_nofault((char *)res, (char *)addr, size); in kdb_getarea_size() 291 int kdb_putarea_size(unsigned long addr, void *res, size_t size) in kdb_putarea_size() argument 293 int ret = copy_to_kernel_nofault((char *)addr, (char *)res, size); in kdb_putarea_size() 317 static int kdb_getphys(void *res, unsigned long addr, size_t size) in kdb_getphys() argument 328 memcpy(res, vaddr + (addr & (PAGE_SIZE - 1)), size); in kdb_getphys()
|
/kernel/locking/ |
D | rtmutex.c | 1146 int chain_walk = 0, res; in task_blocks_on_rt_mutex() local 1185 res = __ww_mutex_add_waiter(waiter, rtm, ww_ctx); in task_blocks_on_rt_mutex() 1186 if (res) { in task_blocks_on_rt_mutex() 1191 return res; in task_blocks_on_rt_mutex() 1231 res = rt_mutex_adjust_prio_chain(owner, chwalk, lock, in task_blocks_on_rt_mutex() 1236 return res; in task_blocks_on_rt_mutex() 1416 bool res = true; in rtmutex_spin_on_owner() local 1440 res = false; in rtmutex_spin_on_owner() 1446 return res; in rtmutex_spin_on_owner() 1583 static void __sched rt_mutex_handle_deadlock(int res, int detect_deadlock, in rt_mutex_handle_deadlock() argument [all …]
|
/kernel/bpf/ |
D | offload.c | 302 int res; in bpf_prog_offload_info_fill() local 305 res = ns_get_path_cb(&ns_path, bpf_prog_offload_info_fill_ns, &args); in bpf_prog_offload_info_fill() 306 if (res) { in bpf_prog_offload_info_fill() 309 return res; in bpf_prog_offload_info_fill() 526 int res; in bpf_map_offload_info_fill() local 528 res = ns_get_path_cb(&ns_path, bpf_map_offload_info_fill_ns, &args); in bpf_map_offload_info_fill() 529 if (res) { in bpf_map_offload_info_fill() 532 return res; in bpf_map_offload_info_fill()
|
D | disasm.c | 28 const char *res; in __func_get_name() local 30 res = cbs->cb_call(cbs->private_data, insn); in __func_get_name() 31 if (res) in __func_get_name() 32 return res; in __func_get_name()
|
D | helpers.c | 439 unsigned long long *res, bool *is_negative) in __bpf_strtoull() argument 448 if (!buf || !buf_len || !res || !is_negative) in __bpf_strtoull() 475 val_len = _parse_integer(cur_buf, base, res); in __bpf_strtoull() 490 long long *res) in __bpf_strtoll() argument 502 *res = -_res; in __bpf_strtoll() 506 *res = _res; in __bpf_strtoll() 512 long *, res) in BPF_CALL_4() argument 522 *res = _res; in BPF_CALL_4() 537 unsigned long *, res) in BPF_CALL_4() argument 550 *res = _res; in BPF_CALL_4()
|
/kernel/futex/ |
D | requeue.c | 772 int res, ret; in futex_wait_requeue_pi() local 863 res = fixup_pi_owner(uaddr2, &q, !ret); in futex_wait_requeue_pi() 868 if (res) in futex_wait_requeue_pi() 869 ret = (res < 0) ? res : 0; in futex_wait_requeue_pi()
|
D | pi.c | 937 int res, ret; in futex_lock_pi() local 1060 res = fixup_pi_owner(uaddr, &q, !ret); in futex_lock_pi() 1065 if (res) in futex_lock_pi() 1066 ret = (res < 0) ? res : 0; in futex_lock_pi()
|
/kernel/sched/ |
D | psi.c | 1186 int psi_show(struct seq_file *m, struct psi_group *group, enum psi_res res) in psi_show() argument 1204 only_full = res == PSI_IRQ; in psi_show() 1213 if (!(group == &psi_system && res == PSI_CPU && full)) { in psi_show() 1215 avg[w] = group->avg[res * 2 + full][w]; in psi_show() 1216 total = div_u64(group->total[PSI_AVGS][res * 2 + full], in psi_show() 1232 char *buf, enum psi_res res) in psi_trigger_create() argument 1243 state = PSI_IO_SOME + res * 2; in psi_trigger_create() 1245 state = PSI_IO_FULL + res * 2; in psi_trigger_create() 1250 if (res == PSI_IRQ && --state != PSI_IRQ_FULL) in psi_trigger_create() 1429 size_t nbytes, enum psi_res res) in psi_write() argument [all …]
|