Home
last modified time | relevance | path

Searched refs:res (Results 1 – 25 of 50) sorted by relevance

12

/kernel/
Dresource.c153 static void free_resource(struct resource *res) in free_resource() argument
161 if (res && PageSlab(virt_to_head_page(res))) in free_resource()
162 kfree(res); in free_resource()
327 struct resource *res) in find_next_iomem_res() argument
331 if (!res) in find_next_iomem_res()
361 *res = (struct resource) { in find_next_iomem_res()
379 struct resource res; in __walk_iomem_res_desc() local
383 !find_next_iomem_res(start, end, flags, desc, &res)) { in __walk_iomem_res_desc()
384 ret = (*func)(&res, arg); in __walk_iomem_res_desc()
388 start = res.end + 1; in __walk_iomem_res_desc()
[all …]
Dregset.c12 int res; in __regset_get() local
23 res = regset->regset_get(target, regset, in __regset_get()
25 if (res < 0) { in __regset_get()
27 return res; in __regset_get()
30 return size - res; in __regset_get()
Dkallsyms.c168 char *res; in cleanup_symbol_name() local
180 res = strstr(s, ".llvm."); in cleanup_symbol_name()
181 if (res) { in cleanup_symbol_name()
182 *res = '\0'; in cleanup_symbol_name()
460 int res; in lookup_symbol_name() local
475 res = lookup_module_symbol_name(addr, symname); in lookup_symbol_name()
476 if (res) in lookup_symbol_name()
477 return res; in lookup_symbol_name()
487 int res; in lookup_symbol_attrs() local
503 res = lookup_module_symbol_attrs(addr, size, offset, modname, name); in lookup_symbol_attrs()
[all …]
Dacct.c165 struct bsd_acct_struct *res; in acct_get() local
169 res = to_acct(READ_ONCE(ns->bacct)); in acct_get()
170 if (!res) { in acct_get()
174 if (!atomic_long_inc_not_zero(&res->count)) { in acct_get()
180 mutex_lock(&res->lock); in acct_get()
181 if (res != to_acct(READ_ONCE(ns->bacct))) { in acct_get()
182 mutex_unlock(&res->lock); in acct_get()
183 acct_put(res); in acct_get()
186 return res; in acct_get()
Diomem.c129 static void devm_memremap_release(struct device *dev, void *res) in devm_memremap_release() argument
131 memunmap(*(void **)res); in devm_memremap_release()
134 static int devm_memremap_match(struct device *dev, void *res, void *match_data) in devm_memremap_match() argument
136 return *(void **)res == match_data; in devm_memremap_match()
Dkexec_file.c491 static int locate_mem_hole_callback(struct resource *res, void *arg) in locate_mem_hole_callback() argument
494 u64 start = res->start, end = res->end; in locate_mem_hole_callback()
500 if (res->flags & IORESOURCE_SYSRAM_DRIVER_MANAGED) in locate_mem_hole_callback()
525 struct resource res = { }; in kexec_walk_memblock() local
543 res.start = mstart; in kexec_walk_memblock()
544 res.end = mend - 1; in kexec_walk_memblock()
545 ret = func(&res, kbuf); in kexec_walk_memblock()
557 res.start = mstart; in kexec_walk_memblock()
558 res.end = mend - 1; in kexec_walk_memblock()
559 ret = func(&res, kbuf); in kexec_walk_memblock()
Dkcov.c478 int res = 0; in kcov_mmap() local
488 res = -EINVAL; in kcov_mmap()
495 res = vm_insert_page(vma, vma->vm_start + off, page); in kcov_mmap()
496 if (res) { in kcov_mmap()
498 return res; in kcov_mmap()
504 return res; in kcov_mmap()
683 int res; in kcov_ioctl() local
740 res = kcov_ioctl_locked(kcov, cmd, arg); in kcov_ioctl()
743 return res; in kcov_ioctl()
/kernel/power/
Duser.c140 ssize_t res; in snapshot_read() local
146 res = -ENODATA; in snapshot_read()
150 res = snapshot_read_next(&data->handle); in snapshot_read()
151 if (res <= 0) in snapshot_read()
154 res = PAGE_SIZE - pg_offp; in snapshot_read()
157 res = simple_read_from_buffer(buf, count, &pg_offp, in snapshot_read()
158 data_of(data->handle), res); in snapshot_read()
159 if (res > 0) in snapshot_read()
160 *offp += res; in snapshot_read()
165 return res; in snapshot_read()
[all …]
Dswap.c351 int res; in swsusp_swap_check() local
354 res = swap_type_of(swsusp_resume_device, swsusp_resume_block); in swsusp_swap_check()
356 res = find_first_swap(&swsusp_resume_device); in swsusp_swap_check()
357 if (res < 0) in swsusp_swap_check()
358 return res; in swsusp_swap_check()
359 root_swap = res; in swsusp_swap_check()
366 res = set_blocksize(hib_resume_bdev, PAGE_SIZE); in swsusp_swap_check()
367 if (res < 0) in swsusp_swap_check()
370 return res; in swsusp_swap_check()
/kernel/cgroup/
Dmisc.c80 return atomic_long_read(&root_cg.res[type].usage); in misc_cg_res_total_usage()
119 WARN_ONCE(atomic_long_add_negative(-amount, &cg->res[type].usage), in misc_cg_cancel_charge()
145 struct misc_res *res; in misc_cg_try_charge() local
155 res = &i->res[type]; in misc_cg_try_charge()
157 new_usage = atomic_long_add_return(amount, &res->usage); in misc_cg_try_charge()
158 if (new_usage > READ_ONCE(res->max) || in misc_cg_try_charge()
168 atomic_long_inc(&j->res[type].events); in misc_cg_try_charge()
216 max = READ_ONCE(cg->res[i].max); in misc_cg_max_show()
282 WRITE_ONCE(cg->res[type].max, max); in misc_cg_max_write()
304 usage = atomic_long_read(&cg->res[i].usage); in misc_cg_current_show()
[all …]
/kernel/irq/
Ddevres.c18 static void devm_irq_release(struct device *dev, void *res) in devm_irq_release() argument
20 struct irq_devres *this = res; in devm_irq_release()
25 static int devm_irq_match(struct device *dev, void *res, void *data) in devm_irq_match() argument
27 struct irq_devres *this = res, *match = data; in devm_irq_match()
154 static void devm_irq_desc_release(struct device *dev, void *res) in devm_irq_desc_release() argument
156 struct irq_desc_devres *this = res; in devm_irq_desc_release()
240 static void devm_irq_remove_generic_chip(struct device *dev, void *res) in devm_irq_remove_generic_chip() argument
242 struct irq_generic_chip_devres *this = res; in devm_irq_remove_generic_chip()
Dhandle.c148 irqreturn_t res; in __handle_irq_event_percpu() local
158 res = action->handler(irq, action->dev_id); in __handle_irq_event_percpu()
159 trace_irq_handler_exit(irq, action, res); in __handle_irq_event_percpu()
165 switch (res) { in __handle_irq_event_percpu()
183 retval |= res; in __handle_irq_event_percpu()
Dchip.c759 irqreturn_t res; in handle_fasteoi_nmi() local
767 res = action->handler(irq, action->dev_id); in handle_fasteoi_nmi()
768 trace_irq_handler_exit(irq, action, res); in handle_fasteoi_nmi()
932 irqreturn_t res; in handle_percpu_devid_irq() local
945 res = action->handler(irq, raw_cpu_ptr(action->percpu_dev_id)); in handle_percpu_devid_irq()
946 trace_irq_handler_exit(irq, action, res); in handle_percpu_devid_irq()
975 irqreturn_t res; in handle_percpu_devid_fasteoi_nmi() local
980 res = action->handler(irq, raw_cpu_ptr(action->percpu_dev_id)); in handle_percpu_devid_fasteoi_nmi()
981 trace_irq_handler_exit(irq, action, res); in handle_percpu_devid_fasteoi_nmi()
/kernel/time/
Dsched_clock.c86 u64 cyc, res; in sched_clock() local
95 res = rd->epoch_ns + cyc_to_ns(cyc, rd->mult, rd->shift); in sched_clock()
98 return res; in sched_clock()
156 u64 res, wrap, new_mask, new_epoch, cyc, ns; in sched_clock_register() local
213 res = cyc_to_ns(1ULL, new_mult, new_shift); in sched_clock_register()
216 bits, r, r_unit, res, wrap); in sched_clock_register()
Dclockevents.c29 int res; member
403 int res; in __clockevents_unbind() local
406 res = __clockevents_try_unbind(cu->ce, smp_processor_id()); in __clockevents_unbind()
407 if (res == -EAGAIN) in __clockevents_unbind()
408 res = clockevents_replace(cu->ce); in __clockevents_unbind()
409 cu->res = res; in __clockevents_unbind()
419 struct ce_unbind cu = { .ce = ced, .res = -ENODEV }; in clockevents_unbind()
422 return cu.res; in clockevents_unbind()
Dtime.c769 struct timespec64 res; in timespec64_add_safe() local
771 set_normalized_timespec64(&res, (timeu64_t) lhs.tv_sec + rhs.tv_sec, in timespec64_add_safe()
774 if (unlikely(res.tv_sec < lhs.tv_sec || res.tv_sec < rhs.tv_sec)) { in timespec64_add_safe()
775 res.tv_sec = TIME64_MAX; in timespec64_add_safe()
776 res.tv_nsec = 0; in timespec64_add_safe()
779 return res; in timespec64_add_safe()
Dntp.c631 int res = -EAGAIN; in sync_hw_clock() local
651 res = update_persistent_clock64(to_set); in sync_hw_clock()
652 if (res != -ENODEV) in sync_hw_clock()
656 res = update_rtc(&to_set, &offset_nsec); in sync_hw_clock()
657 if (res == -ENODEV) in sync_hw_clock()
660 sched_sync_hw_clock(offset_nsec, res != 0); in sync_hw_clock()
/kernel/debug/kdb/
Dkdb_support.c266 int kdb_getarea_size(void *res, unsigned long addr, size_t size) in kdb_getarea_size() argument
268 int ret = copy_from_kernel_nofault((char *)res, (char *)addr, size); in kdb_getarea_size()
291 int kdb_putarea_size(unsigned long addr, void *res, size_t size) in kdb_putarea_size() argument
293 int ret = copy_to_kernel_nofault((char *)addr, (char *)res, size); in kdb_putarea_size()
317 static int kdb_getphys(void *res, unsigned long addr, size_t size) in kdb_getphys() argument
328 memcpy(res, vaddr + (addr & (PAGE_SIZE - 1)), size); in kdb_getphys()
/kernel/locking/
Drtmutex.c1146 int chain_walk = 0, res; in task_blocks_on_rt_mutex() local
1185 res = __ww_mutex_add_waiter(waiter, rtm, ww_ctx); in task_blocks_on_rt_mutex()
1186 if (res) { in task_blocks_on_rt_mutex()
1191 return res; in task_blocks_on_rt_mutex()
1231 res = rt_mutex_adjust_prio_chain(owner, chwalk, lock, in task_blocks_on_rt_mutex()
1236 return res; in task_blocks_on_rt_mutex()
1416 bool res = true; in rtmutex_spin_on_owner() local
1440 res = false; in rtmutex_spin_on_owner()
1446 return res; in rtmutex_spin_on_owner()
1583 static void __sched rt_mutex_handle_deadlock(int res, int detect_deadlock, in rt_mutex_handle_deadlock() argument
[all …]
/kernel/bpf/
Doffload.c302 int res; in bpf_prog_offload_info_fill() local
305 res = ns_get_path_cb(&ns_path, bpf_prog_offload_info_fill_ns, &args); in bpf_prog_offload_info_fill()
306 if (res) { in bpf_prog_offload_info_fill()
309 return res; in bpf_prog_offload_info_fill()
526 int res; in bpf_map_offload_info_fill() local
528 res = ns_get_path_cb(&ns_path, bpf_map_offload_info_fill_ns, &args); in bpf_map_offload_info_fill()
529 if (res) { in bpf_map_offload_info_fill()
532 return res; in bpf_map_offload_info_fill()
Ddisasm.c28 const char *res; in __func_get_name() local
30 res = cbs->cb_call(cbs->private_data, insn); in __func_get_name()
31 if (res) in __func_get_name()
32 return res; in __func_get_name()
Dhelpers.c439 unsigned long long *res, bool *is_negative) in __bpf_strtoull() argument
448 if (!buf || !buf_len || !res || !is_negative) in __bpf_strtoull()
475 val_len = _parse_integer(cur_buf, base, res); in __bpf_strtoull()
490 long long *res) in __bpf_strtoll() argument
502 *res = -_res; in __bpf_strtoll()
506 *res = _res; in __bpf_strtoll()
512 long *, res) in BPF_CALL_4() argument
522 *res = _res; in BPF_CALL_4()
537 unsigned long *, res) in BPF_CALL_4() argument
550 *res = _res; in BPF_CALL_4()
/kernel/futex/
Drequeue.c772 int res, ret; in futex_wait_requeue_pi() local
863 res = fixup_pi_owner(uaddr2, &q, !ret); in futex_wait_requeue_pi()
868 if (res) in futex_wait_requeue_pi()
869 ret = (res < 0) ? res : 0; in futex_wait_requeue_pi()
Dpi.c937 int res, ret; in futex_lock_pi() local
1060 res = fixup_pi_owner(uaddr, &q, !ret); in futex_lock_pi()
1065 if (res) in futex_lock_pi()
1066 ret = (res < 0) ? res : 0; in futex_lock_pi()
/kernel/sched/
Dpsi.c1186 int psi_show(struct seq_file *m, struct psi_group *group, enum psi_res res) in psi_show() argument
1204 only_full = res == PSI_IRQ; in psi_show()
1213 if (!(group == &psi_system && res == PSI_CPU && full)) { in psi_show()
1215 avg[w] = group->avg[res * 2 + full][w]; in psi_show()
1216 total = div_u64(group->total[PSI_AVGS][res * 2 + full], in psi_show()
1232 char *buf, enum psi_res res) in psi_trigger_create() argument
1243 state = PSI_IO_SOME + res * 2; in psi_trigger_create()
1245 state = PSI_IO_FULL + res * 2; in psi_trigger_create()
1250 if (res == PSI_IRQ && --state != PSI_IRQ_FULL) in psi_trigger_create()
1429 size_t nbytes, enum psi_res res) in psi_write() argument
[all …]

12