/kernel/ |
D | resource_kunit.c | 31 struct resource r; member 37 .r1 = &r1, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true, 39 .r1 = &r2, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true, 41 .r1 = &r3, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true, 43 .r1 = &r4, .r2 = &r0, .r.start = R0_START, .r.end = R0_END, .ret = true, 49 .r1 = &r4, .r2 = &r1, .r.start = R1_START, .r.end = R4_END, .ret = true, 53 .r1 = &r2, .r2 = &r4, .r.start = R4_START, .r.end = R4_END, .ret = true, 55 .r1 = &r3, .r2 = &r4, .r.start = R4_START, .r.end = R3_END, .ret = true, 61 .r1 = &r1, .r2 = &r0, .r.start = R1_START, .r.end = R1_END, .ret = true, 63 .r1 = &r2, .r2 = &r0, .r.start = R2_START, .r.end = R2_END, .ret = true, [all …]
|
D | compat.c | 83 int put_compat_rusage(const struct rusage *r, struct compat_rusage __user *ru) in put_compat_rusage() argument 87 r32.ru_utime.tv_sec = r->ru_utime.tv_sec; in put_compat_rusage() 88 r32.ru_utime.tv_usec = r->ru_utime.tv_usec; in put_compat_rusage() 89 r32.ru_stime.tv_sec = r->ru_stime.tv_sec; in put_compat_rusage() 90 r32.ru_stime.tv_usec = r->ru_stime.tv_usec; in put_compat_rusage() 91 r32.ru_maxrss = r->ru_maxrss; in put_compat_rusage() 92 r32.ru_ixrss = r->ru_ixrss; in put_compat_rusage() 93 r32.ru_idrss = r->ru_idrss; in put_compat_rusage() 94 r32.ru_isrss = r->ru_isrss; in put_compat_rusage() 95 r32.ru_minflt = r->ru_minflt; in put_compat_rusage() [all …]
|
D | sys.c | 1530 struct rlimit r; in COMPAT_SYSCALL_DEFINE2() local 1537 r.rlim_cur = RLIM_INFINITY; in COMPAT_SYSCALL_DEFINE2() 1539 r.rlim_cur = r32.rlim_cur; in COMPAT_SYSCALL_DEFINE2() 1541 r.rlim_max = RLIM_INFINITY; in COMPAT_SYSCALL_DEFINE2() 1543 r.rlim_max = r32.rlim_max; in COMPAT_SYSCALL_DEFINE2() 1544 return do_prlimit(current, resource, &r, NULL); in COMPAT_SYSCALL_DEFINE2() 1550 struct rlimit r; in COMPAT_SYSCALL_DEFINE2() local 1553 ret = do_prlimit(current, resource, NULL, &r); in COMPAT_SYSCALL_DEFINE2() 1556 if (r.rlim_cur > COMPAT_RLIM_INFINITY) in COMPAT_SYSCALL_DEFINE2() 1559 r32.rlim_cur = r.rlim_cur; in COMPAT_SYSCALL_DEFINE2() [all …]
|
D | audit_watch.c | 227 static void audit_watch_log_rule_change(struct audit_krule *r, struct audit_watch *w, char *op) in audit_watch_log_rule_change() argument 239 audit_log_key(ab, r->filterkey); in audit_watch_log_rule_change() 240 audit_log_format(ab, " list=%d res=1", r->listnr); in audit_watch_log_rule_change() 250 struct audit_krule *r, *nextr; in audit_update_watch() local 277 list_for_each_entry_safe(r, nextr, &owatch->rules, rlist) { in audit_update_watch() 279 oentry = container_of(r, struct audit_entry, rule); in audit_update_watch() 325 struct audit_krule *r, *nextr; in audit_remove_parent_watches() local 330 list_for_each_entry_safe(r, nextr, &w->rules, rlist) { in audit_remove_parent_watches() 331 e = container_of(r, struct audit_entry, rule); in audit_remove_parent_watches() 332 audit_watch_log_rule_change(r, w, "remove_rule"); in audit_remove_parent_watches() [all …]
|
D | scftorture.c | 261 unsigned long r = torture_random(this_cpu_ptr(&scf_torture_rand)); in scf_handler() local 271 if (!(r & 0xffc0)) { in scf_handler() 272 udelay(r & 0x3f); in scf_handler() 276 if (r & 0xfff) in scf_handler() 278 r = (r >> 12); in scf_handler() 280 udelay((r & 0xff) + 1); in scf_handler() 283 r = r % longwait + 1; in scf_handler() 284 for (i = 0; i < r; i++) { in scf_handler()
|
D | auditfilter.c | 1081 struct audit_krule *r; in audit_list_rules() local 1087 list_for_each_entry(r, &audit_rules_list[i], list) { in audit_list_rules() 1090 data = audit_krule_to_data(r); in audit_list_rules() 1400 static int update_lsm_rule(struct audit_krule *r) in update_lsm_rule() argument 1402 struct audit_entry *entry = container_of(r, struct audit_entry, rule); in update_lsm_rule() 1406 if (!security_audit_rule_known(r)) in update_lsm_rule() 1409 nentry = audit_dupe_rule(r); in update_lsm_rule() 1417 if (r->watch) in update_lsm_rule() 1418 list_del(&r->rlist); in update_lsm_rule() 1420 list_del(&r->list); in update_lsm_rule() [all …]
|
D | crash_core.c | 375 size_t r; in vmcoreinfo_append_str() local 378 r = vscnprintf(buf, sizeof(buf), fmt, args); in vmcoreinfo_append_str() 381 r = min(r, (size_t)VMCOREINFO_BYTES - vmcoreinfo_size); in vmcoreinfo_append_str() 383 memcpy(&vmcoreinfo_data[vmcoreinfo_size], buf, r); in vmcoreinfo_append_str() 385 vmcoreinfo_size += r; in vmcoreinfo_append_str()
|
D | utsname_sysctl.c | 37 int r; in proc_do_uts_string() local 52 r = proc_dostring(&uts_table, write, buffer, lenp, ppos); in proc_do_uts_string() 68 return r; in proc_do_uts_string()
|
D | resource.c | 111 struct resource *r = v, *p; in r_show() local 116 for (depth = 0, p = r; depth < MAX_IORES_LEVEL; depth++, p = p->parent) in r_show() 121 start = r->start; in r_show() 122 end = r->end; in r_show() 131 r->name ? r->name : "<BAD>"); in r_show() 228 static void __release_child_resources(struct resource *r) in __release_child_resources() argument 233 p = r->child; in __release_child_resources() 234 r->child = NULL; in __release_child_resources() 251 void release_child_resources(struct resource *r) in release_child_resources() argument 254 __release_child_resources(r); in release_child_resources() [all …]
|
D | exit.c | 1716 struct rusage r; in SYSCALL_DEFINE5() local 1718 long err = kernel_waitid(which, upid, &info, options, ru ? &r : NULL); in SYSCALL_DEFINE5() 1724 if (ru && copy_to_user(ru, &r, sizeof(struct rusage))) in SYSCALL_DEFINE5() 1808 struct rusage r; in SYSCALL_DEFINE4() local 1809 long err = kernel_wait4(upid, stat_addr, options, ru ? &r : NULL); in SYSCALL_DEFINE4() 1812 if (ru && copy_to_user(ru, &r, sizeof(struct rusage))) in SYSCALL_DEFINE4() 1838 struct rusage r; in COMPAT_SYSCALL_DEFINE4() local 1839 long err = kernel_wait4(pid, stat_addr, options, ru ? &r : NULL); in COMPAT_SYSCALL_DEFINE4() 1841 if (ru && put_compat_rusage(&r, ru)) in COMPAT_SYSCALL_DEFINE4()
|
/kernel/printk/ |
D | printk.c | 648 int r; in devkmsg_emit() local 651 r = vprintk_emit(facility, level, NULL, fmt, args); in devkmsg_emit() 654 return r; in devkmsg_emit() 723 struct printk_record *r = &user->record; in devkmsg_read() local 734 if (!prb_read_valid(prb, atomic64_read(&user->seq), r)) { in devkmsg_read() 752 atomic64_read(&user->seq), r)); /* LMM(devkmsg_read:A) */ in devkmsg_read() 757 if (r->info->seq != atomic64_read(&user->seq)) { in devkmsg_read() 759 atomic64_set(&user->seq, r->info->seq); in devkmsg_read() 764 len = info_print_ext_header(user->buf, sizeof(user->buf), r->info); in devkmsg_read() 766 &r->text_buf[0], r->info->text_len, in devkmsg_read() [all …]
|
D | printk_ringbuffer.h | 297 static inline void prb_rec_init_wr(struct printk_record *r, in prb_rec_init_wr() argument 300 r->info = NULL; in prb_rec_init_wr() 301 r->text_buf = NULL; in prb_rec_init_wr() 302 r->text_buf_size = text_buf_size; in prb_rec_init_wr() 306 struct printk_record *r); 308 struct printk_record *r, u32 caller_id, unsigned int max_size); 332 static inline void prb_rec_init_rd(struct printk_record *r, in prb_rec_init_rd() argument 336 r->info = info; in prb_rec_init_rd() 337 r->text_buf = text_buf; in prb_rec_init_rd() 338 r->text_buf_size = text_buf_size; in prb_rec_init_rd() [all …]
|
D | printk_ringbuffer.c | 1351 struct printk_record *r, u32 caller_id, unsigned int max_size) in prb_reserve_in_last() argument 1394 if (!data_check_size(&rb->text_data_ring, r->text_buf_size)) in prb_reserve_in_last() 1397 if (r->text_buf_size > max_size) in prb_reserve_in_last() 1400 r->text_buf = data_alloc(rb, r->text_buf_size, in prb_reserve_in_last() 1416 r->text_buf_size += info->text_len; in prb_reserve_in_last() 1418 if (!data_check_size(&rb->text_data_ring, r->text_buf_size)) in prb_reserve_in_last() 1421 if (r->text_buf_size > max_size) in prb_reserve_in_last() 1424 r->text_buf = data_realloc(rb, r->text_buf_size, in prb_reserve_in_last() 1427 if (r->text_buf_size && !r->text_buf) in prb_reserve_in_last() 1430 r->info = info; in prb_reserve_in_last() [all …]
|
/kernel/trace/rv/ |
D | rv_reactors.c | 75 struct rv_reactor_def *r; in get_reactor_rdef_by_name() local 77 list_for_each_entry(r, &rv_reactors_list, list) { in get_reactor_rdef_by_name() 78 if (strcmp(name, r->reactor->name) == 0) in get_reactor_rdef_by_name() 79 return r; in get_reactor_rdef_by_name() 280 struct rv_reactor_def *r; in __rv_register_reactor() local 282 list_for_each_entry(r, &rv_reactors_list, list) { in __rv_register_reactor() 283 if (strcmp(reactor->name, r->reactor->name) == 0) { in __rv_register_reactor() 289 r = kzalloc(sizeof(struct rv_reactor_def), GFP_KERNEL); in __rv_register_reactor() 290 if (!r) in __rv_register_reactor() 293 r->reactor = reactor; in __rv_register_reactor() [all …]
|
D | rv.c | 696 struct rv_monitor_def *r; in rv_register_monitor() local 707 list_for_each_entry(r, &rv_monitors_list, list) { in rv_register_monitor() 708 if (strcmp(monitor->name, r->monitor->name) == 0) { in rv_register_monitor() 715 r = kzalloc(sizeof(struct rv_monitor_def), GFP_KERNEL); in rv_register_monitor() 716 if (!r) { in rv_register_monitor() 721 r->monitor = monitor; in rv_register_monitor() 723 retval = create_monitor_dir(r); in rv_register_monitor() 725 kfree(r); in rv_register_monitor() 729 list_add_tail(&r->list, &rv_monitors_list); in rv_register_monitor()
|
/kernel/time/ |
D | sched_clock.c | 158 unsigned long r, flags; in sched_clock_register() local 201 r = rate; in sched_clock_register() 202 if (r >= 4000000) { in sched_clock_register() 203 r = DIV_ROUND_CLOSEST(r, 1000000); in sched_clock_register() 205 } else if (r >= 4000) { in sched_clock_register() 206 r = DIV_ROUND_CLOSEST(r, 1000); in sched_clock_register() 216 bits, r, r_unit, res, wrap); in sched_clock_register()
|
/kernel/kcsan/ |
D | kcsan_test.c | 157 static bool __report_matches(const struct expect_report *r) in __report_matches() argument 159 const bool is_assert = (r->access[0].type | r->access[1].type) & KCSAN_ACCESS_ASSERT; in __report_matches() 182 if (r->access[1].fn) { in __report_matches() 187 scnprintf(tmp[0], sizeof(tmp[0]), "%pS", r->access[0].fn); in __report_matches() 188 scnprintf(tmp[1], sizeof(tmp[1]), "%pS", r->access[1].fn); in __report_matches() 191 cmp < 0 ? r->access[0].fn : r->access[1].fn, in __report_matches() 192 cmp < 0 ? r->access[1].fn : r->access[0].fn); in __report_matches() 194 scnprintf(cur, end - cur, "%pS", r->access[0].fn); in __report_matches() 204 if (!r->access[1].fn) in __report_matches() 209 const int ty = r->access[i].type; in __report_matches() [all …]
|
/kernel/trace/ |
D | trace_events_filter.c | 891 static int regex_match_full(char *str, struct regex *r, int len) in regex_match_full() argument 895 return strcmp(str, r->pattern) == 0; in regex_match_full() 897 return strncmp(str, r->pattern, len) == 0; in regex_match_full() 900 static int regex_match_front(char *str, struct regex *r, int len) in regex_match_front() argument 902 if (len && len < r->len) in regex_match_front() 905 return strncmp(str, r->pattern, r->len) == 0; in regex_match_front() 908 static int regex_match_middle(char *str, struct regex *r, int len) in regex_match_middle() argument 911 return strstr(str, r->pattern) != NULL; in regex_match_middle() 913 return strnstr(str, r->pattern, len) != NULL; in regex_match_middle() 916 static int regex_match_end(char *str, struct regex *r, int len) in regex_match_end() argument [all …]
|
D | trace_stack.c | 329 int r; in stack_max_size_read() local 331 r = snprintf(buf, sizeof(buf), "%ld\n", *ptr); in stack_max_size_read() 332 if (r > sizeof(buf)) in stack_max_size_read() 333 r = sizeof(buf); in stack_max_size_read() 334 return simple_read_from_buffer(ubuf, count, ppos, buf, r); in stack_max_size_read()
|
D | blktrace.c | 1012 struct blk_io_trace_remap r; in blk_add_trace_bio_remap() local 1021 r.device_from = cpu_to_be32(dev); in blk_add_trace_bio_remap() 1022 r.device_to = cpu_to_be32(bio_dev(bio)); in blk_add_trace_bio_remap() 1023 r.sector_from = cpu_to_be64(from); in blk_add_trace_bio_remap() 1028 sizeof(r), &r, blk_trace_bio_get_cgid(q, bio)); in blk_add_trace_bio_remap() 1048 struct blk_io_trace_remap r; in blk_add_trace_rq_remap() local 1057 r.device_from = cpu_to_be32(dev); in blk_add_trace_rq_remap() 1058 r.device_to = cpu_to_be32(disk_devt(rq->q->disk)); in blk_add_trace_rq_remap() 1059 r.sector_from = cpu_to_be64(from); in blk_add_trace_rq_remap() 1063 sizeof(r), &r, blk_trace_request_get_cgid(rq)); in blk_add_trace_rq_remap()
|
D | trace_events.c | 286 const char *fmt, *c, *r, *a; in test_event_printk() local 413 r = strstr(fmt + i, "REC->"); in test_event_printk() 414 if (r && (!c || r < c)) { in test_event_printk() 424 if ((a && (a < r)) || test_field(r, call)) in test_event_printk() 426 } else if ((r = strstr(fmt + i, "__get_dynamic_array(")) && in test_event_printk() 427 (!c || r < c)) { in test_event_printk() 429 } else if ((r = strstr(fmt + i, "__get_sockaddr(")) && in test_event_printk() 430 (!c || r < c)) { in test_event_printk() 1681 int r = -ENODEV; in event_filter_read() local 1700 r = simple_read_from_buffer(ubuf, cnt, ppos, in event_filter_read() [all …]
|
/kernel/locking/ |
D | rwbase_rt.c | 55 int r; in rwbase_read_trylock() local 61 for (r = atomic_read(&rwb->readers); r < 0;) { in rwbase_read_trylock() 62 if (likely(atomic_try_cmpxchg_acquire(&rwb->readers, &r, r + 1))) in rwbase_read_trylock()
|
D | test-ww_mutex.c | 392 int n, r, tmp; in get_random_order() local 402 r = prandom_u32_max(n + 1); in get_random_order() 403 if (r != n) { in get_random_order() 405 order[n] = order[r]; in get_random_order() 406 order[r] = tmp; in get_random_order()
|
/kernel/sched/ |
D | cpudeadline.c | 27 int l, r, largest; in cpudl_heapify_down() local 40 r = right_child(idx); in cpudl_heapify_down() 49 if ((r < cp->size) && dl_time_before(largest_dl, in cpudl_heapify_down() 50 cp->elements[r].dl)) in cpudl_heapify_down() 51 largest = r; in cpudl_heapify_down()
|
/kernel/rcu/ |
D | refscale.c | 664 int exp, r; in main_func() local 700 for (r = 0; r < nreaders; r++) { in main_func() 701 smp_store_release(&reader_tasks[r].start_reader, 1); in main_func() 702 wake_up(&reader_tasks[r].wq); in main_func()
|