/kernel/time/ |
D | timeconst.bc | 16 define fmul(b,n,d) { 17 return (2^b*n+d-1)/d; 21 (imul * n) + (fmulxx * n + fadjxx) >> xx) */ 22 define fadj(b,n,d) { 24 d = d/gcd(n,d); 33 define fmuls(b,n,d) { 36 m = fmul(s,n,d); 44 print "/* Automatically generated by kernel/time/timeconst.bc */\n" 45 print "/* Time conversion constants for HZ == ", hz, " */\n" 46 print "\n" [all …]
|
/kernel/locking/ |
D | test-ww_mutex.c | 318 unsigned int n, last = nthreads - 1; in __test_cycle() local 325 for (n = 0; n < nthreads; n++) { in __test_cycle() 326 struct test_cycle *cycle = &cycles[n]; in __test_cycle() 329 if (n == last) in __test_cycle() 332 cycle->b_mutex = &cycles[n + 1].a_mutex; in __test_cycle() 334 if (n == 0) in __test_cycle() 337 cycle->a_signal = &cycles[n - 1].b_signal; in __test_cycle() 344 for (n = 0; n < nthreads; n++) in __test_cycle() 345 queue_work(wq, &cycles[n].work); in __test_cycle() 350 for (n = 0; n < nthreads; n++) { in __test_cycle() [all …]
|
D | ww_mutex.h | 96 struct rb_node *n = rb_first(&lock->rtmutex.waiters.rb_root); in __ww_waiter_first() local 97 if (!n) in __ww_waiter_first() 99 return rb_entry(n, struct rt_mutex_waiter, tree_entry); in __ww_waiter_first() 105 struct rb_node *n = rb_next(&w->tree_entry); in __ww_waiter_next() local 106 if (!n) in __ww_waiter_next() 108 return rb_entry(n, struct rt_mutex_waiter, tree_entry); in __ww_waiter_next() 114 struct rb_node *n = rb_prev(&w->tree_entry); in __ww_waiter_prev() local 115 if (!n) in __ww_waiter_prev() 117 return rb_entry(n, struct rt_mutex_waiter, tree_entry); in __ww_waiter_prev() 123 struct rb_node *n = rb_last(&lock->rtmutex.waiters.rb_root); in __ww_waiter_last() local [all …]
|
/kernel/ |
D | notifier.c | 23 struct notifier_block *n, in notifier_chain_register() argument 27 if (unlikely((*nl) == n)) { in notifier_chain_register() 29 n->notifier_call); in notifier_chain_register() 32 if (n->priority > (*nl)->priority) in notifier_chain_register() 34 if (n->priority == (*nl)->priority && unique_priority) in notifier_chain_register() 38 n->next = *nl; in notifier_chain_register() 39 rcu_assign_pointer(*nl, n); in notifier_chain_register() 44 struct notifier_block *n) in notifier_chain_unregister() argument 47 if ((*nl) == n) { in notifier_chain_unregister() 48 rcu_assign_pointer(*nl, n->next); in notifier_chain_unregister() [all …]
|
D | auditsc.c | 150 unsigned n; in audit_match_perm() local 154 n = ctx->major; in audit_match_perm() 156 switch (audit_classify_syscall(ctx->arch, n)) { in audit_match_perm() 159 audit_match_class(AUDIT_CLASS_WRITE, n)) in audit_match_perm() 162 audit_match_class(AUDIT_CLASS_READ, n)) in audit_match_perm() 165 audit_match_class(AUDIT_CLASS_CHATTR, n)) in audit_match_perm() 170 audit_match_class(AUDIT_CLASS_WRITE_32, n)) in audit_match_perm() 173 audit_match_class(AUDIT_CLASS_READ_32, n)) in audit_match_perm() 176 audit_match_class(AUDIT_CLASS_CHATTR_32, n)) in audit_match_perm() 196 struct audit_names *n; in audit_match_filetype() local [all …]
|
D | watch_queue.c | 99 struct watch_notification *n) in post_one_notification() argument 123 len = n->info & WATCH_INFO_LENGTH; in post_one_notification() 125 memcpy(p + offset, n, len); in post_one_notification() 160 const struct watch_notification *n) in filter_watch_notification() argument 164 unsigned int st_index = n->subtype / st_bits; in filter_watch_notification() 165 unsigned int st_bit = 1U << (n->subtype % st_bits); in filter_watch_notification() 168 if (!test_bit(n->type, wf->type_filter)) in filter_watch_notification() 173 if (n->type == wt->type && in filter_watch_notification() 175 (n->info & wt->info_mask) == wt->info_filter) in filter_watch_notification() 196 struct watch_notification *n, in __post_watch_notification() argument [all …]
|
D | Makefile | 26 KCOV_INSTRUMENT_softirq.o := n 29 KCSAN_SANITIZE_softirq.o = n 32 KCOV_INSTRUMENT_extable.o := n 33 KCOV_INSTRUMENT_stacktrace.o := n 35 KCOV_INSTRUMENT_kcov.o := n 38 KASAN_SANITIZE_kcov.o := n 39 KCSAN_SANITIZE_kcov.o := n 40 UBSAN_SANITIZE_kcov.o := n 41 KMSAN_SANITIZE_kcov.o := n 133 KASAN_SANITIZE_stackleak.o := n [all …]
|
D | Kconfig.locks | 96 # - DEBUG_SPINLOCK=n and GENERIC_LOCKBREAK=n and ARCH_INLINE_*LOCK=y 99 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y 102 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y 104 # - DEBUG_SPINLOCK=n and PREEMPTION=n 107 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
|
D | audit_tree.c | 221 unsigned long n = key / L1_CACHE_BYTES; in chunk_hash() local 222 return chunk_hash_heads + n % HASH_SIZE; in chunk_hash() 263 int n; in audit_tree_match() local 264 for (n = 0; n < chunk->count; n++) in audit_tree_match() 265 if (chunk->owners[n].owner == tree) in audit_tree_match() 463 int n; in tag_chunk() local 478 for (n = 0; n < old->count; n++) { in tag_chunk() 479 if (old->owners[n].owner == tree) { in tag_chunk() 1006 int n; in evict_chunk() local 1027 for (n = 0; n < chunk->count; n++) in evict_chunk() [all …]
|
D | seccomp.c | 1056 static void seccomp_handle_addfd(struct seccomp_kaddfd *addfd, struct seccomp_knotif *n) in seccomp_handle_addfd() argument 1074 n->state = SECCOMP_NOTIFY_SENT; in seccomp_handle_addfd() 1077 n->flags = 0; in seccomp_handle_addfd() 1078 n->error = 0; in seccomp_handle_addfd() 1079 n->val = fd; in seccomp_handle_addfd() 1091 struct seccomp_knotif *n) in should_sleep_killable() argument 1093 return match->wait_killable_recv && n->state == SECCOMP_NOTIFY_SENT; in should_sleep_killable() 1103 struct seccomp_knotif n = {}; in seccomp_do_user_notification() local 1111 n.task = current; in seccomp_do_user_notification() 1112 n.state = SECCOMP_NOTIFY_INIT; in seccomp_do_user_notification() [all …]
|
D | user.c | 239 int n; in uid_cache_init() local 244 for(n = 0; n < UIDHASH_SZ; ++n) in uid_cache_init() 245 INIT_HLIST_HEAD(uidhash_table + n); in uid_cache_init()
|
/kernel/power/ |
D | wakeup_reason.c | 100 struct wakeup_irq_node *n; in delete_list() local 103 n = list_first_entry(head, struct wakeup_irq_node, siblings); in delete_list() 104 list_del(&n->siblings); in delete_list() 105 kmem_cache_free(wakeup_irq_nodes_cache, n); in delete_list() 111 struct wakeup_irq_node *n = NULL; in add_sibling_node_sorted() local 118 list_for_each_entry(n, head, siblings) { in add_sibling_node_sorted() 119 if (n->irq < irq) in add_sibling_node_sorted() 120 predecessor = &n->siblings; in add_sibling_node_sorted() 121 else if (n->irq == irq) in add_sibling_node_sorted() 127 n = create_node(irq); in add_sibling_node_sorted() [all …]
|
D | main.c | 110 const char *buf, size_t n) in pm_async_store() argument 121 return n; in pm_async_store() 153 static suspend_state_t decode_suspend_state(const char *buf, size_t n) in decode_suspend_state() argument 159 p = memchr(buf, '\n', n); in decode_suspend_state() 160 len = p ? p - buf : n; in decode_suspend_state() 173 const char *buf, size_t n) in mem_sleep_store() argument 187 state = decode_suspend_state(buf, n); in mem_sleep_store() 195 return error ? error : n; in mem_sleep_store() 216 const char *buf, size_t n) in sync_on_suspend_store() argument 227 return n; in sync_on_suspend_store() [all …]
|
D | hibernate.c | 1116 const char *buf, size_t n) in disk_store() argument 1128 p = memchr(buf, '\n', n); in disk_store() 1129 len = p ? p - buf : n; in disk_store() 1162 return error ? error : n; in disk_store() 1175 const char *buf, size_t n) in resume_store() argument 1178 int len = n; in resume_store() 1201 return n; in resume_store() 1214 size_t n) in resume_offset_store() argument 1224 return n; in resume_offset_store() 1236 const char *buf, size_t n) in image_size_store() argument [all …]
|
/kernel/configs/ |
D | nopm.config | 1 CONFIG_PM=n 2 CONFIG_SUSPEND=n 3 CONFIG_HIBERNATION=n 6 CONFIG_CPU_IDLE=n 9 CONFIG_XEN=n 12 CONFIG_ARCH_OMAP2PLUS_TYPICAL=n 13 CONFIG_ARCH_RENESAS=n 14 CONFIG_ARCH_TEGRA=n 15 CONFIG_ARCH_VEXPRESS=n
|
/kernel/sched/ |
D | loadavg.c | 109 fixed_power_int(unsigned long x, unsigned int frac_bits, unsigned int n) in fixed_power_int() argument 113 if (n) { in fixed_power_int() 115 if (n & 1) { in fixed_power_int() 120 n >>= 1; in fixed_power_int() 121 if (!n) in fixed_power_int() 157 unsigned long active, unsigned int n) in calc_load_n() argument 159 return calc_load(load, fixed_power_int(exp, FSHIFT, n), active); in calc_load_n() 306 long delta, active, n; in calc_global_nohz() local 314 n = 1 + (delta / LOAD_FREQ); in calc_global_nohz() 319 avenrun[0] = calc_load_n(avenrun[0], EXP_1, active, n); in calc_global_nohz() [all …]
|
D | stats.c | 188 unsigned long n = *offset; in schedstat_start() local 190 if (n == 0) in schedstat_start() 193 n--; in schedstat_start() 195 if (n > 0) in schedstat_start() 196 n = cpumask_next(n - 1, cpu_online_mask); in schedstat_start() 198 n = cpumask_first(cpu_online_mask); in schedstat_start() 200 *offset = n + 1; in schedstat_start() 202 if (n < nr_cpu_ids) in schedstat_start() 203 return (void *)(unsigned long)(n + 2); in schedstat_start()
|
D | debug.c | 755 #define P64(n) SEQ_printf(m, " .%-30s: %Ld\n", #n, rq->n); in print_cpu() argument 761 #define P(n) SEQ_printf(m, " .%-30s: %d\n", #n, schedstat_val(rq->n)); in print_cpu() argument 876 unsigned long n = *offset; in sched_debug_start() local 878 if (n == 0) in sched_debug_start() 881 n--; in sched_debug_start() 883 if (n > 0) in sched_debug_start() 884 n = cpumask_next(n - 1, cpu_online_mask); in sched_debug_start() 886 n = cpumask_first(cpu_online_mask); in sched_debug_start() 888 *offset = n + 1; in sched_debug_start() 890 if (n < nr_cpu_ids) in sched_debug_start() [all …]
|
/kernel/module/ |
D | tree_lookup.c | 22 static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n) in __mod_tree_val() argument 24 struct module_layout *layout = container_of(n, struct module_layout, mtn.node); in __mod_tree_val() 29 static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n) in __mod_tree_size() argument 31 struct module_layout *layout = container_of(n, struct module_layout, mtn.node); in __mod_tree_size() 43 mod_tree_comp(void *key, struct latch_tree_node *n) in mod_tree_comp() argument 48 start = __mod_tree_val(n); in mod_tree_comp() 52 end = start + __mod_tree_size(n); in mod_tree_comp()
|
/kernel/bpf/ |
D | tnum.c | 183 size_t n; in tnum_sbin() local 185 for (n = 64; n; n--) { in tnum_sbin() 186 if (n < size) { in tnum_sbin() 188 str[n - 1] = 'x'; in tnum_sbin() 190 str[n - 1] = '1'; in tnum_sbin() 192 str[n - 1] = '0'; in tnum_sbin()
|
/kernel/gcov/ |
D | Kconfig | 9 default n 30 GCOV_PROFILE_main.o := n 32 GCOV_PROFILE := n 38 def_bool n 45 default n
|
/kernel/entry/ |
D | Makefile | 5 KASAN_SANITIZE := n 6 UBSAN_SANITIZE := n 7 KCOV_INSTRUMENT := n
|
/kernel/events/ |
D | uprobes.c | 1266 struct rb_node *n = uprobes_tree.rb_node; in find_node_in_range() local 1268 while (n) { in find_node_in_range() 1269 struct uprobe *u = rb_entry(n, struct uprobe, rb_node); in find_node_in_range() 1272 n = n->rb_left; in find_node_in_range() 1274 n = n->rb_right; in find_node_in_range() 1277 n = n->rb_left; in find_node_in_range() 1279 n = n->rb_right; in find_node_in_range() 1285 return n; in find_node_in_range() 1297 struct rb_node *n, *t; in build_probe_list() local 1305 n = find_node_in_range(inode, min, max); in build_probe_list() [all …]
|
D | internal.h | 177 memcpy_common(void *dst, const void *src, unsigned long n) in memcpy_common() argument 179 memcpy(dst, src, n); in memcpy_common() 186 memcpy_skip(void *dst, const void *src, unsigned long n) in DEFINE_OUTPUT_COPY() 197 arch_perf_out_copy_user(void *dst, const void *src, unsigned long n) in DEFINE_OUTPUT_COPY() 202 ret = __copy_from_user_inatomic(dst, src, n); in DEFINE_OUTPUT_COPY()
|
/kernel/kcsan/ |
D | Makefile | 2 KCSAN_SANITIZE := n 3 KCOV_INSTRUMENT := n 4 UBSAN_SANITIZE := n
|