/kernel/time/ |
D | timeconst.bc | 16 define fmul(b,n,d) { 17 return (2^b*n+d-1)/d; 21 (imul * n) + (fmulxx * n + fadjxx) >> xx) */ 22 define fadj(b,n,d) { 24 d = d/gcd(n,d); 33 define fmuls(b,n,d) { 36 m = fmul(s,n,d); 44 print "/* Automatically generated by kernel/time/timeconst.bc */\n" 45 print "/* Time conversion constants for HZ == ", hz, " */\n" 46 print "\n" [all …]
|
/kernel/locking/ |
D | test-ww_mutex.c | 318 unsigned int n, last = nthreads - 1; in __test_cycle() local 325 for (n = 0; n < nthreads; n++) { in __test_cycle() 326 struct test_cycle *cycle = &cycles[n]; in __test_cycle() 329 if (n == last) in __test_cycle() 332 cycle->b_mutex = &cycles[n + 1].a_mutex; in __test_cycle() 334 if (n == 0) in __test_cycle() 337 cycle->a_signal = &cycles[n - 1].b_signal; in __test_cycle() 344 for (n = 0; n < nthreads; n++) in __test_cycle() 345 queue_work(wq, &cycles[n].work); in __test_cycle() 350 for (n = 0; n < nthreads; n++) { in __test_cycle() [all …]
|
D | ww_mutex.h | 96 struct rb_node *n = rb_first(&lock->rtmutex.waiters.rb_root); in __ww_waiter_first() local 97 if (!n) in __ww_waiter_first() 99 return rb_entry(n, struct rt_mutex_waiter, tree.entry); in __ww_waiter_first() 105 struct rb_node *n = rb_next(&w->tree.entry); in __ww_waiter_next() local 106 if (!n) in __ww_waiter_next() 108 return rb_entry(n, struct rt_mutex_waiter, tree.entry); in __ww_waiter_next() 114 struct rb_node *n = rb_prev(&w->tree.entry); in __ww_waiter_prev() local 115 if (!n) in __ww_waiter_prev() 117 return rb_entry(n, struct rt_mutex_waiter, tree.entry); in __ww_waiter_prev() 123 struct rb_node *n = rb_last(&lock->rtmutex.waiters.rb_root); in __ww_waiter_last() local [all …]
|
/kernel/ |
D | notifier.c | 26 struct notifier_block *n, in notifier_chain_register() argument 30 if (unlikely((*nl) == n)) { in notifier_chain_register() 32 n->notifier_call); in notifier_chain_register() 35 if (n->priority > (*nl)->priority) in notifier_chain_register() 37 if (n->priority == (*nl)->priority && unique_priority) in notifier_chain_register() 41 n->next = *nl; in notifier_chain_register() 42 rcu_assign_pointer(*nl, n); in notifier_chain_register() 43 trace_notifier_register((void *)n->notifier_call); in notifier_chain_register() 48 struct notifier_block *n) in notifier_chain_unregister() argument 51 if ((*nl) == n) { in notifier_chain_unregister() [all …]
|
D | auditsc.c | 153 unsigned n; in audit_match_perm() local 157 n = ctx->major; in audit_match_perm() 159 switch (audit_classify_syscall(ctx->arch, n)) { in audit_match_perm() 162 audit_match_class(AUDIT_CLASS_WRITE, n)) in audit_match_perm() 165 audit_match_class(AUDIT_CLASS_READ, n)) in audit_match_perm() 168 audit_match_class(AUDIT_CLASS_CHATTR, n)) in audit_match_perm() 173 audit_match_class(AUDIT_CLASS_WRITE_32, n)) in audit_match_perm() 176 audit_match_class(AUDIT_CLASS_READ_32, n)) in audit_match_perm() 179 audit_match_class(AUDIT_CLASS_CHATTR_32, n)) in audit_match_perm() 199 struct audit_names *n; in audit_match_filetype() local [all …]
|
D | watch_queue.c | 98 struct watch_notification *n) in post_one_notification() argument 122 len = n->info & WATCH_INFO_LENGTH; in post_one_notification() 124 memcpy(p + offset, n, len); in post_one_notification() 159 const struct watch_notification *n) in filter_watch_notification() argument 163 unsigned int st_index = n->subtype / st_bits; in filter_watch_notification() 164 unsigned int st_bit = 1U << (n->subtype % st_bits); in filter_watch_notification() 167 if (!test_bit(n->type, wf->type_filter)) in filter_watch_notification() 172 if (n->type == wt->type && in filter_watch_notification() 174 (n->info & wt->info_mask) == wt->info_filter) in filter_watch_notification() 195 struct watch_notification *n, in __post_watch_notification() argument [all …]
|
D | Makefile | 31 KCOV_INSTRUMENT_softirq.o := n 34 KCSAN_SANITIZE_softirq.o = n 37 KCOV_INSTRUMENT_extable.o := n 38 KCOV_INSTRUMENT_stacktrace.o := n 40 KCOV_INSTRUMENT_kcov.o := n 43 KASAN_SANITIZE_kcov.o := n 44 KCSAN_SANITIZE_kcov.o := n 45 UBSAN_SANITIZE_kcov.o := n 46 KMSAN_SANITIZE_kcov.o := n 142 KASAN_SANITIZE_stackleak.o := n [all …]
|
D | Kconfig.locks | 96 # - DEBUG_SPINLOCK=n and GENERIC_LOCKBREAK=n and ARCH_INLINE_*LOCK=y 99 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y 102 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y 104 # - DEBUG_SPINLOCK=n and PREEMPTION=n 107 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
|
D | audit_tree.c | 221 unsigned long n = key / L1_CACHE_BYTES; in chunk_hash() local 222 return chunk_hash_heads + n % HASH_SIZE; in chunk_hash() 263 int n; in audit_tree_match() local 264 for (n = 0; n < chunk->count; n++) in audit_tree_match() 265 if (chunk->owners[n].owner == tree) in audit_tree_match() 463 int n; in tag_chunk() local 478 for (n = 0; n < old->count; n++) { in tag_chunk() 479 if (old->owners[n].owner == tree) { in tag_chunk() 1006 int n; in evict_chunk() local 1027 for (n = 0; n < chunk->count; n++) in evict_chunk() [all …]
|
D | seccomp.c | 1065 static void seccomp_handle_addfd(struct seccomp_kaddfd *addfd, struct seccomp_knotif *n) in seccomp_handle_addfd() argument 1083 n->state = SECCOMP_NOTIFY_SENT; in seccomp_handle_addfd() 1086 n->flags = 0; in seccomp_handle_addfd() 1087 n->error = 0; in seccomp_handle_addfd() 1088 n->val = fd; in seccomp_handle_addfd() 1100 struct seccomp_knotif *n) in should_sleep_killable() argument 1102 return match->wait_killable_recv && n->state == SECCOMP_NOTIFY_SENT; in should_sleep_killable() 1112 struct seccomp_knotif n = {}; in seccomp_do_user_notification() local 1120 n.task = current; in seccomp_do_user_notification() 1121 n.state = SECCOMP_NOTIFY_INIT; in seccomp_do_user_notification() [all …]
|
D | user.c | 239 int n; in uid_cache_init() local 244 for(n = 0; n < UIDHASH_SZ; ++n) in uid_cache_init() 245 INIT_HLIST_HEAD(uidhash_table + n); in uid_cache_init()
|
/kernel/power/ |
D | wakeup_reason.c | 100 struct wakeup_irq_node *n; in delete_list() local 103 n = list_first_entry(head, struct wakeup_irq_node, siblings); in delete_list() 104 list_del(&n->siblings); in delete_list() 105 kmem_cache_free(wakeup_irq_nodes_cache, n); in delete_list() 111 struct wakeup_irq_node *n = NULL; in add_sibling_node_sorted() local 118 list_for_each_entry(n, head, siblings) { in add_sibling_node_sorted() 119 if (n->irq < irq) in add_sibling_node_sorted() 120 predecessor = &n->siblings; in add_sibling_node_sorted() 121 else if (n->irq == irq) in add_sibling_node_sorted() 127 n = create_node(irq); in add_sibling_node_sorted() [all …]
|
D | main.c | 151 const char *buf, size_t n) in pm_async_store() argument 162 return n; in pm_async_store() 194 static suspend_state_t decode_suspend_state(const char *buf, size_t n) in decode_suspend_state() argument 200 p = memchr(buf, '\n', n); in decode_suspend_state() 201 len = p ? p - buf : n; in decode_suspend_state() 214 const char *buf, size_t n) in mem_sleep_store() argument 228 state = decode_suspend_state(buf, n); in mem_sleep_store() 236 return error ? error : n; in mem_sleep_store() 257 const char *buf, size_t n) in sync_on_suspend_store() argument 268 return n; in sync_on_suspend_store() [all …]
|
D | hibernate.c | 1101 const char *buf, size_t n) in disk_store() argument 1113 p = memchr(buf, '\n', n); in disk_store() 1114 len = p ? p - buf : n; in disk_store() 1147 return error ? error : n; in disk_store() 1160 const char *buf, size_t n) in resume_store() argument 1163 int len = n; in resume_store() 1169 return n; in resume_store() 1207 return n; in resume_store() 1220 size_t n) in resume_offset_store() argument 1230 return n; in resume_offset_store() [all …]
|
/kernel/configs/ |
D | nopm.config | 3 CONFIG_PM=n 4 CONFIG_SUSPEND=n 5 CONFIG_HIBERNATION=n 8 CONFIG_CPU_IDLE=n 11 CONFIG_XEN=n 14 CONFIG_ARCH_OMAP2PLUS_TYPICAL=n 15 CONFIG_ARCH_RENESAS=n 16 CONFIG_ARCH_TEGRA=n 17 CONFIG_ARCH_VEXPRESS=n
|
/kernel/sched/ |
D | loadavg.c | 109 fixed_power_int(unsigned long x, unsigned int frac_bits, unsigned int n) in fixed_power_int() argument 113 if (n) { in fixed_power_int() 115 if (n & 1) { in fixed_power_int() 120 n >>= 1; in fixed_power_int() 121 if (!n) in fixed_power_int() 157 unsigned long active, unsigned int n) in calc_load_n() argument 159 return calc_load(load, fixed_power_int(exp, FSHIFT, n), active); in calc_load_n() 306 long delta, active, n; in calc_global_nohz() local 314 n = 1 + (delta / LOAD_FREQ); in calc_global_nohz() 319 avenrun[0] = calc_load_n(avenrun[0], EXP_1, active, n); in calc_global_nohz() [all …]
|
D | stats.c | 178 unsigned long n = *offset; in schedstat_start() local 180 if (n == 0) in schedstat_start() 183 n--; in schedstat_start() 185 if (n > 0) in schedstat_start() 186 n = cpumask_next(n - 1, cpu_online_mask); in schedstat_start() 188 n = cpumask_first(cpu_online_mask); in schedstat_start() 190 *offset = n + 1; in schedstat_start() 192 if (n < nr_cpu_ids) in schedstat_start() 193 return (void *)(unsigned long)(n + 2); in schedstat_start()
|
/kernel/bpf/ |
D | log.c | 61 u32 new_n, n; in bpf_verifier_vlog() local 63 n = vscnprintf(log->kbuf, BPF_VERIFIER_TMP_LOG_SIZE, fmt, args); in bpf_verifier_vlog() 66 bool newline = n > 0 && log->kbuf[n - 1] == '\n'; in bpf_verifier_vlog() 72 n += 1; /* include terminating zero */ in bpf_verifier_vlog() 73 bpf_vlog_update_len_max(log, n); in bpf_verifier_vlog() 79 new_n = min_t(u32, log->len_total - log->end_pos, n); in bpf_verifier_vlog() 84 log->end_pos += n - 1; /* don't count terminating '\0' */ in bpf_verifier_vlog() 93 new_end = log->end_pos + n; in bpf_verifier_vlog() 105 new_n = min(n, log->len_total); in bpf_verifier_vlog() 125 log->kbuf + n - new_n, in bpf_verifier_vlog() [all …]
|
D | tnum.c | 183 size_t n; in tnum_sbin() local 185 for (n = 64; n; n--) { in tnum_sbin() 186 if (n < size) { in tnum_sbin() 188 str[n - 1] = 'x'; in tnum_sbin() 190 str[n - 1] = '1'; in tnum_sbin() 192 str[n - 1] = '0'; in tnum_sbin()
|
D | helpers.c | 1941 #define bpf_rbtree_postorder_for_each_entry_safe(pos, n, root) \ argument 1943 pos && ({ n = rb_next_postorder(pos); 1; }); \ 1944 pos = n) 1950 struct rb_node *pos, *n; in bpf_rb_root_free() local 1961 bpf_rbtree_postorder_for_each_entry_safe(pos, n, &orig_root.rb_root) { in bpf_rb_root_free() 2040 struct list_head *n = &node->list_head, *h = (void *)head; in __bpf_list_add() local 2053 __bpf_obj_drop_impl((void *)n - off, rec); in __bpf_list_add() 2057 tail ? list_add_tail(n, h) : list_add(n, h); in __bpf_list_add() 2067 struct bpf_list_node_kern *n = (void *)node; in bpf_list_push_front_impl() local 2070 return __bpf_list_add(n, head, false, meta ? meta->record : NULL, off); in bpf_list_push_front_impl() [all …]
|
/kernel/module/ |
D | tree_lookup.c | 22 static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n) in __mod_tree_val() argument 24 struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node); in __mod_tree_val() 29 static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n) in __mod_tree_size() argument 31 struct module_memory *mod_mem = container_of(n, struct module_memory, mtn.node); in __mod_tree_size() 43 mod_tree_comp(void *key, struct latch_tree_node *n) in mod_tree_comp() argument 48 start = __mod_tree_val(n); in mod_tree_comp() 52 end = start + __mod_tree_size(n); in mod_tree_comp()
|
/kernel/gcov/ |
D | Kconfig | 9 default n 30 GCOV_PROFILE_main.o := n 32 GCOV_PROFILE := n 38 def_bool n 45 default n
|
/kernel/events/ |
D | uprobes.c | 1267 struct rb_node *n = uprobes_tree.rb_node; in find_node_in_range() local 1269 while (n) { in find_node_in_range() 1270 struct uprobe *u = rb_entry(n, struct uprobe, rb_node); in find_node_in_range() 1273 n = n->rb_left; in find_node_in_range() 1275 n = n->rb_right; in find_node_in_range() 1278 n = n->rb_left; in find_node_in_range() 1280 n = n->rb_right; in find_node_in_range() 1286 return n; in find_node_in_range() 1298 struct rb_node *n, *t; in build_probe_list() local 1306 n = find_node_in_range(inode, min, max); in build_probe_list() [all …]
|
D | internal.h | 177 memcpy_common(void *dst, const void *src, unsigned long n) in memcpy_common() argument 179 memcpy(dst, src, n); in memcpy_common() 186 memcpy_skip(void *dst, const void *src, unsigned long n) in DEFINE_OUTPUT_COPY() 197 arch_perf_out_copy_user(void *dst, const void *src, unsigned long n) in DEFINE_OUTPUT_COPY() 202 ret = __copy_from_user_inatomic(dst, src, n); in DEFINE_OUTPUT_COPY()
|
/kernel/entry/ |
D | Makefile | 5 KASAN_SANITIZE := n 6 UBSAN_SANITIZE := n 7 KCOV_INSTRUMENT := n
|