Home
last modified time | relevance | path

Searched refs:n (Results 1 – 25 of 101) sorted by relevance

12345

/kernel/time/
Dtimeconst.bc16 define fmul(b,n,d) {
17 return (2^b*n+d-1)/d;
21 (imul * n) + (fmulxx * n + fadjxx) >> xx) */
22 define fadj(b,n,d) {
24 d = d/gcd(n,d);
33 define fmuls(b,n,d) {
36 m = fmul(s,n,d);
44 print "/* Automatically generated by kernel/time/timeconst.bc */\n"
45 print "/* Time conversion constants for HZ == ", hz, " */\n"
46 print "\n"
[all …]
/kernel/locking/
Dtest-ww_mutex.c318 unsigned int n, last = nthreads - 1; in __test_cycle() local
325 for (n = 0; n < nthreads; n++) { in __test_cycle()
326 struct test_cycle *cycle = &cycles[n]; in __test_cycle()
329 if (n == last) in __test_cycle()
332 cycle->b_mutex = &cycles[n + 1].a_mutex; in __test_cycle()
334 if (n == 0) in __test_cycle()
337 cycle->a_signal = &cycles[n - 1].b_signal; in __test_cycle()
344 for (n = 0; n < nthreads; n++) in __test_cycle()
345 queue_work(wq, &cycles[n].work); in __test_cycle()
350 for (n = 0; n < nthreads; n++) { in __test_cycle()
[all …]
Dww_mutex.h96 struct rb_node *n = rb_first(&lock->rtmutex.waiters.rb_root); in __ww_waiter_first() local
97 if (!n) in __ww_waiter_first()
99 return rb_entry(n, struct rt_mutex_waiter, tree_entry); in __ww_waiter_first()
105 struct rb_node *n = rb_next(&w->tree_entry); in __ww_waiter_next() local
106 if (!n) in __ww_waiter_next()
108 return rb_entry(n, struct rt_mutex_waiter, tree_entry); in __ww_waiter_next()
114 struct rb_node *n = rb_prev(&w->tree_entry); in __ww_waiter_prev() local
115 if (!n) in __ww_waiter_prev()
117 return rb_entry(n, struct rt_mutex_waiter, tree_entry); in __ww_waiter_prev()
123 struct rb_node *n = rb_last(&lock->rtmutex.waiters.rb_root); in __ww_waiter_last() local
[all …]
/kernel/
Dnotifier.c23 struct notifier_block *n, in notifier_chain_register() argument
27 if (unlikely((*nl) == n)) { in notifier_chain_register()
29 n->notifier_call); in notifier_chain_register()
32 if (n->priority > (*nl)->priority) in notifier_chain_register()
34 if (n->priority == (*nl)->priority && unique_priority) in notifier_chain_register()
38 n->next = *nl; in notifier_chain_register()
39 rcu_assign_pointer(*nl, n); in notifier_chain_register()
44 struct notifier_block *n) in notifier_chain_unregister() argument
47 if ((*nl) == n) { in notifier_chain_unregister()
48 rcu_assign_pointer(*nl, n->next); in notifier_chain_unregister()
[all …]
Dauditsc.c150 unsigned n; in audit_match_perm() local
154 n = ctx->major; in audit_match_perm()
156 switch (audit_classify_syscall(ctx->arch, n)) { in audit_match_perm()
159 audit_match_class(AUDIT_CLASS_WRITE, n)) in audit_match_perm()
162 audit_match_class(AUDIT_CLASS_READ, n)) in audit_match_perm()
165 audit_match_class(AUDIT_CLASS_CHATTR, n)) in audit_match_perm()
170 audit_match_class(AUDIT_CLASS_WRITE_32, n)) in audit_match_perm()
173 audit_match_class(AUDIT_CLASS_READ_32, n)) in audit_match_perm()
176 audit_match_class(AUDIT_CLASS_CHATTR_32, n)) in audit_match_perm()
196 struct audit_names *n; in audit_match_filetype() local
[all …]
Dwatch_queue.c99 struct watch_notification *n) in post_one_notification() argument
123 len = n->info & WATCH_INFO_LENGTH; in post_one_notification()
125 memcpy(p + offset, n, len); in post_one_notification()
160 const struct watch_notification *n) in filter_watch_notification() argument
164 unsigned int st_index = n->subtype / st_bits; in filter_watch_notification()
165 unsigned int st_bit = 1U << (n->subtype % st_bits); in filter_watch_notification()
168 if (!test_bit(n->type, wf->type_filter)) in filter_watch_notification()
173 if (n->type == wt->type && in filter_watch_notification()
175 (n->info & wt->info_mask) == wt->info_filter) in filter_watch_notification()
196 struct watch_notification *n, in __post_watch_notification() argument
[all …]
DMakefile26 KCOV_INSTRUMENT_softirq.o := n
29 KCSAN_SANITIZE_softirq.o = n
32 KCOV_INSTRUMENT_extable.o := n
33 KCOV_INSTRUMENT_stacktrace.o := n
35 KCOV_INSTRUMENT_kcov.o := n
38 KASAN_SANITIZE_kcov.o := n
39 KCSAN_SANITIZE_kcov.o := n
40 UBSAN_SANITIZE_kcov.o := n
41 KMSAN_SANITIZE_kcov.o := n
133 KASAN_SANITIZE_stackleak.o := n
[all …]
DKconfig.locks96 # - DEBUG_SPINLOCK=n and GENERIC_LOCKBREAK=n and ARCH_INLINE_*LOCK=y
99 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
102 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
104 # - DEBUG_SPINLOCK=n and PREEMPTION=n
107 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
Daudit_tree.c221 unsigned long n = key / L1_CACHE_BYTES; in chunk_hash() local
222 return chunk_hash_heads + n % HASH_SIZE; in chunk_hash()
263 int n; in audit_tree_match() local
264 for (n = 0; n < chunk->count; n++) in audit_tree_match()
265 if (chunk->owners[n].owner == tree) in audit_tree_match()
463 int n; in tag_chunk() local
478 for (n = 0; n < old->count; n++) { in tag_chunk()
479 if (old->owners[n].owner == tree) { in tag_chunk()
1006 int n; in evict_chunk() local
1027 for (n = 0; n < chunk->count; n++) in evict_chunk()
[all …]
Dseccomp.c1056 static void seccomp_handle_addfd(struct seccomp_kaddfd *addfd, struct seccomp_knotif *n) in seccomp_handle_addfd() argument
1074 n->state = SECCOMP_NOTIFY_SENT; in seccomp_handle_addfd()
1077 n->flags = 0; in seccomp_handle_addfd()
1078 n->error = 0; in seccomp_handle_addfd()
1079 n->val = fd; in seccomp_handle_addfd()
1091 struct seccomp_knotif *n) in should_sleep_killable() argument
1093 return match->wait_killable_recv && n->state == SECCOMP_NOTIFY_SENT; in should_sleep_killable()
1103 struct seccomp_knotif n = {}; in seccomp_do_user_notification() local
1111 n.task = current; in seccomp_do_user_notification()
1112 n.state = SECCOMP_NOTIFY_INIT; in seccomp_do_user_notification()
[all …]
Duser.c239 int n; in uid_cache_init() local
244 for(n = 0; n < UIDHASH_SZ; ++n) in uid_cache_init()
245 INIT_HLIST_HEAD(uidhash_table + n); in uid_cache_init()
/kernel/power/
Dwakeup_reason.c100 struct wakeup_irq_node *n; in delete_list() local
103 n = list_first_entry(head, struct wakeup_irq_node, siblings); in delete_list()
104 list_del(&n->siblings); in delete_list()
105 kmem_cache_free(wakeup_irq_nodes_cache, n); in delete_list()
111 struct wakeup_irq_node *n = NULL; in add_sibling_node_sorted() local
118 list_for_each_entry(n, head, siblings) { in add_sibling_node_sorted()
119 if (n->irq < irq) in add_sibling_node_sorted()
120 predecessor = &n->siblings; in add_sibling_node_sorted()
121 else if (n->irq == irq) in add_sibling_node_sorted()
127 n = create_node(irq); in add_sibling_node_sorted()
[all …]
Dmain.c110 const char *buf, size_t n) in pm_async_store() argument
121 return n; in pm_async_store()
153 static suspend_state_t decode_suspend_state(const char *buf, size_t n) in decode_suspend_state() argument
159 p = memchr(buf, '\n', n); in decode_suspend_state()
160 len = p ? p - buf : n; in decode_suspend_state()
173 const char *buf, size_t n) in mem_sleep_store() argument
187 state = decode_suspend_state(buf, n); in mem_sleep_store()
195 return error ? error : n; in mem_sleep_store()
216 const char *buf, size_t n) in sync_on_suspend_store() argument
227 return n; in sync_on_suspend_store()
[all …]
Dhibernate.c1116 const char *buf, size_t n) in disk_store() argument
1128 p = memchr(buf, '\n', n); in disk_store()
1129 len = p ? p - buf : n; in disk_store()
1162 return error ? error : n; in disk_store()
1175 const char *buf, size_t n) in resume_store() argument
1178 int len = n; in resume_store()
1201 return n; in resume_store()
1214 size_t n) in resume_offset_store() argument
1224 return n; in resume_offset_store()
1236 const char *buf, size_t n) in image_size_store() argument
[all …]
/kernel/configs/
Dnopm.config1 CONFIG_PM=n
2 CONFIG_SUSPEND=n
3 CONFIG_HIBERNATION=n
6 CONFIG_CPU_IDLE=n
9 CONFIG_XEN=n
12 CONFIG_ARCH_OMAP2PLUS_TYPICAL=n
13 CONFIG_ARCH_RENESAS=n
14 CONFIG_ARCH_TEGRA=n
15 CONFIG_ARCH_VEXPRESS=n
/kernel/sched/
Dloadavg.c109 fixed_power_int(unsigned long x, unsigned int frac_bits, unsigned int n) in fixed_power_int() argument
113 if (n) { in fixed_power_int()
115 if (n & 1) { in fixed_power_int()
120 n >>= 1; in fixed_power_int()
121 if (!n) in fixed_power_int()
157 unsigned long active, unsigned int n) in calc_load_n() argument
159 return calc_load(load, fixed_power_int(exp, FSHIFT, n), active); in calc_load_n()
306 long delta, active, n; in calc_global_nohz() local
314 n = 1 + (delta / LOAD_FREQ); in calc_global_nohz()
319 avenrun[0] = calc_load_n(avenrun[0], EXP_1, active, n); in calc_global_nohz()
[all …]
Dstats.c188 unsigned long n = *offset; in schedstat_start() local
190 if (n == 0) in schedstat_start()
193 n--; in schedstat_start()
195 if (n > 0) in schedstat_start()
196 n = cpumask_next(n - 1, cpu_online_mask); in schedstat_start()
198 n = cpumask_first(cpu_online_mask); in schedstat_start()
200 *offset = n + 1; in schedstat_start()
202 if (n < nr_cpu_ids) in schedstat_start()
203 return (void *)(unsigned long)(n + 2); in schedstat_start()
Ddebug.c755 #define P64(n) SEQ_printf(m, " .%-30s: %Ld\n", #n, rq->n); in print_cpu() argument
761 #define P(n) SEQ_printf(m, " .%-30s: %d\n", #n, schedstat_val(rq->n)); in print_cpu() argument
876 unsigned long n = *offset; in sched_debug_start() local
878 if (n == 0) in sched_debug_start()
881 n--; in sched_debug_start()
883 if (n > 0) in sched_debug_start()
884 n = cpumask_next(n - 1, cpu_online_mask); in sched_debug_start()
886 n = cpumask_first(cpu_online_mask); in sched_debug_start()
888 *offset = n + 1; in sched_debug_start()
890 if (n < nr_cpu_ids) in sched_debug_start()
[all …]
/kernel/module/
Dtree_lookup.c22 static __always_inline unsigned long __mod_tree_val(struct latch_tree_node *n) in __mod_tree_val() argument
24 struct module_layout *layout = container_of(n, struct module_layout, mtn.node); in __mod_tree_val()
29 static __always_inline unsigned long __mod_tree_size(struct latch_tree_node *n) in __mod_tree_size() argument
31 struct module_layout *layout = container_of(n, struct module_layout, mtn.node); in __mod_tree_size()
43 mod_tree_comp(void *key, struct latch_tree_node *n) in mod_tree_comp() argument
48 start = __mod_tree_val(n); in mod_tree_comp()
52 end = start + __mod_tree_size(n); in mod_tree_comp()
/kernel/bpf/
Dtnum.c183 size_t n; in tnum_sbin() local
185 for (n = 64; n; n--) { in tnum_sbin()
186 if (n < size) { in tnum_sbin()
188 str[n - 1] = 'x'; in tnum_sbin()
190 str[n - 1] = '1'; in tnum_sbin()
192 str[n - 1] = '0'; in tnum_sbin()
/kernel/gcov/
DKconfig9 default n
30 GCOV_PROFILE_main.o := n
32 GCOV_PROFILE := n
38 def_bool n
45 default n
/kernel/entry/
DMakefile5 KASAN_SANITIZE := n
6 UBSAN_SANITIZE := n
7 KCOV_INSTRUMENT := n
/kernel/events/
Duprobes.c1266 struct rb_node *n = uprobes_tree.rb_node; in find_node_in_range() local
1268 while (n) { in find_node_in_range()
1269 struct uprobe *u = rb_entry(n, struct uprobe, rb_node); in find_node_in_range()
1272 n = n->rb_left; in find_node_in_range()
1274 n = n->rb_right; in find_node_in_range()
1277 n = n->rb_left; in find_node_in_range()
1279 n = n->rb_right; in find_node_in_range()
1285 return n; in find_node_in_range()
1297 struct rb_node *n, *t; in build_probe_list() local
1305 n = find_node_in_range(inode, min, max); in build_probe_list()
[all …]
Dinternal.h177 memcpy_common(void *dst, const void *src, unsigned long n) in memcpy_common() argument
179 memcpy(dst, src, n); in memcpy_common()
186 memcpy_skip(void *dst, const void *src, unsigned long n) in DEFINE_OUTPUT_COPY()
197 arch_perf_out_copy_user(void *dst, const void *src, unsigned long n) in DEFINE_OUTPUT_COPY()
202 ret = __copy_from_user_inatomic(dst, src, n); in DEFINE_OUTPUT_COPY()
/kernel/kcsan/
DMakefile2 KCSAN_SANITIZE := n
3 KCOV_INSTRUMENT := n
4 UBSAN_SANITIZE := n

12345