Home
last modified time | relevance | path

Searched refs:n (Results 1 – 25 of 49) sorted by relevance

12

/kernel/
Dtimeconst.bc14 define fmul(b,n,d) {
15 return (2^b*n+d-1)/d;
19 (imul * n) + (fmulxx * n + fadjxx) >> xx) */
20 define fadj(b,n,d) {
22 d = d/gcd(n,d);
31 define fmuls(b,n,d) {
34 m = fmul(s,n,d);
42 print "/* Automatically generated by kernel/timeconst.bc */\n"
43 print "/* Time conversion constants for HZ == ", hz, " */\n"
44 print "\n"
[all …]
Dnotifier.c22 struct notifier_block *n) in notifier_chain_register() argument
25 if (n->priority > (*nl)->priority) in notifier_chain_register()
29 n->next = *nl; in notifier_chain_register()
30 rcu_assign_pointer(*nl, n); in notifier_chain_register()
35 struct notifier_block *n) in notifier_chain_cond_register() argument
38 if ((*nl) == n) in notifier_chain_cond_register()
40 if (n->priority > (*nl)->priority) in notifier_chain_cond_register()
44 n->next = *nl; in notifier_chain_cond_register()
45 rcu_assign_pointer(*nl, n); in notifier_chain_cond_register()
50 struct notifier_block *n) in notifier_chain_unregister() argument
[all …]
Dauditsc.c139 int n = ACC_MODE(flags); in open_arg() local
141 n |= AUDIT_PERM_WRITE; in open_arg()
142 return n & mask; in open_arg()
147 unsigned n; in audit_match_perm() local
150 n = ctx->major; in audit_match_perm()
152 switch (audit_classify_syscall(ctx->arch, n)) { in audit_match_perm()
155 audit_match_class(AUDIT_CLASS_WRITE, n)) in audit_match_perm()
158 audit_match_class(AUDIT_CLASS_READ, n)) in audit_match_perm()
161 audit_match_class(AUDIT_CLASS_CHATTR, n)) in audit_match_perm()
166 audit_match_class(AUDIT_CLASS_WRITE_32, n)) in audit_match_perm()
[all …]
Daudit_tree.c166 unsigned long n = (unsigned long)inode / L1_CACHE_BYTES; in chunk_hash() local
167 return chunk_hash_heads + n % HASH_SIZE; in chunk_hash()
200 int n; in audit_tree_match() local
201 for (n = 0; n < chunk->count; n++) in audit_tree_match()
202 if (chunk->owners[n].owner == tree) in audit_tree_match()
360 int n; in tag_chunk() local
370 for (n = 0; n < old->count; n++) { in tag_chunk()
371 if (old->owners[n].owner == tree) { in tag_chunk()
422 for (n = 0, p = chunk->owners; n < old->count; n++, p++) { in tag_chunk()
423 struct audit_tree *s = old->owners[n].owner; in tag_chunk()
[all …]
DKconfig.locks95 # - DEBUG_SPINLOCK=n and GENERIC_LOCKBREAK=n and ARCH_INLINE_*LOCK=y
98 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
101 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
103 # - DEBUG_SPINLOCK=n and PREEMPT=n
106 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
Duser.c208 int n; in uid_cache_init() local
213 for(n = 0; n < UIDHASH_SZ; ++n) in uid_cache_init()
214 INIT_HLIST_HEAD(uidhash_table + n); in uid_cache_init()
Dprofile.c158 int task_handoff_register(struct notifier_block *n) in task_handoff_register() argument
160 return atomic_notifier_chain_register(&task_free_notifier, n); in task_handoff_register()
164 int task_handoff_unregister(struct notifier_block *n) in task_handoff_unregister() argument
166 return atomic_notifier_chain_unregister(&task_free_notifier, n); in task_handoff_unregister()
170 int profile_event_register(enum profile_type type, struct notifier_block *n) in profile_event_register() argument
177 &task_exit_notifier, n); in profile_event_register()
181 &munmap_notifier, n); in profile_event_register()
189 int profile_event_unregister(enum profile_type type, struct notifier_block *n) in profile_event_unregister() argument
196 &task_exit_notifier, n); in profile_event_unregister()
200 &munmap_notifier, n); in profile_event_unregister()
Dresource.c940 resource_size_t start, resource_size_t n, in __request_region() argument
951 res->end = start + n - 1; in __request_region()
1003 resource_size_t n) in __check_region() argument
1007 res = __request_region(parent, start, n, "check-region", 0); in __check_region()
1026 resource_size_t n) in __release_region() argument
1032 end = start + n - 1; in __release_region()
1175 resource_size_t n; member
1182 __release_region(this->parent, this->start, this->n); in devm_region_release()
1190 this->start == match->start && this->n == match->n; in devm_region_match()
1195 resource_size_t n, const char *name) in __devm_request_region() argument
[all …]
Daudit.c1484 void audit_log_name(struct audit_context *context, struct audit_names *n, in audit_log_name() argument
1496 else if (n->name) { in audit_log_name()
1497 switch (n->name_len) { in audit_log_name()
1501 audit_log_untrustedstring(ab, n->name->name); in audit_log_name()
1511 audit_log_n_untrustedstring(ab, n->name->name, in audit_log_name()
1512 n->name_len); in audit_log_name()
1517 if (n->ino != (unsigned long)-1) { in audit_log_name()
1521 n->ino, in audit_log_name()
1522 MAJOR(n->dev), in audit_log_name()
1523 MINOR(n->dev), in audit_log_name()
[all …]
Dtime.c663 u64 nsecs_to_jiffies64(u64 n) in nsecs_to_jiffies64() argument
667 return div_u64(n, NSEC_PER_SEC / HZ); in nsecs_to_jiffies64()
670 return div_u64(n * HZ / 512, NSEC_PER_SEC / 512); in nsecs_to_jiffies64()
676 return div_u64(n * 9, (9ull * NSEC_PER_SEC + HZ / 2) / HZ); in nsecs_to_jiffies64()
693 unsigned long nsecs_to_jiffies(u64 n) in nsecs_to_jiffies() argument
695 return (unsigned long)nsecs_to_jiffies64(n); in nsecs_to_jiffies()
Dptrace.c500 struct task_struct *p, *n; in exit_ptrace() local
506 list_for_each_entry_safe(p, n, &tracer->ptraced, ptrace_entry) { in exit_ptrace()
517 list_for_each_entry_safe(p, n, &ptrace_dead, ptrace_entry) { in exit_ptrace()
761 int n; in find_regset() local
763 for (n = 0; n < view->n; ++n) { in find_regset()
764 regset = view->regsets + n; in find_regset()
784 (__kernel_size_t) (regset->n * regset->size)); in ptrace_regset()
Dauditfilter.c163 unsigned n = *list++; in audit_register_class() local
164 if (n >= AUDIT_BITMASK_SIZE * 32 - AUDIT_SYSCALL_CLASSES) { in audit_register_class()
168 p[AUDIT_WORD(n)] |= AUDIT_BIT(n); in audit_register_class()
307 u32 n; in audit_to_op() local
308 for (n = Audit_equal; n < Audit_bad && audit_ops[n] != op; n++) in audit_to_op()
310 return n; in audit_to_op()
1352 struct audit_krule *r, *n; in audit_update_lsm_rules() local
1359 list_for_each_entry_safe(r, n, &audit_rules_list[i], list) { in audit_update_lsm_rules()
Dparams.c70 bool parameqn(const char *a, const char *b, size_t n) in parameqn() argument
74 for (i = 0; i < n; i++) { in parameqn()
503 #define to_module_attr(n) container_of(n, struct module_attribute, attr) argument
504 #define to_module_kobject(n) container_of(n, struct module_kobject, kobj) argument
522 #define to_param_attr(n) container_of(n, struct param_attribute, mattr) argument
/kernel/power/
Dmain.c58 const char *buf, size_t n) in pm_async_store() argument
69 return n; in pm_async_store()
108 const char *buf, size_t n) in pm_test_store() argument
116 p = memchr(buf, '\n', n); in pm_test_store()
117 len = p ? p - buf : n; in pm_test_store()
131 return error ? error : n; in pm_test_store()
255 const char *buf, size_t n) in pm_print_times_store() argument
266 return n; in pm_print_times_store()
313 static suspend_state_t decode_state(const char *buf, size_t n) in decode_state() argument
322 p = memchr(buf, '\n', n); in decode_state()
[all …]
Dhibernate.c918 const char *buf, size_t n) in disk_store() argument
926 p = memchr(buf, '\n', n); in disk_store()
927 len = p ? p - buf : n; in disk_store()
959 return error ? error : n; in disk_store()
972 const char *buf, size_t n) in resume_store() argument
991 ret = n; in resume_store()
1005 const char *buf, size_t n) in image_size_store() argument
1011 return n; in image_size_store()
1027 const char *buf, size_t n) in reserved_size_store() argument
1033 return n; in reserved_size_store()
/kernel/sched/
Dstats.c92 unsigned long n = *offset; in schedstat_start() local
94 if (n == 0) in schedstat_start()
97 n--; in schedstat_start()
99 if (n > 0) in schedstat_start()
100 n = cpumask_next(n - 1, cpu_online_mask); in schedstat_start()
102 n = cpumask_first(cpu_online_mask); in schedstat_start()
104 *offset = n + 1; in schedstat_start()
106 if (n < nr_cpu_ids) in schedstat_start()
107 return (void *)(unsigned long)(n + 2); in schedstat_start()
Ddebug.c301 #define P(n) SEQ_printf(m, " .%-30s: %d\n", #n, rq->n); in print_cpu() argument
302 #define P64(n) SEQ_printf(m, " .%-30s: %Ld\n", #n, rq->n); in print_cpu() argument
418 unsigned long n = *offset; in sched_debug_start() local
420 if (n == 0) in sched_debug_start()
423 n--; in sched_debug_start()
425 if (n > 0) in sched_debug_start()
426 n = cpumask_next(n - 1, cpu_online_mask); in sched_debug_start()
428 n = cpumask_first(cpu_online_mask); in sched_debug_start()
430 *offset = n + 1; in sched_debug_start()
432 if (n < nr_cpu_ids) in sched_debug_start()
[all …]
/kernel/events/
Duprobes.c355 struct rb_node *n = uprobes_tree.rb_node; in __find_uprobe() local
359 while (n) { in __find_uprobe()
360 uprobe = rb_entry(n, struct uprobe, rb_node); in __find_uprobe()
368 n = n->rb_left; in __find_uprobe()
370 n = n->rb_right; in __find_uprobe()
962 struct rb_node *n = uprobes_tree.rb_node; in find_node_in_range() local
964 while (n) { in find_node_in_range()
965 struct uprobe *u = rb_entry(n, struct uprobe, rb_node); in find_node_in_range()
968 n = n->rb_left; in find_node_in_range()
970 n = n->rb_right; in find_node_in_range()
[all …]
Dinternal.h113 static inline int memcpy_common(void *dst, const void *src, size_t n) in memcpy_common() argument
115 memcpy(dst, src, n); in memcpy_common()
116 return n; in memcpy_common()
121 #define MEMCPY_SKIP(dst, src, n) (n) argument
/kernel/gcov/
DKconfig7 default n
28 GCOV_PROFILE_main.o := n
30 GCOV_PROFILE := n
39 default n
/kernel/debug/kdb/
Dkdb_main.c1656 int n, z, num = (symbolic ? 1 : (16 / bytesperword)); in kdb_md() local
1668 n = min(num, repeat); in kdb_md()
1671 addr += bytesperword * n; in kdb_md()
1672 repeat -= n; in kdb_md()
2038 int n = 0; in kdb_dmesg() local
2067 n++; in kdb_dmesg()
2070 if (adjust >= n) in kdb_dmesg()
2072 "printed\n", n); in kdb_dmesg()
2073 else if (adjust - lines >= n) in kdb_dmesg()
2075 "lines printed\n", n, n - adjust); in kdb_dmesg()
[all …]
/kernel/trace/
Dtrace_stack.c256 long n = *pos - 1; in __next() local
258 if (n >= max_stack_trace.nr_entries || stack_dump_trace[n] == ULONG_MAX) in __next()
261 m->private = (void *)n; in __next()
Dtrace_stat.c207 int n = *pos; in stat_seq_start() local
215 if (n == 0) in stat_seq_start()
217 n--; in stat_seq_start()
221 for (i = 0; node && i < n; i++) in stat_seq_start()
Dtrace_kprobe.c44 #define SIZEOF_TRACE_PROBE(n) \ argument
46 (sizeof(struct probe_arg) * (n)))
218 int n = trace_probe_nr_files(tp); in enable_trace_probe() local
222 new = kzalloc((n + 2) * sizeof(struct ftrace_event_file *), in enable_trace_probe()
228 memcpy(new, old, n * sizeof(struct ftrace_event_file *)); in enable_trace_probe()
229 new[n] = file; in enable_trace_probe()
290 int n = trace_probe_nr_files(tp); in disable_trace_probe() local
294 if (n == 0 || trace_probe_file_index(tp, file) < 0) { in disable_trace_probe()
299 if (n == 1) { /* Remove the last file */ in disable_trace_probe()
303 new = kzalloc(n * sizeof(struct ftrace_event_file *), in disable_trace_probe()
[all …]
Dtrace_mmiotrace.c145 unsigned long n; in mmio_read() local
147 n = count_overruns(iter); in mmio_read()
148 if (n) { in mmio_read()
150 trace_seq_printf(s, "MARK 0.000000 Lost %lu events.\n", n); in mmio_read()

12