/kernel/ |
D | timeconst.bc | 14 define fmul(b,n,d) { 15 return (2^b*n+d-1)/d; 19 (imul * n) + (fmulxx * n + fadjxx) >> xx) */ 20 define fadj(b,n,d) { 22 d = d/gcd(n,d); 31 define fmuls(b,n,d) { 34 m = fmul(s,n,d); 42 print "/* Automatically generated by kernel/timeconst.bc */\n" 43 print "/* Time conversion constants for HZ == ", hz, " */\n" 44 print "\n" [all …]
|
D | notifier.c | 22 struct notifier_block *n) in notifier_chain_register() argument 25 if (n->priority > (*nl)->priority) in notifier_chain_register() 29 n->next = *nl; in notifier_chain_register() 30 rcu_assign_pointer(*nl, n); in notifier_chain_register() 35 struct notifier_block *n) in notifier_chain_cond_register() argument 38 if ((*nl) == n) in notifier_chain_cond_register() 40 if (n->priority > (*nl)->priority) in notifier_chain_cond_register() 44 n->next = *nl; in notifier_chain_cond_register() 45 rcu_assign_pointer(*nl, n); in notifier_chain_cond_register() 50 struct notifier_block *n) in notifier_chain_unregister() argument [all …]
|
D | auditsc.c | 139 int n = ACC_MODE(flags); in open_arg() local 141 n |= AUDIT_PERM_WRITE; in open_arg() 142 return n & mask; in open_arg() 147 unsigned n; in audit_match_perm() local 150 n = ctx->major; in audit_match_perm() 152 switch (audit_classify_syscall(ctx->arch, n)) { in audit_match_perm() 155 audit_match_class(AUDIT_CLASS_WRITE, n)) in audit_match_perm() 158 audit_match_class(AUDIT_CLASS_READ, n)) in audit_match_perm() 161 audit_match_class(AUDIT_CLASS_CHATTR, n)) in audit_match_perm() 166 audit_match_class(AUDIT_CLASS_WRITE_32, n)) in audit_match_perm() [all …]
|
D | audit_tree.c | 166 unsigned long n = (unsigned long)inode / L1_CACHE_BYTES; in chunk_hash() local 167 return chunk_hash_heads + n % HASH_SIZE; in chunk_hash() 200 int n; in audit_tree_match() local 201 for (n = 0; n < chunk->count; n++) in audit_tree_match() 202 if (chunk->owners[n].owner == tree) in audit_tree_match() 360 int n; in tag_chunk() local 370 for (n = 0; n < old->count; n++) { in tag_chunk() 371 if (old->owners[n].owner == tree) { in tag_chunk() 422 for (n = 0, p = chunk->owners; n < old->count; n++, p++) { in tag_chunk() 423 struct audit_tree *s = old->owners[n].owner; in tag_chunk() [all …]
|
D | Kconfig.locks | 95 # - DEBUG_SPINLOCK=n and GENERIC_LOCKBREAK=n and ARCH_INLINE_*LOCK=y 98 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y 101 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y 103 # - DEBUG_SPINLOCK=n and PREEMPT=n 106 # - DEBUG_SPINLOCK=n and ARCH_INLINE_*LOCK=y
|
D | user.c | 208 int n; in uid_cache_init() local 213 for(n = 0; n < UIDHASH_SZ; ++n) in uid_cache_init() 214 INIT_HLIST_HEAD(uidhash_table + n); in uid_cache_init()
|
D | profile.c | 158 int task_handoff_register(struct notifier_block *n) in task_handoff_register() argument 160 return atomic_notifier_chain_register(&task_free_notifier, n); in task_handoff_register() 164 int task_handoff_unregister(struct notifier_block *n) in task_handoff_unregister() argument 166 return atomic_notifier_chain_unregister(&task_free_notifier, n); in task_handoff_unregister() 170 int profile_event_register(enum profile_type type, struct notifier_block *n) in profile_event_register() argument 177 &task_exit_notifier, n); in profile_event_register() 181 &munmap_notifier, n); in profile_event_register() 189 int profile_event_unregister(enum profile_type type, struct notifier_block *n) in profile_event_unregister() argument 196 &task_exit_notifier, n); in profile_event_unregister() 200 &munmap_notifier, n); in profile_event_unregister()
|
D | resource.c | 940 resource_size_t start, resource_size_t n, in __request_region() argument 951 res->end = start + n - 1; in __request_region() 1003 resource_size_t n) in __check_region() argument 1007 res = __request_region(parent, start, n, "check-region", 0); in __check_region() 1026 resource_size_t n) in __release_region() argument 1032 end = start + n - 1; in __release_region() 1175 resource_size_t n; member 1182 __release_region(this->parent, this->start, this->n); in devm_region_release() 1190 this->start == match->start && this->n == match->n; in devm_region_match() 1195 resource_size_t n, const char *name) in __devm_request_region() argument [all …]
|
D | audit.c | 1484 void audit_log_name(struct audit_context *context, struct audit_names *n, in audit_log_name() argument 1496 else if (n->name) { in audit_log_name() 1497 switch (n->name_len) { in audit_log_name() 1501 audit_log_untrustedstring(ab, n->name->name); in audit_log_name() 1511 audit_log_n_untrustedstring(ab, n->name->name, in audit_log_name() 1512 n->name_len); in audit_log_name() 1517 if (n->ino != (unsigned long)-1) { in audit_log_name() 1521 n->ino, in audit_log_name() 1522 MAJOR(n->dev), in audit_log_name() 1523 MINOR(n->dev), in audit_log_name() [all …]
|
D | time.c | 663 u64 nsecs_to_jiffies64(u64 n) in nsecs_to_jiffies64() argument 667 return div_u64(n, NSEC_PER_SEC / HZ); in nsecs_to_jiffies64() 670 return div_u64(n * HZ / 512, NSEC_PER_SEC / 512); in nsecs_to_jiffies64() 676 return div_u64(n * 9, (9ull * NSEC_PER_SEC + HZ / 2) / HZ); in nsecs_to_jiffies64() 693 unsigned long nsecs_to_jiffies(u64 n) in nsecs_to_jiffies() argument 695 return (unsigned long)nsecs_to_jiffies64(n); in nsecs_to_jiffies()
|
D | ptrace.c | 500 struct task_struct *p, *n; in exit_ptrace() local 506 list_for_each_entry_safe(p, n, &tracer->ptraced, ptrace_entry) { in exit_ptrace() 517 list_for_each_entry_safe(p, n, &ptrace_dead, ptrace_entry) { in exit_ptrace() 761 int n; in find_regset() local 763 for (n = 0; n < view->n; ++n) { in find_regset() 764 regset = view->regsets + n; in find_regset() 784 (__kernel_size_t) (regset->n * regset->size)); in ptrace_regset()
|
D | auditfilter.c | 163 unsigned n = *list++; in audit_register_class() local 164 if (n >= AUDIT_BITMASK_SIZE * 32 - AUDIT_SYSCALL_CLASSES) { in audit_register_class() 168 p[AUDIT_WORD(n)] |= AUDIT_BIT(n); in audit_register_class() 307 u32 n; in audit_to_op() local 308 for (n = Audit_equal; n < Audit_bad && audit_ops[n] != op; n++) in audit_to_op() 310 return n; in audit_to_op() 1352 struct audit_krule *r, *n; in audit_update_lsm_rules() local 1359 list_for_each_entry_safe(r, n, &audit_rules_list[i], list) { in audit_update_lsm_rules()
|
D | params.c | 70 bool parameqn(const char *a, const char *b, size_t n) in parameqn() argument 74 for (i = 0; i < n; i++) { in parameqn() 503 #define to_module_attr(n) container_of(n, struct module_attribute, attr) argument 504 #define to_module_kobject(n) container_of(n, struct module_kobject, kobj) argument 522 #define to_param_attr(n) container_of(n, struct param_attribute, mattr) argument
|
/kernel/power/ |
D | main.c | 58 const char *buf, size_t n) in pm_async_store() argument 69 return n; in pm_async_store() 108 const char *buf, size_t n) in pm_test_store() argument 116 p = memchr(buf, '\n', n); in pm_test_store() 117 len = p ? p - buf : n; in pm_test_store() 131 return error ? error : n; in pm_test_store() 255 const char *buf, size_t n) in pm_print_times_store() argument 266 return n; in pm_print_times_store() 313 static suspend_state_t decode_state(const char *buf, size_t n) in decode_state() argument 322 p = memchr(buf, '\n', n); in decode_state() [all …]
|
D | hibernate.c | 918 const char *buf, size_t n) in disk_store() argument 926 p = memchr(buf, '\n', n); in disk_store() 927 len = p ? p - buf : n; in disk_store() 959 return error ? error : n; in disk_store() 972 const char *buf, size_t n) in resume_store() argument 991 ret = n; in resume_store() 1005 const char *buf, size_t n) in image_size_store() argument 1011 return n; in image_size_store() 1027 const char *buf, size_t n) in reserved_size_store() argument 1033 return n; in reserved_size_store()
|
/kernel/sched/ |
D | stats.c | 92 unsigned long n = *offset; in schedstat_start() local 94 if (n == 0) in schedstat_start() 97 n--; in schedstat_start() 99 if (n > 0) in schedstat_start() 100 n = cpumask_next(n - 1, cpu_online_mask); in schedstat_start() 102 n = cpumask_first(cpu_online_mask); in schedstat_start() 104 *offset = n + 1; in schedstat_start() 106 if (n < nr_cpu_ids) in schedstat_start() 107 return (void *)(unsigned long)(n + 2); in schedstat_start()
|
D | debug.c | 301 #define P(n) SEQ_printf(m, " .%-30s: %d\n", #n, rq->n); in print_cpu() argument 302 #define P64(n) SEQ_printf(m, " .%-30s: %Ld\n", #n, rq->n); in print_cpu() argument 418 unsigned long n = *offset; in sched_debug_start() local 420 if (n == 0) in sched_debug_start() 423 n--; in sched_debug_start() 425 if (n > 0) in sched_debug_start() 426 n = cpumask_next(n - 1, cpu_online_mask); in sched_debug_start() 428 n = cpumask_first(cpu_online_mask); in sched_debug_start() 430 *offset = n + 1; in sched_debug_start() 432 if (n < nr_cpu_ids) in sched_debug_start() [all …]
|
/kernel/events/ |
D | uprobes.c | 355 struct rb_node *n = uprobes_tree.rb_node; in __find_uprobe() local 359 while (n) { in __find_uprobe() 360 uprobe = rb_entry(n, struct uprobe, rb_node); in __find_uprobe() 368 n = n->rb_left; in __find_uprobe() 370 n = n->rb_right; in __find_uprobe() 962 struct rb_node *n = uprobes_tree.rb_node; in find_node_in_range() local 964 while (n) { in find_node_in_range() 965 struct uprobe *u = rb_entry(n, struct uprobe, rb_node); in find_node_in_range() 968 n = n->rb_left; in find_node_in_range() 970 n = n->rb_right; in find_node_in_range() [all …]
|
D | internal.h | 113 static inline int memcpy_common(void *dst, const void *src, size_t n) in memcpy_common() argument 115 memcpy(dst, src, n); in memcpy_common() 116 return n; in memcpy_common() 121 #define MEMCPY_SKIP(dst, src, n) (n) argument
|
/kernel/gcov/ |
D | Kconfig | 7 default n 28 GCOV_PROFILE_main.o := n 30 GCOV_PROFILE := n 39 default n
|
/kernel/debug/kdb/ |
D | kdb_main.c | 1656 int n, z, num = (symbolic ? 1 : (16 / bytesperword)); in kdb_md() local 1668 n = min(num, repeat); in kdb_md() 1671 addr += bytesperword * n; in kdb_md() 1672 repeat -= n; in kdb_md() 2038 int n = 0; in kdb_dmesg() local 2067 n++; in kdb_dmesg() 2070 if (adjust >= n) in kdb_dmesg() 2072 "printed\n", n); in kdb_dmesg() 2073 else if (adjust - lines >= n) in kdb_dmesg() 2075 "lines printed\n", n, n - adjust); in kdb_dmesg() [all …]
|
/kernel/trace/ |
D | trace_stack.c | 256 long n = *pos - 1; in __next() local 258 if (n >= max_stack_trace.nr_entries || stack_dump_trace[n] == ULONG_MAX) in __next() 261 m->private = (void *)n; in __next()
|
D | trace_stat.c | 207 int n = *pos; in stat_seq_start() local 215 if (n == 0) in stat_seq_start() 217 n--; in stat_seq_start() 221 for (i = 0; node && i < n; i++) in stat_seq_start()
|
D | trace_kprobe.c | 44 #define SIZEOF_TRACE_PROBE(n) \ argument 46 (sizeof(struct probe_arg) * (n))) 218 int n = trace_probe_nr_files(tp); in enable_trace_probe() local 222 new = kzalloc((n + 2) * sizeof(struct ftrace_event_file *), in enable_trace_probe() 228 memcpy(new, old, n * sizeof(struct ftrace_event_file *)); in enable_trace_probe() 229 new[n] = file; in enable_trace_probe() 290 int n = trace_probe_nr_files(tp); in disable_trace_probe() local 294 if (n == 0 || trace_probe_file_index(tp, file) < 0) { in disable_trace_probe() 299 if (n == 1) { /* Remove the last file */ in disable_trace_probe() 303 new = kzalloc(n * sizeof(struct ftrace_event_file *), in disable_trace_probe() [all …]
|
D | trace_mmiotrace.c | 145 unsigned long n; in mmio_read() local 147 n = count_overruns(iter); in mmio_read() 148 if (n) { in mmio_read() 150 trace_seq_printf(s, "MARK 0.000000 Lost %lu events.\n", n); in mmio_read()
|