/kernel/ |
D | notifier.c | 22 struct notifier_block *n) in notifier_chain_register() argument 25 if (n->priority > (*nl)->priority) in notifier_chain_register() 29 n->next = *nl; in notifier_chain_register() 30 rcu_assign_pointer(*nl, n); in notifier_chain_register() 35 struct notifier_block *n) in notifier_chain_cond_register() argument 38 if ((*nl) == n) in notifier_chain_cond_register() 40 if (n->priority > (*nl)->priority) in notifier_chain_cond_register() 44 n->next = *nl; in notifier_chain_cond_register() 45 rcu_assign_pointer(*nl, n); in notifier_chain_cond_register() 50 struct notifier_block *n) in notifier_chain_unregister() argument [all …]
|
D | timeconst.pl | 173 my ($b,$n,$d) = @_; 175 $n = bint($n); 178 return scalar (($n << $b)+$d-bint(1))/$d; 183 my($b,$n,$d) = @_; 185 $n = bint($n); 188 $d = $d/bgcd($n, $d); 193 my($b,$n,$d) = @_; 197 $n = bint($n); 201 $m = fmul($s,$n,$d); 228 my($n, $d) = @_; [all …]
|
D | auditsc.c | 254 int n = ACC_MODE(flags); in open_arg() local 256 n |= AUDIT_PERM_WRITE; in open_arg() 257 return n & mask; in open_arg() 262 unsigned n; in audit_match_perm() local 265 n = ctx->major; in audit_match_perm() 267 switch (audit_classify_syscall(ctx->arch, n)) { in audit_match_perm() 270 audit_match_class(AUDIT_CLASS_WRITE, n)) in audit_match_perm() 273 audit_match_class(AUDIT_CLASS_READ, n)) in audit_match_perm() 276 audit_match_class(AUDIT_CLASS_CHATTR, n)) in audit_match_perm() 281 audit_match_class(AUDIT_CLASS_WRITE_32, n)) in audit_match_perm() [all …]
|
D | user_namespace.c | 26 int n; in create_user_ns() local 34 for (n = 0; n < UIDHASH_SZ; ++n) in create_user_ns() 35 INIT_HLIST_HEAD(ns->uidhash_table + n); in create_user_ns()
|
D | audit_tree.c | 164 unsigned long n = (unsigned long)inode / L1_CACHE_BYTES; in chunk_hash() local 165 return chunk_hash_heads + n % HASH_SIZE; in chunk_hash() 192 int n; in audit_tree_match() local 193 for (n = 0; n < chunk->count; n++) in audit_tree_match() 194 if (chunk->owners[n].owner == tree) in audit_tree_match() 363 int n; in tag_chunk() local 372 for (n = 0; n < old->count; n++) { in tag_chunk() 373 if (old->owners[n].owner == tree) { in tag_chunk() 401 for (n = 0, p = chunk->owners; n < old->count; n++, p++) { in tag_chunk() 402 struct audit_tree *s = old->owners[n].owner; in tag_chunk() [all …]
|
D | resource.c | 626 resource_size_t start, resource_size_t n, in __request_region() argument 636 res->end = start + n - 1; in __request_region() 680 resource_size_t n) in __check_region() argument 684 res = __request_region(parent, start, n, "check-region", 0); in __check_region() 703 resource_size_t n) in __release_region() argument 709 end = start + n - 1; in __release_region() 747 resource_size_t n; member 754 __release_region(this->parent, this->start, this->n); in devm_region_release() 762 this->start == match->start && this->n == match->n; in devm_region_match() 767 resource_size_t n, const char *name) in __devm_request_region() argument [all …]
|
D | profile.c | 165 int task_handoff_register(struct notifier_block *n) in task_handoff_register() argument 167 return atomic_notifier_chain_register(&task_free_notifier, n); in task_handoff_register() 171 int task_handoff_unregister(struct notifier_block *n) in task_handoff_unregister() argument 173 return atomic_notifier_chain_unregister(&task_free_notifier, n); in task_handoff_unregister() 177 int profile_event_register(enum profile_type type, struct notifier_block *n) in profile_event_register() argument 184 &task_exit_notifier, n); in profile_event_register() 188 &munmap_notifier, n); in profile_event_register() 196 int profile_event_unregister(enum profile_type type, struct notifier_block *n) in profile_event_unregister() argument 203 &task_exit_notifier, n); in profile_event_unregister() 207 &munmap_notifier, n); in profile_event_unregister()
|
D | auditfilter.c | 296 unsigned n = *list++; in audit_register_class() local 297 if (n >= AUDIT_BITMASK_SIZE * 32 - AUDIT_SYSCALL_CLASSES) { in audit_register_class() 301 p[AUDIT_WORD(n)] |= AUDIT_BIT(n); in audit_register_class() 438 u32 n; in audit_to_op() local 439 for (n = Audit_equal; n < Audit_bad && audit_ops[n] != op; n++) in audit_to_op() 441 return n; in audit_to_op() 459 u32 n; in audit_rule_to_entry() local 461 n = rule->fields[i] & (AUDIT_NEGATE|AUDIT_OPERATORS); in audit_rule_to_entry() 465 if (n & AUDIT_NEGATE) in audit_rule_to_entry() 467 else if (!n) in audit_rule_to_entry() [all …]
|
D | user.c | 483 int n; in uid_cache_init() local 488 for(n = 0; n < UIDHASH_SZ; ++n) in uid_cache_init() 489 INIT_HLIST_HEAD(init_user_ns.uidhash_table + n); in uid_cache_init()
|
D | params.c | 376 #define to_module_attr(n) container_of(n, struct module_attribute, attr); argument 377 #define to_module_kobject(n) container_of(n, struct module_kobject, kobj); argument 395 #define to_param_attr(n) container_of(n, struct param_attribute, mattr); argument
|
D | exit.c | 767 struct task_struct *p, *n; in ptrace_exit() local 770 list_for_each_entry_safe(p, n, &parent->ptraced, ptrace_entry) { in ptrace_exit() 816 struct task_struct *p, *n; in ptrace_exit_finish() local 820 list_for_each_entry_safe(p, n, dead, ptrace_entry) { in ptrace_exit_finish() 896 struct task_struct *p, *n, *reaper; in forget_original_parent() local 906 list_for_each_entry_safe(p, n, &father->children, sibling) { in forget_original_parent() 991 unsigned long *n = end_of_stack(current); in check_stack_usage() local 994 while (*n == 0) in check_stack_usage() 995 n++; in check_stack_usage() 996 free = (unsigned long)n - (unsigned long)end_of_stack(current); in check_stack_usage()
|
D | rtmutex.h | 22 #define debug_rt_mutex_init(m, n) do { } while (0) argument
|
D | sched_debug.c | 288 #define P(n) SEQ_printf(m, " .%-30s: %d\n", #n, rq->n); in print_cpu() argument
|
D | rtmutex.c | 82 # define rt_mutex_cmpxchg(l,c,n) (cmpxchg(&l->owner, c, n) == c) argument 92 # define rt_mutex_cmpxchg(l,c,n) (0) argument
|
D | cpuset.c | 1043 int i, n, ntasks; in update_tasks_nodemask() local 1075 n = 0; in update_tasks_nodemask() 1082 if (n >= ntasks) { in update_tasks_nodemask() 1090 mmarray[n++] = mm; in update_tasks_nodemask() 1109 for (i = 0; i < n; i++) { in update_tasks_nodemask()
|
D | sysctl.c | 1653 int n; in parse_table() local 1657 if (get_user(n, name)) in parse_table() 1662 if (n == table->ctl_name) { in parse_table() 1911 unsigned int n, npath; in __register_sysctl_paths() local 1935 for (n = 0; n < npath; ++n, ++path) { in __register_sysctl_paths()
|
D | signal.c | 255 struct sigqueue *q, *n; in __flush_itimer_signals() local 260 list_for_each_entry_safe(q, n, &pending->list, list) { in __flush_itimer_signals() 533 struct sigqueue *q, *n; in rm_from_queue_full() local 541 list_for_each_entry_safe(q, n, &s->list, list) { in rm_from_queue_full() 557 struct sigqueue *q, *n; in rm_from_queue() local 563 list_for_each_entry_safe(q, n, &s->list, list) { in rm_from_queue()
|
/kernel/power/ |
D | userwakelock.c | 133 struct rb_node *n; in wake_lock_show() local 138 for (n = rb_first(&user_wake_locks); n != NULL; n = rb_next(n)) { in wake_lock_show() 139 l = rb_entry(n, struct user_wake_lock, node); in wake_lock_show() 151 const char *buf, size_t n) in wake_lock_store() argument 159 n = PTR_ERR(l); in wake_lock_store() 172 return n; in wake_lock_store() 181 struct rb_node *n; in wake_unlock_show() local 186 for (n = rb_first(&user_wake_locks); n != NULL; n = rb_next(n)) { in wake_unlock_show() 187 l = rb_entry(n, struct user_wake_lock, node); in wake_unlock_show() 199 const char *buf, size_t n) in wake_unlock_store() argument [all …]
|
D | main.c | 91 const char *buf, size_t n) in pm_test_store() argument 99 p = memchr(buf, '\n', n); in pm_test_store() 100 len = p ? p - buf : n; in pm_test_store() 114 return error ? error : n; in pm_test_store() 512 const char *buf, size_t n) in state_store() argument 526 p = memchr(buf, '\n', n); in state_store() 527 len = p ? p - buf : n; in state_store() 552 return error ? error : n; in state_store() 568 const char *buf, size_t n) in pm_trace_store() argument 574 return n; in pm_trace_store()
|
D | disk.c | 775 const char *buf, size_t n) in disk_store() argument 783 p = memchr(buf, '\n', n); in disk_store() 784 len = p ? p - buf : n; in disk_store() 815 return error ? error : n; in disk_store() 828 const char *buf, size_t n) in resume_store() argument 847 ret = n; in resume_store() 861 const char *buf, size_t n) in image_size_store() argument 867 return n; in image_size_store()
|
D | snapshot.c | 849 unsigned int n = 0; in count_highmem_pages() local 861 n++; in count_highmem_pages() 863 return n; in count_highmem_pages() 912 unsigned int n = 0; in count_data_pages() local 922 n++; in count_data_pages() 924 return n; in count_data_pages() 932 int n; in do_copy_page() local 934 for (n = PAGE_SIZE / sizeof(long); n; n--) in do_copy_page() 1148 alloc_highmem_image_pages(struct memory_bitmap *bm, unsigned int n) { return 0; } in alloc_highmem_image_pages() argument
|
D | wakelock.c | 93 int n; in print_lock_stat() local 115 n = snprintf(buf, len, in print_lock_stat() 123 return n > len ? len : n; in print_lock_stat() 247 struct wake_lock *lock, *n; in has_wake_lock_locked() local 251 list_for_each_entry_safe(lock, n, &active_wake_locks[type], link) { in has_wake_lock_locked()
|
D | power.h | 238 const char *buf, size_t n); 242 const char *buf, size_t n);
|
D | Kconfig | 33 default n 61 default n 128 default n
|
/kernel/trace/ |
D | trace_mmiotrace.c | 141 unsigned long n; in mmio_read() local 143 n = count_overruns(iter); in mmio_read() 144 if (n) { in mmio_read() 146 trace_seq_printf(s, "MARK 0.000000 Lost %lu events.\n", n); in mmio_read()
|