/kernel/ |
D | sched_cpupri.c | 81 if (cpumask_any_and(&p->cpus_allowed, vec->mask) >= nr_cpu_ids) in cpupri_find() 84 cpumask_and(lowest_mask, &p->cpus_allowed, vec->mask); in cpupri_find() 126 cpumask_clear_cpu(cpu, vec->mask); in cpupri_set() 136 cpumask_set_cpu(cpu, vec->mask); in cpupri_set() 166 alloc_bootmem_cpumask_var(&vec->mask); in cpupri_init() 167 else if (!alloc_cpumask_var(&vec->mask, GFP_KERNEL)) in cpupri_init() 177 free_cpumask_var(cp->pri_to_cpu[i].mask); in cpupri_init() 190 free_cpumask_var(cp->pri_to_cpu[i].mask); in cpupri_cleanup()
|
D | taskstats.c | 293 static int add_del_listener(pid_t pid, const struct cpumask *mask, int isadd) in add_del_listener() argument 299 if (!cpumask_subset(mask, cpu_possible_mask)) in add_del_listener() 303 for_each_cpu(cpu, mask) { in add_del_listener() 322 for_each_cpu(cpu, mask) { in add_del_listener() 337 static int parse(struct nlattr *na, struct cpumask *mask) in parse() argument 354 ret = cpulist_parse(data, mask); in parse() 434 cpumask_var_t mask; in taskstats_user_cmd() local 436 if (!alloc_cpumask_var(&mask, GFP_KERNEL)) in taskstats_user_cmd() 439 rc = parse(info->attrs[TASKSTATS_CMD_ATTR_REGISTER_CPUMASK], mask); in taskstats_user_cmd() 443 rc = add_del_listener(info->snd_pid, mask, REGISTER); in taskstats_user_cmd() [all …]
|
D | rcutree.c | 668 cpu_quiet_msk(unsigned long mask, struct rcu_state *rsp, struct rcu_node *rnp, in cpu_quiet_msk() argument 674 if (!(rnp->qsmask & mask)) { in cpu_quiet_msk() 680 rnp->qsmask &= ~mask; in cpu_quiet_msk() 687 mask = rnp->grpmask; in cpu_quiet_msk() 722 unsigned long mask; in cpu_quiet() local 741 mask = rdp->grpmask; in cpu_quiet() 742 if ((rnp->qsmask & mask) == 0) { in cpu_quiet() 754 cpu_quiet_msk(mask, rsp, rnp, flags); /* releases rnp->lock */ in cpu_quiet() 800 unsigned long mask; in __rcu_offline_cpu() local 810 mask = rdp->grpmask; /* rnp->grplo is constant. */ in __rcu_offline_cpu() [all …]
|
D | smp.c | 320 void smp_call_function_many(const struct cpumask *mask, in smp_call_function_many() argument 332 cpu = cpumask_first_and(mask, cpu_online_mask); in smp_call_function_many() 334 cpu = cpumask_next_and(cpu, mask, cpu_online_mask); in smp_call_function_many() 340 next_cpu = cpumask_next_and(cpu, mask, cpu_online_mask); in smp_call_function_many() 342 next_cpu = cpumask_next_and(next_cpu, mask, cpu_online_mask); in smp_call_function_many() 356 if (cpumask_test_cpu(cpu, mask)) in smp_call_function_many() 368 cpumask_and(to_cpumask(data->cpumask_bits), mask, cpu_online_mask); in smp_call_function_many()
|
D | auditfilter.c | 321 static inline int audit_match_class_bits(int class, u32 *mask) in audit_match_class_bits() argument 327 if (mask[i] & classes[class][i]) in audit_match_class_bits() 341 entry->rule.mask) && in audit_match_signal() 343 entry->rule.mask)); in audit_match_signal() 349 entry->rule.mask)); in audit_match_signal() 352 entry->rule.mask)); in audit_match_signal() 400 entry->rule.mask[i] = rule->mask[i]; in audit_to_entry_common() 404 __u32 *p = &entry->rule.mask[AUDIT_WORD(bit)]; in audit_to_entry_common() 414 entry->rule.mask[j] |= class[j]; in audit_to_entry_common() 727 for (i = 0; i < AUDIT_BITMASK_SIZE; i++) rule->mask[i] = krule->mask[i]; in audit_krule_to_rule() [all …]
|
D | signal.c | 149 int next_signal(struct sigpending *pending, sigset_t *mask) in next_signal() argument 155 m = mask->sig; in next_signal() 334 block_all_signals(int (*notifier)(void *priv), void *priv, sigset_t *mask) in block_all_signals() argument 339 current->notifier_mask = mask; in block_all_signals() 395 static int __dequeue_signal(struct sigpending *pending, sigset_t *mask, in __dequeue_signal() argument 398 int sig = next_signal(pending, mask); in __dequeue_signal() 422 int dequeue_signal(struct task_struct *tsk, sigset_t *mask, siginfo_t *info) in dequeue_signal() argument 429 signr = __dequeue_signal(&tsk->pending, mask, info); in dequeue_signal() 432 mask, info); in dequeue_signal() 504 unsigned int mask; in signal_wake_up() local [all …]
|
D | compat.c | 496 cpumask_var_t mask; in compat_sys_sched_getaffinity() local 506 if (!alloc_cpumask_var(&mask, GFP_KERNEL)) in compat_sys_sched_getaffinity() 509 ret = sched_getaffinity(pid, mask); in compat_sys_sched_getaffinity() 513 k = cpumask_bits(mask); in compat_sys_sched_getaffinity() 519 free_cpumask_var(mask); in compat_sys_sched_getaffinity() 726 long compat_get_bitmap(unsigned long *mask, const compat_ulong_t __user *umask, in compat_get_bitmap() argument 761 *mask++ = m; in compat_get_bitmap() 767 long compat_put_bitmap(compat_ulong_t __user *umask, unsigned long *mask, in compat_put_bitmap() argument 784 m = *mask++; in compat_put_bitmap()
|
D | auditsc.c | 252 static inline int open_arg(int flags, int mask) in open_arg() argument 257 return n & mask; in open_arg() 260 static int audit_match_perm(struct audit_context *ctx, int mask) in audit_match_perm() argument 269 if ((mask & AUDIT_PERM_WRITE) && in audit_match_perm() 272 if ((mask & AUDIT_PERM_READ) && in audit_match_perm() 275 if ((mask & AUDIT_PERM_ATTR) && in audit_match_perm() 280 if ((mask & AUDIT_PERM_WRITE) && in audit_match_perm() 283 if ((mask & AUDIT_PERM_READ) && in audit_match_perm() 286 if ((mask & AUDIT_PERM_ATTR) && in audit_match_perm() 291 return mask & ACC_MODE(ctx->argv[1]); in audit_match_perm() [all …]
|
D | sched_rt.c | 808 cpumask_var_t mask; in check_preempt_equal_prio() local 813 if (!alloc_cpumask_var(&mask, GFP_ATOMIC)) in check_preempt_equal_prio() 817 && cpupri_find(&rq->rd->cpupri, p, mask)) in check_preempt_equal_prio() 820 if (!cpupri_find(&rq->rd->cpupri, rq->curr, mask)) in check_preempt_equal_prio() 831 free_cpumask_var(mask); in check_preempt_equal_prio() 963 static inline int pick_optimal_cpu(int this_cpu, cpumask_t *mask) in pick_optimal_cpu() argument 968 if ((this_cpu != -1) && cpu_isset(this_cpu, *mask)) in pick_optimal_cpu() 971 first = cpumask_first(mask); in pick_optimal_cpu()
|
D | sched_cpupri.h | 17 cpumask_var_t mask; member
|
D | cpuset.c | 1556 nodemask_t mask; in cpuset_sprintf_memlist() local 1559 mask = cs->mems_allowed; in cpuset_sprintf_memlist() 1562 return nodelist_scnprintf(page, PAGE_SIZE, mask); in cpuset_sprintf_memlist() 2180 nodemask_t mask; in cpuset_mems_allowed() local 2184 guarantee_online_mems(task_cs(tsk), &mask); in cpuset_mems_allowed() 2188 return mask; in cpuset_mems_allowed()
|
D | audit.c | 710 if (status_get->mask & AUDIT_STATUS_ENABLED) { in audit_receive_msg() 716 if (status_get->mask & AUDIT_STATUS_FAILURE) { in audit_receive_msg() 722 if (status_get->mask & AUDIT_STATUS_PID) { in audit_receive_msg() 733 if (status_get->mask & AUDIT_STATUS_RATE_LIMIT) { in audit_receive_msg() 739 if (status_get->mask & AUDIT_STATUS_BACKLOG_LIMIT) in audit_receive_msg()
|
D | profile.c | 465 struct cpumask *mask = data; in prof_cpu_mask_write_proc() local 474 cpumask_copy(mask, new_value); in prof_cpu_mask_write_proc()
|
D | rtmutex.c | 55 unsigned long mask) in rt_mutex_set_owner() argument 57 unsigned long val = (unsigned long)owner | mask; in rt_mutex_set_owner()
|
D | relay.c | 906 unsigned int mask = 0; in relay_file_poll() local 915 mask |= POLLIN | POLLRDNORM; in relay_file_poll() 918 return mask; in relay_file_poll()
|
D | sys.c | 1701 SYSCALL_DEFINE1(umask, int, mask) in SYSCALL_DEFINE1() argument 1703 mask = xchg(¤t->fs->umask, mask & S_IRWXUGO); in SYSCALL_DEFINE1() 1704 return mask; in SYSCALL_DEFINE1()
|
D | sched.c | 5682 long sched_getaffinity(pid_t pid, struct cpumask *mask) in sched_getaffinity() argument 5699 cpumask_and(mask, &p->cpus_allowed, cpu_online_mask); in sched_getaffinity() 5718 cpumask_var_t mask; in SYSCALL_DEFINE3() local 5723 if (!alloc_cpumask_var(&mask, GFP_KERNEL)) in SYSCALL_DEFINE3() 5726 ret = sched_getaffinity(pid, mask); in SYSCALL_DEFINE3() 5728 if (copy_to_user(user_mask_ptr, mask, cpumask_size())) in SYSCALL_DEFINE3() 5733 free_cpumask_var(mask); in SYSCALL_DEFINE3() 7311 struct sched_group **sg, struct cpumask *mask) in cpu_to_core_group() argument 7315 cpumask_and(mask, &per_cpu(cpu_sibling_map, cpu), cpu_map); in cpu_to_core_group() 7316 group = cpumask_first(mask); in cpu_to_core_group() [all …]
|
/kernel/irq/ |
D | autoprobe.c | 34 unsigned long mask = 0; in probe_irq_on() local 104 mask |= 1 << i; in probe_irq_on() 109 return mask; in probe_irq_on() 127 unsigned int status, mask = 0; in probe_irq_mask() local 137 mask |= 1 << i; in probe_irq_mask() 146 return mask & val; in probe_irq_mask()
|
D | proc.c | 23 const struct cpumask *mask = &desc->affinity; in irq_affinity_proc_show() local 27 mask = &desc->pending_mask; in irq_affinity_proc_show() 29 seq_cpumask(m, mask); in irq_affinity_proc_show()
|
D | chip.c | 257 desc->chip->mask(irq); in default_shutdown() 292 desc->chip->mask(irq); in mask_ack_irq() 419 if (desc->chip->mask) in handle_fasteoi_irq() 420 desc->chip->mask(irq); in handle_fasteoi_irq() 490 desc->chip->mask(irq); in handle_edge_irq()
|
D | migration.c | 60 desc->chip->mask(irq); in move_native_irq()
|
/kernel/time/ |
D | tick-broadcast.c | 132 static void tick_do_broadcast(struct cpumask *mask) in tick_do_broadcast() argument 140 if (cpumask_test_cpu(cpu, mask)) { in tick_do_broadcast() 141 cpumask_clear_cpu(cpu, mask); in tick_do_broadcast() 146 if (!cpumask_empty(mask)) { in tick_do_broadcast() 153 td = &per_cpu(tick_cpu_device, cpumask_first(mask)); in tick_do_broadcast() 154 td->evtdev->broadcast(mask); in tick_do_broadcast() 509 static void tick_broadcast_init_next_event(struct cpumask *mask, in tick_broadcast_init_next_event() argument 515 for_each_cpu(cpu, mask) { in tick_broadcast_init_next_event()
|
D | timekeeping.c | 76 cycle_delta = (cycle_now - clock->cycle_last) & clock->mask; in clocksource_forward_now() 109 cycle_delta = (cycle_now - clock->cycle_last) & clock->mask; in getnstimeofday() 236 cycle_delta = (cycle_now - clock->cycle_last) & clock->mask; in getrawmonotonic() 489 offset = (clocksource_read(clock) - clock->cycle_last) & clock->mask; in update_wall_time()
|
D | jiffies.c | 62 .mask = 0xffffffff, /*32bits*/
|
D | clocksource.c | 109 wd_nsec = cyc2ns(watchdog, (wdnow - watchdog_last) & watchdog->mask); in clocksource_watchdog() 136 cs_nsec = cyc2ns(cs, (csnow - cs->wd_last) & cs->mask); in clocksource_watchdog()
|