Home
last modified time | relevance | path

Searched refs:mask (Results 1 – 25 of 29) sorted by relevance

12

/kernel/
Dsched_cpupri.c81 if (cpumask_any_and(&p->cpus_allowed, vec->mask) >= nr_cpu_ids) in cpupri_find()
84 cpumask_and(lowest_mask, &p->cpus_allowed, vec->mask); in cpupri_find()
126 cpumask_clear_cpu(cpu, vec->mask); in cpupri_set()
136 cpumask_set_cpu(cpu, vec->mask); in cpupri_set()
166 alloc_bootmem_cpumask_var(&vec->mask); in cpupri_init()
167 else if (!alloc_cpumask_var(&vec->mask, GFP_KERNEL)) in cpupri_init()
177 free_cpumask_var(cp->pri_to_cpu[i].mask); in cpupri_init()
190 free_cpumask_var(cp->pri_to_cpu[i].mask); in cpupri_cleanup()
Dtaskstats.c293 static int add_del_listener(pid_t pid, const struct cpumask *mask, int isadd) in add_del_listener() argument
299 if (!cpumask_subset(mask, cpu_possible_mask)) in add_del_listener()
303 for_each_cpu(cpu, mask) { in add_del_listener()
322 for_each_cpu(cpu, mask) { in add_del_listener()
337 static int parse(struct nlattr *na, struct cpumask *mask) in parse() argument
354 ret = cpulist_parse(data, mask); in parse()
434 cpumask_var_t mask; in taskstats_user_cmd() local
436 if (!alloc_cpumask_var(&mask, GFP_KERNEL)) in taskstats_user_cmd()
439 rc = parse(info->attrs[TASKSTATS_CMD_ATTR_REGISTER_CPUMASK], mask); in taskstats_user_cmd()
443 rc = add_del_listener(info->snd_pid, mask, REGISTER); in taskstats_user_cmd()
[all …]
Drcutree.c668 cpu_quiet_msk(unsigned long mask, struct rcu_state *rsp, struct rcu_node *rnp, in cpu_quiet_msk() argument
674 if (!(rnp->qsmask & mask)) { in cpu_quiet_msk()
680 rnp->qsmask &= ~mask; in cpu_quiet_msk()
687 mask = rnp->grpmask; in cpu_quiet_msk()
722 unsigned long mask; in cpu_quiet() local
741 mask = rdp->grpmask; in cpu_quiet()
742 if ((rnp->qsmask & mask) == 0) { in cpu_quiet()
754 cpu_quiet_msk(mask, rsp, rnp, flags); /* releases rnp->lock */ in cpu_quiet()
800 unsigned long mask; in __rcu_offline_cpu() local
810 mask = rdp->grpmask; /* rnp->grplo is constant. */ in __rcu_offline_cpu()
[all …]
Dsmp.c320 void smp_call_function_many(const struct cpumask *mask, in smp_call_function_many() argument
332 cpu = cpumask_first_and(mask, cpu_online_mask); in smp_call_function_many()
334 cpu = cpumask_next_and(cpu, mask, cpu_online_mask); in smp_call_function_many()
340 next_cpu = cpumask_next_and(cpu, mask, cpu_online_mask); in smp_call_function_many()
342 next_cpu = cpumask_next_and(next_cpu, mask, cpu_online_mask); in smp_call_function_many()
356 if (cpumask_test_cpu(cpu, mask)) in smp_call_function_many()
368 cpumask_and(to_cpumask(data->cpumask_bits), mask, cpu_online_mask); in smp_call_function_many()
Dauditfilter.c321 static inline int audit_match_class_bits(int class, u32 *mask) in audit_match_class_bits() argument
327 if (mask[i] & classes[class][i]) in audit_match_class_bits()
341 entry->rule.mask) && in audit_match_signal()
343 entry->rule.mask)); in audit_match_signal()
349 entry->rule.mask)); in audit_match_signal()
352 entry->rule.mask)); in audit_match_signal()
400 entry->rule.mask[i] = rule->mask[i]; in audit_to_entry_common()
404 __u32 *p = &entry->rule.mask[AUDIT_WORD(bit)]; in audit_to_entry_common()
414 entry->rule.mask[j] |= class[j]; in audit_to_entry_common()
727 for (i = 0; i < AUDIT_BITMASK_SIZE; i++) rule->mask[i] = krule->mask[i]; in audit_krule_to_rule()
[all …]
Dsignal.c149 int next_signal(struct sigpending *pending, sigset_t *mask) in next_signal() argument
155 m = mask->sig; in next_signal()
334 block_all_signals(int (*notifier)(void *priv), void *priv, sigset_t *mask) in block_all_signals() argument
339 current->notifier_mask = mask; in block_all_signals()
395 static int __dequeue_signal(struct sigpending *pending, sigset_t *mask, in __dequeue_signal() argument
398 int sig = next_signal(pending, mask); in __dequeue_signal()
422 int dequeue_signal(struct task_struct *tsk, sigset_t *mask, siginfo_t *info) in dequeue_signal() argument
429 signr = __dequeue_signal(&tsk->pending, mask, info); in dequeue_signal()
432 mask, info); in dequeue_signal()
504 unsigned int mask; in signal_wake_up() local
[all …]
Dcompat.c496 cpumask_var_t mask; in compat_sys_sched_getaffinity() local
506 if (!alloc_cpumask_var(&mask, GFP_KERNEL)) in compat_sys_sched_getaffinity()
509 ret = sched_getaffinity(pid, mask); in compat_sys_sched_getaffinity()
513 k = cpumask_bits(mask); in compat_sys_sched_getaffinity()
519 free_cpumask_var(mask); in compat_sys_sched_getaffinity()
726 long compat_get_bitmap(unsigned long *mask, const compat_ulong_t __user *umask, in compat_get_bitmap() argument
761 *mask++ = m; in compat_get_bitmap()
767 long compat_put_bitmap(compat_ulong_t __user *umask, unsigned long *mask, in compat_put_bitmap() argument
784 m = *mask++; in compat_put_bitmap()
Dauditsc.c252 static inline int open_arg(int flags, int mask) in open_arg() argument
257 return n & mask; in open_arg()
260 static int audit_match_perm(struct audit_context *ctx, int mask) in audit_match_perm() argument
269 if ((mask & AUDIT_PERM_WRITE) && in audit_match_perm()
272 if ((mask & AUDIT_PERM_READ) && in audit_match_perm()
275 if ((mask & AUDIT_PERM_ATTR) && in audit_match_perm()
280 if ((mask & AUDIT_PERM_WRITE) && in audit_match_perm()
283 if ((mask & AUDIT_PERM_READ) && in audit_match_perm()
286 if ((mask & AUDIT_PERM_ATTR) && in audit_match_perm()
291 return mask & ACC_MODE(ctx->argv[1]); in audit_match_perm()
[all …]
Dsched_rt.c808 cpumask_var_t mask; in check_preempt_equal_prio() local
813 if (!alloc_cpumask_var(&mask, GFP_ATOMIC)) in check_preempt_equal_prio()
817 && cpupri_find(&rq->rd->cpupri, p, mask)) in check_preempt_equal_prio()
820 if (!cpupri_find(&rq->rd->cpupri, rq->curr, mask)) in check_preempt_equal_prio()
831 free_cpumask_var(mask); in check_preempt_equal_prio()
963 static inline int pick_optimal_cpu(int this_cpu, cpumask_t *mask) in pick_optimal_cpu() argument
968 if ((this_cpu != -1) && cpu_isset(this_cpu, *mask)) in pick_optimal_cpu()
971 first = cpumask_first(mask); in pick_optimal_cpu()
Dsched_cpupri.h17 cpumask_var_t mask; member
Dcpuset.c1556 nodemask_t mask; in cpuset_sprintf_memlist() local
1559 mask = cs->mems_allowed; in cpuset_sprintf_memlist()
1562 return nodelist_scnprintf(page, PAGE_SIZE, mask); in cpuset_sprintf_memlist()
2180 nodemask_t mask; in cpuset_mems_allowed() local
2184 guarantee_online_mems(task_cs(tsk), &mask); in cpuset_mems_allowed()
2188 return mask; in cpuset_mems_allowed()
Daudit.c710 if (status_get->mask & AUDIT_STATUS_ENABLED) { in audit_receive_msg()
716 if (status_get->mask & AUDIT_STATUS_FAILURE) { in audit_receive_msg()
722 if (status_get->mask & AUDIT_STATUS_PID) { in audit_receive_msg()
733 if (status_get->mask & AUDIT_STATUS_RATE_LIMIT) { in audit_receive_msg()
739 if (status_get->mask & AUDIT_STATUS_BACKLOG_LIMIT) in audit_receive_msg()
Dprofile.c465 struct cpumask *mask = data; in prof_cpu_mask_write_proc() local
474 cpumask_copy(mask, new_value); in prof_cpu_mask_write_proc()
Drtmutex.c55 unsigned long mask) in rt_mutex_set_owner() argument
57 unsigned long val = (unsigned long)owner | mask; in rt_mutex_set_owner()
Drelay.c906 unsigned int mask = 0; in relay_file_poll() local
915 mask |= POLLIN | POLLRDNORM; in relay_file_poll()
918 return mask; in relay_file_poll()
Dsys.c1701 SYSCALL_DEFINE1(umask, int, mask) in SYSCALL_DEFINE1() argument
1703 mask = xchg(&current->fs->umask, mask & S_IRWXUGO); in SYSCALL_DEFINE1()
1704 return mask; in SYSCALL_DEFINE1()
Dsched.c5682 long sched_getaffinity(pid_t pid, struct cpumask *mask) in sched_getaffinity() argument
5699 cpumask_and(mask, &p->cpus_allowed, cpu_online_mask); in sched_getaffinity()
5718 cpumask_var_t mask; in SYSCALL_DEFINE3() local
5723 if (!alloc_cpumask_var(&mask, GFP_KERNEL)) in SYSCALL_DEFINE3()
5726 ret = sched_getaffinity(pid, mask); in SYSCALL_DEFINE3()
5728 if (copy_to_user(user_mask_ptr, mask, cpumask_size())) in SYSCALL_DEFINE3()
5733 free_cpumask_var(mask); in SYSCALL_DEFINE3()
7311 struct sched_group **sg, struct cpumask *mask) in cpu_to_core_group() argument
7315 cpumask_and(mask, &per_cpu(cpu_sibling_map, cpu), cpu_map); in cpu_to_core_group()
7316 group = cpumask_first(mask); in cpu_to_core_group()
[all …]
/kernel/irq/
Dautoprobe.c34 unsigned long mask = 0; in probe_irq_on() local
104 mask |= 1 << i; in probe_irq_on()
109 return mask; in probe_irq_on()
127 unsigned int status, mask = 0; in probe_irq_mask() local
137 mask |= 1 << i; in probe_irq_mask()
146 return mask & val; in probe_irq_mask()
Dproc.c23 const struct cpumask *mask = &desc->affinity; in irq_affinity_proc_show() local
27 mask = &desc->pending_mask; in irq_affinity_proc_show()
29 seq_cpumask(m, mask); in irq_affinity_proc_show()
Dchip.c257 desc->chip->mask(irq); in default_shutdown()
292 desc->chip->mask(irq); in mask_ack_irq()
419 if (desc->chip->mask) in handle_fasteoi_irq()
420 desc->chip->mask(irq); in handle_fasteoi_irq()
490 desc->chip->mask(irq); in handle_edge_irq()
Dmigration.c60 desc->chip->mask(irq); in move_native_irq()
/kernel/time/
Dtick-broadcast.c132 static void tick_do_broadcast(struct cpumask *mask) in tick_do_broadcast() argument
140 if (cpumask_test_cpu(cpu, mask)) { in tick_do_broadcast()
141 cpumask_clear_cpu(cpu, mask); in tick_do_broadcast()
146 if (!cpumask_empty(mask)) { in tick_do_broadcast()
153 td = &per_cpu(tick_cpu_device, cpumask_first(mask)); in tick_do_broadcast()
154 td->evtdev->broadcast(mask); in tick_do_broadcast()
509 static void tick_broadcast_init_next_event(struct cpumask *mask, in tick_broadcast_init_next_event() argument
515 for_each_cpu(cpu, mask) { in tick_broadcast_init_next_event()
Dtimekeeping.c76 cycle_delta = (cycle_now - clock->cycle_last) & clock->mask; in clocksource_forward_now()
109 cycle_delta = (cycle_now - clock->cycle_last) & clock->mask; in getnstimeofday()
236 cycle_delta = (cycle_now - clock->cycle_last) & clock->mask; in getrawmonotonic()
489 offset = (clocksource_read(clock) - clock->cycle_last) & clock->mask; in update_wall_time()
Djiffies.c62 .mask = 0xffffffff, /*32bits*/
Dclocksource.c109 wd_nsec = cyc2ns(watchdog, (wdnow - watchdog_last) & watchdog->mask); in clocksource_watchdog()
136 cs_nsec = cyc2ns(cs, (csnow - cs->wd_last) & cs->mask); in clocksource_watchdog()

12