Home
last modified time | relevance | path

Searched refs:q (Results 1 – 19 of 19) sorted by relevance

/kernel/sched/
Dswait.c7 void __init_swait_queue_head(struct swait_queue_head *q, const char *name, in __init_swait_queue_head() argument
10 raw_spin_lock_init(&q->lock); in __init_swait_queue_head()
11 lockdep_set_class_and_name(&q->lock, key, name); in __init_swait_queue_head()
12 INIT_LIST_HEAD(&q->task_list); in __init_swait_queue_head()
22 void swake_up_locked(struct swait_queue_head *q) in swake_up_locked() argument
26 if (list_empty(&q->task_list)) in swake_up_locked()
29 curr = list_first_entry(&q->task_list, typeof(*curr), task_list); in swake_up_locked()
35 void swake_up_one(struct swait_queue_head *q) in swake_up_one() argument
39 raw_spin_lock_irqsave(&q->lock, flags); in swake_up_one()
40 swake_up_locked(q); in swake_up_one()
[all …]
/kernel/trace/
Dblktrace.c343 static int __blk_trace_remove(struct request_queue *q) in __blk_trace_remove() argument
347 bt = xchg(&q->blk_trace, NULL); in __blk_trace_remove()
357 int blk_trace_remove(struct request_queue *q) in blk_trace_remove() argument
361 mutex_lock(&q->blk_trace_mutex); in blk_trace_remove()
362 ret = __blk_trace_remove(q); in blk_trace_remove()
363 mutex_unlock(&q->blk_trace_mutex); in blk_trace_remove()
474 static int do_blk_trace_setup(struct request_queue *q, char *name, dev_t dev, in do_blk_trace_setup() argument
547 if (cmpxchg(&q->blk_trace, NULL, bt)) in do_blk_trace_setup()
561 static int __blk_trace_setup(struct request_queue *q, char *name, dev_t dev, in __blk_trace_setup() argument
571 ret = do_blk_trace_setup(q, name, dev, bdev, &buts); in __blk_trace_setup()
[all …]
Dtrace_events_filter.c1161 char q; in parse_pred() local
1244 q = str[i]; in parse_pred()
1246 q = 0; in parse_pred()
1249 if (q && str[i] == q) in parse_pred()
1251 if (!q && (str[i] == ')' || str[i] == '&' || in parse_pred()
1256 if (q) in parse_pred()
1270 char q = str[i]; in parse_pred() local
1292 if (str[i] == q) in parse_pred()
/kernel/
Dfutex.c1504 static void __unqueue_futex(struct futex_q *q) in __unqueue_futex() argument
1508 if (WARN_ON_SMP(!q->lock_ptr) || WARN_ON(plist_node_empty(&q->list))) in __unqueue_futex()
1510 lockdep_assert_held(q->lock_ptr); in __unqueue_futex()
1512 hb = container_of(q->lock_ptr, struct futex_hash_bucket, lock); in __unqueue_futex()
1513 plist_del(&q->list, &hb->chain); in __unqueue_futex()
1523 static void mark_wake_futex(struct wake_q_head *wake_q, struct futex_q *q) in mark_wake_futex() argument
1525 struct task_struct *p = q->task; in mark_wake_futex()
1527 if (WARN(q->pi_state || q->rt_waiter, "refusing to wake PI futex\n")) in mark_wake_futex()
1531 __unqueue_futex(q); in mark_wake_futex()
1539 smp_store_release(&q->lock_ptr, NULL); in mark_wake_futex()
[all …]
Dlatencytop.c97 int q, same = 1; in account_global_scheduler_latency() local
105 for (q = 0; q < LT_BACKTRACEDEPTH; q++) { in account_global_scheduler_latency()
106 unsigned long record = lat->backtrace[q]; in account_global_scheduler_latency()
108 if (latency_record[i].backtrace[q] != record) { in account_global_scheduler_latency()
154 int i, q; in __account_scheduler_latency() local
182 for (q = 0; q < LT_BACKTRACEDEPTH; q++) { in __account_scheduler_latency()
183 unsigned long record = lat.backtrace[q]; in __account_scheduler_latency()
185 if (mylat->backtrace[q] != record) { in __account_scheduler_latency()
227 int q; in lstats_show() local
230 for (q = 0; q < LT_BACKTRACEDEPTH; q++) { in lstats_show()
[all …]
Dsignal.c414 struct sigqueue *q = NULL; in __sigqueue_alloc() local
429 q = kmem_cache_alloc(sigqueue_cachep, flags); in __sigqueue_alloc()
434 if (unlikely(q == NULL)) { in __sigqueue_alloc()
438 INIT_LIST_HEAD(&q->list); in __sigqueue_alloc()
439 q->flags = 0; in __sigqueue_alloc()
440 q->user = user; in __sigqueue_alloc()
443 return q; in __sigqueue_alloc()
446 static void __sigqueue_free(struct sigqueue *q) in __sigqueue_free() argument
448 if (q->flags & SIGQUEUE_PREALLOC) in __sigqueue_free()
450 atomic_dec(&q->user->sigpending); in __sigqueue_free()
[all …]
Dcrash_core.c196 char *q; in get_last_crashkernel() local
206 q = end_p - strlen(suffix_tbl[i]); in get_last_crashkernel()
207 if (!strncmp(q, suffix_tbl[i], in get_last_crashkernel()
213 q = end_p - strlen(suffix); in get_last_crashkernel()
214 if (!strncmp(q, suffix, strlen(suffix))) in get_last_crashkernel()
Dauditsc.c251 struct audit_tree_refs *q; in unroll_tree_refs() local
262 for (q = p; q != ctx->trees; q = q->next, n = 31) { in unroll_tree_refs()
264 audit_put_chunk(q->c[n]); in unroll_tree_refs()
265 q->c[n] = NULL; in unroll_tree_refs()
269 audit_put_chunk(q->c[n]); in unroll_tree_refs()
270 q->c[n] = NULL; in unroll_tree_refs()
278 struct audit_tree_refs *p, *q; in free_tree_refs() local
279 for (p = ctx->first_trees; p; p = q) { in free_tree_refs()
280 q = p->next; in free_tree_refs()
Dauditfilter.c1062 static void audit_list_rules(int seq, struct sk_buff_head *q) in audit_list_rules() argument
1081 skb_queue_tail(q, skb); in audit_list_rules()
1087 skb_queue_tail(q, skb); in audit_list_rules()
1174 skb_queue_head_init(&dest->q); in audit_list_rules_send()
1177 audit_list_rules(seq, &dest->q); in audit_list_rules_send()
1182 skb_queue_purge(&dest->q); in audit_list_rules_send()
Daudit_tree.c613 struct list_head *p, *q; in trim_marked() local
620 for (p = tree->chunks.next; p != &tree->chunks; p = q) { in trim_marked()
622 q = p->next; in trim_marked()
Dptrace.c712 struct sigqueue *q; in ptrace_peek_siginfo() local
741 list_for_each_entry(q, &pending->list, list) { in ptrace_peek_siginfo()
744 copy_siginfo(&info, &q->info); in ptrace_peek_siginfo()
Daudit.h229 struct sk_buff_head q; member
Daudit.c892 while ((skb = __skb_dequeue(&dest->q)) != NULL) in audit_send_list()
/kernel/cgroup/
Dpids.c145 struct pids_cgroup *p, *q; in pids_try_charge() local
163 for (q = pids; q != p; q = parent_pids(q)) in pids_try_charge()
164 pids_cancel(q, num); in pids_try_charge()
Dcpuset.c444 static int is_cpuset_subset(const struct cpuset *p, const struct cpuset *q) in is_cpuset_subset() argument
446 return cpumask_subset(p->cpus_allowed, q->cpus_allowed) && in is_cpuset_subset()
447 nodes_subset(p->mems_allowed, q->mems_allowed) && in is_cpuset_subset()
448 is_cpu_exclusive(p) <= is_cpu_exclusive(q) && in is_cpuset_subset()
449 is_mem_exclusive(p) <= is_mem_exclusive(q); in is_cpuset_subset()
/kernel/bpf/
Dcpumap.c45 void *q[CPU_MAP_BULK_SIZE]; member
607 struct ptr_ring *q; in bq_flush_to_queue() local
613 q = rcpu->queue; in bq_flush_to_queue()
614 spin_lock(&q->producer_lock); in bq_flush_to_queue()
617 struct xdp_frame *xdpf = bq->q[i]; in bq_flush_to_queue()
620 err = __ptr_ring_produce(q, xdpf); in bq_flush_to_queue()
631 spin_unlock(&q->producer_lock); in bq_flush_to_queue()
660 bq->q[bq->count++] = xdpf; in bq_enqueue()
Ddevmap.c59 struct xdp_frame *q[DEV_MAP_BULK_SIZE]; member
360 struct xdp_frame *xdpf = bq->q[i]; in bq_xmit_all()
365 sent = dev->netdev_ops->ndo_xdp_xmit(dev, bq->count, bq->q, flags); in bq_xmit_all()
385 struct xdp_frame *xdpf = bq->q[i]; in bq_xmit_all()
452 bq->q[bq->count++] = xdpf; in bq_enqueue()
/kernel/time/
Dtimeconst.bc55 print "#error \qinclude/generated/timeconst.h has the wrong HZ value!\q\n"
/kernel/events/
Duprobes.c330 struct list_head *pos, *q; in delayed_uprobe_remove() local
336 list_for_each_safe(pos, q, &delayed_uprobe_list) { in delayed_uprobe_remove()
1331 struct list_head *pos, *q; in delayed_ref_ctr_inc() local
1337 list_for_each_safe(pos, q, &delayed_uprobe_list) { in delayed_ref_ctr_inc()