Home
last modified time | relevance | path

Searched refs:bit (Results 1 – 20 of 20) sorted by relevance

/kernel/irq/
Dmatrix.c181 void irq_matrix_assign_system(struct irq_matrix *m, unsigned int bit, in irq_matrix_assign_system() argument
186 BUG_ON(bit > m->matrix_bits); in irq_matrix_assign_system()
189 set_bit(bit, m->system_map); in irq_matrix_assign_system()
191 BUG_ON(!test_and_clear_bit(bit, cm->alloc_map)); in irq_matrix_assign_system()
195 if (bit >= m->alloc_start && bit < m->alloc_end) in irq_matrix_assign_system()
198 trace_irq_matrix_assign_system(bit, m); in irq_matrix_assign_system()
216 unsigned int bit; in irq_matrix_reserve_managed() local
218 bit = matrix_alloc_area(m, cm, 1, true); in irq_matrix_reserve_managed()
219 if (bit >= m->alloc_end) in irq_matrix_reserve_managed()
226 trace_irq_matrix_reserve_managed(bit, cpu, m, cm); in irq_matrix_reserve_managed()
[all …]
/kernel/sched/
Dwait_bit.c12 wait_queue_head_t *bit_waitqueue(void *word, int bit) in bit_waitqueue() argument
15 unsigned long val = (unsigned long)word << shift | bit; in bit_waitqueue()
58 int __sched out_of_line_wait_on_bit(void *word, int bit, in out_of_line_wait_on_bit() argument
61 struct wait_queue_head *wq_head = bit_waitqueue(word, bit); in out_of_line_wait_on_bit()
62 DEFINE_WAIT_BIT(wq_entry, word, bit); in out_of_line_wait_on_bit()
69 void *word, int bit, wait_bit_action_f *action, in out_of_line_wait_on_bit_timeout() argument
72 struct wait_queue_head *wq_head = bit_waitqueue(word, bit); in out_of_line_wait_on_bit_timeout()
73 DEFINE_WAIT_BIT(wq_entry, word, bit); in out_of_line_wait_on_bit_timeout()
111 int __sched out_of_line_wait_on_bit_lock(void *word, int bit, in out_of_line_wait_on_bit_lock() argument
114 struct wait_queue_head *wq_head = bit_waitqueue(word, bit); in out_of_line_wait_on_bit_lock()
[all …]
/kernel/trace/
Dtrace.h473 u32 bit; /* Mask assigned in val field in tracer_flags */ member
487 #define TRACER_OPT(s, b) .name = #s, .bit = b
545 u32 old_flags, u32 bit, int set);
632 #define trace_recursion_set(bit) do { (current)->trace_recursion |= (1<<(bit)); } while (0) argument
633 #define trace_recursion_clear(bit) do { (current)->trace_recursion &= ~(1<<(bit)); } while (0) argument
634 #define trace_recursion_test(bit) ((current)->trace_recursion & (1<<(bit))) argument
664 int bit; in trace_get_context_bit() local
668 bit = TRACE_CTX_NMI; in trace_get_context_bit()
671 bit = TRACE_CTX_IRQ; in trace_get_context_bit()
673 bit = TRACE_CTX_SOFTIRQ; in trace_get_context_bit()
[all …]
Dtrace_nop.c64 static int nop_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in nop_set_flag() argument
70 if (bit == TRACE_NOP_OPT_ACCEPT) { in nop_set_flag()
77 if (bit == TRACE_NOP_OPT_REFUSE) { in nop_set_flag()
Dtrace_functions.c137 int bit; in function_trace_call() local
147 bit = trace_test_and_set_recursion(TRACE_FTRACE_START); in function_trace_call()
148 if (bit < 0) in function_trace_call()
157 trace_clear_recursion(bit); in function_trace_call()
242 func_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in func_set_flag() argument
244 switch (bit) { in func_set_flag()
Dtrace_functions_graph.c1253 func_graph_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in func_graph_set_flag() argument
1255 if (bit == TRACE_GRAPH_PRINT_IRQS) in func_graph_set_flag()
1258 if (bit == TRACE_GRAPH_SLEEP_TIME) in func_graph_set_flag()
1261 if (bit == TRACE_GRAPH_GRAPH_TIME) in func_graph_set_flag()
Dring_buffer.c3174 int bit; in trace_recursive_lock() local
3177 bit = RB_CTX_NORMAL; in trace_recursive_lock()
3179 bit = pc & NMI_MASK ? RB_CTX_NMI : in trace_recursive_lock()
3182 if (unlikely(val & (1 << (bit + cpu_buffer->nest)))) { in trace_recursive_lock()
3188 bit = RB_CTX_TRANSITION; in trace_recursive_lock()
3189 if (val & (1 << (bit + cpu_buffer->nest))) in trace_recursive_lock()
3193 val |= (1 << (bit + cpu_buffer->nest)); in trace_recursive_lock()
Dftrace.c7010 int bit; in __ftrace_ops_list_func() local
7012 bit = trace_test_and_set_recursion(TRACE_LIST_START); in __ftrace_ops_list_func()
7013 if (bit < 0) in __ftrace_ops_list_func()
7046 trace_clear_recursion(bit); in __ftrace_ops_list_func()
7085 int bit; in ftrace_ops_assist_func() local
7087 bit = trace_test_and_set_recursion(TRACE_LIST_START); in ftrace_ops_assist_func()
7088 if (bit < 0) in ftrace_ops_assist_func()
7097 trace_clear_recursion(bit); in ftrace_ops_assist_func()
DKconfig444 Either of the above profilers adds a bit of overhead to the system.
450 No branch profiling. Branch profiling adds a bit of overhead.
813 with the event enabled. This adds a bit more time for kernel boot
Dblktrace.c1613 blk_tracer_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in blk_tracer_set_flag() argument
1616 if (bit == TRACE_BLK_OPT_CLASSIC) { in blk_tracer_set_flag()
Dtrace.c99 dummy_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set) in dummy_set_flag() argument
4880 if (tracer_flags & trace_opts[i].bit) in tracing_trace_options_show()
4897 ret = trace->set_flag(tr, tracer_flags->val, opts->bit, !neg); in __set_tracer_option()
4902 tracer_flags->val &= ~opts->bit; in __set_tracer_option()
4904 tracer_flags->val |= opts->bit; in __set_tracer_option()
8245 if (topt->flags->val & topt->opt->bit) in trace_options_read()
8268 if (!!(topt->flags->val & topt->opt->bit) != val) { in trace_options_write()
/kernel/time/
Dtick-sched.c349 enum tick_dep_bits bit) in tick_nohz_dep_set_all() argument
353 prev = atomic_fetch_or(BIT(bit), dep); in tick_nohz_dep_set_all()
362 void tick_nohz_dep_set(enum tick_dep_bits bit) in tick_nohz_dep_set() argument
364 tick_nohz_dep_set_all(&tick_dep_mask, bit); in tick_nohz_dep_set()
367 void tick_nohz_dep_clear(enum tick_dep_bits bit) in tick_nohz_dep_clear() argument
369 atomic_andnot(BIT(bit), &tick_dep_mask); in tick_nohz_dep_clear()
376 void tick_nohz_dep_set_cpu(int cpu, enum tick_dep_bits bit) in tick_nohz_dep_set_cpu() argument
383 prev = atomic_fetch_or(BIT(bit), &ts->tick_dep_mask); in tick_nohz_dep_set_cpu()
399 void tick_nohz_dep_clear_cpu(int cpu, enum tick_dep_bits bit) in tick_nohz_dep_clear_cpu() argument
403 atomic_andnot(BIT(bit), &ts->tick_dep_mask); in tick_nohz_dep_clear_cpu()
[all …]
/kernel/power/
Dsnapshot.c775 unsigned int bit; in memory_bm_set_bit() local
778 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in memory_bm_set_bit()
780 set_bit(bit, addr); in memory_bm_set_bit()
786 unsigned int bit; in mem_bm_set_bit_check() local
789 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in mem_bm_set_bit_check()
791 set_bit(bit, addr); in mem_bm_set_bit_check()
799 unsigned int bit; in memory_bm_clear_bit() local
802 error = memory_bm_find_bit(bm, pfn, &addr, &bit); in memory_bm_clear_bit()
804 clear_bit(bit, addr); in memory_bm_clear_bit()
809 int bit; in memory_bm_clear_current() local
[all …]
/kernel/
Dwatch_queue.c63 unsigned int bit; in watch_queue_pipe_buf_release() local
69 bit = buf->offset + buf->len; in watch_queue_pipe_buf_release()
70 if ((bit & (WATCH_QUEUE_NOTE_SIZE - 1)) == 0) in watch_queue_pipe_buf_release()
71 bit -= WATCH_QUEUE_NOTE_SIZE; in watch_queue_pipe_buf_release()
72 bit /= WATCH_QUEUE_NOTE_SIZE; in watch_queue_pipe_buf_release()
75 bit += page->index; in watch_queue_pipe_buf_release()
77 set_bit(bit, wqueue->notes_bitmap); in watch_queue_pipe_buf_release()
Dauditfilter.c281 int bit = AUDIT_BITMASK_SIZE * 32 - i - 1; in audit_to_entry_common() local
282 __u32 *p = &entry->rule.mask[AUDIT_WORD(bit)]; in audit_to_entry_common()
285 if (!(*p & AUDIT_BIT(bit))) in audit_to_entry_common()
287 *p &= ~AUDIT_BIT(bit); in audit_to_entry_common()
Dauditsc.c790 int word, bit; in audit_in_mask() local
799 bit = AUDIT_BIT(val); in audit_in_mask()
801 return rule->mask[word] & bit; in audit_in_mask()
/kernel/locking/
Dlockdep.c636 static inline unsigned long lock_flag(enum lock_usage_bit bit) in lock_flag() argument
638 return 1UL << bit; in lock_flag()
641 static char get_usage_char(struct lock_class *class, enum lock_usage_bit bit) in get_usage_char() argument
657 if (class->usage_mask & lock_flag(bit + LOCK_USAGE_DIR_MASK)) { in get_usage_char()
659 if (class->usage_mask & lock_flag(bit)) in get_usage_char()
661 } else if (class->usage_mask & lock_flag(bit)) in get_usage_char()
2278 int bit; in print_lock_class_header() local
2287 for (bit = 0; bit < LOCK_TRACE_STATES; bit++) { in print_lock_class_header()
2288 if (class->usage_mask & (1 << bit)) { in print_lock_class_header()
2291 len += printk("%*s %s", depth, "", usage_str[bit]); in print_lock_class_header()
[all …]
/kernel/dma/
DKconfig89 # The only thing that is really required is a way to set an uncached bit
/kernel/cgroup/
Dcpuset.c1983 static int update_flag(cpuset_flagbits_t bit, struct cpuset *cs, in update_flag() argument
1996 set_bit(bit, &trialcs->flags); in update_flag()
1998 clear_bit(bit, &trialcs->flags); in update_flag()
/kernel/events/
Dcore.c6480 int bit; in perf_output_sample_regs() local
6484 for_each_set_bit(bit, _mask, sizeof(mask) * BITS_PER_BYTE) { in perf_output_sample_regs()
6487 val = perf_reg_value(regs, bit); in perf_output_sample_regs()