Home
last modified time | relevance | path

Searched refs:base (Results 1 – 25 of 42) sorted by relevance

12

/kernel/time/
Dtimer.c578 trigger_dyntick_cpu(struct timer_base *base, struct timer_list *timer) in trigger_dyntick_cpu() argument
588 if (tick_nohz_full_cpu(base->cpu)) in trigger_dyntick_cpu()
589 wake_up_nohz_cpu(base->cpu); in trigger_dyntick_cpu()
598 if (base->is_idle) in trigger_dyntick_cpu()
599 wake_up_nohz_cpu(base->cpu); in trigger_dyntick_cpu()
607 static void enqueue_timer(struct timer_base *base, struct timer_list *timer, in enqueue_timer() argument
611 hlist_add_head(&timer->entry, base->vectors + idx); in enqueue_timer()
612 __set_bit(idx, base->pending_map); in enqueue_timer()
622 if (time_before(bucket_expiry, base->next_expiry)) { in enqueue_timer()
627 base->next_expiry = bucket_expiry; in enqueue_timer()
[all …]
Dhrtimer.c147 static inline bool is_migration_base(struct hrtimer_clock_base *base) in is_migration_base() argument
149 return base == &migration_base; in is_migration_base()
168 struct hrtimer_clock_base *base; in lock_hrtimer_base() local
171 base = READ_ONCE(timer->base); in lock_hrtimer_base()
172 if (likely(base != &migration_base)) { in lock_hrtimer_base()
173 raw_spin_lock_irqsave(&base->cpu_base->lock, *flags); in lock_hrtimer_base()
174 if (likely(base == timer->base)) in lock_hrtimer_base()
175 return base; in lock_hrtimer_base()
177 raw_spin_unlock_irqrestore(&base->cpu_base->lock, *flags); in lock_hrtimer_base()
202 struct hrtimer_cpu_base *get_target_base(struct hrtimer_cpu_base *base, in get_target_base() argument
[all …]
Dalarmtimer.c160 static void alarmtimer_enqueue(struct alarm_base *base, struct alarm *alarm) in alarmtimer_enqueue() argument
163 timerqueue_del(&base->timerqueue, &alarm->node); in alarmtimer_enqueue()
165 timerqueue_add(&base->timerqueue, &alarm->node); in alarmtimer_enqueue()
178 static void alarmtimer_dequeue(struct alarm_base *base, struct alarm *alarm) in alarmtimer_dequeue() argument
183 timerqueue_del(&base->timerqueue, &alarm->node); in alarmtimer_dequeue()
200 struct alarm_base *base = &alarm_bases[alarm->type]; in alarmtimer_fired() local
205 spin_lock_irqsave(&base->lock, flags); in alarmtimer_fired()
206 alarmtimer_dequeue(base, alarm); in alarmtimer_fired()
207 spin_unlock_irqrestore(&base->lock, flags); in alarmtimer_fired()
210 restart = alarm->function(alarm, base->get_ktime()); in alarmtimer_fired()
[all …]
Dtimekeeping.c73 struct tk_read_base base[2]; member
107 .base[0] = FAST_TK_INIT,
108 .base[1] = FAST_TK_INIT,
113 .base[0] = FAST_TK_INIT,
114 .base[1] = FAST_TK_INIT,
419 struct tk_read_base *base = tkf->base; in update_fast_timekeeper() local
425 memcpy(base, tkr, sizeof(*base)); in update_fast_timekeeper()
431 memcpy(base + 1, base, sizeof(*base)); in update_fast_timekeeper()
450 tkr = tkf->base + (seq & 0x01); in __ktime_get_fast_ns()
451 now = ktime_to_ns(tkr->base); in __ktime_get_fast_ns()
[all …]
Dtimer_list.c60 print_active_timers(struct seq_file *m, struct hrtimer_clock_base *base, in print_active_timers() argument
73 raw_spin_lock_irqsave(&base->cpu_base->lock, flags); in print_active_timers()
75 curr = timerqueue_getnext(&base->active); in print_active_timers()
89 raw_spin_unlock_irqrestore(&base->cpu_base->lock, flags); in print_active_timers()
95 raw_spin_unlock_irqrestore(&base->cpu_base->lock, flags); in print_active_timers()
99 print_base(struct seq_file *m, struct hrtimer_clock_base *base, u64 now) in print_base() argument
101 SEQ_printf(m, " .base: %pK\n", base); in print_base()
102 SEQ_printf(m, " .index: %d\n", base->index); in print_base()
106 SEQ_printf(m, " .get_time: %ps\n", base->get_time); in print_base()
109 (unsigned long long) ktime_to_ns(base->offset)); in print_base()
[all …]
Dposix-cpu-timers.c437 struct posix_cputimer_base *base = timer_base(timer, tsk); in trigger_base_recalc_expires() local
439 base->nextevt = 0; in trigger_base_recalc_expires()
455 struct posix_cputimer_base *base; in disarm_timer() local
460 base = timer_base(timer, p); in disarm_timer()
461 if (cpu_timer_getexpires(ctmr) == base->nextevt) in disarm_timer()
559 struct posix_cputimer_base *base = timer_base(timer, p); in arm_timer() local
563 if (!cpu_timer_enqueue(&base->tqhead, ctmr)) in arm_timer()
572 if (newexp < base->nextevt) in arm_timer()
573 base->nextevt = newexp; in arm_timer()
862 struct posix_cputimer_base *base = pct->bases; in collect_posix_cputimers() local
[all …]
/kernel/trace/
Dtrace_probe_kernel.h40 static nokprobe_inline void set_data_loc(int ret, void *dest, void *__dest, void *base) in set_data_loc() argument
44 *(u32 *)dest = make_data_loc(ret, __dest - base); in set_data_loc()
52 kern_fetch_store_string_user(unsigned long addr, void *dest, void *base) in kern_fetch_store_string_user() argument
62 __dest = get_loc_data(dest, base); in kern_fetch_store_string_user()
65 set_data_loc(ret, dest, __dest, base); in kern_fetch_store_string_user()
75 kern_fetch_store_string(unsigned long addr, void *dest, void *base) in kern_fetch_store_string() argument
83 return kern_fetch_store_string_user(addr, dest, base); in kern_fetch_store_string()
89 __dest = get_loc_data(dest, base); in kern_fetch_store_string()
96 set_data_loc(ret, dest, __dest, base); in kern_fetch_store_string()
Dtrace_probe_tmpl.h58 void *dest, void *base);
61 fetch_store_string(unsigned long addr, void *dest, void *base);
64 fetch_store_string_user(unsigned long addr, void *dest, void *base);
88 fetch_store_symstring(unsigned long addr, void *dest, void *base) in fetch_store_symstring() argument
96 __dest = get_loc_data(dest, base); in fetch_store_symstring()
104 void *dest, void *base) in process_fetch_insn_bottom() argument
163 ret = fetch_store_string(val + code->offset, dest, base); in process_fetch_insn_bottom()
167 ret = fetch_store_string_user(val + code->offset, dest, base); in process_fetch_insn_bottom()
171 ret = fetch_store_symstring(val + code->offset, dest, base); in process_fetch_insn_bottom()
238 void *base = data - header_size; in store_trace_args() local
[all …]
Dtrace_eprobe.c428 void *base) in process_fetch_insn() argument
453 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
476 fetch_store_string_user(unsigned long addr, void *dest, void *base) in fetch_store_string_user() argument
478 return kern_fetch_store_string_user(addr, dest, base); in fetch_store_string_user()
486 fetch_store_string(unsigned long addr, void *dest, void *base) in fetch_store_string() argument
488 return kern_fetch_store_string(addr, dest, base); in fetch_store_string()
Dtrace_uprobe.c147 fetch_store_string(unsigned long addr, void *dest, void *base) in fetch_store_string() argument
152 u8 *dst = get_loc_data(dest, base); in fetch_store_string()
172 *(u32 *)dest = make_data_loc(ret, (void *)dst - base); in fetch_store_string()
174 *(u32 *)dest = make_data_loc(0, (void *)dst - base); in fetch_store_string()
180 fetch_store_string_user(unsigned long addr, void *dest, void *base) in fetch_store_string_user() argument
182 return fetch_store_string(addr, dest, base); in fetch_store_string_user()
220 void *base) in process_fetch_insn() argument
256 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
/kernel/module/
Dstrict_rwx.c38 set_memory((unsigned long)layout->base, in frob_text()
45 set_memory((unsigned long)layout->base + layout->text_size, in frob_rodata()
52 set_memory((unsigned long)layout->base + layout->ro_size, in frob_ro_after_init()
59 set_memory((unsigned long)layout->base + layout->ro_after_init_size, in frob_writable_data()
65 return WARN_ON(!PAGE_ALIGNED(layout->base)) || in layout_check_misalignment()
84 if (!PAGE_ALIGNED(mod->core_layout.base) || in module_enable_x()
85 !PAGE_ALIGNED(mod->init_layout.base)) in module_enable_x()
101 set_vm_flush_reset_perms(mod->core_layout.base); in module_enable_ro()
102 set_vm_flush_reset_perms(mod->init_layout.base); in module_enable_ro()
Dmain.c104 static void __mod_update_bounds(void *base, unsigned int size, struct mod_tree_root *tree) in __mod_update_bounds() argument
106 unsigned long min = (unsigned long)base; in __mod_update_bounds()
117 __mod_update_bounds(mod->core_layout.base, mod->core_layout.size, &mod_tree); in mod_update_bounds()
119 __mod_update_bounds(mod->init_layout.base, mod->init_layout.size, &mod_tree); in mod_update_bounds()
121 __mod_update_bounds(mod->data_layout.base, mod->data_layout.size, &mod_data_tree); in mod_update_bounds()
245 #define symversion(base, idx) NULL argument
247 #define symversion(base, idx) ((base != NULL) ? ((base) + (idx)) : NULL) argument
1215 module_memfree(mod->init_layout.base); in free_module()
1220 lockdep_free_key_range(mod->data_layout.base, mod->data_layout.size); in free_module()
1224 module_memfree(mod->core_layout.base); in free_module()
[all …]
Dkallsyms.c176 mod->kallsyms = (void __rcu *)mod->init_layout.base + in add_kallsyms()
186 rcu_dereference(mod->kallsyms)->typetab = mod->init_layout.base + info->init_typeoffs; in add_kallsyms()
192 mod->core_kallsyms.symtab = dst = mod->data_layout.base + info->symoffs; in add_kallsyms()
193 mod->core_kallsyms.strtab = s = mod->data_layout.base + info->stroffs; in add_kallsyms()
194 mod->core_kallsyms.typetab = mod->data_layout.base + info->core_typeoffs; in add_kallsyms()
273 nextval = (unsigned long)mod->init_layout.base + mod->init_layout.text_size; in find_kallsyms_symbol()
275 nextval = (unsigned long)mod->core_layout.base + mod->core_layout.text_size; in find_kallsyms_symbol()
Dkdb.c43 kdb_printf(" 0x%px", mod->core_layout.base); in kdb_lsmod()
45 kdb_printf("/0x%px", mod->data_layout.base); in kdb_lsmod()
/kernel/locking/
Dmutex.c318 ww = container_of(lock, struct ww_mutex, base); in ww_mutex_spin_on_owner()
583 mutex_unlock(&lock->base); in ww_mutex_unlock()
606 ww = container_of(lock, struct ww_mutex, base); in __mutex_lock_common()
800 return mutex_trylock(&ww->base); in ww_mutex_trylock()
802 MUTEX_WARN_ON(ww->base.magic != &ww->base); in ww_mutex_trylock()
812 if (__mutex_trylock(&ww->base)) { in ww_mutex_trylock()
814 mutex_acquire_nest(&ww->base.dep_map, 0, 1, &ww_ctx->dep_map, _RET_IP_); in ww_mutex_trylock()
898 ret = __ww_mutex_lock(&lock->base, TASK_UNINTERRUPTIBLE, in ww_mutex_lock()
913 ret = __ww_mutex_lock(&lock->base, TASK_INTERRUPTIBLE, in ww_mutex_lock_interruptible()
1081 return __ww_mutex_lock(&lock->base, TASK_UNINTERRUPTIBLE, 0, in __ww_mutex_lock_slowpath()
[all …]
Dww_mutex.h400 if (likely(!__ww_mutex_has_waiters(&lock->base))) in ww_mutex_set_context_fastpath()
407 lock_wait_lock(&lock->base); in ww_mutex_set_context_fastpath()
408 __ww_mutex_check_waiters(&lock->base, ctx); in ww_mutex_set_context_fastpath()
409 unlock_wait_lock(&lock->base); in ww_mutex_set_context_fastpath()
419 ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_kill()
444 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_check_kill()
545 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_add_waiter()
Dww_rt_mutex.c14 struct rt_mutex *rtm = &lock->base; in ww_mutex_trylock()
42 struct rt_mutex *rtm = &lock->base; in __ww_rt_mutex_lock()
94 struct rt_mutex *rtm = &lock->base; in ww_mutex_unlock()
Dqspinlock.c135 struct mcs_spinlock *grab_mcs_node(struct mcs_spinlock *base, int idx) in grab_mcs_node() argument
137 return &((struct qnode *)base + idx)->mcs; in grab_mcs_node()
/kernel/irq/
Ddevres.c183 int base; in __devm_irq_alloc_descs() local
189 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs()
190 if (base < 0) { in __devm_irq_alloc_descs()
192 return base; in __devm_irq_alloc_descs()
195 dr->from = base; in __devm_irq_alloc_descs()
199 return base; in __devm_irq_alloc_descs()
Dipi.c290 unsigned int base = data->irq; in __ipi_send_mask() local
293 unsigned irq = base + cpu - data->common->ipi_offset; in __ipi_send_mask()
/kernel/dma/
Dcontiguous.c207 dma_contiguous_early_fixup(phys_addr_t base, unsigned long size) in dma_contiguous_early_fixup() argument
228 int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, in dma_contiguous_reserve_area() argument
234 ret = cma_declare_contiguous(base, size, limit, 0, 0, fixed, in dma_contiguous_reserve_area()
418 if (!IS_ALIGNED(rmem->base | rmem->size, CMA_MIN_ALIGNMENT_BYTES)) { in rmem_cma_setup()
423 err = cma_init_reserved_mem(rmem->base, rmem->size, 0, rmem->name, &cma); in rmem_cma_setup()
429 dma_contiguous_early_fixup(rmem->base, rmem->size); in rmem_cma_setup()
438 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_cma_setup()
Dcoherent.c341 mem = dma_init_coherent_memory(rmem->base, rmem->base, in rmem_dma_device_init()
387 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_dma_setup()
396 return dma_init_global_coherent(dma_reserved_default_memory->base, in dma_init_reserved_memory()
/kernel/
Dresource.c1775 static resource_size_t gfr_start(struct resource *base, resource_size_t size, in gfr_start() argument
1781 end = min_t(resource_size_t, base->end, in gfr_start()
1786 return ALIGN(base->start, align); in gfr_start()
1789 static bool gfr_continue(struct resource *base, resource_size_t addr, in gfr_continue() argument
1793 return addr > size && addr >= base->start; in gfr_continue()
1799 addr <= min_t(resource_size_t, base->end, in gfr_continue()
1821 get_free_mem_region(struct device *dev, struct resource *base, in get_free_mem_region() argument
1849 for (addr = gfr_start(base, size, align, flags); in get_free_mem_region()
1850 gfr_continue(base, addr, size, flags); in get_free_mem_region()
1852 if (__region_intersects(base, addr, size, 0, IORES_DESC_NONE) != in get_free_mem_region()
[all …]
/kernel/gcov/
DMakefile4 obj-y := base.o fs.o
/kernel/events/
Dring_buffer.c891 void *base; in rb_free_work() local
897 base = rb->user_page; in rb_free_work()
900 perf_mmap_unmark_page(base + (i * PAGE_SIZE)); in rb_free_work()
902 vfree(base); in rb_free_work()

12