/kernel/time/ |
D | timer.c | 555 trigger_dyntick_cpu(struct timer_base *base, struct timer_list *timer) in trigger_dyntick_cpu() argument 565 if (tick_nohz_full_cpu(base->cpu)) in trigger_dyntick_cpu() 566 wake_up_nohz_cpu(base->cpu); in trigger_dyntick_cpu() 575 if (base->is_idle) in trigger_dyntick_cpu() 576 wake_up_nohz_cpu(base->cpu); in trigger_dyntick_cpu() 584 static void enqueue_timer(struct timer_base *base, struct timer_list *timer, in enqueue_timer() argument 588 hlist_add_head(&timer->entry, base->vectors + idx); in enqueue_timer() 589 __set_bit(idx, base->pending_map); in enqueue_timer() 599 if (time_before(bucket_expiry, base->next_expiry)) { in enqueue_timer() 604 base->next_expiry = bucket_expiry; in enqueue_timer() [all …]
|
D | hrtimer.c | 147 static inline bool is_migration_base(struct hrtimer_clock_base *base) in is_migration_base() argument 149 return base == &migration_base; in is_migration_base() 168 struct hrtimer_clock_base *base; in lock_hrtimer_base() local 171 base = READ_ONCE(timer->base); in lock_hrtimer_base() 172 if (likely(base != &migration_base)) { in lock_hrtimer_base() 173 raw_spin_lock_irqsave(&base->cpu_base->lock, *flags); in lock_hrtimer_base() 174 if (likely(base == timer->base)) in lock_hrtimer_base() 175 return base; in lock_hrtimer_base() 177 raw_spin_unlock_irqrestore(&base->cpu_base->lock, *flags); in lock_hrtimer_base() 202 struct hrtimer_cpu_base *get_target_base(struct hrtimer_cpu_base *base, in get_target_base() argument [all …]
|
D | alarmtimer.c | 160 static void alarmtimer_enqueue(struct alarm_base *base, struct alarm *alarm) in alarmtimer_enqueue() argument 163 timerqueue_del(&base->timerqueue, &alarm->node); in alarmtimer_enqueue() 165 timerqueue_add(&base->timerqueue, &alarm->node); in alarmtimer_enqueue() 178 static void alarmtimer_dequeue(struct alarm_base *base, struct alarm *alarm) in alarmtimer_dequeue() argument 183 timerqueue_del(&base->timerqueue, &alarm->node); in alarmtimer_dequeue() 200 struct alarm_base *base = &alarm_bases[alarm->type]; in alarmtimer_fired() local 205 spin_lock_irqsave(&base->lock, flags); in alarmtimer_fired() 206 alarmtimer_dequeue(base, alarm); in alarmtimer_fired() 207 spin_unlock_irqrestore(&base->lock, flags); in alarmtimer_fired() 210 restart = alarm->function(alarm, base->get_ktime()); in alarmtimer_fired() [all …]
|
D | timekeeping.c | 74 struct tk_read_base base[2]; member 108 .base[0] = FAST_TK_INIT, 109 .base[1] = FAST_TK_INIT, 114 .base[0] = FAST_TK_INIT, 115 .base[1] = FAST_TK_INIT, 420 struct tk_read_base *base = tkf->base; in update_fast_timekeeper() local 426 memcpy(base, tkr, sizeof(*base)); in update_fast_timekeeper() 432 memcpy(base + 1, base, sizeof(*base)); in update_fast_timekeeper() 443 tkr = tkf->base + (seq & 0x01); in __ktime_get_fast_ns() 444 now = ktime_to_ns(tkr->base); in __ktime_get_fast_ns() [all …]
|
D | timer_list.c | 60 print_active_timers(struct seq_file *m, struct hrtimer_clock_base *base, in print_active_timers() argument 73 raw_spin_lock_irqsave(&base->cpu_base->lock, flags); in print_active_timers() 75 curr = timerqueue_getnext(&base->active); in print_active_timers() 89 raw_spin_unlock_irqrestore(&base->cpu_base->lock, flags); in print_active_timers() 95 raw_spin_unlock_irqrestore(&base->cpu_base->lock, flags); in print_active_timers() 99 print_base(struct seq_file *m, struct hrtimer_clock_base *base, u64 now) in print_base() argument 101 SEQ_printf(m, " .base: %pK\n", base); in print_base() 102 SEQ_printf(m, " .index: %d\n", base->index); in print_base() 106 SEQ_printf(m, " .get_time: %ps\n", base->get_time); in print_base() 109 (unsigned long long) ktime_to_ns(base->offset)); in print_base() [all …]
|
D | posix-cpu-timers.c | 430 struct posix_cputimer_base *base = timer_base(timer, tsk); in trigger_base_recalc_expires() local 432 base->nextevt = 0; in trigger_base_recalc_expires() 448 struct posix_cputimer_base *base; in disarm_timer() local 453 base = timer_base(timer, p); in disarm_timer() 454 if (cpu_timer_getexpires(ctmr) == base->nextevt) in disarm_timer() 552 struct posix_cputimer_base *base = timer_base(timer, p); in arm_timer() local 556 if (!cpu_timer_enqueue(&base->tqhead, ctmr)) in arm_timer() 565 if (newexp < base->nextevt) in arm_timer() 566 base->nextevt = newexp; in arm_timer() 855 struct posix_cputimer_base *base = pct->bases; in collect_posix_cputimers() local [all …]
|
/kernel/trace/ |
D | trace_probe_kernel.h | 40 static nokprobe_inline void set_data_loc(int ret, void *dest, void *__dest, void *base) in set_data_loc() argument 44 *(u32 *)dest = make_data_loc(ret, __dest - base); in set_data_loc() 52 kern_fetch_store_string_user(unsigned long addr, void *dest, void *base) in kern_fetch_store_string_user() argument 62 __dest = get_loc_data(dest, base); in kern_fetch_store_string_user() 65 set_data_loc(ret, dest, __dest, base); in kern_fetch_store_string_user() 75 kern_fetch_store_string(unsigned long addr, void *dest, void *base) in kern_fetch_store_string() argument 83 return kern_fetch_store_string_user(addr, dest, base); in kern_fetch_store_string() 89 __dest = get_loc_data(dest, base); in kern_fetch_store_string() 96 set_data_loc(ret, dest, __dest, base); in kern_fetch_store_string()
|
D | trace_probe_tmpl.h | 58 void *dest, void *base); 61 fetch_store_string(unsigned long addr, void *dest, void *base); 64 fetch_store_string_user(unsigned long addr, void *dest, void *base); 88 fetch_store_symstring(unsigned long addr, void *dest, void *base) in fetch_store_symstring() argument 96 __dest = get_loc_data(dest, base); in fetch_store_symstring() 104 void *dest, void *base) in process_fetch_insn_bottom() argument 163 ret = fetch_store_string(val + code->offset, dest, base); in process_fetch_insn_bottom() 167 ret = fetch_store_string_user(val + code->offset, dest, base); in process_fetch_insn_bottom() 171 ret = fetch_store_symstring(val + code->offset, dest, base); in process_fetch_insn_bottom() 238 void *base = data - header_size; in store_trace_args() local [all …]
|
D | trace_eprobe.c | 418 void *base) in process_fetch_insn() argument 443 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn() 466 fetch_store_string_user(unsigned long addr, void *dest, void *base) in fetch_store_string_user() argument 468 return kern_fetch_store_string_user(addr, dest, base); in fetch_store_string_user() 476 fetch_store_string(unsigned long addr, void *dest, void *base) in fetch_store_string() argument 478 return kern_fetch_store_string(addr, dest, base); in fetch_store_string()
|
D | trace_uprobe.c | 145 fetch_store_string(unsigned long addr, void *dest, void *base) in fetch_store_string() argument 150 u8 *dst = get_loc_data(dest, base); in fetch_store_string() 170 *(u32 *)dest = make_data_loc(ret, (void *)dst - base); in fetch_store_string() 172 *(u32 *)dest = make_data_loc(0, (void *)dst - base); in fetch_store_string() 178 fetch_store_string_user(unsigned long addr, void *dest, void *base) in fetch_store_string_user() argument 180 return fetch_store_string(addr, dest, base); in fetch_store_string_user() 218 void *base) in process_fetch_insn() argument 254 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
|
/kernel/ |
D | cfi.c | 45 unsigned long base; member 76 if (unlikely(page < s->base)) in ptr_to_shadow() 79 index = page - s->base; in ptr_to_shadow() 94 return (s->base + index) << PAGE_SHIFT; in shadow_to_ptr() 108 return (s->base + s->shadow[index]) << PAGE_SHIFT; in shadow_to_check_fn() 123 if (prev->base == next->base) { in prepare_next_shadow() 211 next->base = base_addr >> PAGE_SHIFT; in update_shadow() 214 min_addr = (unsigned long)mod->core_layout.base; in update_shadow()
|
D | module.c | 118 return (unsigned long)layout->base; in __mod_tree_val() 241 static void __mod_update_bounds(void *base, unsigned int size) in __mod_update_bounds() argument 243 unsigned long min = (unsigned long)base; in __mod_update_bounds() 254 __mod_update_bounds(mod->core_layout.base, mod->core_layout.size); in mod_update_bounds() 256 __mod_update_bounds(mod->init_layout.base, mod->init_layout.size); in mod_update_bounds() 424 #define symversion(base, idx) NULL argument 426 #define symversion(base, idx) ((base != NULL) ? ((base) + (idx)) : NULL) argument 1969 BUG_ON((unsigned long)layout->base & (PAGE_SIZE-1)); in frob_text() 1971 set_memory((unsigned long)layout->base, in frob_text() 1988 BUG_ON((unsigned long)layout->base & (PAGE_SIZE-1)); in frob_rodata() [all …]
|
D | resource.c | 1740 struct resource *base, unsigned long size, const char *name) in __request_free_mem_region() argument 1747 end = min_t(unsigned long, base->end, (1UL << MAX_PHYSMEM_BITS) - 1); in __request_free_mem_region() 1764 for (; addr > size && addr >= base->start; addr -= size) { in __request_free_mem_region() 1810 struct resource *base, unsigned long size) in devm_request_free_mem_region() argument 1812 return __request_free_mem_region(dev, base, size, dev_name(dev)); in devm_request_free_mem_region() 1816 struct resource *request_free_mem_region(struct resource *base, in request_free_mem_region() argument 1819 return __request_free_mem_region(NULL, base, size, name); in request_free_mem_region()
|
/kernel/irq/ |
D | devres.c | 183 int base; in __devm_irq_alloc_descs() local 189 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs() 190 if (base < 0) { in __devm_irq_alloc_descs() 192 return base; in __devm_irq_alloc_descs() 195 dr->from = base; in __devm_irq_alloc_descs() 199 return base; in __devm_irq_alloc_descs()
|
D | ipi.c | 284 unsigned int base = data->irq; in __ipi_send_mask() local 287 unsigned irq = base + cpu - data->common->ipi_offset; in __ipi_send_mask()
|
/kernel/dma/ |
D | contiguous.c | 207 dma_contiguous_early_fixup(phys_addr_t base, unsigned long size) in dma_contiguous_early_fixup() argument 228 int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, in dma_contiguous_reserve_area() argument 234 ret = cma_declare_contiguous(base, size, limit, 0, 0, fixed, in dma_contiguous_reserve_area() 420 if ((rmem->base & mask) || (rmem->size & mask)) { in rmem_cma_setup() 425 err = cma_init_reserved_mem(rmem->base, rmem->size, 0, rmem->name, &cma); in rmem_cma_setup() 431 dma_contiguous_early_fixup(rmem->base, rmem->size); in rmem_cma_setup() 440 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_cma_setup()
|
D | coherent.c | 342 mem = dma_init_coherent_memory(rmem->base, rmem->base, in rmem_dma_device_init() 388 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_dma_setup() 397 return dma_init_global_coherent(dma_reserved_default_memory->base, in dma_init_reserved_memory()
|
D | swiotlb.c | 811 if (PageHighMem(pfn_to_page(PHYS_PFN(rmem->base)))) { in rmem_swiotlb_device_init() 833 set_memory_decrypted((unsigned long)phys_to_virt(rmem->base), in rmem_swiotlb_device_init() 835 swiotlb_init_io_tlb_mem(mem, rmem->base, nslabs, false); in rmem_swiotlb_device_init() 872 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_swiotlb_setup()
|
/kernel/locking/ |
D | ww_mutex.h | 400 if (likely(!__ww_mutex_has_waiters(&lock->base))) in ww_mutex_set_context_fastpath() 407 lock_wait_lock(&lock->base); in ww_mutex_set_context_fastpath() 408 __ww_mutex_check_waiters(&lock->base, ctx); in ww_mutex_set_context_fastpath() 409 unlock_wait_lock(&lock->base); in ww_mutex_set_context_fastpath() 419 ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_kill() 444 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_check_kill() 545 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_add_waiter()
|
D | mutex.c | 311 ww = container_of(lock, struct ww_mutex, base); in ww_mutex_spin_on_owner() 575 mutex_unlock(&lock->base); in ww_mutex_unlock() 598 ww = container_of(lock, struct ww_mutex, base); in __mutex_lock_common() 841 ret = __ww_mutex_lock(&lock->base, TASK_UNINTERRUPTIBLE, in ww_mutex_lock() 856 ret = __ww_mutex_lock(&lock->base, TASK_INTERRUPTIBLE, in ww_mutex_lock_interruptible() 1024 return __ww_mutex_lock(&lock->base, TASK_UNINTERRUPTIBLE, 0, in __ww_mutex_lock_slowpath() 1032 return __ww_mutex_lock(&lock->base, TASK_INTERRUPTIBLE, 0, in __ww_mutex_lock_interruptible_slowpath() 1074 if (__mutex_trylock_fast(&lock->base)) { in ww_mutex_lock() 1089 if (__mutex_trylock_fast(&lock->base)) { in ww_mutex_lock_interruptible()
|
D | ww_rt_mutex.c | 17 struct rt_mutex *rtm = &lock->base; in __ww_rt_mutex_lock() 69 struct rt_mutex *rtm = &lock->base; in ww_mutex_unlock()
|
D | qspinlock.c | 134 struct mcs_spinlock *grab_mcs_node(struct mcs_spinlock *base, int idx) in grab_mcs_node() argument 136 return &((struct qnode *)base + idx)->mcs; in grab_mcs_node()
|
/kernel/bpf/ |
D | helpers.c | 444 unsigned int base = flags & BPF_STRTOX_BASE_MASK; in __bpf_strtoull() local 454 if (base != 0 && base != 8 && base != 10 && base != 16) in __bpf_strtoull() 477 cur_buf = _parse_integer_fixup_radix(cur_buf, &base); in __bpf_strtoull() 478 val_len = _parse_integer(cur_buf, base, res); in __bpf_strtoull()
|
/kernel/gcov/ |
D | Makefile | 4 obj-y := base.o fs.o
|
/kernel/events/ |
D | ring_buffer.c | 888 void *base; in rb_free_work() local 894 base = rb->user_page; in rb_free_work() 897 perf_mmap_unmark_page(base + (i * PAGE_SIZE)); in rb_free_work() 899 vfree(base); in rb_free_work()
|