| /kernel/time/ |
| D | timer.c | 626 trigger_dyntick_cpu(struct timer_base *base, struct timer_list *timer) in trigger_dyntick_cpu() argument 646 if (base->is_idle) { in trigger_dyntick_cpu() 648 tick_nohz_full_cpu(base->cpu))); in trigger_dyntick_cpu() 649 wake_up_nohz_cpu(base->cpu); in trigger_dyntick_cpu() 658 static void enqueue_timer(struct timer_base *base, struct timer_list *timer, in enqueue_timer() argument 662 hlist_add_head(&timer->entry, base->vectors + idx); in enqueue_timer() 663 __set_bit(idx, base->pending_map); in enqueue_timer() 673 if (time_before(bucket_expiry, base->next_expiry)) { in enqueue_timer() 678 WRITE_ONCE(base->next_expiry, bucket_expiry); in enqueue_timer() 679 base->timers_pending = true; in enqueue_timer() [all …]
|
| D | hrtimer.c | 124 static inline bool hrtimer_base_is_online(struct hrtimer_cpu_base *base) in hrtimer_base_is_online() argument 129 return likely(base->online); in hrtimer_base_is_online() 168 __acquires(&timer->base->lock) in lock_hrtimer_base() 170 struct hrtimer_clock_base *base; in lock_hrtimer_base() local 173 base = READ_ONCE(timer->base); in lock_hrtimer_base() 174 if (likely(base != &migration_base)) { in lock_hrtimer_base() 175 raw_spin_lock_irqsave(&base->cpu_base->lock, *flags); in lock_hrtimer_base() 176 if (likely(base == timer->base)) in lock_hrtimer_base() 177 return base; in lock_hrtimer_base() 179 raw_spin_unlock_irqrestore(&base->cpu_base->lock, *flags); in lock_hrtimer_base() [all …]
|
| D | alarmtimer.c | 159 static void alarmtimer_enqueue(struct alarm_base *base, struct alarm *alarm) in alarmtimer_enqueue() argument 162 timerqueue_del(&base->timerqueue, &alarm->node); in alarmtimer_enqueue() 164 timerqueue_add(&base->timerqueue, &alarm->node); in alarmtimer_enqueue() 177 static void alarmtimer_dequeue(struct alarm_base *base, struct alarm *alarm) in alarmtimer_dequeue() argument 182 timerqueue_del(&base->timerqueue, &alarm->node); in alarmtimer_dequeue() 199 struct alarm_base *base = &alarm_bases[alarm->type]; in alarmtimer_fired() local 204 spin_lock_irqsave(&base->lock, flags); in alarmtimer_fired() 205 alarmtimer_dequeue(base, alarm); in alarmtimer_fired() 206 spin_unlock_irqrestore(&base->lock, flags); in alarmtimer_fired() 209 restart = alarm->function(alarm, base->get_ktime()); in alarmtimer_fired() [all …]
|
| D | timekeeping.c | 73 struct tk_read_base base[2]; member 107 .base[0] = FAST_TK_INIT, 108 .base[1] = FAST_TK_INIT, 113 .base[0] = FAST_TK_INIT, 114 .base[1] = FAST_TK_INIT, 325 struct tk_read_base *base = tkf->base; in update_fast_timekeeper() local 331 memcpy(base, tkr, sizeof(*base)); in update_fast_timekeeper() 337 memcpy(base + 1, base, sizeof(*base)); in update_fast_timekeeper() 348 tkr = tkf->base + (seq & 0x01); in __ktime_get_fast_ns() 349 now = ktime_to_ns(tkr->base); in __ktime_get_fast_ns() [all …]
|
| D | timer_list.c | 60 print_active_timers(struct seq_file *m, struct hrtimer_clock_base *base, in print_active_timers() argument 73 raw_spin_lock_irqsave(&base->cpu_base->lock, flags); in print_active_timers() 75 curr = timerqueue_getnext(&base->active); in print_active_timers() 89 raw_spin_unlock_irqrestore(&base->cpu_base->lock, flags); in print_active_timers() 95 raw_spin_unlock_irqrestore(&base->cpu_base->lock, flags); in print_active_timers() 99 print_base(struct seq_file *m, struct hrtimer_clock_base *base, u64 now) in print_base() argument 101 SEQ_printf(m, " .base: %p\n", base); in print_base() 102 SEQ_printf(m, " .index: %d\n", base->index); in print_base() 106 SEQ_printf(m, " .get_time: %ps\n", base->get_time); in print_base() 109 (unsigned long long) ktime_to_ns(base->offset)); in print_base() [all …]
|
| D | posix-cpu-timers.c | 436 struct posix_cputimer_base *base = timer_base(timer, tsk); in trigger_base_recalc_expires() local 438 base->nextevt = 0; in trigger_base_recalc_expires() 454 struct posix_cputimer_base *base; in disarm_timer() local 460 base = timer_base(timer, p); in disarm_timer() 461 if (cpu_timer_getexpires(ctmr) == base->nextevt) in disarm_timer() 559 struct posix_cputimer_base *base = timer_base(timer, p); in arm_timer() local 564 if (!cpu_timer_enqueue(&base->tqhead, ctmr)) in arm_timer() 573 if (newexp < base->nextevt) in arm_timer() 574 base->nextevt = newexp; in arm_timer() 824 struct posix_cputimer_base *base = pct->bases; in collect_posix_cputimers() local [all …]
|
| /kernel/trace/ |
| D | trace_probe_kernel.h | 40 static nokprobe_inline void set_data_loc(int ret, void *dest, void *__dest, void *base) in set_data_loc() argument 44 *(u32 *)dest = make_data_loc(ret, __dest - base); in set_data_loc() 52 fetch_store_string_user(unsigned long addr, void *dest, void *base) in fetch_store_string_user() argument 62 __dest = get_loc_data(dest, base); in fetch_store_string_user() 65 set_data_loc(ret, dest, __dest, base); in fetch_store_string_user() 75 fetch_store_string(unsigned long addr, void *dest, void *base) in fetch_store_string() argument 83 return fetch_store_string_user(addr, dest, base); in fetch_store_string() 89 __dest = get_loc_data(dest, base); in fetch_store_string() 96 set_data_loc(ret, dest, __dest, base); in fetch_store_string()
|
| D | trace_probe_tmpl.h | 58 void *dest, void *base); 61 fetch_store_string(unsigned long addr, void *dest, void *base); 64 fetch_store_string_user(unsigned long addr, void *dest, void *base); 88 fetch_store_symstring(unsigned long addr, void *dest, void *base) in fetch_store_symstring() argument 96 __dest = get_loc_data(dest, base); in fetch_store_symstring() 124 void *dest, void *base) in process_fetch_insn_bottom() argument 183 ret = fetch_store_string(val + code->offset, dest, base); in process_fetch_insn_bottom() 187 ret = fetch_store_string_user(val + code->offset, dest, base); in process_fetch_insn_bottom() 191 ret = fetch_store_symstring(val + code->offset, dest, base); in process_fetch_insn_bottom() 258 void *base = data - header_size; in store_trace_args() local [all …]
|
| D | trace_uprobe.c | 143 fetch_store_string(unsigned long addr, void *dest, void *base) in fetch_store_string() argument 148 u8 *dst = get_loc_data(dest, base); in fetch_store_string() 168 *(u32 *)dest = make_data_loc(ret, (void *)dst - base); in fetch_store_string() 170 *(u32 *)dest = make_data_loc(0, (void *)dst - base); in fetch_store_string() 176 fetch_store_string_user(unsigned long addr, void *dest, void *base) in fetch_store_string_user() argument 178 return fetch_store_string(addr, dest, base); in fetch_store_string_user() 216 void *dest, void *base) in process_fetch_insn() argument 249 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
|
| /kernel/locking/ |
| D | mutex.c | 308 ww = container_of(lock, struct ww_mutex, base); in ww_mutex_spin_on_owner() 578 mutex_unlock(&lock->base); in ww_mutex_unlock() 603 ww = container_of(lock, struct ww_mutex, base); in __mutex_lock_common() 829 return mutex_trylock(&ww->base); in ww_mutex_trylock() 831 MUTEX_WARN_ON(ww->base.magic != &ww->base); in ww_mutex_trylock() 841 if (__mutex_trylock(&ww->base)) { in ww_mutex_trylock() 843 mutex_acquire_nest(&ww->base.dep_map, 0, 1, &ww_ctx->dep_map, _RET_IP_); in ww_mutex_trylock() 927 ret = __ww_mutex_lock(&lock->base, TASK_UNINTERRUPTIBLE, in ww_mutex_lock() 942 ret = __ww_mutex_lock(&lock->base, TASK_INTERRUPTIBLE, in ww_mutex_lock_interruptible() 1152 return __ww_mutex_lock(&lock->base, TASK_UNINTERRUPTIBLE, 0, in __ww_mutex_lock_slowpath() [all …]
|
| D | ww_mutex.h | 418 if (likely(!__ww_mutex_has_waiters(&lock->base))) in ww_mutex_set_context_fastpath() 425 lock_wait_lock(&lock->base, &flags); in ww_mutex_set_context_fastpath() 426 __ww_mutex_check_waiters(&lock->base, ctx, &wake_q); in ww_mutex_set_context_fastpath() 428 unlock_wait_lock(&lock->base, &flags); in ww_mutex_set_context_fastpath() 440 ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_kill() 465 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_check_kill() 567 struct ww_mutex *ww = container_of(lock, struct ww_mutex, base); in __ww_mutex_add_waiter()
|
| D | ww_rt_mutex.c | 14 struct rt_mutex *rtm = &lock->base; in ww_mutex_trylock() 42 struct rt_mutex *rtm = &lock->base; in __ww_rt_mutex_lock() 94 struct rt_mutex *rtm = &lock->base; in ww_mutex_unlock()
|
| D | qspinlock.c | 135 struct mcs_spinlock *grab_mcs_node(struct mcs_spinlock *base, int idx) in grab_mcs_node() argument 137 return &((struct qnode *)base + idx)->mcs; in grab_mcs_node()
|
| /kernel/module/ |
| D | kdb.c | 44 kdb_printf(" 0x%px", mod->mem[MOD_TEXT].base); in kdb_lsmod() 45 kdb_printf("/0x%px", mod->mem[MOD_RODATA].base); in kdb_lsmod() 46 kdb_printf("/0x%px", mod->mem[MOD_RO_AFTER_INIT].base); in kdb_lsmod() 47 kdb_printf("/0x%px", mod->mem[MOD_DATA].base); in kdb_lsmod()
|
| D | strict_rwx.c | 19 if (!mod_mem->base) in module_set_memory() 22 set_vm_flush_reset_perms(mod_mem->base); in module_set_memory() 23 return set_memory((unsigned long)mod_mem->base, mod_mem->size >> PAGE_SHIFT); in module_set_memory()
|
| D | main.c | 107 static void __mod_update_bounds(enum mod_mem_type type __maybe_unused, void *base, in __mod_update_bounds() argument 110 unsigned long min = (unsigned long)base; in __mod_update_bounds() 134 __mod_update_bounds(type, mod_mem->base, mod_mem->size, &mod_tree); in mod_update_bounds() 290 #define symversion(base, idx) NULL argument 292 #define symversion(base, idx) ((base != NULL) ? ((base) + (idx)) : NULL) argument 1302 mod->mem[type].base = ptr; in module_memory_alloc() 1309 execmem_free(mod->mem[type].base); in module_memory_free() 1322 lockdep_free_key_range(mod_mem->base, mod_mem->size); in free_mod_mem() 1328 lockdep_free_key_range(mod->mem[MOD_DATA].base, mod->mem[MOD_DATA].size); in free_mod_mem() 2682 mod->mem[type].base = NULL; in move_module() [all …]
|
| D | debug_kmemleak.c | 18 kmemleak_no_scan(mod->mem[type].base); in kmemleak_load_module()
|
| /kernel/irq/ |
| D | devres.c | 184 int base; in __devm_irq_alloc_descs() local 190 base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity); in __devm_irq_alloc_descs() 191 if (base < 0) { in __devm_irq_alloc_descs() 193 return base; in __devm_irq_alloc_descs() 196 dr->from = base; in __devm_irq_alloc_descs() 200 return base; in __devm_irq_alloc_descs()
|
| D | ipi.c | 290 unsigned int base = data->irq; in __ipi_send_mask() local 293 unsigned irq = base + cpu - data->common->ipi_offset; in __ipi_send_mask()
|
| /kernel/dma/ |
| D | contiguous.c | 253 dma_contiguous_early_fixup(phys_addr_t base, unsigned long size) in dma_contiguous_early_fixup() argument 274 int __init dma_contiguous_reserve_area(phys_addr_t size, phys_addr_t base, in dma_contiguous_reserve_area() argument 281 ret = cma_declare_contiguous(base, size, limit, 0, 0, fixed, in dma_contiguous_reserve_area() 480 if (!IS_ALIGNED(rmem->base | rmem->size, CMA_MIN_ALIGNMENT_BYTES)) { in rmem_cma_setup() 492 err = cma_init_reserved_mem(rmem->base, rmem->size, 0, rmem->name, in rmem_cma_setup() 506 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_cma_setup()
|
| D | coherent.c | 342 mem = dma_init_coherent_memory(rmem->base, rmem->base, in rmem_dma_device_init() 382 dma_reserved_default_memory_base = rmem->base; in rmem_dma_setup() 389 &rmem->base, (unsigned long)rmem->size / SZ_1M); in rmem_dma_setup()
|
| /kernel/sched/ |
| D | TEST_MAPPING | 4 "path": "frameworks/base/packages/PackageInstaller" 7 "path": "frameworks/base/core/java/android/content/pm" 10 "path": "frameworks/base/services/core/java/com/android/server" 13 "path": "frameworks/base/core/java/com/android/internal/app" 16 "path": "frameworks/base/apex/jobscheduler/service/java/com/android/server/job"
|
| /kernel/ |
| D | resource.c | 1867 static resource_size_t gfr_start(struct resource *base, resource_size_t size, in gfr_start() argument 1873 end = min_t(resource_size_t, base->end, PHYSMEM_END); in gfr_start() 1877 return ALIGN(max(base->start, align), align); in gfr_start() 1880 static bool gfr_continue(struct resource *base, resource_size_t addr, in gfr_continue() argument 1884 return addr > size && addr >= base->start; in gfr_continue() 1890 addr <= min_t(resource_size_t, base->end, PHYSMEM_END); in gfr_continue() 1911 get_free_mem_region(struct device *dev, struct resource *base, in get_free_mem_region() argument 1939 for (addr = gfr_start(base, size, align, flags); in get_free_mem_region() 1940 gfr_continue(base, addr, align, flags); in get_free_mem_region() 1942 if (__region_intersects(base, addr, size, 0, IORES_DESC_NONE) != in get_free_mem_region() [all …]
|
| /kernel/events/ |
| D | ring_buffer.c | 914 void *base; in rb_free_work() local 920 base = rb->user_page; in rb_free_work() 923 perf_mmap_unmark_page(base + (i * PAGE_SIZE)); in rb_free_work() 925 vfree(base); in rb_free_work()
|
| /kernel/gcov/ |
| D | Makefile | 4 obj-y := base.o fs.o
|