/kernel/ |
D | memremap.c | 24 __weak void __iomem *ioremap_cache(resource_size_t offset, unsigned long size) in ioremap_cache() argument 26 return ioremap(offset, size); in ioremap_cache() 31 static void *arch_memremap_wb(resource_size_t offset, unsigned long size) in arch_memremap_wb() argument 33 return (__force void *)ioremap_cache(offset, size); in arch_memremap_wb() 37 static void *try_ram_remap(resource_size_t offset, size_t size) in try_ram_remap() argument 39 unsigned long pfn = PHYS_PFN(offset); in try_ram_remap() 43 return __va(offset); in try_ram_remap() 74 void *memremap(resource_size_t offset, size_t size, unsigned long flags) in memremap() argument 76 int is_ram = region_intersects(offset, size, in memremap() 85 &offset, (unsigned long) size); in memremap() [all …]
|
D | pid.c | 109 int offset = nr & BITS_PER_PAGE_MASK; in free_pidmap() local 111 clear_bit(offset, map->page); in free_pidmap() 156 int i, offset, max_scan, pid, last = pid_ns->last_pid; in alloc_pidmap() local 162 offset = pid & BITS_PER_PAGE_MASK; in alloc_pidmap() 169 max_scan = DIV_ROUND_UP(pid_max, BITS_PER_PAGE) - !offset; in alloc_pidmap() 189 if (!test_and_set_bit(offset, map->page)) { in alloc_pidmap() 194 offset = find_next_offset(map, offset); in alloc_pidmap() 195 if (offset >= BITS_PER_PAGE) in alloc_pidmap() 197 pid = mk_pid(pid_ns, map, offset); in alloc_pidmap() 204 offset = 0; in alloc_pidmap() [all …]
|
D | kallsyms.c | 238 unsigned long *offset) in get_symbol_pos() argument 290 if (offset) in get_symbol_pos() 291 *offset = addr - symbol_start; in get_symbol_pos() 300 unsigned long *offset) in kallsyms_lookup_size_offset() argument 304 return !!get_symbol_pos(addr, symbolsize, offset); in kallsyms_lookup_size_offset() 306 return !!module_address_lookup(addr, symbolsize, offset, NULL, namebuf); in kallsyms_lookup_size_offset() 336 unsigned long *offset, in kallsyms_lookup() argument 345 pos = get_symbol_pos(addr, symbolsize, offset); in kallsyms_lookup() 355 if (!module_address_lookup(addr, symbolsize, offset, modname, in kallsyms_lookup() 391 unsigned long *offset, char *modname, char *name) in lookup_symbol_attrs() argument [all …]
|
D | relay.c | 367 buf->offset = 0; in __relay_reset() 746 if (buf->offset != buf->chan->subbuf_size + 1) { in relay_switch_subbuf() 747 buf->prev_padding = buf->chan->subbuf_size - buf->offset; in relay_switch_subbuf() 773 buf->offset = 0; in relay_switch_subbuf() 775 buf->offset = buf->chan->subbuf_size + 1; in relay_switch_subbuf() 781 if (unlikely(length + buf->offset > buf->chan->subbuf_size)) in relay_switch_subbuf() 966 buf->offset == buf->bytes_consumed) in relay_file_read_consume() 981 (buf->offset == subbuf_size)) in relay_file_read_consume() 1002 if (unlikely(buf->offset > subbuf_size)) { in relay_file_read_avail() 1014 produced = (produced % n_subbufs) * subbuf_size + buf->offset; in relay_file_read_avail() [all …]
|
D | configs.c | 59 size_t len, loff_t * offset) in ikconfig_read_current() argument 61 return simple_read_from_buffer(buf, len, offset, in ikconfig_read_current()
|
/kernel/events/ |
D | uprobes.c | 73 loff_t offset; member 131 static unsigned long offset_to_vaddr(struct vm_area_struct *vma, loff_t offset) in offset_to_vaddr() argument 133 return vma->vm_start + offset - ((loff_t)vma->vm_pgoff << PAGE_SHIFT); in offset_to_vaddr() 384 if (l->offset < r->offset) in match_uprobe() 387 if (l->offset > r->offset) in match_uprobe() 393 static struct uprobe *__find_uprobe(struct inode *inode, loff_t offset) in __find_uprobe() argument 395 struct uprobe u = { .inode = inode, .offset = offset }; in __find_uprobe() 418 static struct uprobe *find_uprobe(struct inode *inode, loff_t offset) in find_uprobe() argument 423 uprobe = __find_uprobe(inode, offset); in find_uprobe() 478 static struct uprobe *alloc_uprobe(struct inode *inode, loff_t offset) in alloc_uprobe() argument [all …]
|
D | ring_buffer.c | 122 unsigned long tail, offset, head; in __perf_output_begin() local 161 offset = head = local_read(&rb->head); in __perf_output_begin() 185 } while (local_cmpxchg(&rb->head, offset, head) != offset); in __perf_output_begin() 188 offset = head; in __perf_output_begin() 202 handle->page = (offset >> page_shift) & (rb->nr_pages - 1); in __perf_output_begin() 203 offset &= (1UL << page_shift) - 1; in __perf_output_begin() 204 handle->addr = rb->data_pages[handle->page] + offset; in __perf_output_begin() 205 handle->size = (1UL << page_shift) - offset; in __perf_output_begin()
|
/kernel/power/ |
D | user.c | 205 sector_t offset; in snapshot_ioctl() local 316 offset = alloc_swapdev_block(data->swap); in snapshot_ioctl() 317 if (offset) { in snapshot_ioctl() 318 offset <<= PAGE_SHIFT; in snapshot_ioctl() 319 error = put_user(offset, (loff_t __user *)arg); in snapshot_ioctl() 375 offset = swap_area.offset; in snapshot_ioctl() 376 data->swap = swap_type_of(swdev, offset, NULL); in snapshot_ioctl() 400 compat_loff_t offset; member 414 loff_t offset; in snapshot_compat_ioctl() local 420 err = snapshot_ioctl(file, cmd, (unsigned long) &offset); in snapshot_compat_ioctl() [all …]
|
D | swap.c | 178 unsigned long offset; in alloc_swapdev_block() local 180 offset = swp_offset(get_swap_page_of_type(swap)); in alloc_swapdev_block() 181 if (offset) { in alloc_swapdev_block() 182 if (swsusp_extents_insert(offset)) in alloc_swapdev_block() 183 swap_free(swp_entry(swap, offset)); in alloc_swapdev_block() 185 return swapdev_block(swap, offset); in alloc_swapdev_block() 202 unsigned long offset; in free_all_swap_pages() local 206 for (offset = ext->start; offset <= ext->end; offset++) in free_all_swap_pages() 207 swap_free(swp_entry(swap, offset)); in free_all_swap_pages() 369 static int write_page(void *buf, sector_t offset, struct hib_bio_batch *hb) in write_page() argument [all …]
|
/kernel/time/ |
D | ntp.c | 123 static inline s64 ntp_offset_chunk(s64 offset) in ntp_offset_chunk() argument 126 return offset; in ntp_offset_chunk() 128 return shift_right(offset, SHIFT_PLL + time_constant); in ntp_offset_chunk() 208 static inline s64 ntp_offset_chunk(s64 offset) in ntp_offset_chunk() argument 210 return shift_right(offset, SHIFT_PLL + time_constant); in ntp_offset_chunk() 294 static void ntp_update_offset(long offset) in ntp_update_offset() argument 305 offset = clamp(offset, -USEC_PER_SEC, USEC_PER_SEC); in ntp_update_offset() 306 offset *= NSEC_PER_USEC; in ntp_update_offset() 313 offset = clamp(offset, -MAXPHASE, MAXPHASE); in ntp_update_offset() 325 offset64 = offset; in ntp_update_offset() [all …]
|
D | timekeeping.c | 146 static void timekeeping_check_update(struct timekeeper *tk, cycle_t offset) in timekeeping_check_update() argument 152 if (offset > max_cycles) { in timekeeping_check_update() 154 offset, name, max_cycles); in timekeeping_check_update() 157 if (offset > (max_cycles >> 1)) { in timekeeping_check_update() 159 offset, name, max_cycles >> 1); in timekeeping_check_update() 226 static inline void timekeeping_check_update(struct timekeeper *tk, cycle_t offset) in timekeeping_check_update() argument 794 ktime_t base, *offset = offsets[offs]; in ktime_get_with_offset() local 801 base = ktime_add(tk->tkr_mono.base, *offset); in ktime_get_with_offset() 818 ktime_t *offset = offsets[offs]; in ktime_mono_to_any() local 824 tconv = ktime_add(tmono, *offset); in ktime_mono_to_any() [all …]
|
D | timer_list.c | 141 (unsigned long long) ktime_to_ns(base->offset)); in print_base() 144 print_active_timers(m, base, now + ktime_to_ns(base->offset)); in print_base() 337 static void *move_iter(struct timer_list_iter *iter, loff_t offset) in move_iter() argument 339 for (; offset; offset--) { in move_iter() 356 static void *timer_list_start(struct seq_file *file, loff_t *offset) in timer_list_start() argument 360 if (!*offset) in timer_list_start() 364 return move_iter(iter, *offset); in timer_list_start() 367 static void *timer_list_next(struct seq_file *file, void *v, loff_t *offset) in timer_list_next() argument 370 ++*offset; in timer_list_next()
|
D | timeconv.c | 77 void time64_to_tm(time64_t totalsecs, int offset, struct tm *result) in time64_to_tm() argument 85 rem += offset; in time64_to_tm()
|
/kernel/trace/ |
D | trace_probe.c | 90 void FETCH_FUNC_NAME(reg, type)(struct pt_regs *regs, void *offset, void *dest) \ 93 (unsigned int)((unsigned long)offset)); \ 116 long offset; member 129 addr += dprm->offset; \ 146 addr += dprm->offset; in DEFINE_FETCH_deref() 322 int traceprobe_split_symbol_offset(char *symbol, long *offset) in traceprobe_split_symbol_offset() argument 327 if (!offset) in traceprobe_split_symbol_offset() 332 ret = kstrtol(tmp, 0, offset); in traceprobe_split_symbol_offset() 337 *offset = 0; in traceprobe_split_symbol_offset() 392 long offset; in parse_probe_arg() local [all …]
|
D | trace_probe.h | 255 struct symbol_cache *alloc_symbol_cache(const char *sym, long offset); 277 alloc_symbol_cache(const char *sym, long offset) in alloc_symbol_cache() argument 286 unsigned int offset; /* Offset from argument entry */ member 357 extern int traceprobe_split_symbol_offset(char *symbol, long *offset); 396 dl = (u32 *)(data + tp->args[i].offset); in store_trace_args() 397 *dl = make_data_rloc(maxlen, end - tp->args[i].offset); in store_trace_args() 405 ent_size + tp->args[i].offset); in store_trace_args() 409 data + tp->args[i].offset); in store_trace_args()
|
D | trace_kprobe.c | 55 return tk->rp.kp.offset; in trace_kprobe_offset() 89 long offset; member 98 sc->addr += sc->offset; in update_symbol_cache() 109 struct symbol_cache *alloc_symbol_cache(const char *sym, long offset) in alloc_symbol_cache() argument 125 sc->offset = offset; in alloc_symbol_cache() 136 void *offset, void *dest) \ 139 (unsigned int)((unsigned long)offset)); \ 290 tk->rp.kp.offset = offs; in alloc_trace_kprobe() 610 long offset = 0; in create_trace_kprobe() local 677 ret = traceprobe_split_symbol_offset(symbol, &offset); in create_trace_kprobe() [all …]
|
D | trace_uprobe.c | 58 unsigned long offset; member 112 void *offset, void *dest) \ 115 ((unsigned long)offset)); \ 182 base_addr = udd->bp_addr - udd->tu->offset; in translate_user_vaddr() 188 void *offset, void *dest)\ 190 void *vaddr = (void *)translate_user_vaddr(offset); \ 366 unsigned long offset; in create_trace_uprobe() local 454 ret = kstrtoul(arg, 0, &offset); in create_trace_uprobe() 476 snprintf(buf, MAX_EVENT_NAME_LEN, "%c_%s_0x%lx", 'p', tail, offset); in create_trace_uprobe() 487 tu->offset = offset; in create_trace_uprobe() [all …]
|
D | trace_events_hist.c | 35 unsigned int offset; member 50 char *addr = (char *)(event + hist_field->field->offset); in hist_field_string() 57 u32 str_item = *(u32 *)(event + hist_field->field->offset); in hist_field_dynstring() 66 char **addr = (char **)(event + hist_field->field->offset); in hist_field_pstring() 73 u64 val = *(u64 *)(event + hist_field->field->offset); in hist_field_log2() 81 type *addr = (type *)(event + hist_field->field->offset); \ 569 hist_data->fields[key_idx]->offset = key_offset; in create_key_field() 757 hist_field->offset, in create_tracing_map_fields() 865 size = *(u32 *)(rec + field->offset) >> 16; in add_to_key() 876 memcpy(compound_key + key_field->offset, key, size); in add_to_key() [all …]
|
/kernel/locking/ |
D | qspinlock_paravirt.h | 201 #define for_each_hash_entry(he, offset, hash) \ argument 202 for (hash &= ~(PV_HE_PER_LINE - 1), he = &pv_lock_hash[hash], offset = 0; \ 203 offset < (1 << pv_lock_hash_bits); \ 204 offset++, he = &pv_lock_hash[(hash + offset) & ((1 << pv_lock_hash_bits) - 1)]) 208 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_hash() local 212 for_each_hash_entry(he, offset, hash) { in pv_hash() 235 unsigned long offset, hash = hash_ptr(lock, pv_lock_hash_bits); in pv_unhash() local 239 for_each_hash_entry(he, offset, hash) { in pv_unhash()
|
/kernel/sched/ |
D | stats.c | 109 static void *schedstat_start(struct seq_file *file, loff_t *offset) in schedstat_start() argument 111 unsigned long n = *offset; in schedstat_start() 123 *offset = n + 1; in schedstat_start() 130 static void *schedstat_next(struct seq_file *file, void *data, loff_t *offset) in schedstat_next() argument 132 (*offset)++; in schedstat_next() 133 return schedstat_start(file, offset); in schedstat_next()
|
D | loadavg.c | 74 void get_avenrun(unsigned long *loads, unsigned long offset, int shift) in get_avenrun() argument 76 loads[0] = (avenrun[0] + offset) << shift; in get_avenrun() 77 loads[1] = (avenrun[1] + offset) << shift; in get_avenrun() 78 loads[2] = (avenrun[2] + offset) << shift; in get_avenrun()
|
/kernel/rcu/ |
D | rcu.h | 108 unsigned long offset = (unsigned long)head->func; in __rcu_reclaim() local 111 if (__is_kfree_rcu_offset(offset)) { in __rcu_reclaim() 112 RCU_TRACE(trace_rcu_invoke_kfree_callback(rn, head, offset)); in __rcu_reclaim() 113 kfree((void *)head - offset); in __rcu_reclaim()
|
/kernel/irq/ |
D | ipi.c | 27 unsigned int nr_irqs, offset; in irq_reserve_ipi() local 55 offset = 0; in irq_reserve_ipi() 65 offset = cpumask_first(dest); in irq_reserve_ipi() 70 next = cpumask_next_zero(offset, dest); in irq_reserve_ipi() 96 data->common->ipi_offset = offset; in irq_reserve_ipi()
|
/kernel/bpf/ |
D | cgroup.c | 191 unsigned int offset = skb->data - skb_network_header(skb); in __cgroup_bpf_run_filter() local 195 __skb_push(skb, offset); in __cgroup_bpf_run_filter() 197 __skb_pull(skb, offset); in __cgroup_bpf_run_filter()
|
/kernel/gcov/ |
D | gcc_3_4.c | 321 unsigned int offset; member 409 iter->type_info[iter->num_types++].offset = 0; in gcov_iter_start() 454 get_type(iter)->offset += iter->count; in gcov_iter_next() 558 values[iter->count + get_type(iter)->offset]); in gcov_iter_write()
|