Home
last modified time | relevance | path

Searched refs:j (Results 1 – 25 of 48) sorted by relevance

12

/kernel/
Drange.c67 int i, j; in subtract_range() local
72 for (j = 0; j < az; j++) { in subtract_range()
73 if (!range[j].end) in subtract_range()
76 if (start <= range[j].start && end >= range[j].end) { in subtract_range()
77 range[j].start = 0; in subtract_range()
78 range[j].end = 0; in subtract_range()
82 if (start <= range[j].start && end < range[j].end && in subtract_range()
83 range[j].start < end) { in subtract_range()
84 range[j].start = end; in subtract_range()
89 if (start > range[j].start && end >= range[j].end && in subtract_range()
[all …]
Dprofile.c233 int i, j, cpu; in profile_flip_buffers() local
236 j = per_cpu(cpu_profile_flip, get_cpu()); in profile_flip_buffers()
240 struct profile_hit *hits = per_cpu(cpu_profile_hits, cpu)[j]; in profile_flip_buffers()
272 int i, j, cpu; in do_profile_hits() local
291 for (j = 0; j < PROFILE_GRPSZ; ++j) { in do_profile_hits()
292 if (hits[i + j].pc == pc) { in do_profile_hits()
293 hits[i + j].hits += nr_hits; in do_profile_hits()
295 } else if (!hits[i + j].hits) { in do_profile_hits()
296 hits[i + j].pc = pc; in do_profile_hits()
297 hits[i + j].hits = nr_hits; in do_profile_hits()
Dkexec_file.c678 int ret = 0, i, j, zero_buf_sz, sha_region_sz; in kexec_calculate_store_digests() local
723 for (j = i = 0; i < image->nr_segments; i++) { in kexec_calculate_store_digests()
758 sha_regions[j].start = ksegment->mem; in kexec_calculate_store_digests()
759 sha_regions[j].len = ksegment->memsz; in kexec_calculate_store_digests()
760 j++; in kexec_calculate_store_digests()
1154 int i, j; in crash_exclude_mem_range() local
1179 for (j = i; j < mem->nr_ranges - 1; j++) { in crash_exclude_mem_range()
1180 mem->ranges[j].start = in crash_exclude_mem_range()
1181 mem->ranges[j+1].start; in crash_exclude_mem_range()
1182 mem->ranges[j].end = in crash_exclude_mem_range()
[all …]
Dsmp.c254 unsigned int i, j, k; in cfd_seq_data_add() local
262 for (j = 0; j < *n_data; j++) { in cfd_seq_data_add()
263 if (new[i].u.cnt == data[j].u.cnt) { in cfd_seq_data_add()
266 data[j].val = new[i].val; in cfd_seq_data_add()
269 if (new[i].u.cnt < data[j].u.cnt) { in cfd_seq_data_add()
270 for (k = *n_data; k > j; k--) in cfd_seq_data_add()
272 data[j].val = new[i].val; in cfd_seq_data_add()
277 if (j == *n_data) { in cfd_seq_data_add()
278 data[j].val = new[i].val; in cfd_seq_data_add()
Daudit_tree.c296 int i, j; in replace_chunk() local
302 for (i = j = 0; j < old->count; i++, j++) { in replace_chunk()
303 if (!old->owners[j].owner) { in replace_chunk()
307 owner = old->owners[j].owner; in replace_chunk()
309 new->owners[i].index = old->owners[j].index - j + i; in replace_chunk()
313 list_replace_init(&old->owners[j].list, &new->owners[i].list); in replace_chunk()
/kernel/rcu/
Dtree_stall.h92 unsigned long j = jiffies; in rcu_gp_might_be_stalled() local
103 return !time_before(j, READ_ONCE(rcu_state.gp_start) + d); in rcu_gp_might_be_stalled()
172 unsigned long j = jiffies; in record_gp_stall_check_time() local
175 WRITE_ONCE(rcu_state.gp_start, j); in record_gp_stall_check_time()
179 WRITE_ONCE(rcu_state.jiffies_stall, j + j1); in record_gp_stall_check_time()
180 rcu_state.jiffies_resched = j + j1 / 2; in record_gp_stall_check_time()
198 unsigned long j; in rcu_stall_kick_kthreads() local
202 j = READ_ONCE(rcu_state.jiffies_kick_kthreads); in rcu_stall_kick_kthreads()
203 if (time_after(jiffies, j) && rcu_state.gp_kthread && in rcu_stall_kick_kthreads()
209 WRITE_ONCE(rcu_state.jiffies_kick_kthreads, j + HZ); in rcu_stall_kick_kthreads()
[all …]
Drcu_segcblist.c482 int i, j; in rcu_segcblist_advance() local
504 for (j = RCU_WAIT_TAIL; j < i; j++) in rcu_segcblist_advance()
505 WRITE_ONCE(rsclp->tails[j], rsclp->tails[RCU_DONE_TAIL]); in rcu_segcblist_advance()
513 for (j = RCU_WAIT_TAIL; i < RCU_NEXT_TAIL; i++, j++) { in rcu_segcblist_advance()
514 if (rsclp->tails[j] == rsclp->tails[RCU_NEXT_TAIL]) in rcu_segcblist_advance()
516 WRITE_ONCE(rsclp->tails[j], rsclp->tails[i]); in rcu_segcblist_advance()
517 rcu_segcblist_move_seglen(rsclp, i, j); in rcu_segcblist_advance()
518 rsclp->gp_seq[j] = rsclp->gp_seq[i]; in rcu_segcblist_advance()
539 int i, j; in rcu_segcblist_accelerate() local
583 for (j = i + 1; j <= RCU_NEXT_TAIL; j++) in rcu_segcblist_accelerate()
[all …]
Dupdate.c484 int j; in __wait_rcu_gp() local
493 for (j = 0; j < i; j++) in __wait_rcu_gp()
494 if (crcu_array[j] == crcu_array[i]) in __wait_rcu_gp()
496 if (j == i) { in __wait_rcu_gp()
508 for (j = 0; j < i; j++) in __wait_rcu_gp()
509 if (crcu_array[j] == crcu_array[i]) in __wait_rcu_gp()
511 if (j == i) { in __wait_rcu_gp()
Dtree.c447 unsigned long j; in adjust_jiffies_till_sched_qs() local
455 j = READ_ONCE(jiffies_till_first_fqs) + in adjust_jiffies_till_sched_qs()
457 if (j < HZ / 10 + nr_cpu_ids / RCU_JIFFIES_FQS_DIV) in adjust_jiffies_till_sched_qs()
458 j = HZ / 10 + nr_cpu_ids / RCU_JIFFIES_FQS_DIV; in adjust_jiffies_till_sched_qs()
459 pr_info("RCU calculated value of scheduler-enlistment delay is %ld jiffies.\n", j); in adjust_jiffies_till_sched_qs()
460 WRITE_ONCE(jiffies_to_sched_qs, j); in adjust_jiffies_till_sched_qs()
465 ulong j; in param_set_first_fqs_jiffies() local
466 int ret = kstrtoul(val, 0, &j); in param_set_first_fqs_jiffies()
469 WRITE_ONCE(*(ulong *)kp->arg, (j > HZ) ? HZ : j); in param_set_first_fqs_jiffies()
477 ulong j; in param_set_next_fqs_jiffies() local
[all …]
Dtree_nocb.h331 unsigned long j, bool lazy) in rcu_nocb_do_flush_bypass() argument
361 WRITE_ONCE(rdp->nocb_bypass_first, j); in rcu_nocb_do_flush_bypass()
375 unsigned long j, bool lazy) in rcu_nocb_flush_bypass() argument
381 return rcu_nocb_do_flush_bypass(rdp, rhp, j, lazy); in rcu_nocb_flush_bypass()
388 static void rcu_nocb_try_flush_bypass(struct rcu_data *rdp, unsigned long j) in rcu_nocb_try_flush_bypass() argument
394 WARN_ON_ONCE(!rcu_nocb_do_flush_bypass(rdp, NULL, j, false)); in rcu_nocb_try_flush_bypass()
421 unsigned long j = jiffies; in rcu_nocb_try_bypass() local
452 if (j == rdp->nocb_nobypass_last) { in rcu_nocb_try_bypass()
455 WRITE_ONCE(rdp->nocb_nobypass_last, j); in rcu_nocb_try_bypass()
476 WARN_ON_ONCE(!rcu_nocb_flush_bypass(rdp, NULL, j, false)); in rcu_nocb_try_bypass()
[all …]
Drcuscale.c721 int j; in rcu_scale_cleanup() local
762 j = writer_n_durations[i]; in rcu_scale_cleanup()
764 scale_type, SCALE_FLAG, i, j); in rcu_scale_cleanup()
765 ngps += j; in rcu_scale_cleanup()
783 for (j = 0; j < writer_n_durations[i]; j++) { in rcu_scale_cleanup()
784 wdp = &wdpp[j]; in rcu_scale_cleanup()
787 i, j, *wdp); in rcu_scale_cleanup()
788 if (j % 100 == 0) in rcu_scale_cleanup()
Dtasks.h215 int j = READ_ONCE(i); // Prevent the compiler from reading twice in tasks_gp_state_getname() local
217 if (j >= ARRAY_SIZE(rcu_tasks_gp_state_names)) in tasks_gp_state_getname()
219 return rcu_tasks_gp_state_names[j]; in tasks_gp_state_getname()
290 unsigned long j; in call_rcu_tasks_generic() local
304 j = jiffies; in call_rcu_tasks_generic()
305 if (rtpcp->rtp_jiffies != j) { in call_rcu_tasks_generic()
306 rtpcp->rtp_jiffies = j; in call_rcu_tasks_generic()
667 unsigned long j; in rcu_tasks_wait_gp() local
737 j = jiffies; in rcu_tasks_wait_gp()
738 if (rtsi > 0 && !reported && time_after(j, lastinfo + rtsi)) { in rcu_tasks_wait_gp()
[all …]
Drcu.h485 void rcu_lazy_set_jiffies_till_flush(unsigned long j);
488 static inline void rcu_lazy_set_jiffies_till_flush(unsigned long j) { } in rcu_lazy_set_jiffies_till_flush() argument
554 static inline void rcu_fwd_progress_check(unsigned long j) { } in rcu_fwd_progress_check() argument
565 void rcu_fwd_progress_check(unsigned long j);
Dsrcutree.c338 unsigned long j; in spin_lock_irqsave_check_contention() local
342 j = jiffies; in spin_lock_irqsave_check_contention()
343 if (ssp->srcu_size_jiffies != j) { in spin_lock_irqsave_check_contention()
344 ssp->srcu_size_jiffies = j; in spin_lock_irqsave_check_contention()
568 unsigned long j; in srcu_get_delay() local
574 j = jiffies - 1; in srcu_get_delay()
576 if (time_after(j, gpstart)) in srcu_get_delay()
577 jbase += j - gpstart; in srcu_get_delay()
1645 unsigned long j; in process_srcu() local
1655 j = jiffies; in process_srcu()
[all …]
/kernel/sched/
Dtopology.c1689 int i,j; in sched_numa_warn() local
1700 for (j = 0; j < nr_node_ids; j++) { in sched_numa_warn()
1701 if (!node_state(i, N_CPU) || !node_state(j, N_CPU)) in sched_numa_warn()
1702 printk(KERN_CONT "(%02d) ", node_distance(i,j)); in sched_numa_warn()
1704 printk(KERN_CONT " %02d ", node_distance(i,j)); in sched_numa_warn()
1804 int i, j; in sched_init_numa() local
1818 for_each_cpu_node_but(j, offline_node) { in sched_init_numa()
1819 int distance = node_distance(i, j); in sched_init_numa()
1842 for (i = 0, j = 0; i < nr_levels; i++, j++) { in sched_init_numa()
1843 j = find_next_bit(distance_map, NR_DISTANCE_VALUES, j); in sched_init_numa()
[all …]
/kernel/cgroup/
Dmisc.c143 struct misc_cg *i, *j; in misc_cg_try_charge() local
167 for (j = i; j; j = parent_misc(j)) { in misc_cg_try_charge()
168 atomic_long_inc(&j->res[type].events); in misc_cg_try_charge()
169 cgroup_file_notify(&j->events_file); in misc_cg_try_charge()
172 for (j = cg; j != i; j = parent_misc(j)) in misc_cg_try_charge()
173 misc_cg_cancel_charge(type, j, amount); in misc_cg_try_charge()
/kernel/irq/
Daffinity.c73 int j; in irq_create_affinity_masks() local
81 for (j = 0; j < this_vecs; j++) in irq_create_affinity_masks()
82 cpumask_copy(&masks[curvec + j].mask, &result[j]); in irq_create_affinity_masks()
Dproc.c465 int i = *(loff_t *) v, j; in show_interrupts() local
477 for (prec = 3, j = 1000; prec < 10 && j <= nr_irqs; ++prec) in show_interrupts()
478 j *= 10; in show_interrupts()
481 for_each_online_cpu(j) in show_interrupts()
482 seq_printf(p, "CPU%-8d", j); in show_interrupts()
492 for_each_online_cpu(j) in show_interrupts()
493 any_count |= data_race(*per_cpu_ptr(desc->kstat_irqs, j)); in show_interrupts()
500 for_each_online_cpu(j) in show_interrupts()
502 *per_cpu_ptr(desc->kstat_irqs, j) : 0); in show_interrupts()
/kernel/time/
Dtimer.c303 static unsigned long round_jiffies_common(unsigned long j, int cpu, in round_jiffies_common() argument
307 unsigned long original = j; in round_jiffies_common()
317 j += cpu * 3; in round_jiffies_common()
319 rem = j % HZ; in round_jiffies_common()
329 j = j - rem; in round_jiffies_common()
331 j = j - rem + HZ; in round_jiffies_common()
334 j -= cpu * 3; in round_jiffies_common()
340 return time_is_after_jiffies(j) ? j : original; in round_jiffies_common()
363 unsigned long __round_jiffies(unsigned long j, int cpu) in __round_jiffies() argument
365 return round_jiffies_common(j, cpu, false); in __round_jiffies()
[all …]
Dtime.c374 unsigned int jiffies_to_msecs(const unsigned long j) in jiffies_to_msecs() argument
377 return (MSEC_PER_SEC / HZ) * j; in jiffies_to_msecs()
379 return (j + (HZ / MSEC_PER_SEC) - 1)/(HZ / MSEC_PER_SEC); in jiffies_to_msecs()
382 return (HZ_TO_MSEC_MUL32 * j + (1ULL << HZ_TO_MSEC_SHR32) - 1) >> in jiffies_to_msecs()
385 return DIV_ROUND_UP(j * HZ_TO_MSEC_NUM, HZ_TO_MSEC_DEN); in jiffies_to_msecs()
391 unsigned int jiffies_to_usecs(const unsigned long j) in jiffies_to_usecs() argument
400 return (USEC_PER_SEC / HZ) * j; in jiffies_to_usecs()
403 return (HZ_TO_USEC_MUL32 * j) >> HZ_TO_USEC_SHR32; in jiffies_to_usecs()
405 return (j * HZ_TO_USEC_NUM) / HZ_TO_USEC_DEN; in jiffies_to_usecs()
691 u64 jiffies64_to_nsecs(u64 j) in jiffies64_to_nsecs() argument
[all …]
/kernel/trace/
Dtrace_probe_tmpl.h262 int i, j; in print_probe_args() local
275 for (j = 0; j < a->count; j++) { in print_probe_args()
278 trace_seq_putc(s, j == a->count - 1 ? '}' : ','); in print_probe_args()
Dpid_list.c461 int i, j; in trace_pid_list_free() local
487 for (j = 0; j < UPPER2_SIZE; j++) { in trace_pid_list_free()
488 lower = upper->data[j]; in trace_pid_list_free()
/kernel/bpf/
Dpercpu_freelist.c103 unsigned int cpu, cpu_idx, i, j, n, m; in pcpu_freelist_populate() local
111 j = n + (cpu_idx < m ? 1 : 0); in pcpu_freelist_populate()
112 for (i = 0; i < j; i++) { in pcpu_freelist_populate()
/kernel/power/
Dsnapshot.c2166 int j; in pack_pfns() local
2168 for (j = 0; j < PAGE_SIZE / sizeof(long); j++) { in pack_pfns()
2169 buf[j] = memory_bm_next_pfn(bm); in pack_pfns()
2170 if (unlikely(buf[j] == BM_END_OF_MAP)) in pack_pfns()
2172 if (memory_bm_test_bit(zero_bm, buf[j])) in pack_pfns()
2173 buf[j] |= ENCODED_PFN_ZERO_FLAG; in pack_pfns()
2322 int j; in unpack_orig_pfns() local
2324 for (j = 0; j < PAGE_SIZE / sizeof(long); j++) { in unpack_orig_pfns()
2325 if (unlikely(buf[j] == BM_END_OF_MAP)) in unpack_orig_pfns()
2328 zero = !!(buf[j] & ENCODED_PFN_ZERO_FLAG); in unpack_orig_pfns()
[all …]
/kernel/locking/
Dlocktorture.c902 int i, j; in lock_torture_init() local
1077 for (i = 0, j = 0; i < cxt.nrealwriters_stress || in lock_torture_init()
1078 j < cxt.nrealreaders_stress; i++, j++) { in lock_torture_init()
1089 if (cxt.cur_ops->readlock == NULL || (j >= cxt.nrealreaders_stress)) in lock_torture_init()
1092 firsterr = torture_create_kthread(lock_torture_reader, &cxt.lrsa[j], in lock_torture_init()
1093 reader_tasks[j]); in lock_torture_init()

12