Home
last modified time | relevance | path

Searched refs:j (Results 1 – 25 of 35) sorted by relevance

12

/kernel/
Drange.c66 int i, j; in subtract_range() local
71 for (j = 0; j < az; j++) { in subtract_range()
72 if (!range[j].end) in subtract_range()
75 if (start <= range[j].start && end >= range[j].end) { in subtract_range()
76 range[j].start = 0; in subtract_range()
77 range[j].end = 0; in subtract_range()
81 if (start <= range[j].start && end < range[j].end && in subtract_range()
82 range[j].start < end) { in subtract_range()
83 range[j].start = end; in subtract_range()
88 if (start > range[j].start && end >= range[j].end && in subtract_range()
[all …]
Dprofile.c249 int i, j, cpu; in profile_flip_buffers() local
252 j = per_cpu(cpu_profile_flip, get_cpu()); in profile_flip_buffers()
256 struct profile_hit *hits = per_cpu(cpu_profile_hits, cpu)[j]; in profile_flip_buffers()
288 int i, j, cpu; in do_profile_hits() local
307 for (j = 0; j < PROFILE_GRPSZ; ++j) { in do_profile_hits()
308 if (hits[i + j].pc == pc) { in do_profile_hits()
309 hits[i + j].hits += nr_hits; in do_profile_hits()
311 } else if (!hits[i + j].hits) { in do_profile_hits()
312 hits[i + j].pc = pc; in do_profile_hits()
313 hits[i + j].hits = nr_hits; in do_profile_hits()
Dkexec_file.c710 int ret = 0, i, j, zero_buf_sz, sha_region_sz; in kexec_calculate_store_digests() local
753 for (j = i = 0; i < image->nr_segments; i++) { in kexec_calculate_store_digests()
788 sha_regions[j].start = ksegment->mem; in kexec_calculate_store_digests()
789 sha_regions[j].len = ksegment->memsz; in kexec_calculate_store_digests()
790 j++; in kexec_calculate_store_digests()
1172 int i, j; in crash_exclude_mem_range() local
1195 for (j = i; j < mem->nr_ranges - 1; j++) { in crash_exclude_mem_range()
1196 mem->ranges[j].start = in crash_exclude_mem_range()
1197 mem->ranges[j+1].start; in crash_exclude_mem_range()
1198 mem->ranges[j].end = in crash_exclude_mem_range()
[all …]
Daudit_tree.c298 int i, j; in replace_chunk() local
304 for (i = j = 0; j < old->count; i++, j++) { in replace_chunk()
305 if (!old->owners[j].owner) { in replace_chunk()
309 owner = old->owners[j].owner; in replace_chunk()
311 new->owners[i].index = old->owners[j].index - j + i; in replace_chunk()
315 list_replace_init(&old->owners[j].list, &new->owners[i].list); in replace_chunk()
Dkexec_core.c188 unsigned long j; in sanity_check_segment_list() local
192 for (j = 0; j < i; j++) { in sanity_check_segment_list()
195 pstart = image->segment[j].mem; in sanity_check_segment_list()
196 pend = pstart + image->segment[j].memsz; in sanity_check_segment_list()
Dauditfilter.c290 int j; in audit_to_entry_common() local
291 for (j = 0; j < AUDIT_BITMASK_SIZE; j++) in audit_to_entry_common()
292 entry->rule.mask[j] |= class[j]; in audit_to_entry_common()
/kernel/rcu/
Dtree_stall.h102 unsigned long j = jiffies; in record_gp_stall_check_time() local
105 rcu_state.gp_start = j; in record_gp_stall_check_time()
108 smp_store_release(&rcu_state.jiffies_stall, j + j1); /* ^^^ */ in record_gp_stall_check_time()
109 rcu_state.jiffies_resched = j + j1 / 2; in record_gp_stall_check_time()
127 unsigned long j; in rcu_stall_kick_kthreads() local
131 j = READ_ONCE(rcu_state.jiffies_kick_kthreads); in rcu_stall_kick_kthreads()
132 if (time_after(jiffies, j) && rcu_state.gp_kthread && in rcu_stall_kick_kthreads()
138 WRITE_ONCE(rcu_state.jiffies_kick_kthreads, j + HZ); in rcu_stall_kick_kthreads()
336 unsigned long j; in rcu_check_gp_kthread_starvation() local
338 j = jiffies - READ_ONCE(rcu_state.gp_activity); in rcu_check_gp_kthread_starvation()
[all …]
Dupdate.c347 int j; in __wait_rcu_gp() local
358 for (j = 0; j < i; j++) in __wait_rcu_gp()
359 if (crcu_array[j] == crcu_array[i]) in __wait_rcu_gp()
361 if (j == i) in __wait_rcu_gp()
370 for (j = 0; j < i; j++) in __wait_rcu_gp()
371 if (crcu_array[j] == crcu_array[i]) in __wait_rcu_gp()
373 if (j == i) in __wait_rcu_gp()
Drcu_segcblist.c413 int i, j; in rcu_segcblist_advance() local
434 for (j = RCU_WAIT_TAIL; j < i; j++) in rcu_segcblist_advance()
435 WRITE_ONCE(rsclp->tails[j], rsclp->tails[RCU_DONE_TAIL]); in rcu_segcblist_advance()
443 for (j = RCU_WAIT_TAIL; i < RCU_NEXT_TAIL; i++, j++) { in rcu_segcblist_advance()
444 if (rsclp->tails[j] == rsclp->tails[RCU_NEXT_TAIL]) in rcu_segcblist_advance()
446 WRITE_ONCE(rsclp->tails[j], rsclp->tails[i]); in rcu_segcblist_advance()
447 rsclp->gp_seq[j] = rsclp->gp_seq[i]; in rcu_segcblist_advance()
Drcuperf.c487 int j; in rcu_perf_cleanup() local
523 j = writer_n_durations[i]; in rcu_perf_cleanup()
525 perf_type, PERF_FLAG, i, j); in rcu_perf_cleanup()
526 ngps += j; in rcu_perf_cleanup()
544 for (j = 0; j <= writer_n_durations[i]; j++) { in rcu_perf_cleanup()
545 wdp = &wdpp[j]; in rcu_perf_cleanup()
548 i, j, *wdp); in rcu_perf_cleanup()
549 if (j % 100 == 0) in rcu_perf_cleanup()
Dtree.c444 unsigned long j; in adjust_jiffies_till_sched_qs() local
452 j = READ_ONCE(jiffies_till_first_fqs) + in adjust_jiffies_till_sched_qs()
454 if (j < HZ / 10 + nr_cpu_ids / RCU_JIFFIES_FQS_DIV) in adjust_jiffies_till_sched_qs()
455 j = HZ / 10 + nr_cpu_ids / RCU_JIFFIES_FQS_DIV; in adjust_jiffies_till_sched_qs()
456 pr_info("RCU calculated value of scheduler-enlistment delay is %ld jiffies.\n", j); in adjust_jiffies_till_sched_qs()
457 WRITE_ONCE(jiffies_to_sched_qs, j); in adjust_jiffies_till_sched_qs()
462 ulong j; in param_set_first_fqs_jiffies() local
463 int ret = kstrtoul(val, 0, &j); in param_set_first_fqs_jiffies()
466 WRITE_ONCE(*(ulong *)kp->arg, (j > HZ) ? HZ : j); in param_set_first_fqs_jiffies()
474 ulong j; in param_set_next_fqs_jiffies() local
[all …]
Dtree_plugin.h1691 unsigned long j) in rcu_nocb_do_flush_bypass() argument
1707 WRITE_ONCE(rdp->nocb_bypass_first, j); in rcu_nocb_do_flush_bypass()
1721 unsigned long j) in rcu_nocb_flush_bypass() argument
1727 return rcu_nocb_do_flush_bypass(rdp, rhp, j); in rcu_nocb_flush_bypass()
1734 static void rcu_nocb_try_flush_bypass(struct rcu_data *rdp, unsigned long j) in rcu_nocb_try_flush_bypass() argument
1740 WARN_ON_ONCE(!rcu_nocb_do_flush_bypass(rdp, NULL, j)); in rcu_nocb_try_flush_bypass()
1766 unsigned long j = jiffies; in rcu_nocb_try_bypass() local
1785 if (j == rdp->nocb_nobypass_last) { in rcu_nocb_try_bypass()
1788 WRITE_ONCE(rdp->nocb_nobypass_last, j); in rcu_nocb_try_bypass()
1807 WARN_ON_ONCE(!rcu_nocb_flush_bypass(rdp, NULL, j)); in rcu_nocb_try_bypass()
[all …]
Drcutorture.c1239 int j; in rcutorture_loop_extend() local
1248 for (j = 0; j < i; j++) { in rcutorture_loop_extend()
1250 rcutorture_one_extend(readstate, mask, trsp, &rtrsp[j]); in rcutorture_loop_extend()
1252 return &rtrsp[j]; in rcutorture_loop_extend()
1684 int j; in rcu_torture_fwd_cb_hist() local
1692 for (j = 0; j <= i; j++) { in rcu_torture_fwd_cb_hist()
1693 gps = n_launders_hist[j].launder_gp_seq; in rcu_torture_fwd_cb_hist()
1695 j + 1, FWD_CBS_HIST_DIV, n_launders_hist[j].n_launders, in rcu_torture_fwd_cb_hist()
/kernel/sched/
Dtopology.c1473 int i,j; in sched_numa_warn() local
1484 for (j = 0; j < nr_node_ids; j++) in sched_numa_warn()
1485 printk(KERN_CONT "%02d ", node_distance(i,j)); in sched_numa_warn()
1563 int i, j, k; in sched_init_numa() local
1582 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa()
1648 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa()
1653 sched_domains_numa_masks[i][j] = mask; in sched_init_numa()
1656 if (node_distance(j, k) > sched_domains_numa_distance[i]) in sched_init_numa()
1690 for (j = 1; j < level; i++, j++) { in sched_init_numa()
1695 .numa_level = j, in sched_init_numa()
[all …]
/kernel/time/
Dtimer.c273 static unsigned long round_jiffies_common(unsigned long j, int cpu, in round_jiffies_common() argument
277 unsigned long original = j; in round_jiffies_common()
287 j += cpu * 3; in round_jiffies_common()
289 rem = j % HZ; in round_jiffies_common()
299 j = j - rem; in round_jiffies_common()
301 j = j - rem + HZ; in round_jiffies_common()
304 j -= cpu * 3; in round_jiffies_common()
310 return time_is_after_jiffies(j) ? j : original; in round_jiffies_common()
333 unsigned long __round_jiffies(unsigned long j, int cpu) in __round_jiffies() argument
335 return round_jiffies_common(j, cpu, false); in __round_jiffies()
[all …]
Dtime.c376 unsigned int jiffies_to_msecs(const unsigned long j) in jiffies_to_msecs() argument
379 return (MSEC_PER_SEC / HZ) * j; in jiffies_to_msecs()
381 return (j + (HZ / MSEC_PER_SEC) - 1)/(HZ / MSEC_PER_SEC); in jiffies_to_msecs()
384 return (HZ_TO_MSEC_MUL32 * j + (1ULL << HZ_TO_MSEC_SHR32) - 1) >> in jiffies_to_msecs()
387 return DIV_ROUND_UP(j * HZ_TO_MSEC_NUM, HZ_TO_MSEC_DEN); in jiffies_to_msecs()
393 unsigned int jiffies_to_usecs(const unsigned long j) in jiffies_to_usecs() argument
402 return (USEC_PER_SEC / HZ) * j; in jiffies_to_usecs()
405 return (HZ_TO_USEC_MUL32 * j) >> HZ_TO_USEC_SHR32; in jiffies_to_usecs()
407 return (j * HZ_TO_USEC_NUM) / HZ_TO_USEC_DEN; in jiffies_to_usecs()
780 u64 jiffies64_to_nsecs(u64 j) in jiffies64_to_nsecs() argument
[all …]
/kernel/irq/
Dproc.c443 int i = *(loff_t *) v, j; in show_interrupts() local
455 for (prec = 3, j = 1000; prec < 10 && j <= nr_irqs; ++prec) in show_interrupts()
456 j *= 10; in show_interrupts()
459 for_each_online_cpu(j) in show_interrupts()
460 seq_printf(p, "CPU%-8d", j); in show_interrupts()
470 for_each_online_cpu(j) in show_interrupts()
471 any_count |= *per_cpu_ptr(desc->kstat_irqs, j); in show_interrupts()
477 for_each_online_cpu(j) in show_interrupts()
479 *per_cpu_ptr(desc->kstat_irqs, j) : 0); in show_interrupts()
/kernel/gcov/
Dgcc_3_4.c217 unsigned int j; in gcov_info_add() local
220 for (j = 0; j < dest->counts[i].num; j++) { in gcov_info_add()
221 dest->counts[i].values[j] += in gcov_info_add()
222 source->counts[i].values[j]; in gcov_info_add()
/kernel/trace/
Dtrace_probe_tmpl.h221 int i, j; in print_probe_args() local
234 for (j = 0; j < a->count; j++) { in print_probe_args()
237 trace_seq_putc(s, j == a->count - 1 ? '}' : ','); in print_probe_args()
Dtrace_probe.c845 int i, j; in __set_print_fmt() local
865 for (j = 1; j < parg->count; j++) in __set_print_fmt()
883 for (j = 0; j < parg->count; j++) in __set_print_fmt()
885 fmt, parg->name, j); in __set_print_fmt()
Dtrace_events_hist.c3166 unsigned int n, i, j; in compatible_keys() local
3172 j = target_hist_data->n_vals; in compatible_keys()
3176 target_hist_field = target_hist_data->fields[j + n]; in compatible_keys()
3426 unsigned int i, j, var_idx; in __update_field_vars() local
3429 for (i = 0, j = field_var_str_start; i < n_field_vars; i++) { in __update_field_vars()
3438 char *str = elt_data->field_var_str[j++]; in __update_field_vars()
4528 unsigned int i, j = 1; in create_val_fields() local
4543 for (i = 0, j = 1; i < TRACING_MAP_VALS_MAX && in create_val_fields()
4544 j < TRACING_MAP_VALS_MAX; i++) { in create_val_fields()
4552 ret = create_val_field(hist_data, j++, file, field_str); in create_val_fields()
[all …]
/kernel/power/
Dsnapshot.c2068 int j; in pack_pfns() local
2070 for (j = 0; j < PAGE_SIZE / sizeof(long); j++) { in pack_pfns()
2071 buf[j] = memory_bm_next_pfn(bm); in pack_pfns()
2072 if (unlikely(buf[j] == BM_END_OF_MAP)) in pack_pfns()
2075 page_key_read(buf + j); in pack_pfns()
2219 int j; in unpack_orig_pfns() local
2221 for (j = 0; j < PAGE_SIZE / sizeof(long); j++) { in unpack_orig_pfns()
2222 if (unlikely(buf[j] == BM_END_OF_MAP)) in unpack_orig_pfns()
2226 page_key_memorize(buf + j); in unpack_orig_pfns()
2228 if (pfn_valid(buf[j]) && memory_bm_pfn_present(bm, buf[j])) in unpack_orig_pfns()
[all …]
/kernel/bpf/
Dverifier.c504 int j; in print_verifier_state() local
506 for (j = 0; j < BPF_REG_SIZE; j++) { in print_verifier_state()
507 if (state->stack[i].slot_type[j] != STACK_INVALID) in print_verifier_state()
509 types_buf[j] = slot_type_char[ in print_verifier_state()
510 state->stack[i].slot_type[j]]; in print_verifier_state()
1647 int i, j; in mark_all_scalars_precise() local
1655 for (j = 0; j < BPF_REG_FP; j++) { in mark_all_scalars_precise()
1656 reg = &func->regs[j]; in mark_all_scalars_precise()
1661 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_precise()
1662 if (func->stack[j].slot_type[0] != STACK_SPILL) in mark_all_scalars_precise()
[all …]
/kernel/locking/
Dlocktorture.c842 int i, j; in lock_torture_init() local
1015 for (i = 0, j = 0; i < cxt.nrealwriters_stress || in lock_torture_init()
1016 j < cxt.nrealreaders_stress; i++, j++) { in lock_torture_init()
1027 if (cxt.cur_ops->readlock == NULL || (j >= cxt.nrealreaders_stress)) in lock_torture_init()
1030 firsterr = torture_create_kthread(lock_torture_reader, &cxt.lrsa[j], in lock_torture_init()
1031 reader_tasks[j]); in lock_torture_init()
Dlockdep.c2737 int i, j, id; in check_no_collision() local
2746 for (j = 0; j < chain->depth - 1; j++, i++) { in check_no_collision()
2749 if (DEBUG_LOCKS_WARN_ON(chain_hlocks[chain->base + j] != id)) { in check_no_collision()
2799 int i, j; in add_chain_cache() local
2829 for (j = 0; j < chain->depth - 1; j++, i++) { in add_chain_cache()
2831 chain_hlocks[chain->base + j] = lock_id; in add_chain_cache()
2833 chain_hlocks[chain->base + j] = class - lock_classes; in add_chain_cache()
5077 int i, j; in lock_class_cache_is_registered() local
5082 for (j = 0; j < NR_LOCKDEP_CACHING_CLASSES; j++) in lock_class_cache_is_registered()
5083 if (lock->class_cache[j] == class) in lock_class_cache_is_registered()
[all …]

12