Home
last modified time | relevance | path

Searched refs:i (Results 1 – 25 of 48) sorted by relevance

12

/block/
Dblk-cgroup-rwstat.c10 int i, ret; in blkg_rwstat_init() local
12 for (i = 0; i < BLKG_RWSTAT_NR; i++) { in blkg_rwstat_init()
13 ret = percpu_counter_init(&rwstat->cpu_cnt[i], 0, gfp); in blkg_rwstat_init()
15 while (--i >= 0) in blkg_rwstat_init()
16 percpu_counter_destroy(&rwstat->cpu_cnt[i]); in blkg_rwstat_init()
19 atomic64_set(&rwstat->aux_cnt[i], 0); in blkg_rwstat_init()
27 int i; in blkg_rwstat_exit() local
29 for (i = 0; i < BLKG_RWSTAT_NR; i++) in blkg_rwstat_exit()
30 percpu_counter_destroy(&rwstat->cpu_cnt[i]); in blkg_rwstat_exit()
54 int i; in __blkg_prfill_rwstat() local
[all …]
Dblk-cgroup-rwstat.h92 int i; in blkg_rwstat_read() local
94 for (i = 0; i < BLKG_RWSTAT_NR; i++) in blkg_rwstat_read()
95 result->cnt[i] = in blkg_rwstat_read()
96 percpu_counter_sum_positive(&rwstat->cpu_cnt[i]); in blkg_rwstat_read()
121 int i; in blkg_rwstat_reset() local
123 for (i = 0; i < BLKG_RWSTAT_NR; i++) { in blkg_rwstat_reset()
124 percpu_counter_set(&rwstat->cpu_cnt[i], 0); in blkg_rwstat_reset()
125 atomic64_set(&rwstat->aux_cnt[i], 0); in blkg_rwstat_reset()
140 int i; in blkg_rwstat_add_aux() local
142 for (i = 0; i < BLKG_RWSTAT_NR; i++) in blkg_rwstat_add_aux()
[all …]
Dblk-cgroup.c75 int i; in blkg_free() local
80 for (i = 0; i < BLKCG_MAX_POLS; i++) in blkg_free()
81 if (blkg->pd[i]) in blkg_free()
82 blkcg_policy[i]->pd_free_fn(blkg->pd[i]); in blkg_free()
155 int i, cpu; in blkg_alloc() local
180 for (i = 0; i < BLKCG_MAX_POLS; i++) { in blkg_alloc()
181 struct blkcg_policy *pol = blkcg_policy[i]; in blkg_alloc()
192 blkg->pd[i] = pd; in blkg_alloc()
194 pd->plid = i; in blkg_alloc()
237 int i, ret; in blkg_create() local
[all …]
Dkyber-iosched.c363 int i; in kyber_queue_data_alloc() local
379 for (i = 0; i < KYBER_NUM_DOMAINS; i++) { in kyber_queue_data_alloc()
380 WARN_ON(!kyber_depth[i]); in kyber_queue_data_alloc()
381 WARN_ON(!kyber_batch_size[i]); in kyber_queue_data_alloc()
382 ret = sbitmap_queue_init_node(&kqd->domain_tokens[i], in kyber_queue_data_alloc()
383 kyber_depth[i], -1, false, in kyber_queue_data_alloc()
386 while (--i >= 0) in kyber_queue_data_alloc()
387 sbitmap_queue_free(&kqd->domain_tokens[i]); in kyber_queue_data_alloc()
392 for (i = 0; i < KYBER_OTHER; i++) { in kyber_queue_data_alloc()
393 kqd->domain_p99[i] = -1; in kyber_queue_data_alloc()
[all …]
Dblk-crypto-fallback.c146 int i; in blk_crypto_fallback_encrypt_endio() local
148 for (i = 0; i < enc_bio->bi_vcnt; i++) in blk_crypto_fallback_encrypt_endio()
149 mempool_free(enc_bio->bi_io_vec[i].bv_page, in blk_crypto_fallback_encrypt_endio()
214 unsigned int i = 0; in blk_crypto_fallback_split_bio_if_needed() local
221 if (++i == BIO_MAX_VECS) in blk_crypto_fallback_split_bio_if_needed()
249 int i; in blk_crypto_dun_to_iv() local
251 for (i = 0; i < BLK_CRYPTO_DUN_ARRAY_SIZE; i++) in blk_crypto_dun_to_iv()
252 iv->dun[i] = cpu_to_le64(dun[i]); in blk_crypto_dun_to_iv()
273 unsigned int i, j; in blk_crypto_fallback_encrypt_bio() local
317 for (i = 0; i < enc_bio->bi_vcnt; i++) { in blk_crypto_fallback_encrypt_bio()
[all …]
Dblk-crypto.c69 size_t i; in bio_crypt_ctx_init() local
87 for (i = 0; i < BLK_ENCRYPTION_MODE_MAX; i++) { in bio_crypt_ctx_init()
88 BUG_ON(blk_crypto_modes[i].keysize > in bio_crypt_ctx_init()
90 BUG_ON(blk_crypto_modes[i].security_strength > in bio_crypt_ctx_init()
91 blk_crypto_modes[i].keysize); in bio_crypt_ctx_init()
92 BUG_ON(blk_crypto_modes[i].ivsize > BLK_CRYPTO_MAX_IV_SIZE); in bio_crypt_ctx_init()
141 int i; in bio_crypt_dun_increment() local
143 for (i = 0; inc && i < BLK_CRYPTO_DUN_ARRAY_SIZE; i++) { in bio_crypt_dun_increment()
144 dun[i] += inc; in bio_crypt_dun_increment()
149 if (dun[i] < inc) in bio_crypt_dun_increment()
[all …]
Dblk-mq-sysfs.c112 unsigned int i, first = 1; in blk_mq_hw_sysfs_cpus_show() local
115 for_each_cpu(i, hctx->cpumask) { in blk_mq_hw_sysfs_cpus_show()
117 ret = snprintf(pos + page, size - pos, "%u", i); in blk_mq_hw_sysfs_cpus_show()
119 ret = snprintf(pos + page, size - pos, ", %u", i); in blk_mq_hw_sysfs_cpus_show()
175 int i; in blk_mq_unregister_hctx() local
180 hctx_for_each_ctx(hctx, ctx, i) in blk_mq_unregister_hctx()
190 int i, j, ret; in blk_mq_register_hctx() local
199 hctx_for_each_ctx(hctx, ctx, i) { in blk_mq_register_hctx()
208 if (j < i) in blk_mq_register_hctx()
218 int i; in blk_mq_unregister_dev() local
[all …]
Dblk-mq.c234 unsigned int i; in blk_mq_quiesce_queue() local
239 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_quiesce_queue()
269 unsigned int i; in blk_mq_wake_waiters() local
271 queue_for_each_hw_ctx(q, hctx, i) in blk_mq_wake_waiters()
958 int i; in blk_mq_timeout_work() local
987 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_timeout_work()
1688 int i; in blk_mq_run_hw_queues() local
1693 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_run_hw_queues()
1716 int i; in blk_mq_delay_run_hw_queues() local
1721 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_delay_run_hw_queues()
[all …]
Dblk-mq-debugfs.c79 int i; in blk_flags_show() local
81 for (i = 0; i < sizeof(flags) * BITS_PER_BYTE; i++) { in blk_flags_show()
82 if (!(flags & BIT(i))) in blk_flags_show()
87 if (i < flag_name_count && flag_name[i]) in blk_flags_show()
88 seq_puts(m, flag_name[i]); in blk_flags_show()
90 seq_printf(m, "%d", i); in blk_flags_show()
187 int i; in queue_write_hint_show() local
189 for (i = 0; i < BLK_MAX_WRITE_HINTS; i++) in queue_write_hint_show()
190 seq_printf(m, "hint%d: %llu\n", i, q->write_hints[i]); in queue_write_hint_show()
199 int i; in queue_write_hint_store() local
[all …]
Dkeyslot-manager.c84 unsigned int i; in blk_ksm_init() local
125 for (i = 0; i < slot_hashtable_size; i++) in blk_ksm_init()
126 INIT_HLIST_HEAD(&ksm->slot_hashtable[i]); in blk_ksm_init()
470 unsigned int i; in blk_ksm_intersect_modes() local
475 for (i = 0; i < ARRAY_SIZE(child->crypto_modes_supported); in blk_ksm_intersect_modes()
476 i++) { in blk_ksm_intersect_modes()
477 parent->crypto_modes_supported[i] &= in blk_ksm_intersect_modes()
478 child->crypto_modes_supported[i]; in blk_ksm_intersect_modes()
503 int i; in blk_ksm_is_superset() local
511 for (i = 0; i < ARRAY_SIZE(ksm_superset->crypto_modes_supported); i++) { in blk_ksm_is_superset()
[all …]
Dblk-map.c135 int i = 0, ret; in bio_copy_user_iov() local
162 i = map_data->offset / PAGE_SIZE; in bio_copy_user_iov()
173 if (i == map_data->nr_entries * nr_pages) { in bio_copy_user_iov()
178 page = map_data->pages[i / nr_pages]; in bio_copy_user_iov()
179 page += (i % nr_pages); in bio_copy_user_iov()
181 i++; in bio_copy_user_iov()
314 unsigned long i, len = 0; in bio_invalidate_vmalloc_pages() local
316 for (i = 0; i < bio->bi_vcnt; i++) in bio_invalidate_vmalloc_pages()
317 len += bio->bi_io_vec[i].bv_len; in bio_invalidate_vmalloc_pages()
348 int offset, i; in bio_map_kern() local
[all …]
Dblk-crypto-profile.c77 unsigned int i; in blk_crypto_profile_init() local
131 for (i = 0; i < slot_hashtable_size; i++) in blk_crypto_profile_init()
132 INIT_HLIST_HEAD(&profile->slot_hashtable[i]); in blk_crypto_profile_init()
533 unsigned int i; in blk_crypto_intersect_capabilities() local
538 for (i = 0; i < ARRAY_SIZE(child->modes_supported); i++) in blk_crypto_intersect_capabilities()
539 parent->modes_supported[i] &= child->modes_supported[i]; in blk_crypto_intersect_capabilities()
561 int i; in blk_crypto_has_capabilities() local
569 for (i = 0; i < ARRAY_SIZE(target->modes_supported); i++) { in blk_crypto_has_capabilities()
570 if (reference->modes_supported[i] & ~target->modes_supported[i]) in blk_crypto_has_capabilities()
Dblk-mq-debugfs-zoned.c12 unsigned int i; in queue_zone_wlock_show() local
17 for (i = 0; i < q->nr_zones; i++) in queue_zone_wlock_show()
18 if (test_bit(i, q->seq_zones_wlock)) in queue_zone_wlock_show()
19 seq_printf(m, "%u\n", i); in queue_zone_wlock_show()
Dbadblocks.c131 int i; in badblocks_update_acked() local
137 for (i = 0; i < bb->count ; i++) { in badblocks_update_acked()
138 if (!BB_ACK(p[i])) { in badblocks_update_acked()
442 int i; in ack_all_badblocks() local
444 for (i = 0; i < bb->count ; i++) { in ack_all_badblocks()
445 if (!BB_ACK(p[i])) { in ack_all_badblocks()
446 sector_t start = BB_OFFSET(p[i]); in ack_all_badblocks()
447 int len = BB_LEN(p[i]); in ack_all_badblocks()
449 p[i] = BB_MAKE(start, len, 1); in ack_all_badblocks()
470 int i; in badblocks_show() local
[all …]
Ddisk-events.c174 int nr_events = 0, i; in disk_event_uevent() local
176 for (i = 0; i < ARRAY_SIZE(disk_uevents); i++) in disk_event_uevent()
177 if (events & disk->events & (1 << i)) in disk_event_uevent()
178 envp[nr_events++] = disk_uevents[i]; in disk_event_uevent()
344 int i; in __disk_events_show() local
346 for (i = 0; i < ARRAY_SIZE(disk_events_strs); i++) in __disk_events_show()
347 if (events & (1 << i)) { in __disk_events_show()
349 delim, disk_events_strs[i]); in __disk_events_show()
Dblk-crypto-sysfs.c159 int i; in blk_crypto_sysfs_init() local
162 for (i = 1; i < BLK_ENCRYPTION_MODE_MAX; i++) { in blk_crypto_sysfs_init()
163 struct blk_crypto_attr *attr = &__blk_crypto_mode_attrs[i]; in blk_crypto_sysfs_init()
165 attr->attr.name = blk_crypto_modes[i].name; in blk_crypto_sysfs_init()
168 blk_crypto_mode_attrs[i - 1] = &attr->attr; in blk_crypto_sysfs_init()
Dblk-mq-sched.c545 int i; in blk_mq_sched_tags_teardown() local
547 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_sched_tags_teardown()
560 int ret, i; in blk_mq_init_sched_shared_sbitmap() local
573 queue_for_each_hw_ctx(queue, hctx, i) { in blk_mq_init_sched_shared_sbitmap()
596 unsigned int i; in blk_mq_init_sched() local
613 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_init_sched()
614 ret = blk_mq_sched_alloc_tags(q, hctx, i); in blk_mq_init_sched()
631 queue_for_each_hw_ctx(q, hctx, i) { in blk_mq_init_sched()
633 ret = e->ops.init_hctx(hctx, i); in blk_mq_init_sched()
664 int i; in blk_mq_sched_free_requests() local
[all …]
Dblk-wbt.h80 unsigned int i, ret = 0; in wbt_inflight() local
82 for (i = 0; i < WBT_NUM_RWQ; i++) in wbt_inflight()
83 ret += atomic_read(&rwb->rq_wait[i].inflight); in wbt_inflight()
/block/partitions/
Defi.c175 int i, part = 0, ret = 0; /* invalid by default */ in is_pmbr_valid() local
180 for (i = 0; i < 4; i++) { in is_pmbr_valid()
181 ret = pmbr_part_valid(&mbr->partition_record[i]); in is_pmbr_valid()
183 part = i; in is_pmbr_valid()
196 for (i = 0; i < 4; i++) in is_pmbr_valid()
197 if ((mbr->partition_record[i].os_type != in is_pmbr_valid()
199 (mbr->partition_record[i].os_type != 0x00)) in is_pmbr_valid()
680 unsigned int i = 0; in utf16_le_to_7bit() local
684 while (i < size) { in utf16_le_to_7bit()
685 u8 c = le16_to_cpu(in[i]) & 0xff; in utf16_le_to_7bit()
[all …]
Dacorn.c330 int i; in valid_ics_sector() local
332 for (i = 0, sum = 0x50617274; i < 508; i++) in valid_ics_sector()
333 sum += data[i]; in valid_ics_sector()
421 int i; in valid_ptec_sector() local
430 for (i = 0; i < 511; i++) in valid_ptec_sector()
431 checksum += data[i]; in valid_ptec_sector()
452 int i; in adfspart_check_POWERTEC() local
465 for (i = 0, p = (const struct ptec_part *)data; i < 12; i++, p++) { in adfspart_check_POWERTEC()
514 int i, slot = 1; in adfspart_check_EESOX() local
523 for (i = 0; i < 256; i++) in adfspart_check_EESOX()
[all …]
Daix.c207 int i; in aix_partition() local
213 for (i = 0; foundlvs < numlvs && i < state->limit; i += 1) { in aix_partition()
214 lvip[i].pps_per_lv = be16_to_cpu(p[i].num_lps); in aix_partition()
215 if (lvip[i].pps_per_lv) in aix_partition()
226 int i; in aix_partition() local
231 for (i = 0; i < numpps; i += 1) { in aix_partition()
232 struct ppe *p = pvd->ppe + i; in aix_partition()
257 (i + 1 - lp_ix) * pp_blocks_size + psn_part1, in aix_partition()
268 for (i = 0; i < state->limit; i += 1) in aix_partition()
269 if (lvip[i].pps_found && !lvip[i].lv_is_contiguous) { in aix_partition()
[all …]
Dultrix.c14 int i; in ultrix_partition() local
36 for (i=0; i<8; i++) in ultrix_partition()
37 if (label->pt_part[i].pi_nblocks) in ultrix_partition()
38 put_partition(state, i+1, in ultrix_partition()
39 label->pt_part[i].pi_blkoff, in ultrix_partition()
40 label->pt_part[i].pi_nblocks); in ultrix_partition()
Dmac.c25 int i; in mac_fix_string() local
27 for (i = len - 1; i >= 0 && stg[i] == ' '; i--) in mac_fix_string()
28 stg[i] = 0; in mac_fix_string()
110 int i, l; in mac_partition() local
116 for (i = 0; i <= l - 4; ++i) { in mac_partition()
117 if (strncasecmp(part->name + i, "root", in mac_partition()
Dsysv68.c51 int i, slices; in sysv68_partition() local
69 i = be32_to_cpu(b->dk_ios.ios_slcblk); in sysv68_partition()
72 data = read_part_sector(state, i, &sect); in sysv68_partition()
80 for (i = 0; i < slices; i++, slice++) { in sysv68_partition()
87 snprintf(tmp, sizeof(tmp), "(s%u)", i); in sysv68_partition()
Dmsdos.c141 int i; in parse_extended() local
173 for (i = 0; i < 4; i++, p++) { in parse_extended()
184 if (i >= 2) { in parse_extended()
209 for (i = 0; i < 4; i++, p++) in parse_extended()
212 if (i == 4) in parse_extended()
256 int i; in parse_solaris_x86() local
283 for (i = 0; i < max_nparts && state->next < state->limit; i++) { in parse_solaris_x86()
284 struct solaris_x86_slice *s = &v->v_slice[i]; in parse_solaris_x86()
289 snprintf(tmp, sizeof(tmp), " [s%d]", i); in parse_solaris_x86()
536 int i; in parse_minix() local
[all …]

12