/kernel/linux/linux-5.10/drivers/lightnvm/ |
D | pblk-rl.c | 22 static void pblk_rl_kick_u_timer(struct pblk_rl *rl) in pblk_rl_kick_u_timer() argument 24 mod_timer(&rl->u_timer, jiffies + msecs_to_jiffies(5000)); in pblk_rl_kick_u_timer() 27 int pblk_rl_is_limit(struct pblk_rl *rl) in pblk_rl_is_limit() argument 31 rb_space = atomic_read(&rl->rb_space); in pblk_rl_is_limit() 36 int pblk_rl_user_may_insert(struct pblk_rl *rl, int nr_entries) in pblk_rl_user_may_insert() argument 38 int rb_user_cnt = atomic_read(&rl->rb_user_cnt); in pblk_rl_user_may_insert() 39 int rb_space = atomic_read(&rl->rb_space); in pblk_rl_user_may_insert() 44 if (rb_user_cnt >= rl->rb_user_max) in pblk_rl_user_may_insert() 50 void pblk_rl_inserted(struct pblk_rl *rl, int nr_entries) in pblk_rl_inserted() argument 52 int rb_space = atomic_read(&rl->rb_space); in pblk_rl_inserted() [all …]
|
D | pblk.h | 623 struct pblk_rl rl; member 902 void pblk_rl_init(struct pblk_rl *rl, int budget, int threshold); 903 void pblk_rl_free(struct pblk_rl *rl); 904 void pblk_rl_update_rates(struct pblk_rl *rl); 905 int pblk_rl_high_thrs(struct pblk_rl *rl); 906 unsigned long pblk_rl_nr_free_blks(struct pblk_rl *rl); 907 unsigned long pblk_rl_nr_user_free_blks(struct pblk_rl *rl); 908 int pblk_rl_user_may_insert(struct pblk_rl *rl, int nr_entries); 909 void pblk_rl_inserted(struct pblk_rl *rl, int nr_entries); 910 void pblk_rl_user_in(struct pblk_rl *rl, int nr_entries); [all …]
|
/kernel/linux/linux-5.10/fs/ntfs/ |
D | runlist.c | 60 static inline runlist_element *ntfs_rl_realloc(runlist_element *rl, in ntfs_rl_realloc() argument 65 old_size = PAGE_ALIGN(old_size * sizeof(*rl)); in ntfs_rl_realloc() 66 new_size = PAGE_ALIGN(new_size * sizeof(*rl)); in ntfs_rl_realloc() 68 return rl; in ntfs_rl_realloc() 74 if (likely(rl != NULL)) { in ntfs_rl_realloc() 77 memcpy(new_rl, rl, old_size); in ntfs_rl_realloc() 78 ntfs_free(rl); in ntfs_rl_realloc() 106 static inline runlist_element *ntfs_rl_realloc_nofail(runlist_element *rl, in ntfs_rl_realloc_nofail() argument 111 old_size = PAGE_ALIGN(old_size * sizeof(*rl)); in ntfs_rl_realloc_nofail() 112 new_size = PAGE_ALIGN(new_size * sizeof(*rl)); in ntfs_rl_realloc_nofail() [all …]
|
D | lcnalloc.c | 37 const runlist_element *rl) in ntfs_cluster_free_from_rl_nolock() argument 43 if (!rl) in ntfs_cluster_free_from_rl_nolock() 45 for (; rl->length; rl++) { in ntfs_cluster_free_from_rl_nolock() 48 if (rl->lcn < 0) in ntfs_cluster_free_from_rl_nolock() 50 err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length); in ntfs_cluster_free_from_rl_nolock() 142 runlist_element *rl = NULL; in ntfs_cluster_alloc() local 320 if ((rlpos + 2) * sizeof(*rl) > rlsize) { in ntfs_cluster_alloc() 324 if (!rl) in ntfs_cluster_alloc() 336 memcpy(rl2, rl, rlsize); in ntfs_cluster_alloc() 337 ntfs_free(rl); in ntfs_cluster_alloc() [all …]
|
D | debug.c | 120 void ntfs_debug_dump_runlist(const runlist_element *rl) in ntfs_debug_dump_runlist() argument 129 if (!rl) { in ntfs_debug_dump_runlist() 135 LCN lcn = (rl + i)->lcn; in ntfs_debug_dump_runlist() 143 (long long)(rl + i)->vcn, lcn_str[index], in ntfs_debug_dump_runlist() 144 (long long)(rl + i)->length, in ntfs_debug_dump_runlist() 145 (rl + i)->length ? "" : in ntfs_debug_dump_runlist() 149 (long long)(rl + i)->vcn, in ntfs_debug_dump_runlist() 150 (long long)(rl + i)->lcn, in ntfs_debug_dump_runlist() 151 (long long)(rl + i)->length, in ntfs_debug_dump_runlist() 152 (rl + i)->length ? "" : in ntfs_debug_dump_runlist() [all …]
|
D | runlist.h | 41 runlist_element *rl; member 45 static inline void ntfs_init_runlist(runlist *rl) in ntfs_init_runlist() argument 47 rl->rl = NULL; in ntfs_init_runlist() 48 init_rwsem(&rl->lock); in ntfs_init_runlist() 65 extern LCN ntfs_rl_vcn_to_lcn(const runlist_element *rl, const VCN vcn); 69 extern runlist_element *ntfs_rl_find_vcn_nolock(runlist_element *rl, 73 const runlist_element *rl, const VCN first_vcn, 77 const int dst_len, const runlist_element *rl,
|
D | attrib.c | 77 runlist_element *rl; in ntfs_map_runlist_nolock() local 173 rl = ntfs_mapping_pairs_decompress(ni->vol, a, ni->runlist.rl); in ntfs_map_runlist_nolock() 174 if (IS_ERR(rl)) in ntfs_map_runlist_nolock() 175 err = PTR_ERR(rl); in ntfs_map_runlist_nolock() 177 ni->runlist.rl = rl; in ntfs_map_runlist_nolock() 290 if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <= in ntfs_map_runlist() 340 if (!ni->runlist.rl) { in ntfs_attr_vcn_to_lcn_nolock() 350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock() 364 if (unlikely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) != in ntfs_attr_vcn_to_lcn_nolock() 454 runlist_element *rl; in ntfs_attr_find_vcn_nolock() local [all …]
|
D | logfile.c | 719 runlist_element *rl; in ntfs_empty_logfile() local 746 rl = log_ni->runlist.rl; in ntfs_empty_logfile() 747 if (unlikely(!rl || vcn < rl->vcn || !rl->length)) { in ntfs_empty_logfile() 755 rl = log_ni->runlist.rl; in ntfs_empty_logfile() 756 BUG_ON(!rl || vcn < rl->vcn || !rl->length); in ntfs_empty_logfile() 759 while (rl->length && vcn >= rl[1].vcn) in ntfs_empty_logfile() 760 rl++; in ntfs_empty_logfile() 770 lcn = rl->lcn; in ntfs_empty_logfile() 772 vcn = rl->vcn; in ntfs_empty_logfile() 776 if (unlikely(!rl->length || lcn < LCN_HOLE)) in ntfs_empty_logfile() [all …]
|
D | mft.c | 463 runlist_element *rl; in ntfs_sync_mft_mirror() local 511 rl = NULL; in ntfs_sync_mft_mirror() 535 if (!rl) { in ntfs_sync_mft_mirror() 538 rl = NTFS_I(vol->mftmirr_ino)->runlist.rl; in ntfs_sync_mft_mirror() 543 BUG_ON(!rl); in ntfs_sync_mft_mirror() 546 while (rl->length && rl[1].vcn <= vcn) in ntfs_sync_mft_mirror() 547 rl++; in ntfs_sync_mft_mirror() 548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror() 573 if (unlikely(rl)) in ntfs_sync_mft_mirror() 669 runlist_element *rl; in write_mft_record_nolock() local [all …]
|
D | aops.c | 173 runlist_element *rl; in ntfs_read_block() local 186 BUG_ON(!ni->runlist.rl && !ni->mft_no && !NInoAttr(ni)); in ntfs_read_block() 225 rl = NULL; in ntfs_read_block() 246 if (!rl) { in ntfs_read_block() 249 rl = ni->runlist.rl; in ntfs_read_block() 251 if (likely(rl != NULL)) { in ntfs_read_block() 253 while (rl->length && rl[1].vcn <= vcn) in ntfs_read_block() 254 rl++; in ntfs_read_block() 255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block() 286 rl = NULL; in ntfs_read_block() [all …]
|
D | lcnalloc.h | 100 const runlist_element *rl); 119 const runlist_element *rl) in ntfs_cluster_free_from_rl() argument 124 ret = ntfs_cluster_free_from_rl_nolock(vol, rl); in ntfs_cluster_free_from_rl()
|
D | file.c | 581 runlist_element *rl, *rl2; in ntfs_prepare_pages_for_non_resident_write() local 626 rl = NULL; in ntfs_prepare_pages_for_non_resident_write() 829 if (!rl) { in ntfs_prepare_pages_for_non_resident_write() 832 rl = ni->runlist.rl; in ntfs_prepare_pages_for_non_resident_write() 834 if (likely(rl != NULL)) { in ntfs_prepare_pages_for_non_resident_write() 836 while (rl->length && rl[1].vcn <= bh_cpos) in ntfs_prepare_pages_for_non_resident_write() 837 rl++; in ntfs_prepare_pages_for_non_resident_write() 838 lcn = ntfs_rl_vcn_to_lcn(rl, bh_cpos); in ntfs_prepare_pages_for_non_resident_write() 846 vcn_len = rl[1].vcn - vcn; in ntfs_prepare_pages_for_non_resident_write() 863 rl = NULL; in ntfs_prepare_pages_for_non_resident_write() [all …]
|
D | compress.c | 470 runlist_element *rl; in ntfs_read_compressed_block() local 595 rl = NULL; in ntfs_read_compressed_block() 600 if (!rl) { in ntfs_read_compressed_block() 603 rl = ni->runlist.rl; in ntfs_read_compressed_block() 605 if (likely(rl != NULL)) { in ntfs_read_compressed_block() 607 while (rl->length && rl[1].vcn <= vcn) in ntfs_read_compressed_block() 608 rl++; in ntfs_read_compressed_block() 609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block() 646 if (rl) in ntfs_read_compressed_block()
|
D | debug.h | 33 extern void ntfs_debug_dump_runlist(const runlist_element *rl); 43 #define ntfs_debug_dump_runlist(rl) do {} while (0) argument
|
/kernel/linux/linux-5.10/drivers/s390/scsi/ |
D | zfcp_reqlist.h | 41 struct zfcp_reqlist *rl; in zfcp_reqlist_alloc() local 43 rl = kzalloc(sizeof(struct zfcp_reqlist), GFP_KERNEL); in zfcp_reqlist_alloc() 44 if (!rl) in zfcp_reqlist_alloc() 47 spin_lock_init(&rl->lock); in zfcp_reqlist_alloc() 50 INIT_LIST_HEAD(&rl->buckets[i]); in zfcp_reqlist_alloc() 52 return rl; in zfcp_reqlist_alloc() 61 static inline int zfcp_reqlist_isempty(struct zfcp_reqlist *rl) in zfcp_reqlist_isempty() argument 66 if (!list_empty(&rl->buckets[i])) in zfcp_reqlist_isempty() 75 static inline void zfcp_reqlist_free(struct zfcp_reqlist *rl) in zfcp_reqlist_free() argument 78 BUG_ON(!zfcp_reqlist_isempty(rl)); in zfcp_reqlist_free() [all …]
|
/kernel/linux/linux-5.10/crypto/ |
D | vmac.c | 103 #define ADD128(rh, rl, ih, il) \ argument 106 (rl) += (_il); \ 107 if ((rl) < (_il)) \ 114 #define PMUL64(rh, rl, i1, i2) /* Assumes m doesn't overflow */ \ argument 119 rl = MUL32(_i1, _i2); \ 120 ADD128(rh, rl, (m >> 32), (m << 32)); \ 123 #define MUL64(rh, rl, i1, i2) \ argument 129 rl = MUL32(_i1, _i2); \ 130 ADD128(rh, rl, (m1 >> 32), (m1 << 32)); \ 131 ADD128(rh, rl, (m2 >> 32), (m2 << 32)); \ [all …]
|
D | crypto_user_stat.c | 198 struct crypto_stat_larval rl; in crypto_reportstat_one() local 200 memset(&rl, 0, sizeof(rl)); in crypto_reportstat_one() 201 strscpy(rl.type, "larval", sizeof(rl.type)); in crypto_reportstat_one() 202 if (nla_put(skb, CRYPTOCFGA_STAT_LARVAL, sizeof(rl), &rl)) in crypto_reportstat_one()
|
/kernel/linux/linux-5.10/tools/testing/selftests/kvm/ |
D | kvm_create_max_vcpus.c | 49 struct rlimit rl; in main() local 58 TEST_ASSERT(!getrlimit(RLIMIT_NOFILE, &rl), "getrlimit() failed!"); in main() 60 if (rl.rlim_cur < nr_fds_wanted) { in main() 61 rl.rlim_cur = nr_fds_wanted; in main() 62 if (rl.rlim_max < nr_fds_wanted) { in main() 63 int old_rlim_max = rl.rlim_max; in main() 64 rl.rlim_max = nr_fds_wanted; in main() 66 int r = setrlimit(RLIMIT_NOFILE, &rl); in main() 73 TEST_ASSERT(!setrlimit(RLIMIT_NOFILE, &rl), "setrlimit() failed!"); in main()
|
/kernel/linux/linux-5.10/include/linux/ |
D | jump_label_ratelimit.h | 41 jump_label_rate_limit(struct static_key_deferred *key, unsigned long rl); 45 #define DEFINE_STATIC_KEY_DEFERRED_TRUE(name, rl) \ argument 48 .timeout = (rl), \ 54 #define DEFINE_STATIC_KEY_DEFERRED_FALSE(name, rl) \ argument 57 .timeout = (rl), \ 73 #define DEFINE_STATIC_KEY_DEFERRED_TRUE(name, rl) \ argument 75 #define DEFINE_STATIC_KEY_DEFERRED_FALSE(name, rl) \ argument 91 unsigned long rl) in jump_label_rate_limit() argument
|
D | math64.h | 203 } rl, rm, rn, rh, a0, b0; in mul_u64_u64_shr() local 209 rl.ll = mul_u32_u32(a0.l.low, b0.l.low); in mul_u64_u64_shr() 219 rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low; in mul_u64_u64_shr() 228 return rl.ll; in mul_u64_u64_shr() 230 return (rl.ll >> shift) | (rh.ll << (64 - shift)); in mul_u64_u64_shr() 249 } u, rl, rh; in mul_u64_u32_div() local 252 rl.ll = mul_u32_u32(u.l.low, mul); in mul_u64_u32_div() 253 rh.ll = mul_u32_u32(u.l.high, mul) + rl.l.high; in mul_u64_u32_div() 256 rl.l.high = do_div(rh.ll, divisor); in mul_u64_u32_div() 259 do_div(rl.ll, divisor); in mul_u64_u32_div() [all …]
|
/kernel/linux/linux-5.10/fs/dlm/ |
D | rcom.c | 374 struct rcom_lock *rl) in pack_rcom_lock() argument 376 memset(rl, 0, sizeof(*rl)); in pack_rcom_lock() 378 rl->rl_ownpid = cpu_to_le32(lkb->lkb_ownpid); in pack_rcom_lock() 379 rl->rl_lkid = cpu_to_le32(lkb->lkb_id); in pack_rcom_lock() 380 rl->rl_exflags = cpu_to_le32(lkb->lkb_exflags); in pack_rcom_lock() 381 rl->rl_flags = cpu_to_le32(lkb->lkb_flags); in pack_rcom_lock() 382 rl->rl_lvbseq = cpu_to_le32(lkb->lkb_lvbseq); in pack_rcom_lock() 383 rl->rl_rqmode = lkb->lkb_rqmode; in pack_rcom_lock() 384 rl->rl_grmode = lkb->lkb_grmode; in pack_rcom_lock() 385 rl->rl_status = lkb->lkb_status; in pack_rcom_lock() [all …]
|
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlx5/core/ |
D | rl.c | 247 struct mlx5_rate_limit *rl) in mlx5_rl_add_rate() argument 251 MLX5_SET(set_pp_rate_limit_context, rl_raw, rate_limit, rl->rate); in mlx5_rl_add_rate() 253 rl->max_burst_sz); in mlx5_rl_add_rate() 255 rl->typical_pkt_sz); in mlx5_rl_add_rate() 264 void mlx5_rl_remove_rate(struct mlx5_core_dev *dev, struct mlx5_rate_limit *rl) in mlx5_rl_remove_rate() argument 271 if (rl->rate == 0) in mlx5_rl_remove_rate() 274 MLX5_SET(set_pp_rate_limit_context, rl_raw, rate_limit, rl->rate); in mlx5_rl_remove_rate() 276 rl->max_burst_sz); in mlx5_rl_remove_rate() 278 rl->typical_pkt_sz); in mlx5_rl_remove_rate() 286 rl->rate, rl->max_burst_sz, rl->typical_pkt_sz); in mlx5_rl_remove_rate()
|
/kernel/linux/linux-5.10/arch/arm/mm/ |
D | proc-v7-3level.S | 57 #define rl r3 macro 60 #define rl r2 macro 73 tst rl, #L_PTE_VALID 76 bicne rl, #L_PTE_VALID 82 orrne rl, #PTE_AP2 83 biceq rl, #PTE_AP2
|
/kernel/linux/linux-5.10/drivers/slimbus/ |
D | slimbus.h | 128 u8 rl; member 140 #define DEFINE_SLIM_LDEST_TXN(name, mc, rl, la, msg) \ argument 141 struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_LOGICALADDR, 0,\ 144 #define DEFINE_SLIM_BCAST_TXN(name, mc, rl, la, msg) \ argument 145 struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_BROADCAST, 0,\ 148 #define DEFINE_SLIM_EDEST_TXN(name, mc, rl, la, msg) \ argument 149 struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_ENUMADDR, 0,\
|
/kernel/linux/linux-5.10/arch/arm/vfp/ |
D | vfp.h | 73 u64 rh, rma, rmb, rl; in mul64to128() local 77 rl = (u64)nl * ml; in mul64to128() 90 rl += rma; in mul64to128() 91 rh += (rl < rma); in mul64to128() 93 *resl = rl; in mul64to128() 105 u64 rh, rl; in vfp_hi64multiply64() local 106 mul64to128(&rh, &rl, n, m); in vfp_hi64multiply64() 107 return rh | (rl != 0); in vfp_hi64multiply64()
|