Searched refs:curr (Results 1 – 10 of 10) sorted by relevance
/lib/ |
D | sg_split.c | 27 struct sg_splitter *curr = splitters; in sg_calculate_split() local 43 if (!curr->in_sg0) { in sg_calculate_split() 44 curr->in_sg0 = sg; in sg_calculate_split() 45 curr->skip_sg0 = skip; in sg_calculate_split() 48 curr->nents++; in sg_calculate_split() 49 curr->length_last_sg = len; in sg_calculate_split() 52 curr++; in sg_calculate_split() 57 curr->in_sg0 = sg; in sg_calculate_split() 58 curr->skip_sg0 = skip; in sg_calculate_split() 59 curr->nents = 1; in sg_calculate_split() [all …]
|
D | xarray.c | 1236 void *curr; in xas_find_conflict() local 1245 curr = xas_start(xas); in xas_find_conflict() 1246 if (!curr) in xas_find_conflict() 1248 while (xa_is_node(curr)) { in xas_find_conflict() 1249 struct xa_node *node = xa_to_node(curr); in xas_find_conflict() 1250 curr = xas_descend(xas, node); in xas_find_conflict() 1252 if (curr) in xas_find_conflict() 1253 return curr; in xas_find_conflict() 1270 curr = xa_entry_locked(xas->xa, xas->xa_node, ++xas->xa_offset); in xas_find_conflict() 1271 if (xa_is_sibling(curr)) in xas_find_conflict() [all …]
|
D | test_xarray.c | 72 void *curr; in xa_store_order() local 76 curr = xas_store(&xas, entry); in xa_store_order() 80 return curr; in xa_store_order()
|
/lib/zlib_inflate/ |
D | inftrees.c | 30 unsigned curr; /* number of index bits for current table */ in zlib_inflate_table() local 191 curr = root; /* current table index bits */ in zlib_inflate_table() 220 fill = 1U << curr; in zlib_inflate_table() 255 curr = len - drop; in zlib_inflate_table() 256 left = (int)(1 << curr); in zlib_inflate_table() 257 while (curr + drop < max) { in zlib_inflate_table() 258 left -= count[curr + drop]; in zlib_inflate_table() 260 curr++; in zlib_inflate_table() 265 used += 1U << curr; in zlib_inflate_table() 271 (*table)[low].op = (unsigned char)curr; in zlib_inflate_table()
|
/lib/dim/ |
D | net_dim.c | 137 static int net_dim_stats_compare(struct dim_stats *curr, in net_dim_stats_compare() argument 141 return curr->bpms ? DIM_STATS_BETTER : DIM_STATS_SAME; in net_dim_stats_compare() 143 if (IS_SIGNIFICANT_DIFF(curr->bpms, prev->bpms)) in net_dim_stats_compare() 144 return (curr->bpms > prev->bpms) ? DIM_STATS_BETTER : in net_dim_stats_compare() 148 return curr->ppms ? DIM_STATS_BETTER : in net_dim_stats_compare() 151 if (IS_SIGNIFICANT_DIFF(curr->ppms, prev->ppms)) in net_dim_stats_compare() 152 return (curr->ppms > prev->ppms) ? DIM_STATS_BETTER : in net_dim_stats_compare() 158 if (IS_SIGNIFICANT_DIFF(curr->epms, prev->epms)) in net_dim_stats_compare() 159 return (curr->epms < prev->epms) ? DIM_STATS_BETTER : in net_dim_stats_compare()
|
D | rdma_dim.c | 26 static int rdma_dim_stats_compare(struct dim_stats *curr, in rdma_dim_stats_compare() argument 33 if (IS_SIGNIFICANT_DIFF(curr->cpms, prev->cpms)) in rdma_dim_stats_compare() 34 return (curr->cpms > prev->cpms) ? DIM_STATS_BETTER : in rdma_dim_stats_compare() 37 if (IS_SIGNIFICANT_DIFF(curr->cpe_ratio, prev->cpe_ratio)) in rdma_dim_stats_compare() 38 return (curr->cpe_ratio > prev->cpe_ratio) ? DIM_STATS_BETTER : in rdma_dim_stats_compare()
|
/lib/zstd/ |
D | compress.c | 1038 U32 const curr = (U32)(ip - base); in ZSTD_compressBlock_fast_generic() local 1041 hashTable[h] = curr; /* update hash table */ in ZSTD_compressBlock_fast_generic() 1072 …hashTable[ZSTD_hashPtr(base + curr + 2, hBits, mls)] = curr + 2; /* here because curr+2 could be >… in ZSTD_compressBlock_fast_generic() 1141 const U32 curr = (U32)(ip - base); in ZSTD_compressBlock_fast_extDict_generic() local 1142 const U32 repIndex = curr + 1 - offset_1; /* offset_1 expected <= curr +1 */ in ZSTD_compressBlock_fast_extDict_generic() 1146 hashTable[h] = curr; /* update hash table */ in ZSTD_compressBlock_fast_extDict_generic() 1169 offset = curr - matchIndex; in ZSTD_compressBlock_fast_extDict_generic() 1182 hashTable[ZSTD_hashPtr(base + curr + 2, hBits, mls)] = curr + 2; in ZSTD_compressBlock_fast_extDict_generic() 1287 U32 const curr = (U32)(ip - base); in ZSTD_compressBlock_doubleFast_generic() local 1292 hashLong[h2] = hashSmall[h] = curr; /* update hash tables */ in ZSTD_compressBlock_doubleFast_generic() [all …]
|
D | zstd_opt.h | 239 const U32 curr = (U32)(ip - base); in ZSTD_insertBtAndGetAllMatches() local 252 const U32 btLow = btMask >= curr ? 0 : curr - btMask; in ZSTD_insertBtAndGetAllMatches() 254 U32 *smallerPtr = bt + 2 * (curr & btMask); in ZSTD_insertBtAndGetAllMatches() 255 U32 *largerPtr = bt + 2 * (curr & btMask) + 1; in ZSTD_insertBtAndGetAllMatches() 256 U32 matchEndIdx = curr + 8; in ZSTD_insertBtAndGetAllMatches() 265 if (matchIndex3 > windowLow && (curr - matchIndex3 < (1 << 18))) { in ZSTD_insertBtAndGetAllMatches() 282 matches[mnum].off = ZSTD_REP_MOVE_OPT + curr - matchIndex3; in ZSTD_insertBtAndGetAllMatches() 293 hashTable[h] = curr; /* Update Hash Table */ in ZSTD_insertBtAndGetAllMatches() 316 matches[mnum].off = ZSTD_REP_MOVE_OPT + curr - matchIndex; in ZSTD_insertBtAndGetAllMatches() 351 zc->nextToUpdate = (matchEndIdx > curr + 8) ? matchEndIdx - 8 : curr + 1; in ZSTD_insertBtAndGetAllMatches() [all …]
|
D | huf_compress.c | 237 U32 curr = nextRankStart; in HUF_readCTable_wksp() local 239 rankVal[n] = curr; in HUF_readCTable_wksp() 386 U32 curr; member 402 rank[n].curr = rank[n].base; in HUF_sort() 406 U32 pos = rank[r].curr++; in HUF_sort()
|
D | huf_decompress.c | 133 U32 const curr = nextRankStart; in HUF_readDTableX2_wksp() local 135 rankVal[n] = curr; in HUF_readDTableX2_wksp() 541 U32 curr = nextRankStart; in HUF_readDTableX4_wksp() local 543 rankStart[w] = curr; in HUF_readDTableX4_wksp() 569 U32 curr = nextRankVal; in HUF_readDTableX4_wksp() local 571 rankVal0[w] = curr; in HUF_readDTableX4_wksp()
|