Home
last modified time | relevance | path

Searched refs:curr (Results 1 – 10 of 10) sorted by relevance

/lib/
Dsg_split.c27 struct sg_splitter *curr = splitters; in sg_calculate_split() local
43 if (!curr->in_sg0) { in sg_calculate_split()
44 curr->in_sg0 = sg; in sg_calculate_split()
45 curr->skip_sg0 = skip; in sg_calculate_split()
48 curr->nents++; in sg_calculate_split()
49 curr->length_last_sg = len; in sg_calculate_split()
52 curr++; in sg_calculate_split()
57 curr->in_sg0 = sg; in sg_calculate_split()
58 curr->skip_sg0 = skip; in sg_calculate_split()
59 curr->nents = 1; in sg_calculate_split()
[all …]
Dxarray.c1236 void *curr; in xas_find_conflict() local
1245 curr = xas_start(xas); in xas_find_conflict()
1246 if (!curr) in xas_find_conflict()
1248 while (xa_is_node(curr)) { in xas_find_conflict()
1249 struct xa_node *node = xa_to_node(curr); in xas_find_conflict()
1250 curr = xas_descend(xas, node); in xas_find_conflict()
1252 if (curr) in xas_find_conflict()
1253 return curr; in xas_find_conflict()
1270 curr = xa_entry_locked(xas->xa, xas->xa_node, ++xas->xa_offset); in xas_find_conflict()
1271 if (xa_is_sibling(curr)) in xas_find_conflict()
[all …]
Dtest_xarray.c72 void *curr; in xa_store_order() local
76 curr = xas_store(&xas, entry); in xa_store_order()
80 return curr; in xa_store_order()
/lib/zlib_inflate/
Dinftrees.c30 unsigned curr; /* number of index bits for current table */ in zlib_inflate_table() local
191 curr = root; /* current table index bits */ in zlib_inflate_table()
220 fill = 1U << curr; in zlib_inflate_table()
255 curr = len - drop; in zlib_inflate_table()
256 left = (int)(1 << curr); in zlib_inflate_table()
257 while (curr + drop < max) { in zlib_inflate_table()
258 left -= count[curr + drop]; in zlib_inflate_table()
260 curr++; in zlib_inflate_table()
265 used += 1U << curr; in zlib_inflate_table()
271 (*table)[low].op = (unsigned char)curr; in zlib_inflate_table()
/lib/dim/
Dnet_dim.c137 static int net_dim_stats_compare(struct dim_stats *curr, in net_dim_stats_compare() argument
141 return curr->bpms ? DIM_STATS_BETTER : DIM_STATS_SAME; in net_dim_stats_compare()
143 if (IS_SIGNIFICANT_DIFF(curr->bpms, prev->bpms)) in net_dim_stats_compare()
144 return (curr->bpms > prev->bpms) ? DIM_STATS_BETTER : in net_dim_stats_compare()
148 return curr->ppms ? DIM_STATS_BETTER : in net_dim_stats_compare()
151 if (IS_SIGNIFICANT_DIFF(curr->ppms, prev->ppms)) in net_dim_stats_compare()
152 return (curr->ppms > prev->ppms) ? DIM_STATS_BETTER : in net_dim_stats_compare()
158 if (IS_SIGNIFICANT_DIFF(curr->epms, prev->epms)) in net_dim_stats_compare()
159 return (curr->epms < prev->epms) ? DIM_STATS_BETTER : in net_dim_stats_compare()
Drdma_dim.c26 static int rdma_dim_stats_compare(struct dim_stats *curr, in rdma_dim_stats_compare() argument
33 if (IS_SIGNIFICANT_DIFF(curr->cpms, prev->cpms)) in rdma_dim_stats_compare()
34 return (curr->cpms > prev->cpms) ? DIM_STATS_BETTER : in rdma_dim_stats_compare()
37 if (IS_SIGNIFICANT_DIFF(curr->cpe_ratio, prev->cpe_ratio)) in rdma_dim_stats_compare()
38 return (curr->cpe_ratio > prev->cpe_ratio) ? DIM_STATS_BETTER : in rdma_dim_stats_compare()
/lib/zstd/
Dcompress.c1038 U32 const curr = (U32)(ip - base); in ZSTD_compressBlock_fast_generic() local
1041 hashTable[h] = curr; /* update hash table */ in ZSTD_compressBlock_fast_generic()
1072 …hashTable[ZSTD_hashPtr(base + curr + 2, hBits, mls)] = curr + 2; /* here because curr+2 could be >… in ZSTD_compressBlock_fast_generic()
1141 const U32 curr = (U32)(ip - base); in ZSTD_compressBlock_fast_extDict_generic() local
1142 const U32 repIndex = curr + 1 - offset_1; /* offset_1 expected <= curr +1 */ in ZSTD_compressBlock_fast_extDict_generic()
1146 hashTable[h] = curr; /* update hash table */ in ZSTD_compressBlock_fast_extDict_generic()
1169 offset = curr - matchIndex; in ZSTD_compressBlock_fast_extDict_generic()
1182 hashTable[ZSTD_hashPtr(base + curr + 2, hBits, mls)] = curr + 2; in ZSTD_compressBlock_fast_extDict_generic()
1287 U32 const curr = (U32)(ip - base); in ZSTD_compressBlock_doubleFast_generic() local
1292 hashLong[h2] = hashSmall[h] = curr; /* update hash tables */ in ZSTD_compressBlock_doubleFast_generic()
[all …]
Dzstd_opt.h239 const U32 curr = (U32)(ip - base); in ZSTD_insertBtAndGetAllMatches() local
252 const U32 btLow = btMask >= curr ? 0 : curr - btMask; in ZSTD_insertBtAndGetAllMatches()
254 U32 *smallerPtr = bt + 2 * (curr & btMask); in ZSTD_insertBtAndGetAllMatches()
255 U32 *largerPtr = bt + 2 * (curr & btMask) + 1; in ZSTD_insertBtAndGetAllMatches()
256 U32 matchEndIdx = curr + 8; in ZSTD_insertBtAndGetAllMatches()
265 if (matchIndex3 > windowLow && (curr - matchIndex3 < (1 << 18))) { in ZSTD_insertBtAndGetAllMatches()
282 matches[mnum].off = ZSTD_REP_MOVE_OPT + curr - matchIndex3; in ZSTD_insertBtAndGetAllMatches()
293 hashTable[h] = curr; /* Update Hash Table */ in ZSTD_insertBtAndGetAllMatches()
316 matches[mnum].off = ZSTD_REP_MOVE_OPT + curr - matchIndex; in ZSTD_insertBtAndGetAllMatches()
351 zc->nextToUpdate = (matchEndIdx > curr + 8) ? matchEndIdx - 8 : curr + 1; in ZSTD_insertBtAndGetAllMatches()
[all …]
Dhuf_compress.c237 U32 curr = nextRankStart; in HUF_readCTable_wksp() local
239 rankVal[n] = curr; in HUF_readCTable_wksp()
386 U32 curr; member
402 rank[n].curr = rank[n].base; in HUF_sort()
406 U32 pos = rank[r].curr++; in HUF_sort()
Dhuf_decompress.c133 U32 const curr = nextRankStart; in HUF_readDTableX2_wksp() local
135 rankVal[n] = curr; in HUF_readDTableX2_wksp()
541 U32 curr = nextRankStart; in HUF_readDTableX4_wksp() local
543 rankStart[w] = curr; in HUF_readDTableX4_wksp()
569 U32 curr = nextRankVal; in HUF_readDTableX4_wksp() local
571 rankVal0[w] = curr; in HUF_readDTableX4_wksp()