| /third_party/grpc/src/core/telemetry/ |
| D | stats_data.yaml | 25 buckets: 26 49 buckets: 20 53 buckets: 10 61 buckets: 20 65 buckets: 20 69 buckets: 10 74 buckets: 20 88 buckets: 26 97 buckets: 40 101 buckets: 20 [all …]
|
| D | stats_data.cc | 530 call_initial_size.buckets()}; in histogram() 533 tcp_write_size.buckets()}; in histogram() 536 tcp_write_iov_size.buckets()}; in histogram() 539 tcp_read_size.buckets()}; in histogram() 542 tcp_read_offer.buckets()}; in histogram() 545 tcp_read_offer_iov_size.buckets()}; in histogram() 548 http2_send_message_size.buckets()}; in histogram() 551 http2_metadata_size.buckets()}; in histogram() 554 http2_hpack_entry_lifetime.buckets()}; in histogram() 557 http2_header_table_size.buckets()}; in histogram() [all …]
|
| /third_party/cares/src/lib/dsa/ |
| D | ares_htable.c | 43 /* NOTE: if we converted buckets into ares_slist_t we could guarantee on 49 ares_llist_t **buckets; member 71 static void ares_htable_buckets_destroy(ares_llist_t **buckets, in ares_htable_buckets_destroy() argument 77 if (buckets == NULL) { in ares_htable_buckets_destroy() 82 if (buckets[i] == NULL) { in ares_htable_buckets_destroy() 87 ares_llist_replace_destructor(buckets[i], NULL); in ares_htable_buckets_destroy() 90 ares_llist_destroy(buckets[i]); in ares_htable_buckets_destroy() 93 ares_free(buckets); in ares_htable_buckets_destroy() 101 ares_htable_buckets_destroy(htable->buckets, htable->size, ARES_TRUE); in ares_htable_destroy() 128 htable->buckets = ares_malloc_zero(sizeof(*htable->buckets) * htable->size); in ares_htable_create() [all …]
|
| /third_party/grpc/tools/distrib/python/xds_protos/google/logging/v2/ |
| D | logging_config_pb2.py | 24 …buckets/{bucket}\x12\x42organizations/{organization}/locations/{location}/buckets/{bucket}\x12\x36… 47 …buckets/{bucket}\022Borganizations/{organization}/locations/{location}/buckets/{bucket}\0226folder… 53 …buckets/{bucket}/views/{view}\022Oorganizations/{organization}/locations/{location}/buckets/{bucke… 87 …buckets/{bucket}/links/{link}\022Oorganizations/{organization}/locations/{location}/buckets/{bucke… 255 …v2/{parent=folders/*/locations/*}/bucketsZ4\0222/v2/{parent=billingAccounts/*/locations/*}/buckets' 257 …buckets/*}Z-\022+/v2/{name=projects/*/locations/*/buckets/*}Z2\0220/v2/{name=organizations/*/locat… 259 …buckets:createAsync:\006bucketZA\"7/v2/{parent=projects/*/locations/*}/buckets:createAsync:\006buc… 261 …buckets/*}:updateAsync:\006bucketZA\"7/v2/{name=projects/*/locations/*/buckets/*}:updateAsync:\006… 263 …buckets:\006bucketZ5\"+/v2/{parent=projects/*/locations/*}/buckets:\006bucketZ:\"0/v2/{parent=orga… 265 …buckets/*}:\006bucketZ52+/v2/{name=projects/*/locations/*/buckets/*}:\006bucketZ:20/v2/{name=organ… [all …]
|
| /third_party/rust/crates/aho-corasick/src/packed/teddy/ |
| D | compile.rs | 19 /// slim Teddy is used (8 buckets) and `true` means fat Teddy is used 20 /// (16 buckets). Fat Teddy requires AVX2, so if that CPU feature isn't 54 /// 16 buckets where as Slim Teddy uses 8 buckets. More buckets are useful 115 let Compiler { buckets, masks, .. } = compiler; in build_imp() 122 buckets, in build_imp() 131 buckets, in build_imp() 140 buckets, in build_imp() 149 buckets, in build_imp() 159 buckets, in build_imp() 169 buckets, in build_imp() [all …]
|
| /third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/Support/ |
| D | OnDiskHashTable.h | 85 Bucket *Buckets; variable 89 void insert(Bucket *Buckets, size_t Size, Item *E) { in insert() argument 90 Bucket &B = Buckets[E->Hash & (Size - 1)]; in insert() 96 /// Resize the hash table, moving the old entries into the new buckets. 102 for (Item *E = Buckets[I].Head; E;) { in resize() 109 free(Buckets); in resize() 111 Buckets = NewBuckets; in resize() 130 insert(Buckets, NumBuckets, new (BA.Allocate()) Item(Key, Data, InfoObj)); in insert() 136 for (Item *I = Buckets[Hash & (NumBuckets - 1)].Head; I; I = I->Next) in contains() 158 // 64 buckets.) We aim for an occupancy ratio in [3/8, 3/4). in Emit() [all …]
|
| /third_party/grpc/tools/codegen/core/ |
| D | gen_stats_data.py | 57 make_type("Histogram", ["max", "buckets"]), 59 Shape = collections.namedtuple("Shape", "max buckets") 168 while len(bounds) < shape.buckets + 1: 169 if len(bounds) == shape.buckets: 174 1.0 / (shape.buckets + 1 - len(bounds)), 255 shapes.add(Shape(max=histogram.max, buckets=histogram.buckets)) 298 "class HistogramCollector_%d_%d;" % (shape.max, shape.buckets), 301 print("class Histogram_%d_%d {" % (shape.max, shape.buckets), file=H) 304 print(" const uint64_t* buckets() const { return buckets_; }", file=H) 306 " size_t bucket_count() const { return %d; }" % shape.buckets, [all …]
|
| /third_party/icu/ohos_icu4j/src/main/java/ohos/global/icu/text/ |
| D | AlphabeticIndex.java | 49 * The class also supports having buckets for strings before the first (underflow), 71 * // Show index at top. We could skip or gray out empty buckets 79 * // Show the buckets with their contents, skipping empty buckets 91 * if its bucket is empty. Small buckets could also be combined based on size, such as: 103 * they "lazily" build the index buckets. 106 * buckets and their labels and label types. 128 * Prefix string for Chinese index buckets. 156 private BucketList<V> buckets; field in AlphabeticIndex 165 * and random access to buckets and their properties, 172 private final BucketList<V> buckets; field in AlphabeticIndex.ImmutableIndex [all …]
|
| /third_party/icu/icu4j/main/classes/collate/src/com/ibm/icu/text/ |
| D | AlphabeticIndex.java | 48 * The class also supports having buckets for strings before the first (underflow), 70 * // Show index at top. We could skip or gray out empty buckets 78 * // Show the buckets with their contents, skipping empty buckets 90 * if its bucket is empty. Small buckets could also be combined based on size, such as: 102 * they "lazily" build the index buckets. 105 * buckets and their labels and label types. 128 * Prefix string for Chinese index buckets. 156 private BucketList<V> buckets; field in AlphabeticIndex 165 * and random access to buckets and their properties, 173 private final BucketList<V> buckets; field in AlphabeticIndex.ImmutableIndex [all …]
|
| /third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/include/llvm/DebugInfo/PDB/Native/ |
| D | HashTable.h | 74 return Map->Buckets[Index]; 83 while (Index < Map->Buckets.size()) { 115 HashTable() { Buckets.resize(8); } in HashTable() 117 Buckets.resize(Capacity); in HashTable() 131 Buckets.resize(H->Capacity); in load() 146 if (auto EC = Stream.readInteger(Buckets[P].first)) in load() 151 Buckets[P].second = *Value; in load() 207 Buckets.resize(8); in clear() 213 uint32_t capacity() const { return Buckets.size(); } in capacity() 228 if (Traits.storageKeyToLookupKey(Buckets[I].first) == K) in find_as() [all …]
|
| /third_party/grpc/test/core/test_util/ |
| D | histogram.cc | 32 // There are sufficient buckets to reach max_bucket_start 50 // number of buckets 52 // the buckets themselves 53 uint32_t* buckets; member 91 h->buckets = in grpc_histogram_create() 97 gpr_free(h->buckets); in grpc_histogram_destroy() 111 h->buckets[bucket_for(h, x)]++; in grpc_histogram_add() 120 grpc_histogram_merge_contents(dst, src->buckets, src->num_buckets, in grpc_histogram_merge() 142 histogram->buckets[i] += data[i]; in grpc_histogram_merge_contents() 167 count_so_far += h->buckets[lower_idx]; in threshold_for_count_below() [all …]
|
| /third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Support/ |
| D | FoldingSet.cpp | 187 /// The problem with this is that the start of the hash buckets are not 208 static void **GetBucketFor(unsigned Hash, void **Buckets, unsigned NumBuckets) { in GetBucketFor() argument 211 return Buckets + BucketNum; in GetBucketFor() 216 void **Buckets = static_cast<void**>(safe_calloc(NumBuckets + 1, in AllocateBuckets() local 219 Buckets[NumBuckets] = reinterpret_cast<void*>(-1); in AllocateBuckets() 220 return Buckets; in AllocateBuckets() 232 Buckets = AllocateBuckets(NumBuckets); in FoldingSetBase() 237 : Buckets(Arg.Buckets), NumBuckets(Arg.NumBuckets), NumNodes(Arg.NumNodes) { in FoldingSetBase() 238 Arg.Buckets = nullptr; in FoldingSetBase() 244 free(Buckets); // This may be null if the set is in a moved-from state. in operator =() [all …]
|
| /third_party/skia/m133/infra/project-config/ |
| D | cr-buildbucket.cfg | 1 # Defines buckets on cr-buildbucket.appspot.com, used to schedule builds 2 # on buildbot. In particular, CQ uses some of these buckets to schedule tryjobs. 9 buckets { 29 buckets { 49 buckets { 69 buckets {
|
| /third_party/skia/infra/project-config/ |
| D | cr-buildbucket.cfg | 1 # Defines buckets on cr-buildbucket.appspot.com, used to schedule builds 2 # on buildbot. In particular, CQ uses some of these buckets to schedule tryjobs. 9 buckets { 29 buckets { 49 buckets { 69 buckets {
|
| /third_party/mesa3d/src/virtio/vulkan/ |
| D | vn_renderer_internal.c | 24 const struct vn_renderer_shmem_bucket *bucket = &cache->buckets[idx]; in vn_renderer_shmem_cache_dump() 30 vn_log(NULL, " buckets[%d]: %d shmems", idx, count); in vn_renderer_shmem_cache_dump() 42 static_assert(ARRAY_SIZE(cache->buckets) <= 32, ""); in vn_renderer_shmem_cache_init() 49 for (uint32_t i = 0; i < ARRAY_SIZE(cache->buckets); i++) { in vn_renderer_shmem_cache_init() 50 struct vn_renderer_shmem_bucket *bucket = &cache->buckets[i]; in vn_renderer_shmem_cache_init() 68 struct vn_renderer_shmem_bucket *bucket = &cache->buckets[idx]; in vn_renderer_shmem_cache_fini() 88 if (unlikely(idx >= ARRAY_SIZE(cache->buckets))) in choose_bucket() 92 return &cache->buckets[idx]; in choose_bucket() 102 struct vn_renderer_shmem_bucket *bucket = &cache->buckets[idx]; in vn_renderer_shmem_cache_remove_expired_locked()
|
| /third_party/python/Python/ |
| D | hashtable.c | 58 ((_Py_hashtable_entry_t *)_Py_SLIST_HEAD(&(HT)->buckets[BUCKET])) 105 /* makes sure the real size of the buckets array is a power of 2 */ 123 /* buckets */ in _Py_hashtable_size() 193 _Py_slist_remove(&ht->buckets[index], (_Py_slist_item_t *)previous, in _Py_hashtable_steal() 242 _Py_slist_prepend(&ht->buckets[index], (_Py_slist_item_t*)entry); in _Py_hashtable_set() 287 size_t buckets_size = new_size * sizeof(ht->buckets[0]); in hashtable_rehash() 296 _Py_hashtable_entry_t *entry = BUCKETS_HEAD(ht->buckets[bucket]); in hashtable_rehash() 308 ht->alloc.free(ht->buckets); in hashtable_rehash() 310 ht->buckets = new_buckets; in hashtable_rehash() 339 size_t buckets_size = ht->nbuckets * sizeof(ht->buckets[0]); in _Py_hashtable_new_full() [all …]
|
| /third_party/mindspore/mindspore-src/source/mindspore/ccsrc/plugin/device/cpu/kernel/ |
| D | sparse_optimizer_cpu_kernel.h | 111 std::vector<std::shared_ptr<BucketSparseGradient<T>>> buckets; in BucketReduceSparseGradient() local 112 … GatherSegmentIndicesToOutputBucket(multi_thread_param, segments, segment_bucket_sizes, &buckets); in BucketReduceSparseGradient() 115 ReduceBucketSparseGradientToWorkspace(multi_thread_param, buckets, &reduced_buckets); in BucketReduceSparseGradient() 207 … const std::vector<std::shared_ptr<BucketSparseGradient<T>>> &buckets) { in CopySegmentIndicesToBucket() argument 220 buckets[bucket_id]->indices_[bucket_index] = index; in CopySegmentIndicesToBucket() 221 buckets[bucket_id]->global_indices_[bucket_index] = bucket_offset + i; in CopySegmentIndicesToBucket() 238 auto &buckets = *buckets_ptr; in GatherSegmentIndicesToOutputBucket() local 251 (void)buckets.emplace_back(std::make_shared<BucketSparseGradient<T>>()); in GatherSegmentIndicesToOutputBucket() 252 … buckets[i]->value_ = param.output_grad_->value_ + current_indices_offset * param.value_stride_; in GatherSegmentIndicesToOutputBucket() 253 buckets[i]->indices_ = param.output_grad_->indices_ + current_indices_offset; in GatherSegmentIndicesToOutputBucket() [all …]
|
| D | unique_cpu_kernel.h | 271 … const std::vector<std::shared_ptr<UniqueParam<DataType, IndexType>>> &buckets) { in SegmentToBuckets() argument 276 auto bucket_size = buckets.size(); in SegmentToBuckets() 289 auto &bucket = buckets[bucket_id]; in SegmentToBuckets() 320 auto &buckets = *buckets_ptr; in GatherSegmentsToBuckets() local 322 buckets.reserve(thread_num); in GatherSegmentsToBuckets() 340 (void)buckets.emplace_back(bucket); in GatherSegmentsToBuckets() 348 bucket->input_ = buckets[j]->input_ + tmp_bucket_data_size[j]; in GatherSegmentsToBuckets() 349 bucket->input_size_ = buckets[j]->input_size_ - tmp_bucket_data_size[j]; in GatherSegmentsToBuckets() 350 bucket->workspace_idx_ = buckets[j]->workspace_idx_ + tmp_bucket_data_size[j]; in GatherSegmentsToBuckets() 428 …d UniqueEachBucket(const std::vector<std::shared_ptr<UniqueParam<DataType, IndexType>>> &buckets) { in UniqueEachBucket() argument [all …]
|
| /third_party/skia/third_party/externals/freetype/src/cache/ |
| D | ftccache.c | 100 return cache->buckets + idx; in ftc_get_top_node_for_hash() 107 * buckets array appropriately, we simply degrade the hash table's 118 FT_UFast count = mask + p + 1; /* number of buckets */ in ftc_cache_resize() 121 /* do we need to expand the buckets array? */ in ftc_cache_resize() 127 /* try to expand the buckets array _before_ splitting in ftc_cache_resize() 137 if ( FT_RENEW_ARRAY( cache->buckets, in ftc_cache_resize() 143 pnode = cache->buckets + p; in ftc_cache_resize() 161 cache->buckets[p + mask + 1] = new_list; in ftc_cache_resize() 174 /* do we need to shrink the buckets array? */ in ftc_cache_resize() 191 if ( FT_QRENEW_ARRAY( cache->buckets, in ftc_cache_resize() [all …]
|
| /third_party/libbpf/src/ |
| D | hashmap.c | 21 /* start with 4 buckets */ 45 map->buckets = NULL; in hashmap__init() 71 free(map->buckets); in hashmap__clear() 72 map->buckets = NULL; in hashmap__clear() 124 free(map->buckets); in hashmap_grow() 125 map->buckets = new_buckets; in hashmap_grow() 137 if (!map->buckets) in hashmap_find_entry() 140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry() 200 hashmap_add_entry(&map->buckets[h], entry); in hashmap_insert()
|
| /third_party/skia/third_party/externals/swiftshader/third_party/llvm-subzero/include/llvm/ADT/ |
| D | DenseMap.h | 287 /// somewhere into the DenseMap's array of buckets (i.e. either to a key or 293 /// getPointerIntoBucketsArray() - Return an opaque pointer into the buckets 319 "# initial buckets must be a power of two!"); in initEmpty() 325 /// Returns the number of buckets to allocate to ensure that the DenseMap can 473 // the buckets are empty (meaning that many are filled with tombstones), in InsertIntoBucketImpl() 495 // so that when growing buckets we have self-consistent entry count. in InsertIntoBucketImpl() 590 BucketT *Buckets; variable 618 operator delete(Buckets); in ~DenseMap() 624 std::swap(Buckets, RHS.Buckets); in swap() 638 operator delete(Buckets); [all …]
|
| /third_party/mesa3d/src/gallium/auxiliary/pipebuffer/ |
| D | pb_bufmgr_slab.c | 157 struct pb_manager **buckets; member 493 return mgr->buckets[i]->create_buffer(mgr->buckets[i], size, desc); in pb_slab_range_manager_create_buffer() 522 mgr->buckets[i]->destroy(mgr->buckets[i]); in pb_slab_range_manager_destroy() 523 FREE(mgr->buckets); in pb_slab_range_manager_destroy() 562 mgr->buckets = CALLOC(mgr->numBuckets, sizeof(*mgr->buckets)); in pb_slab_range_manager_create() 563 if (!mgr->buckets) in pb_slab_range_manager_create() 568 mgr->buckets[i] = pb_slab_manager_create(provider, bufSize, slabSize, desc); in pb_slab_range_manager_create() 569 if(!mgr->buckets[i]) in pb_slab_range_manager_create() 578 if(mgr->buckets[i]) in pb_slab_range_manager_create() 579 mgr->buckets[i]->destroy(mgr->buckets[i]); in pb_slab_range_manager_create() [all …]
|
| /third_party/brotli/c/enc/ |
| D | hash_longest_match_quickly_inc.h | 58 uint32_t* BROTLI_RESTRICT buckets = self->buckets_; in FN() local 66 buckets[key] = 0; in FN() 70 buckets[(key + (j << 3)) & BUCKET_MASK] = 0; in FN() 79 memset(buckets, 0, sizeof(uint32_t) * BUCKET_SIZE); in FN() 155 uint32_t* BROTLI_RESTRICT buckets = self->buckets_; in FN() local 179 buckets[key] = (uint32_t)cur_ix; in FN() 194 prev_ix = buckets[key]; in FN() 195 buckets[key] = (uint32_t)cur_ix; in FN() 226 prev_ix = buckets[keys[i]]; in FN() 257 buckets[key_out] = (uint32_t)cur_ix; in FN()
|
| /third_party/skia/third_party/externals/brotli/c/enc/ |
| D | hash_longest_match_quickly_inc.h | 58 uint32_t* BROTLI_RESTRICT buckets = self->buckets_; in FN() local 66 buckets[key] = 0; in FN() 70 buckets[(key + (j << 3)) & BUCKET_MASK] = 0; in FN() 79 memset(buckets, 0, sizeof(uint32_t) * BUCKET_SIZE); in FN() 155 uint32_t* BROTLI_RESTRICT buckets = self->buckets_; in FN() local 179 buckets[key] = (uint32_t)cur_ix; in FN() 194 prev_ix = buckets[key]; in FN() 195 buckets[key] = (uint32_t)cur_ix; in FN() 226 prev_ix = buckets[keys[i]]; in FN() 257 buckets[key_out] = (uint32_t)cur_ix; in FN()
|
| /third_party/grpc/src/ruby/qps/ |
| D | histogram.rb | 41 @buckets=Array.new(bucket_for(max_possible)+1, 0) 56 @buckets[bucket_for(value)] += 1 74 @buckets 85 @buckets = @buckets.map.with_index{ |m,i| m + received_bucket[i].to_i }
|