Home
last modified time | relevance | path

Searched refs:total_size_ (Results 1 – 24 of 24) sorted by relevance

/third_party/protobuf/src/google/protobuf/
Drepeated_field.cc52 if (total_size_ >= new_size) { in InternalExtend()
60 std::max(total_size_ * 2, new_size)); in InternalExtend()
71 const int old_total_size = total_size_; in InternalExtend()
73 total_size_ = new_size; in InternalExtend()
112 if (!rep_ || rep_->allocated_size == total_size_) { in AddWeak()
113 Reserve(total_size_ + 1); in AddWeak()
Drepeated_field.h312 return (total_size_ == 0) ? static_cast<Arena*>(arena_or_elements_) in GetArena()
334 int total_size_; variable
353 GOOGLE_DCHECK_GT(total_size_, 0); in elements()
447 capacity_ = repeated_field_->total_size_; in FastAdderImpl()
456 capacity_ = repeated_field_->total_size_; in Add()
741 int total_size_;
1186 : current_size_(0), total_size_(0), arena_or_elements_(nullptr) {}
1190 : current_size_(0), total_size_(0), arena_or_elements_(arena) {}
1194 : current_size_(0), total_size_(0), arena_or_elements_(nullptr) {
1205 : current_size_(0), total_size_(0), arena_or_elements_(nullptr) {
[all …]
/third_party/grpc/include/grpcpp/impl/codegen/
Dproto_buffer_writer.h62 total_size_(total_size), in ProtoBufferWriter()
83 GPR_CODEGEN_ASSERT(byte_count_ < total_size_); in Next()
89 size_t remain = static_cast<size_t>(total_size_ - byte_count_); in Next()
155 const int total_size_; ///< byte size of proto being serialized variable
/third_party/mindspore/mindspore/ccsrc/runtime/device/ascend/
Dascend_bucket.cc55 total_size_ = total_size; in AllocateAllReduceAddr()
145 if (type_size == 0 || total_size_ % type_size != 0) { in LaunchAllReduce()
146 …MS_LOG(EXCEPTION) << "Total_size[" << total_size_ << "],Type_size[" << type_size << "] != 0, fail!… in LaunchAllReduce()
148 auto hccl_count = total_size_ / type_size; in LaunchAllReduce()
173 auto launch_mul = std::make_shared<AscendLaunchMul>(stream_, tensor_type_list_[0], total_size_); in CreateLaunchMul()
183 std::make_shared<AscendLaunchAtomicClean>(compute_stream_, tensor_type_list_[0], total_size_); in CreateLaunchAtomicClean()
Dascend_launch_atomic_clean.cc59 input->size = total_size_; in LaunchOpKernel()
85 auto shape = total_size_ / dtype_size; in ObtainAtomicCleanKernelGraph()
111 std::vector<size_t> clean_size = {total_size_}; in ConstructKernelGraphAndSetAttr()
Dascend_launch_atomic_clean.h30 total_size_(total_size), in AscendLaunchAtomicClean()
47 size_t total_size_; variable
Dascend_launch_transdata.h32 total_size_(total_size), in AscendLaunchTransData()
53 size_t total_size_; variable
Dascend_launch_transdata.cc61 input->size = total_size_; in LaunchOpKernel()
/third_party/skia/third_party/externals/brotli/c/enc/
Dringbuffer.h37 const uint32_t total_size_; member
63 *(uint32_t*)&rb->total_size_ = rb->size_ + rb->tail_size_; in RingBufferSetup()
120 if (rb->cur_size_ < rb->total_size_) { in RingBufferWrite()
122 RingBufferInitBuffer(m, rb->total_size_, rb); in RingBufferWrite()
144 BROTLI_MIN(size_t, n, rb->total_size_ - masked_pos)); in RingBufferWrite()
/third_party/node/deps/brotli/c/enc/
Dringbuffer.h37 const uint32_t total_size_; member
63 *(uint32_t*)&rb->total_size_ = rb->size_ + rb->tail_size_; in RingBufferSetup()
120 if (rb->cur_size_ < rb->total_size_) { in RingBufferWrite()
122 RingBufferInitBuffer(m, rb->total_size_, rb); in RingBufferWrite()
144 BROTLI_MIN(size_t, n, rb->total_size_ - masked_pos)); in RingBufferWrite()
/third_party/mindspore/mindspore/lite/src/train/
Dstatic_allocator.h29 size_t total_size() { return total_size_; } in total_size()
32 total_size_ += size; in Malloc()
49 size_t total_size_ = 0; variable
/third_party/mindspore/mindspore/ccsrc/runtime/device/gpu/
Dgpu_bucket.cc69 total_size_ = total_size; in AllocateAllReduceAddr()
106 cudaMemsetAsync(ar_input_addr_, 0, total_size_, static_cast<cudaStream_t>(compute_stream_)), in CopyTensorToContiguousMemory()
150 …(*all_reduce_funcptr)(ar_input_addr_, ar_output_addr_, total_size_ / type_size, nccl_data_type_ite… in LaunchAllReduce()
163 auto launch_mul = std::make_shared<GPULaunchMul>(stream_, tensor_type_list_[0], total_size_); in CreateLaunchMul()
/third_party/mindspore/mindspore/ccsrc/runtime/device/
Dlaunch_mul.h29 total_size_(total_size), in LaunchMul()
50 size_t total_size_; variable
Dbucket.h43 total_size_(0), in Bucket()
67 size_t total_size_; variable
Dlaunch_mul.cc31 int64_t shape = SizeToLong(total_size_ / dtype_size); in ObtainMulKernelGraph()
/third_party/mindspore/mindspore/lite/src/runtime/
Dinner_allocator.h47 size_t total_size() const { return this->total_size_; } in total_size()
61 size_t total_size_ = 0; variable
Dinner_allocator.cc58 if (this->total_size_ >= max_malloc_size_) { in Malloc()
79 this->total_size_ += size; in Malloc()
/third_party/boost/libs/asio/include/boost/asio/detail/
Dconsuming_buffers.hpp86 total_size_ = buffer_size(buffers); in consuming_buffers()
92 return total_consumed_ >= total_size_; in empty()
154 std::size_t total_size_; member in boost::asio::detail::consuming_buffers
/third_party/boost/boost/asio/detail/
Dconsuming_buffers.hpp86 total_size_ = buffer_size(buffers); in consuming_buffers()
92 return total_consumed_ >= total_size_; in empty()
154 std::size_t total_size_; member in boost::asio::detail::consuming_buffers
/third_party/mindspore/mindspore/lite/src/runtime/gpu/opencl/
Dopencl_allocator.cc214 total_size_ += size; in _Malloc()
216 UNLOCK_AND_RETURN_NULL(total_size_ >= max_size, nullptr); in _Malloc()
268 << ", total size: " << total_size_; in _Malloc()
Dopencl_allocator.h116 uint64_t total_size_{0};
/third_party/boost/libs/smart_ptr/test/
Dshared_ptr_test.cpp3318 long total_size_; member in n_spt_pimpl::file::impl
3322 impl(char const *, char const *): total_size_(0) in impl()
3334 total_size_ += size; in read()
3339 return total_size_; in total_size()
/third_party/chromium/patch/
D0004-3.2-Beta5.patch40984 OriginInfo::OriginInfo() : total_size_(0) {}
D0003-3.2-Beta3-1115.patch71079 - : total_size_(0) {}
71087 +OriginInfo::OriginInfo() : total_size_(0) {}