/external/flatbuffers/tests/ |
D | monster_test_generated.lobster | 72 b_:flatbuffers_builder 74 b_.StartObject(0) 77 return b_.EndObject() 86 b_:flatbuffers_builder 88 b_.StartObject(0) 91 return b_.EndObject() 101 def CreateTest(b_:flatbuffers_builder, a:int, b:int): 102 b_.Prep(2, 4) 103 b_.Pad(1) 104 b_.PrependInt8(b) [all …]
|
/external/flatbuffers/samples/ |
D | monster_generated.lobster | 29 def CreateVec3(b_:flatbuffers_builder, x:float, y:float, z:float): 30 b_.Prep(4, 12) 31 b_.PrependFloat32(z) 32 b_.PrependFloat32(y) 33 b_.PrependFloat32(x) 34 return b_.Offset() 68 b_:flatbuffers_builder 70 b_.StartObject(11) 73 b_.PrependStructSlot(0, pos) 76 b_.PrependInt16Slot(1, mana, 150) [all …]
|
/external/sfntly/cpp/src/sfntly/data/ |
D | memory_byte_array.cc | 24 : ByteArray(0, length), b_(NULL), allocated_(true) { in MemoryByteArray() 28 : ByteArray(filled_length, filled_length), b_(b), allocated_(false) { in MemoryByteArray() 40 os->Write(b_, offset, length); in CopyTo() 45 if (allocated_ && b_ == NULL) { in Init() 46 b_ = new uint8_t[Size()]; in Init() 47 memset(b_, 0, Size()); in Init() 53 b_[index] = b; in InternalPut() 62 memcpy(b_ + index, b + offset, length); in InternalPut() 68 return b_[index]; in InternalGet() 77 memcpy(b + offset, b_ + index, length); in InternalGet() [all …]
|
D | growable_memory_byte_array.cc | 38 os->Write(&b_, offset, length); in CopyTo() 43 if ((size_t)index >= b_.size()) { in InternalPut() 44 b_.resize((size_t)(index + 1)); in InternalPut() 46 b_[index] = b; in InternalPut() 53 if ((size_t)index + length >= b_.size()) { in InternalPut() 56 b_.resize((size_t)(index + length + 1)); in InternalPut() 58 std::copy(b + offset, b + offset + length, b_.begin() + index); in InternalPut() 63 return b_[index]; in InternalGet() 70 memcpy(b + offset, &(b_[0]) + index, length); in InternalGet() 75 b_.clear(); in Close() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
D | ir_emitter.cc | 101 b_(llvm_module->getContext()), in IrEmitter() 110 b_.setFastMathFlags(llvm_ir::GetCpuFastMathFlags(hlo_module_config_)); in IrEmitter() 141 &b_); in EmitThreadLocalFunctionEpilogue() 147 root_value.GetBasePointer(), &b_); in EmitThreadLocalFunctionEpilogue() 220 hlo_module_config_, module_, &b_, in InitializeIrFunction() 365 GetEmittedValueFor(operand), &b_); in HandleGetTupleElement() 385 GetEmittedValueFor(on_false), &b_); in HandleTupleSelect() 408 llvm_ir::EmitTuple(GetIrArrayFor(infeed), {data_address, token_address}, &b_); in HandleInfeed() 439 tuple_element_addresses, &b_); in HandleInfeed() 462 llvm_ir::EncodeSelfDescribingShapeConstant(shape, &shape_length, &b_)); in EmitXfeedTransfer() [all …]
|
D | dot_op_emitter.cc | 214 *b_->GetInsertBlock()->getParent(), in GetMlirGemmTileSize() 217 *b_->GetInsertBlock()->getParent()); in GetMlirGemmTileSize() 228 llvm::IRBuilder<>* b_; member in xla::cpu::__anonf3b678a30111::DotOpEmitter 249 b_(b), in DotOpEmitter() 262 b_->CreateMemSet(target_ptr, b_->getInt8(0), /*Size=*/size_bytes, in EmitLinalgMatmul() 271 mlir_context_, b_, dot_info_.result_shape, operand_shapes, target_ptr, in EmitLinalgMatmul() 359 b_->CreateMemSet(target, b_->getInt8(0), /*Size=*/size_bytes, in EmitTiledLlvmIrGemm() 364 *b_->GetInsertBlock()->getParent(), primitive_type); in EmitTiledLlvmIrGemm() 377 /*rhs=*/rhs, /*result=*/target, b_, hlo_module_config_); in EmitTiledLlvmIrGemm() 461 *b_->GetInsertBlock()->getParent(), primitive_type); in EmitTiledLlvmIrGemv() [all …]
|
D | tiled_dot_emitter.cc | 39 : vsl_(vsl), b_(b) { in MemoryTile() 88 pointers_[i], b_->CreateAdd(minor_dim_offset, b_->getInt64(j)))); in LoadBroadcastTile() 96 llvm::IRBuilder<>* b_; member in xla::cpu::__anon47412fb60111::MemoryTile 241 b_(b), in ColumnMajorMatrixVectorProductEmitter() 242 ksl_(b_), in ColumnMajorMatrixVectorProductEmitter() 243 vsl_(config.scalar_type(), /*vector_size=*/config.tile_rows(), b_, "") { in ColumnMajorMatrixVectorProductEmitter() 257 return MemoryTile(&vsl_, b_, /*matrix=*/lhs_, in GetLhsMemoryTile() 287 llvm::IRBuilder<>* b_; member in xla::cpu::__anon47412fb60111::ColumnMajorMatrixVectorProductEmitter 316 EmitOuterLoopBody(b_->getInt64(column_limit), column_remainder, in Emit() 348 llvm::Value* columns_llvm = b_->getInt64(columns); in EmitInnerLoopEpilogue() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
D | parallel_loop_emitter.cc | 79 EmitCallToTargetIntrinsic(TargetIntrinsicID::kBlockIdx, {}, {}, b_); in EmitIndexAndSetExitBasicBlock() 82 block_id = b_->CreateZExtOrTrunc(block_id, index_type, "block_id"); in EmitIndexAndSetExitBasicBlock() 89 EmitCallToTargetIntrinsic(TargetIntrinsicID::kThreadIdx, {}, {}, b_); in EmitIndexAndSetExitBasicBlock() 92 thread_id = b_->CreateZExtOrTrunc(thread_id, index_type, "thread_id"); in EmitIndexAndSetExitBasicBlock() 94 llvm::Value* linear_index_base = b_->CreateAdd( in EmitIndexAndSetExitBasicBlock() 95 b_->CreateMul( in EmitIndexAndSetExitBasicBlock() 113 {b_->CreateICmpULT( in EmitIndexAndSetExitBasicBlock() 119 {}, b_); in EmitIndexAndSetExitBasicBlock() 122 linear_index_base = b_->CreateMul( in EmitIndexAndSetExitBasicBlock() 129 b_->CreateAdd(linear_index_base, base_index, "linear_index_plus_base", in EmitIndexAndSetExitBasicBlock() [all …]
|
D | ir_emitter_unnested.cc | 570 std::vector<llvm::Type*>(args.size(), b_.getInt8PtrTy()), in BuildKernelPrototype() 607 AnnotateFunctionAsGpuKernel(module, kernel, &b_); in BuildKernelPrototype() 618 b_.SetInsertPoint(llvm::ReturnInst::Create(context, entry_bb)); in BuildKernelPrototype() 755 b_.getInt8Ty(), literal.getType().getNumElements() * element_bytes); in EmitConstant() 840 b_.CreateBitCast(source_buffer, b_.getInt8Ty()->getPointerTo()); in EmitPadToStaticFromMlir() 859 llvm::Value* metadata = b_.CreateConstInBoundsGEP1_32( in EmitPadToStaticFromMlir() 860 b_.getInt8Ty(), raw_buffer, raw_data_size + dim_index * sizeof(int32)); in EmitPadToStaticFromMlir() 861 llvm::Value* dyn_dim_size = b_.CreateLoad( in EmitPadToStaticFromMlir() 862 b_.CreateBitCast(metadata, b_.getInt32Ty()->getPointerTo()), in EmitPadToStaticFromMlir() 874 KernelSupportLibrary{&b_}.If("is_thred_0", IsBlock0Thread0(&b_), [&] { in EmitPadToStaticFromMlir() [all …]
|
D | ir_emitter.cc | 83 b_(module_->getContext()), in IrEmitter() 85 &ir_emitter_context->buffer_assignment(), &b_, module_, in IrEmitter() 95 .EmitReadArrayElement(index, &b_, operand->name()); in DefaultAction() 99 *hlo, GpuElementalIrEmitter(hlo_module_config_, module_, &b_, in DefaultAction() 113 llvm::ArrayType::get(b_.getInt8Ty(), literal.size_bytes()); in EmitConstants() 191 /*alignment=*/1, GetBasePointer(*operand), &b_)); in HandleGetTupleElement() 220 llvm_ir::EmitTuple(GetIrArray(*tuple, *tuple), base_ptrs, &b_); in HandleTuple() 245 [this](llvm::Value* arg) { return AddrCastToDefault(arg, b_); }); in EmitCallToNestedComputation() 247 llvm::Value* casted_output = AddrCastToDefault(output, b_); in EmitCallToNestedComputation() 403 llvm::Type* atomic_type = b_.getIntNTy(atomic_size); in EmitAtomicOperationUsingCAS() [all …]
|
D | ir_emitter_nested.cc | 82 llvm::FunctionType::get(b_.getVoidTy(), argument_types, false); in CodegenNestedComputation() 107 b_.SetInsertPoint(ret_instr); in CodegenNestedComputation() 120 b_.SetInsertPoint(ret_instr); in CodegenNestedComputation() 148 /*alignment=*/1, tuple_ptr, &b_); in CodegenNestedComputation() 152 /*alignment=*/1, root_value, &b_); in CodegenNestedComputation() 157 b_.SetInsertPoint(ret_instr); in CodegenNestedComputation() 175 llvm_ir::LoopEmitter(element_generator, target_arrays, &b_).EmitLoop()); in EmitTargetElementLoop() 176 llvm_ir::EmitTuple(GetIrArray(hlo, hlo), target_arrays, &b_); in EmitTargetElementLoop() 179 return llvm_ir::LoopEmitter(element_generator, GetIrArray(hlo, hlo), &b_) in EmitTargetElementLoop()
|
/external/flatbuffers/tests/namespace_test/ |
D | namespace_test2_generated.lobster | 29 b_:flatbuffers_builder 31 b_.StartObject(3) 34 b_.PrependUOffsetTRelativeSlot(0, foo_table) 37 b_.PrependInt8Slot(1, foo_enum, 0) 40 b_.PrependStructSlot(2, foo_struct) 43 return b_.EndObject() 58 b_:flatbuffers_builder 60 b_.StartObject(2) 63 b_.PrependUOffsetTRelativeSlot(0, refer_to_a1) 66 b_.PrependUOffsetTRelativeSlot(1, refer_to_a2) [all …]
|
D | namespace_test1_generated.lobster | 22 b_:flatbuffers_builder 24 b_.StartObject(1) 27 b_.PrependInt32Slot(0, foo, 0) 30 return b_.EndObject() 38 def CreateStructInNestedNS(b_:flatbuffers_builder, a:int, b:int): 39 b_.Prep(4, 8) 40 b_.PrependInt32(b) 41 b_.PrependInt32(a) 42 return b_.Offset()
|
/external/zstd/contrib/pzstd/utils/ |
D | Range.h | 49 Iter b_; 60 constexpr Range() : b_(), e_() {} 61 constexpr Range(Iter begin, Iter end) : b_(begin), e_(end) {} 63 constexpr Range(Iter begin, size_type size) : b_(begin), e_(begin + size) {} 66 /* implicit */ Range(Iter str) : b_(str), e_(str + std::strlen(str)) {} 70 : b_(str.data()), e_(b_ + str.size()) {} 81 : b_(other.begin()), e_(other.end()) {} 90 return e_ - b_; 93 return b_ == e_; 96 return b_; [all …]
|
/external/fmtlib/test/ |
D | gtest-extra-test.cc | 39 b_ = 0; in SingleEvaluationTest() 46 static int b_; member in __anon043ecb0d0111::SingleEvaluationTest 52 int SingleEvaluationTest::b_; member in __anon043ecb0d0111::SingleEvaluationTest 86 std::exception, (b_++, "test")); in TEST_F() 88 EXPECT_EQ(1, b_); in TEST_F() 96 std::logic_error, (b_++, "test")), in TEST_F() 99 EXPECT_EQ(2, b_); in TEST_F() 107 std::exception, (b_++, "other")), in TEST_F() 110 EXPECT_EQ(3, b_); in TEST_F() 114 EXPECT_THROW_MSG(a_++, std::exception, (b_++, "test")), "throws nothing"); in TEST_F() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
D | elemental_ir_emitter.cc | 260 return b_->CreateZExt( in EmitIntegerUnaryOp() 273 F32, module_, b_), in EmitIntegerUnaryOp() 274 b_); in EmitIntegerUnaryOp() 277 module_, b_); in EmitIntegerUnaryOp() 328 auto is_zero_undef = b_->getFalse(); in EmitIntegerUnaryOp() 331 {operand_value->getType()}, b_); in EmitIntegerUnaryOp() 350 return b_->CreateZExt(Not(Trunc(operand_value, b_->getInt1Ty())), in EmitIntegerUnaryOp() 360 {operand_value->getType()}, b_); in EmitIntegerUnaryOp() 380 operand_value = EmitBF16ToF32(operand_value, b_); in EmitFloatUnaryOp() 402 operand_value = b_->CreateFPCast( in EmitFloatUnaryOp() [all …]
|
/external/protobuf/src/google/protobuf/stubs/ |
D | common_unittest.cc | 175 void SetABMethod(int a, const char* b) { a_ = a; b_ = b; } in SetABMethod() 178 current_instance_->b_ = b; in SetABFunction() 184 b_ = nullptr; in SetUp() 194 const char* b_; member in google::protobuf::__anon257d15b70111::ClosureTest 252 EXPECT_NE(cstr, b_); in TEST_F() 255 EXPECT_EQ(cstr, b_); in TEST_F() 263 EXPECT_NE(cstr, b_); in TEST_F() 266 EXPECT_EQ(cstr, b_); in TEST_F() 321 EXPECT_NE(cstr, b_); in TEST_F() 324 EXPECT_EQ(cstr, b_); in TEST_F() [all …]
|
/external/tensorflow/tensorflow/compiler/xla/service/llvm_ir/ |
D | kernel_support_library.cc | 25 return IfWithStatus(b_->CreateICmpSLT(start, end), [&]() -> Status { in ForWithStatus() 28 name, b_->CreateAdd(start, step), end, step, in ForWithStatus() 42 return for_body_generator(indvar, b_->getInt1(is_first_iteration)); in ForWithStatus() 46 name, start, end, step, b_, in ForWithStatus() 49 b_->SetInsertPoint(&loop->GetBodyBasicBlock()->back()); in ForWithStatus() 52 /*is_first_iteration=*/b_->CreateICmpEQ( in ForWithStatus() 54 llvm_ir::SetToLastInsertPoint(loop->GetExitBasicBlock(), b_); in ForWithStatus() 64 llvm_ir::EmitIfThenElse(condition, name, b_, in IfWithStatus() 66 b_->SetInsertPoint(&if_data.true_block->back()); in IfWithStatus() 69 b_->SetInsertPoint(&if_data.false_block->back()); in IfWithStatus() [all …]
|
D | kernel_support_library.h | 40 : b_(b), in b_() function 75 return ForWithStatus(name, /*start=*/b_->getInt64(start), in ForWithStatus() 76 /*end=*/b_->getInt64(end), in ForWithStatus() 77 /*step=*/b_->getInt64(step), for_body_generator); in ForWithStatus() 84 For(name, /*start=*/b_->getInt64(start), in For() 85 /*end=*/b_->getInt64(end), in For() 86 /*step=*/b_->getInt64(step), for_body_generator); in For() 187 return ForWithStatus(name, /*start=*/b_->getInt64(start), in ForWithStatus() 188 /*end=*/b_->getInt64(end), in ForWithStatus() 189 /*step=*/b_->getInt64(step), for_body_generator); in ForWithStatus() [all …]
|
/external/boringssl/src/crypto/chacha/asm/ |
D | chacha-x86.pl | 54 ($b,$b_)=("ebx","ebp"); 92 &mov (&DWP(4*$bp,"esp"),$b_) if ($i!=0); 98 &mov ($b_,&DWP(4*$bn,"esp")) if ($i<7); 99 &mov ($b_,&DWP(128,"esp")) if ($i==7); # loop counter 109 &add ($a,$b_) if ($i<7); # elsewhere 112 ($b,$b_)=($b_,$b); 180 &mov ($b_,&DWP(64+4*6,"esp")); 186 &mov (&DWP(4*6,"esp"),$b_); 195 &mov ($b_,&DWP(64+4*4,"esp")); 207 &add ($a,$b_); # elsewhere [all …]
|
/external/rust/crates/ring/crypto/chacha/asm/ |
D | chacha-x86.pl | 54 ($b,$b_)=("ebx","ebp"); 92 &mov (&DWP(4*$bp,"esp"),$b_) if ($i!=0); 98 &mov ($b_,&DWP(4*$bn,"esp")) if ($i<7); 99 &mov ($b_,&DWP(128,"esp")) if ($i==7); # loop counter 109 &add ($a,$b_) if ($i<7); # elsewhere 112 ($b,$b_)=($b_,$b); 180 &mov ($b_,&DWP(64+4*6,"esp")); 186 &mov (&DWP(4*6,"esp"),$b_); 195 &mov ($b_,&DWP(64+4*4,"esp")); 207 &add ($a,$b_); # elsewhere [all …]
|
/external/freetype/src/base/ |
D | ftcalc.c | 174 FT_Long b_, in FT_MulDiv() argument 183 b = (FT_UInt64)b_; in FT_MulDiv() 187 FT_MOVE_SIGN( b_, b, s ); in FT_MulDiv() 203 FT_Long b_, in FT_MulDiv_No_Round() argument 212 b = (FT_UInt64)b_; in FT_MulDiv_No_Round() 216 FT_MOVE_SIGN( b_, b, s ); in FT_MulDiv_No_Round() 232 FT_Long b_ ) in FT_MulFix() argument 236 return FT_MULFIX_ASSEMBLER( (FT_Int32)a_, (FT_Int32)b_ ); in FT_MulFix() 240 FT_Int64 ab = (FT_Int64)a_ * (FT_Int64)b_; in FT_MulFix() 253 FT_Long b_ ) in FT_DivFix() argument [all …]
|
/external/bzip2/ |
D | bzip2.1.preformatted | 19 _b_z_i_p_2 compresses files using the Burrows‐Wheeler block 29 _b_z_i_p_2 expects a list of file names to accompany the com 41 _b_z_i_p_2 and _b_u_n_z_i_p_2 will by default not overwrite existing 44 If no file names are specified, _b_z_i_p_2 compresses from 45 standard input to standard output. In this case, _b_z_i_p_2 50 _b_u_n_z_i_p_2 (or _b_z_i_p_2 _−_d_) decompresses all specified files. 51 Files which were not created by _b_z_i_p_2 will be detected and 52 ignored, and a warning issued. _b_z_i_p_2 attempts to guess 63 …_._b_z_2_, _._b_z_, _._t_b_z_2 or _._t_b_z_, _b_z_i_p_2 complains that i… 70 _b_u_n_z_i_p_2 will correctly decompress a file which is the con [all …]
|
/external/tensorflow/tensorflow/core/lib/gtl/ |
D | flatset.h | 111 const_iterator() : b_(nullptr), end_(nullptr), i_(0) {} in const_iterator() 114 const_iterator(Bucket* b, Bucket* end) : b_(b), end_(end), i_(0) { in const_iterator() 120 : b_(b), end_(end), i_(i) {} in const_iterator() 125 return b_ == x.b_ && i_ == x.i_; 129 DCHECK(b_ != end_); 142 Bucket* b_; variable 146 reference key() const { return b_->key(i_); } in key() 148 while (b_ < end_) { in SkipUnused() 151 b_++; in SkipUnused() 152 } else if (b_->marker[i_] < 2) { in SkipUnused() [all …]
|
/external/libtextclassifier/abseil-cpp/absl/random/ |
D | beta_distribution.h | 72 b_ = alpha; in param_type() 76 b_ = beta; in param_type() 78 if (a_ <= 1 && b_ >= ThresholdForLargeA()) { in param_type() 86 if ((b_ < 1.0 && a_ + b_ <= 1.2) || a_ <= ThresholdForSmallA()) { in param_type() 91 b_ = result_type(1) / beta_; in param_type() 92 if (std::isinf(a_) || std::isinf(b_)) { in param_type() 102 result_type r = a_ / b_; in param_type() 106 x_ = a_ + b_; in param_type() 115 result_type r = (a_ - 1) / (b_ - 1); in param_type() 116 y_ = std::sqrt((1 + r) / (b_ * r * 2 - r + 1)); in param_type() [all …]
|