/external/tensorflow/tensorflow/compiler/xla/service/ |
D | call_graph.cc | 37 string CallContextToString(CallContext context) { in CallContextToString() 39 case CallContext::kNone: in CallContextToString() 41 case CallContext::kSequential: in CallContextToString() 43 case CallContext::kParallel: in CallContextToString() 45 case CallContext::kBoth: in CallContextToString() 50 std::ostream& operator<<(std::ostream& out, const CallContext& context) { in operator <<() 55 CallContext GetInstructionCallContext(HloOpcode opcode) { in GetInstructionCallContext() 60 return CallContext::kSequential; in GetInstructionCallContext() 69 return CallContext::kParallel; in GetInstructionCallContext() 71 return CallContext::kNone; in GetInstructionCallContext() [all …]
|
D | call_graph.h | 32 enum class CallContext { enum 53 string CallContextToString(CallContext context); 54 std::ostream& operator<<(std::ostream& out, const CallContext& context); 56 CallContext GetInstructionCallContext(HloOpcode opcode); 63 CallContext context) in CallSite() 77 CallContext context() const { return context_; } in context() 89 const CallContext context_; 122 CallContext context() const { return context_; } in context() 136 void set_context(CallContext value) { context_ = value; } in set_context() 174 CallContext context_ = CallContext::kNone;
|
D | flatten_call_graph.cc | 77 if (call_site.context() == CallContext::kParallel) { in FlattenNode() 80 CHECK_EQ(call_site.context(), CallContext::kSequential); in FlattenNode() 84 if (node.context() != CallContext::kBoth && i == 0) { in FlattenNode() 100 CallContext::kSequential) { in FlattenNode()
|
D | call_graph_test.cc | 111 EXPECT_EQ(CallContext::kSequential, node.context()); in TEST_F() 129 EXPECT_EQ(CallContext::kSequential, entry_node.context()); in TEST_F() 135 EXPECT_EQ(CallContext::kSequential, unreachable_node.context()); in TEST_F() 153 EXPECT_EQ(CallContext::kSequential, entry_node.context()); in TEST_F() 162 EXPECT_EQ(CallContext::kParallel, map_node.context()); in TEST_F() 187 EXPECT_EQ(CallContext::kSequential, entry_node.context()); in TEST_F() 195 EXPECT_EQ(CallContext::kSequential, called_node.context()); in TEST_F() 232 EXPECT_EQ(CallContext::kSequential, call_callsite.context()); in TEST_F() 239 EXPECT_EQ(CallContext::kParallel, map_callsite.context()); in TEST_F() 244 EXPECT_EQ(CallContext::kBoth, sub_node.context()); in TEST_F() [all …]
|
D | hlo_liveness_analysis.cc | 201 if (call_graph_node.context() == CallContext::kSequential) { in PropagateLivenessToParameterCallers() 231 if (call_graph_node.context() == CallContext::kSequential) { in PropagateLivenessThroughControlFlow()
|
D | hlo_dataflow_analysis.cc | 476 if (call_graph_node.context() == CallContext::kParallel || in UpdateParameterValueSet() 480 CHECK_EQ(call_graph_node.context(), CallContext::kSequential); in UpdateParameterValueSet() 692 if (call_graph_node.context() == CallContext::kSequential) { in Propagate() 785 if (call_graph_node.context() == CallContext::kBoth) { in InitializeInstructionValueSets() 797 call_graph_node.context() == CallContext::kParallel) { in InitializeInstructionValueSets()
|
D | hlo_rematerialization.cc | 960 if (callsite == nullptr || callsite->context() == CallContext::kParallel) { in CalledComputationsMemoryUsage() 1133 callsite->context() == CallContext::kSequential && in RematerializeComputation() 1240 if (node.context() == CallContext::kSequential) { in Run()
|
D | copy_insertion.cc | 1027 if (node.context() == CallContext::kParallel) { in AddSpecialCaseCopies() 1030 TF_RET_CHECK(node.context() == CallContext::kSequential); in AddSpecialCaseCopies() 1221 if (node.context() == CallContext::kSequential && in IsWhileBody()
|
/external/llvm/lib/Target/X86/ |
D | X86CallFrameOptimization.cpp | 55 struct CallContext { struct in __anone1509d7a0111::X86CallFrameOptimization 56 CallContext() in CallContext() function 82 typedef SmallVector<CallContext, 8> ContextVector; 89 MachineBasicBlock::iterator I, CallContext &Context); 91 void adjustCallSequence(MachineFunction &MF, const CallContext &Context); 240 CallContext Context; in runOnMachineFunction() 320 CallContext &Context) { in collectCallInfo() 452 const CallContext &Context) { in adjustCallSequence()
|
/external/swiftshader/third_party/llvm-7.0/llvm/lib/Target/X86/ |
D | X86CallFrameOptimization.cpp | 78 struct CallContext { struct in __anon26f419d20111::X86CallFrameOptimization 79 CallContext() : FrameSetup(nullptr), ArgStoreVector(4, nullptr) {} in CallContext() argument 103 typedef SmallVector<CallContext, 8> ContextVector; 110 MachineBasicBlock::iterator I, CallContext &Context); 112 void adjustCallSequence(MachineFunction &MF, const CallContext &Context); 258 CallContext Context; in runOnMachineFunction() 354 CallContext &Context) { in collectCallInfo() 492 const CallContext &Context) { in adjustCallSequence()
|
/external/v8/gnparser/ |
D | gnParser.py | 189 return self.getTypedRuleContext(gnParser.CallContext,0) 362 class CallContext(ParserRuleContext): class in gnParser 365 super(gnParser.CallContext, self).__init__(parent, invokingState) 395 localctx = gnParser.CallContext(self, self._ctx, self.state) 879 return self.getTypedRuleContext(gnParser.CallContext,0)
|
/external/harfbuzz_ng/src/ |
D | hb-cff-interp-cs-common.hh | 42 struct CallContext struct 60 struct CallStack : Stack<CallContext, kMaxCallLimit> {}; 205 CallContext context;
|
D | hb-subset-cff-common.hh | 495 ParsedCStr *get_parsed_str_for_context (CallContext &context) in get_parsed_str_for_context()
|