/external/libtextclassifier/native/annotator/grammar/ |
D | grammar-annotator.cc | 116 const grammar::TextContext& input_context, in InstantiateAnnotatedSpanFromDerivation() argument 123 input_context, parse_tree, interpretation, &classification)) { in InstantiateAnnotatedSpanFromDerivation() 132 const grammar::TextContext& input_context, in InstantiateClassificationFromDerivation() argument 181 input_context.Span(capturing_match->codepoint_span); in InstantiateClassificationFromDerivation() 204 grammar::TextContext input_context = in Annotate() local 210 analyzer_.parser().Parse(input_context, &arena))) { in Annotate() 218 input_context, derivation.parse_tree, interpretation, in Annotate() 235 grammar::TextContext input_context = in SuggestSelection() local 244 selection, analyzer_.parser().Parse(input_context, &arena), in SuggestSelection() 263 return InstantiateAnnotatedSpanFromDerivation(input_context, best_match, in SuggestSelection() [all …]
|
D | grammar-annotator.h | 68 const grammar::TextContext& input_context, 75 const grammar::TextContext& input_context,
|
/external/tensorflow/tensorflow/python/keras/utils/ |
D | dataset_creator_test.py | 64 def dataset_fn(input_context): argument 66 batch_size = input_context.get_per_replica_batch_size(global_batch_size) 68 dataset = dataset.shard(input_context.num_input_pipelines, 69 input_context.input_pipeline_id)
|
/external/tensorflow/tensorflow/python/distribute/ |
D | input_lib_test.py | 71 input_context=None): argument 77 input_context, 102 input_context=input_context) 111 input_context=None): argument 119 input_context=input_context) 126 input_context=input_context) 172 input_context=None): argument 193 input_context=input_context) 202 input_context=input_context) 659 input_context=distribution.extended._make_input_context()) [all …]
|
D | distribute_lib_test.py | 56 def _test_input_fn(input_context): argument 57 del input_context 562 input_context = distribute_lib.InputContext( 564 self.assertEqual(6, input_context.num_replicas_in_sync) 565 self.assertEqual(1, input_context.input_pipeline_id) 566 self.assertEqual(2, input_context.num_input_pipelines) 569 input_context = distribute_lib.InputContext( 571 self.assertEqual(2, input_context.get_per_replica_batch_size(12)) 573 input_context.get_per_replica_batch_size(13) 576 input_context = distribute_lib.InputContext( [all …]
|
D | strategy_common_test.py | 200 def dataset_fn(input_context): argument 202 batch_size = input_context.get_per_replica_batch_size(global_batch_size) 204 return d.shard(input_context.num_input_pipelines, 205 input_context.input_pipeline_id) 244 def dataset_fn(input_context): argument 246 batch_size = input_context.get_per_replica_batch_size(global_batch_size) 249 return dataset.shard(input_context.num_input_pipelines, 250 input_context.input_pipeline_id)
|
D | collective_all_reduce_strategy.py | 587 input_context = distribute_lib.InputContext( 591 return input_context 601 input_context = self._make_input_context() 607 input_context=input_context) 617 input_context = self._make_input_context() 621 input_contexts=[input_context], 637 input_context = self._make_input_context() 643 input_context=input_context) 650 input_context = self._make_input_context() 652 [input_context],
|
D | input_lib.py | 68 input_context=None): argument 101 input_context=input_context) 108 input_context=input_context) 967 input_context=None): argument 1001 num_workers = input_context.num_input_pipelines if input_context else len( 1009 if input_context: 1013 dataset = rebatch_fn(dataset, input_context.input_pipeline_id) 1015 input_context.num_input_pipelines, 1016 input_context.input_pipeline_id, 1152 input_context=None): argument [all …]
|
D | parameter_server_strategy.py | 368 input_context = distribute_lib.InputContext( 373 [input_context], 390 input_context = distribute_lib.InputContext( 398 [input_context],
|
D | input_lib_type_spec_test.py | 288 def dataset_fn(input_context): argument 290 dataset = dataset.shard(input_context.num_input_pipelines, 291 input_context.input_pipeline_id) 294 input_context.get_per_replica_batch_size(4))
|
D | strategy_test_lib.py | 322 def _input_fn(input_context): argument 324 self.assertIsNotNone(input_context) 326 input_context.num_replicas_in_sync) 328 input_context.num_input_pipelines) 331 input_context.input_pipeline_id) 333 self.assertEqual(worker_id_counter[0], input_context.input_pipeline_id)
|
/external/tensorflow/tensorflow/python/keras/integration_test/ |
D | multi_worker_tutorial_test.py | 97 def dataset_fn(self, global_batch_size, input_context): argument 98 batch_size = input_context.get_per_replica_batch_size(global_batch_size) 100 dataset = dataset.shard(input_context.num_input_pipelines, 101 input_context.input_pipeline_id) 156 lambda input_context: self.dataset_fn(global_batch_size, input_context 255 … lambda input_context: self.dataset_fn(global_batch_size, # pylint: disable=g-long-lambda 256 input_context))
|
/external/libtextclassifier/native/actions/ |
D | tflite-sensitive-model.cc | 78 if (model_config_->model_spec()->input_context() >= 0) { in EvalConversation() 83 model_config_->model_spec()->input_context(), context, in EvalConversation()
|
D | actions-suggestions.cc | 604 if (model_->tflite_model_spec()->input_context() >= 0) { in AllocateInput() 606 interpreter->inputs()[model_->tflite_model_spec()->input_context()], in AllocateInput() 684 if (model_->tflite_model_spec()->input_context() >= 0) { in SetupModelInput() 687 model_->tflite_model_spec()->input_context(), in SetupModelInput() 694 model_->tflite_model_spec()->input_context(), context, interpreter); in SetupModelInput()
|
D | actions_model.fbs | 66 input_context:int = 1; 120 // If greater than 0, pad or truncate the input_user_id and input_context
|
/external/tensorflow/tensorflow/python/keras/distribute/ |
D | parameter_server_training_test.py | 281 def dataset_fn(input_context): argument 282 del input_context
|