Home
last modified time | relevance | path

Searched refs:input_context (Results 1 – 16 of 16) sorted by relevance

/external/libtextclassifier/native/annotator/grammar/
Dgrammar-annotator.cc116 const grammar::TextContext& input_context, in InstantiateAnnotatedSpanFromDerivation() argument
123 input_context, parse_tree, interpretation, &classification)) { in InstantiateAnnotatedSpanFromDerivation()
132 const grammar::TextContext& input_context, in InstantiateClassificationFromDerivation() argument
181 input_context.Span(capturing_match->codepoint_span); in InstantiateClassificationFromDerivation()
204 grammar::TextContext input_context = in Annotate() local
210 analyzer_.parser().Parse(input_context, &arena))) { in Annotate()
218 input_context, derivation.parse_tree, interpretation, in Annotate()
235 grammar::TextContext input_context = in SuggestSelection() local
244 selection, analyzer_.parser().Parse(input_context, &arena), in SuggestSelection()
263 return InstantiateAnnotatedSpanFromDerivation(input_context, best_match, in SuggestSelection()
[all …]
Dgrammar-annotator.h68 const grammar::TextContext& input_context,
75 const grammar::TextContext& input_context,
/external/tensorflow/tensorflow/python/keras/utils/
Ddataset_creator_test.py64 def dataset_fn(input_context): argument
66 batch_size = input_context.get_per_replica_batch_size(global_batch_size)
68 dataset = dataset.shard(input_context.num_input_pipelines,
69 input_context.input_pipeline_id)
/external/tensorflow/tensorflow/python/distribute/
Dinput_lib_test.py71 input_context=None): argument
77 input_context,
102 input_context=input_context)
111 input_context=None): argument
119 input_context=input_context)
126 input_context=input_context)
172 input_context=None): argument
193 input_context=input_context)
202 input_context=input_context)
659 input_context=distribution.extended._make_input_context())
[all …]
Ddistribute_lib_test.py56 def _test_input_fn(input_context): argument
57 del input_context
562 input_context = distribute_lib.InputContext(
564 self.assertEqual(6, input_context.num_replicas_in_sync)
565 self.assertEqual(1, input_context.input_pipeline_id)
566 self.assertEqual(2, input_context.num_input_pipelines)
569 input_context = distribute_lib.InputContext(
571 self.assertEqual(2, input_context.get_per_replica_batch_size(12))
573 input_context.get_per_replica_batch_size(13)
576 input_context = distribute_lib.InputContext(
[all …]
Dstrategy_common_test.py200 def dataset_fn(input_context): argument
202 batch_size = input_context.get_per_replica_batch_size(global_batch_size)
204 return d.shard(input_context.num_input_pipelines,
205 input_context.input_pipeline_id)
244 def dataset_fn(input_context): argument
246 batch_size = input_context.get_per_replica_batch_size(global_batch_size)
249 return dataset.shard(input_context.num_input_pipelines,
250 input_context.input_pipeline_id)
Dcollective_all_reduce_strategy.py587 input_context = distribute_lib.InputContext(
591 return input_context
601 input_context = self._make_input_context()
607 input_context=input_context)
617 input_context = self._make_input_context()
621 input_contexts=[input_context],
637 input_context = self._make_input_context()
643 input_context=input_context)
650 input_context = self._make_input_context()
652 [input_context],
Dinput_lib.py68 input_context=None): argument
101 input_context=input_context)
108 input_context=input_context)
967 input_context=None): argument
1001 num_workers = input_context.num_input_pipelines if input_context else len(
1009 if input_context:
1013 dataset = rebatch_fn(dataset, input_context.input_pipeline_id)
1015 input_context.num_input_pipelines,
1016 input_context.input_pipeline_id,
1152 input_context=None): argument
[all …]
Dparameter_server_strategy.py368 input_context = distribute_lib.InputContext(
373 [input_context],
390 input_context = distribute_lib.InputContext(
398 [input_context],
Dinput_lib_type_spec_test.py288 def dataset_fn(input_context): argument
290 dataset = dataset.shard(input_context.num_input_pipelines,
291 input_context.input_pipeline_id)
294 input_context.get_per_replica_batch_size(4))
Dstrategy_test_lib.py322 def _input_fn(input_context): argument
324 self.assertIsNotNone(input_context)
326 input_context.num_replicas_in_sync)
328 input_context.num_input_pipelines)
331 input_context.input_pipeline_id)
333 self.assertEqual(worker_id_counter[0], input_context.input_pipeline_id)
/external/tensorflow/tensorflow/python/keras/integration_test/
Dmulti_worker_tutorial_test.py97 def dataset_fn(self, global_batch_size, input_context): argument
98 batch_size = input_context.get_per_replica_batch_size(global_batch_size)
100 dataset = dataset.shard(input_context.num_input_pipelines,
101 input_context.input_pipeline_id)
156 lambda input_context: self.dataset_fn(global_batch_size, input_context
255 … lambda input_context: self.dataset_fn(global_batch_size, # pylint: disable=g-long-lambda
256 input_context))
/external/libtextclassifier/native/actions/
Dtflite-sensitive-model.cc78 if (model_config_->model_spec()->input_context() >= 0) { in EvalConversation()
83 model_config_->model_spec()->input_context(), context, in EvalConversation()
Dactions-suggestions.cc604 if (model_->tflite_model_spec()->input_context() >= 0) { in AllocateInput()
606 interpreter->inputs()[model_->tflite_model_spec()->input_context()], in AllocateInput()
684 if (model_->tflite_model_spec()->input_context() >= 0) { in SetupModelInput()
687 model_->tflite_model_spec()->input_context(), in SetupModelInput()
694 model_->tflite_model_spec()->input_context(), context, interpreter); in SetupModelInput()
Dactions_model.fbs66 input_context:int = 1;
120 // If greater than 0, pad or truncate the input_user_id and input_context
/external/tensorflow/tensorflow/python/keras/distribute/
Dparameter_server_training_test.py281 def dataset_fn(input_context): argument
282 del input_context