Home
last modified time | relevance | path

Searched refs:num_tokens (Results 1 – 25 of 58) sorted by relevance

123

/external/tensorflow/tensorflow/python/keras/layers/preprocessing/
Dcategory_encoding_test.py59 num_tokens = 6
63 num_tokens=num_tokens, output_mode=category_encoding.COUNT, sparse=True)
73 num_tokens=num_tokens,
91 num_tokens = 6
92 expected_output_shape = [None, num_tokens]
97 num_tokens=num_tokens, output_mode=category_encoding.BINARY)
115 num_tokens = 6
116 expected_output_shape = [None, num_tokens]
122 num_tokens=num_tokens, output_mode=category_encoding.COUNT)
145 num_tokens = 6
[all …]
Dcategory_encoding.py106 num_tokens=None, argument
115 num_tokens = kwargs["max_tokens"]
127 if num_tokens is None:
131 if num_tokens < 1:
134 self.num_tokens = num_tokens
139 return tensor_shape.TensorShape([input_shape[0], self.num_tokens])
152 "num_tokens": self.num_tokens,
169 out_depth = self.num_tokens
Dcategory_encoding_distribution_test.py67 num_tokens = 6
73 num_tokens=num_tokens, output_mode=category_encoding.BINARY)
/external/jsmn/
Djsmn.h106 jsmntok_t *tokens, const unsigned int num_tokens);
113 const size_t num_tokens) { in jsmn_alloc_token() argument
115 if (parser->toknext >= num_tokens) { in jsmn_alloc_token()
143 const size_t num_tokens) { in jsmn_parse_primitive() argument
180 token = jsmn_alloc_token(parser, tokens, num_tokens); in jsmn_parse_primitive()
198 const size_t num_tokens) { in jsmn_parse_string() argument
214 token = jsmn_alloc_token(parser, tokens, num_tokens); in jsmn_parse_string()
272 jsmntok_t *tokens, const unsigned int num_tokens) { in jsmn_parse() argument
290 token = jsmn_alloc_token(parser, tokens, num_tokens); in jsmn_parse()
365 r = jsmn_parse_string(parser, js, len, tokens, num_tokens); in jsmn_parse()
[all …]
/external/tensorflow/tensorflow/python/training/
Dsync_replicas_optimizer.py408 def get_init_tokens_op(self, num_tokens=-1): argument
433 if num_tokens == -1:
434 num_tokens = self._replicas_to_aggregate
435 elif num_tokens < tokens_needed:
438 (num_tokens, tokens_needed))
440 if num_tokens > 0:
442 tokens = array_ops.fill([num_tokens], self._global_step)
449 def make_session_run_hook(self, is_chief, num_tokens=-1): argument
451 return _SyncReplicasOptimizerHook(self, is_chief, num_tokens)
457 def __init__(self, sync_optimizer, is_chief, num_tokens): argument
[all …]
/external/icing/icing/util/
Dtokenized-document.h49 int32_t num_tokens() const { in num_tokens() function
50 int32_t num_tokens = 0; in num_tokens() local
52 num_tokens += section.token_sequence.size(); in num_tokens()
54 return num_tokens; in num_tokens()
/external/libtextclassifier/native/actions/
Dngram-model.cc92 bool NGramSensitiveModel::IsNGramMatch(const uint32* tokens, size_t num_tokens, in IsNGramMatch() argument
97 for (; token_idx < num_tokens && ngram_token_idx < num_ngram_tokens;) { in IsNGramMatch()
117 uint64 NGramSensitiveModel::GetNumSkipGrams(int num_tokens, in GetNumSkipGrams() argument
121 uint64 total = num_tokens; in GetNumSkipGrams()
123 ngram_len <= max_ngram_length && ngram_len <= num_tokens; ++ngram_len) { in GetNumSkipGrams()
134 if (len_i > num_tokens) continue; in GetNumSkipGrams()
137 const int num_start_offsets = num_tokens - len_i + 1; in GetNumSkipGrams()
Dngram-model.h48 static uint64 GetNumSkipGrams(int num_tokens, int max_ngram_length,
61 bool IsNGramMatch(const uint32* tokens, size_t num_tokens,
/external/libtextclassifier/native/utils/tflite/
Dtoken_encoder.cc76 const TfLiteTensor& num_tokens = in Prepare() local
78 TF_LITE_ENSURE_EQ(context, num_tokens.dims->size, kEncoderInputRank); in Prepare()
79 TF_LITE_ENSURE_EQ(context, num_tokens.dims->data[0], kEncoderBatchSize); in Prepare()
120 const TfLiteTensor& num_tokens = in Eval() local
143 num_tokens.data.i32[i] + 2; /* num_tokens + start and end token. */ in Eval()
/external/tensorflow/tensorflow/core/kernels/
Dstring_ngrams_op.cc152 int num_tokens = ngram_width - (left_padding + right_padding); in CreateNgrams() local
161 for (int n = 0; n < num_tokens; ++n) { in CreateNgrams()
167 int num_separators = left_padding + right_padding + num_tokens - 1; in CreateNgrams()
177 for (int n = 0; n < num_tokens - 1; ++n) { in CreateNgrams()
181 ngram->append(data[data_start_index + num_tokens - 1]); in CreateNgrams()
/external/e2fsprogs/contrib/android/
De2fsdroid.c71 size_t num_tokens; in parse_ugid_map_entry() local
75 for (token = strtok_r(line, " ", &token_saveptr), num_tokens = 0; in parse_ugid_map_entry()
76 token && num_tokens < 3; in parse_ugid_map_entry()
77 token = strtok_r(NULL, " ", &token_saveptr), ++num_tokens) { in parse_ugid_map_entry()
84 *parsed[num_tokens] = (unsigned int) t; in parse_ugid_map_entry()
86 if (num_tokens < 3 || strtok_r(NULL, " ", &token_saveptr) != NULL) { in parse_ugid_map_entry()
/external/icing/icing/index/
Dindex-processor.cc68 uint32_t num_tokens = 0; in IndexDocument() local
76 if (++num_tokens > options_.max_tokens_per_document) { in IndexDocument()
123 num_tokens); in IndexDocument()
/external/mesa3d/src/mesa/state_tracker/
Dst_shader_cache.c68 unsigned num_tokens = tgsi_num_tokens(tokens); in write_tgsi_to_cache() local
70 blob_write_uint32(blob, num_tokens); in write_tgsi_to_cache()
71 blob_write_bytes(blob, tokens, num_tokens * sizeof(struct tgsi_token)); in write_tgsi_to_cache()
167 unsigned num_tokens = blob_read_uint32(blob_reader); in read_tgsi_from_cache() local
168 unsigned tokens_size = num_tokens * sizeof(struct tgsi_token); in read_tgsi_from_cache()
/external/libtextclassifier/native/annotator/pod_ner/
Dutils_test.cc701 int first_token_index, num_tokens; in TEST() local
705 &num_tokens); in TEST()
708 EXPECT_EQ(num_tokens, 4); in TEST()
713 &num_tokens); in TEST()
716 EXPECT_EQ(num_tokens, 3); in TEST()
721 int first_token_index, num_tokens; in TEST() local
725 &num_tokens); in TEST()
728 EXPECT_EQ(num_tokens, 4); in TEST()
733 int first_token_index, num_tokens; in TEST() local
737 &num_tokens); in TEST()
[all …]
/external/autotest/client/site_tests/platform_Pkcs11Events/
Dplatform_Pkcs11Events.py13 def run_once(self, num_tokens, num_events): argument
15 token_list = ['/tmp/chaps%d' % x for x in range(num_tokens)]
Dcontrol19 job.run_test('platform_Pkcs11Events', num_tokens=2, num_events=20)
Dcontrol.regression19 job.run_test('platform_Pkcs11Events', num_tokens=7, num_events=200)
/external/tensorflow/tensorflow/python/keras/layers/preprocessing/benchmarks/
Dcategory_encoding_benchmark.py75 for num_tokens in [100, 1000, 20000]:
80 max_tokens=num_tokens)
/external/tensorflow/tensorflow/tools/api/golden/v1/
Dtensorflow.train.-sync-replicas-optimizer.pbtxt37 argspec: "args=[\'self\', \'num_tokens\'], varargs=None, keywords=None, defaults=[\'-1\'], "
53 …argspec: "args=[\'self\', \'is_chief\', \'num_tokens\'], varargs=None, keywords=None, defaults=[\'…
/external/libtextclassifier/native/annotator/
Dcached-features_test.cc40 std::unique_ptr<std::vector<float>> MakeFeatures(int num_tokens) { in MakeFeatures() argument
42 for (int i = 1; i <= num_tokens; ++i) { in MakeFeatures()
/external/virglrenderer/src/gallium/auxiliary/tgsi/
Dtgsi_text.h43 uint num_tokens );
Dtgsi_parse.c292 tgsi_alloc_tokens(unsigned num_tokens) in tgsi_alloc_tokens() argument
294 unsigned bytes = num_tokens * sizeof(struct tgsi_token); in tgsi_alloc_tokens()
/external/mesa3d/src/gallium/auxiliary/tgsi/
Dtgsi_text.h43 uint num_tokens );
/external/mesa3d/src/gallium/drivers/nouveau/codegen/
Dnv50_ir_serialize.cpp35 unsigned int num_tokens = tgsi_num_tokens(tokens); in nv50_ir_prog_info_serialize() local
37 blob_write_uint32(blob, num_tokens); in nv50_ir_prog_info_serialize()
38 blob_write_bytes(blob, tokens, num_tokens * sizeof(struct tgsi_token)); in nv50_ir_prog_info_serialize()
/external/icing/icing/store/
Ddocument-store.h165 const DocumentProto& document, int32_t num_tokens = 0,
168 DocumentProto&& document, int32_t num_tokens = 0,

123