/external/tensorflow/tensorflow/python/keras/layers/preprocessing/ |
D | category_encoding_test.py | 59 num_tokens = 6 63 num_tokens=num_tokens, output_mode=category_encoding.COUNT, sparse=True) 73 num_tokens=num_tokens, 91 num_tokens = 6 92 expected_output_shape = [None, num_tokens] 97 num_tokens=num_tokens, output_mode=category_encoding.BINARY) 115 num_tokens = 6 116 expected_output_shape = [None, num_tokens] 122 num_tokens=num_tokens, output_mode=category_encoding.COUNT) 145 num_tokens = 6 [all …]
|
D | category_encoding.py | 106 num_tokens=None, argument 115 num_tokens = kwargs["max_tokens"] 127 if num_tokens is None: 131 if num_tokens < 1: 134 self.num_tokens = num_tokens 139 return tensor_shape.TensorShape([input_shape[0], self.num_tokens]) 152 "num_tokens": self.num_tokens, 169 out_depth = self.num_tokens
|
D | category_encoding_distribution_test.py | 67 num_tokens = 6 73 num_tokens=num_tokens, output_mode=category_encoding.BINARY)
|
/external/jsmn/ |
D | jsmn.h | 106 jsmntok_t *tokens, const unsigned int num_tokens); 113 const size_t num_tokens) { in jsmn_alloc_token() argument 115 if (parser->toknext >= num_tokens) { in jsmn_alloc_token() 143 const size_t num_tokens) { in jsmn_parse_primitive() argument 180 token = jsmn_alloc_token(parser, tokens, num_tokens); in jsmn_parse_primitive() 198 const size_t num_tokens) { in jsmn_parse_string() argument 214 token = jsmn_alloc_token(parser, tokens, num_tokens); in jsmn_parse_string() 272 jsmntok_t *tokens, const unsigned int num_tokens) { in jsmn_parse() argument 290 token = jsmn_alloc_token(parser, tokens, num_tokens); in jsmn_parse() 365 r = jsmn_parse_string(parser, js, len, tokens, num_tokens); in jsmn_parse() [all …]
|
/external/tensorflow/tensorflow/python/training/ |
D | sync_replicas_optimizer.py | 408 def get_init_tokens_op(self, num_tokens=-1): argument 433 if num_tokens == -1: 434 num_tokens = self._replicas_to_aggregate 435 elif num_tokens < tokens_needed: 438 (num_tokens, tokens_needed)) 440 if num_tokens > 0: 442 tokens = array_ops.fill([num_tokens], self._global_step) 449 def make_session_run_hook(self, is_chief, num_tokens=-1): argument 451 return _SyncReplicasOptimizerHook(self, is_chief, num_tokens) 457 def __init__(self, sync_optimizer, is_chief, num_tokens): argument [all …]
|
/external/icing/icing/util/ |
D | tokenized-document.h | 49 int32_t num_tokens() const { in num_tokens() function 50 int32_t num_tokens = 0; in num_tokens() local 52 num_tokens += section.token_sequence.size(); in num_tokens() 54 return num_tokens; in num_tokens()
|
/external/libtextclassifier/native/actions/ |
D | ngram-model.cc | 92 bool NGramSensitiveModel::IsNGramMatch(const uint32* tokens, size_t num_tokens, in IsNGramMatch() argument 97 for (; token_idx < num_tokens && ngram_token_idx < num_ngram_tokens;) { in IsNGramMatch() 117 uint64 NGramSensitiveModel::GetNumSkipGrams(int num_tokens, in GetNumSkipGrams() argument 121 uint64 total = num_tokens; in GetNumSkipGrams() 123 ngram_len <= max_ngram_length && ngram_len <= num_tokens; ++ngram_len) { in GetNumSkipGrams() 134 if (len_i > num_tokens) continue; in GetNumSkipGrams() 137 const int num_start_offsets = num_tokens - len_i + 1; in GetNumSkipGrams()
|
D | ngram-model.h | 48 static uint64 GetNumSkipGrams(int num_tokens, int max_ngram_length, 61 bool IsNGramMatch(const uint32* tokens, size_t num_tokens,
|
/external/libtextclassifier/native/utils/tflite/ |
D | token_encoder.cc | 76 const TfLiteTensor& num_tokens = in Prepare() local 78 TF_LITE_ENSURE_EQ(context, num_tokens.dims->size, kEncoderInputRank); in Prepare() 79 TF_LITE_ENSURE_EQ(context, num_tokens.dims->data[0], kEncoderBatchSize); in Prepare() 120 const TfLiteTensor& num_tokens = in Eval() local 143 num_tokens.data.i32[i] + 2; /* num_tokens + start and end token. */ in Eval()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | string_ngrams_op.cc | 152 int num_tokens = ngram_width - (left_padding + right_padding); in CreateNgrams() local 161 for (int n = 0; n < num_tokens; ++n) { in CreateNgrams() 167 int num_separators = left_padding + right_padding + num_tokens - 1; in CreateNgrams() 177 for (int n = 0; n < num_tokens - 1; ++n) { in CreateNgrams() 181 ngram->append(data[data_start_index + num_tokens - 1]); in CreateNgrams()
|
/external/e2fsprogs/contrib/android/ |
D | e2fsdroid.c | 71 size_t num_tokens; in parse_ugid_map_entry() local 75 for (token = strtok_r(line, " ", &token_saveptr), num_tokens = 0; in parse_ugid_map_entry() 76 token && num_tokens < 3; in parse_ugid_map_entry() 77 token = strtok_r(NULL, " ", &token_saveptr), ++num_tokens) { in parse_ugid_map_entry() 84 *parsed[num_tokens] = (unsigned int) t; in parse_ugid_map_entry() 86 if (num_tokens < 3 || strtok_r(NULL, " ", &token_saveptr) != NULL) { in parse_ugid_map_entry()
|
/external/icing/icing/index/ |
D | index-processor.cc | 68 uint32_t num_tokens = 0; in IndexDocument() local 76 if (++num_tokens > options_.max_tokens_per_document) { in IndexDocument() 123 num_tokens); in IndexDocument()
|
/external/mesa3d/src/mesa/state_tracker/ |
D | st_shader_cache.c | 68 unsigned num_tokens = tgsi_num_tokens(tokens); in write_tgsi_to_cache() local 70 blob_write_uint32(blob, num_tokens); in write_tgsi_to_cache() 71 blob_write_bytes(blob, tokens, num_tokens * sizeof(struct tgsi_token)); in write_tgsi_to_cache() 167 unsigned num_tokens = blob_read_uint32(blob_reader); in read_tgsi_from_cache() local 168 unsigned tokens_size = num_tokens * sizeof(struct tgsi_token); in read_tgsi_from_cache()
|
/external/libtextclassifier/native/annotator/pod_ner/ |
D | utils_test.cc | 701 int first_token_index, num_tokens; in TEST() local 705 &num_tokens); in TEST() 708 EXPECT_EQ(num_tokens, 4); in TEST() 713 &num_tokens); in TEST() 716 EXPECT_EQ(num_tokens, 3); in TEST() 721 int first_token_index, num_tokens; in TEST() local 725 &num_tokens); in TEST() 728 EXPECT_EQ(num_tokens, 4); in TEST() 733 int first_token_index, num_tokens; in TEST() local 737 &num_tokens); in TEST() [all …]
|
/external/autotest/client/site_tests/platform_Pkcs11Events/ |
D | platform_Pkcs11Events.py | 13 def run_once(self, num_tokens, num_events): argument 15 token_list = ['/tmp/chaps%d' % x for x in range(num_tokens)]
|
D | control | 19 job.run_test('platform_Pkcs11Events', num_tokens=2, num_events=20)
|
D | control.regression | 19 job.run_test('platform_Pkcs11Events', num_tokens=7, num_events=200)
|
/external/tensorflow/tensorflow/python/keras/layers/preprocessing/benchmarks/ |
D | category_encoding_benchmark.py | 75 for num_tokens in [100, 1000, 20000]: 80 max_tokens=num_tokens)
|
/external/tensorflow/tensorflow/tools/api/golden/v1/ |
D | tensorflow.train.-sync-replicas-optimizer.pbtxt | 37 argspec: "args=[\'self\', \'num_tokens\'], varargs=None, keywords=None, defaults=[\'-1\'], " 53 …argspec: "args=[\'self\', \'is_chief\', \'num_tokens\'], varargs=None, keywords=None, defaults=[\'…
|
/external/libtextclassifier/native/annotator/ |
D | cached-features_test.cc | 40 std::unique_ptr<std::vector<float>> MakeFeatures(int num_tokens) { in MakeFeatures() argument 42 for (int i = 1; i <= num_tokens; ++i) { in MakeFeatures()
|
/external/virglrenderer/src/gallium/auxiliary/tgsi/ |
D | tgsi_text.h | 43 uint num_tokens );
|
D | tgsi_parse.c | 292 tgsi_alloc_tokens(unsigned num_tokens) in tgsi_alloc_tokens() argument 294 unsigned bytes = num_tokens * sizeof(struct tgsi_token); in tgsi_alloc_tokens()
|
/external/mesa3d/src/gallium/auxiliary/tgsi/ |
D | tgsi_text.h | 43 uint num_tokens );
|
/external/mesa3d/src/gallium/drivers/nouveau/codegen/ |
D | nv50_ir_serialize.cpp | 35 unsigned int num_tokens = tgsi_num_tokens(tokens); in nv50_ir_prog_info_serialize() local 37 blob_write_uint32(blob, num_tokens); in nv50_ir_prog_info_serialize() 38 blob_write_bytes(blob, tokens, num_tokens * sizeof(struct tgsi_token)); in nv50_ir_prog_info_serialize()
|
/external/icing/icing/store/ |
D | document-store.h | 165 const DocumentProto& document, int32_t num_tokens = 0, 168 DocumentProto&& document, int32_t num_tokens = 0,
|