/external/pigweed/pw_tokenizer/py/ |
D | detokenize_test.py | 28 from pw_tokenizer import detokenize 109 detok = detokenize.Detokenizer( 119 str(detok.detokenize(b'\xab\xcd\0\0\x02\x03Two\x66')), '01 Two 3%' 123 detok = detokenize.Detokenizer( 133 result = detok.detokenize(b'\x01\0\0\0\x04args') 143 detok = detokenize.Detokenizer( 149 str(detok.detokenize(b'\xab\xcd')), 'This token is 16 bits' 153 detok = detokenize.Detokenizer( 163 result = detok.detokenize(b'\x02\0\0\0') 172 detok = detokenize.Detokenizer( [all …]
|
D | detokenize_proto_test.py | 22 from pw_tokenizer import detokenize, encode, tokens 32 _DETOKENIZER = detokenize.Detokenizer(_DATABASE) 97 self.detok = detokenize.Detokenizer(
|
D | BUILD.bazel | 25 "pw_tokenizer/detokenize.py", 39 name = "detokenize",
|
D | BUILD.gn | 40 "pw_tokenizer/detokenize.py",
|
/external/pigweed/pw_metric/py/ |
D | metric_parser_test.py | 21 from pw_tokenizer import detokenize, tokens 40 self.detokenize = detokenize.Detokenizer(DATABASE) 85 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s), 102 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s), 127 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s), 151 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s), 179 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s), 206 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s), 246 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s), 268 parse_metrics(self.rpcs, self.detokenize, self.rpc_timeout_s) [all …]
|
/external/pigweed/pw_tokenizer/py/pw_tokenizer/proto/ |
D | __init__.py | 22 from pw_tokenizer import detokenize, encode 37 detokenizer: detokenize.Detokenizer, 43 result = detokenizer.detokenize(data) 57 detokenized = detokenize.detokenize_base64(detokenizer, data, prefix) 71 detokenizer: detokenize.Detokenizer,
|
/external/tflite-support/tensorflow_lite_support/custom_ops/python/ |
D | sentencepiece_tokenizer.py | 83 def detokenize(self, input): # pylint: disable=redefined-builtin member in SentencepieceTokenizer 103 tokens = self.detokenize(input_tensor.flat_values) 107 tokens = self.detokenize(input_tensor.values) 116 return self.detokenize( 120 tokens = self.detokenize(tf.stack([input_tensor]))
|
D | sentencepiece_tokenizer_test.py | 85 tftext_detokenized = tftext_sp.detokenize(tftext_tokenized) 86 opt_detokenized = opt_sp.detokenize(tftext_tokenized) 157 return self.sp.detokenize(input_tensor)
|
/external/pigweed/pw_tokenizer/py/pw_tokenizer/ |
D | serial_detokenizer.py | 25 from pw_tokenizer import database, detokenize, tokens 62 default=detokenize.BASE64_PREFIX, 92 detokenizer = detokenize.Detokenizer(
|
D | __main__.py | 16 from pw_tokenizer import detokenize 18 detokenize.main()
|
D | __init__.py | 16 from pw_tokenizer.detokenize import AutoUpdatingDetokenizer, Detokenizer
|
D | detokenize.py | 210 def detokenize(self, encoded_message: bytes) -> DetokenizedString: member in Detokenizer 303 detokenized_string = self.detokenize(
|
/external/pigweed/pw_metric/py/pw_metric/ |
D | metric_parser.py | 19 from pw_tokenizer import detokenize 46 detokenizer: Optional[detokenize.Detokenizer], 67 detokenize.DetokenizedString(
|
/external/pigweed/pw_tokenizer/ts/ |
D | detokenizer_test.ts | 48 expect(detokenizer.detokenize(frame)).toEqual('regular token'); 51 expect(detokenizer.detokenize(generateFrame('aabbcc'))).toEqual('aabbcc');
|
D | detokenizer.ts | 52 detokenize(tokenizedFrame: Frame): string { method in Detokenizer
|
/external/libvpx/vp8/ |
D | vp8dx.mk | 26 VP8_DX_SRCS-yes += decoder/detokenize.c 33 VP8_DX_SRCS-yes += decoder/detokenize.h
|
/external/mesa3d/src/gallium/frontends/clover/llvm/codegen/ |
D | common.cpp | 129 s += detokenize(work_group_size_hint, ","); in kernel_attributes() 137 s += detokenize(reqd_work_group_size, ","); in kernel_attributes() 146 return detokenize(attributes, " "); in kernel_attributes()
|
/external/mesa3d/src/gallium/frontends/clover/util/ |
D | algorithm.hpp | 270 detokenize(const std::vector<T> &ss, const std::string &sep) { in detokenize() function 284 detokenize(const std::vector<std::string> &ss, const std::string &sep) { in detokenize() function
|
/external/pigweed/pw_system/py/pw_system/ |
D | device.py | 30 from pw_tokenizer import detokenize 52 detokenizer: Optional[detokenize.Detokenizer],
|
/external/pigweed/pw_tokenizer/java/dev/pigweed/tokenizer/ |
D | Detokenizer.java | 54 public String detokenize(String message) { in detokenize() method in Detokenizer
|
/external/pigweed/pw_tokenizer/ |
D | Android.bp | 25 "detokenize.cc",
|
D | BUILD.bazel | 95 "detokenize.cc", 99 "public/pw_tokenizer/detokenize.h",
|
D | CMakeLists.txt | 90 public/pw_tokenizer/detokenize.h 99 detokenize.cc
|
D | BUILD.gn | 136 "public/pw_tokenizer/detokenize.h", 141 "detokenize.cc",
|
/external/pigweed/pw_snapshot/py/pw_snapshot_metadata/ |
D | metadata.py | 86 self._reason_token = self._tokenizer_db.detokenize(
|