| /external/tensorflow/tensorflow/tools/compatibility/ |
| D | all_renames_v2.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 18 # pylint: disable=line-too-long 29 "tf.compat.v1.batch_gather", 35 "tf.compat.v1.estimator.inputs", 84 "tf.contrib.data.AUTOTUNE": 85 "tf.data.experimental.AUTOTUNE", 86 "tf.contrib.data.Counter": 87 "tf.data.experimental.Counter", 88 "tf.contrib.data.CheckpointInputPipelineHook": 89 "tf.data.experimental.CheckpointInputPipelineHook", [all …]
|
| D | tf_upgrade_v2_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 23 import tensorflow.compat.v1 as tf 88 if hasattr(tf.compat, "v2"): 98 visitor.private_map["tf.compat"] = ["v1", "v2"] 99 traverse.traverse(tf.compat.v2, visitor) 101 if hasattr(tf.compat, "v1"): 111 visitor.private_map["tf.compat"] = ["v1", "v2"] 112 traverse.traverse(tf.compat.v1, visitor) 143 self.assertNotEqual(report.find("Failed to parse"), -1) 163 if not hasattr(tf.compat, "v2"): [all …]
|
| D | renames_v2.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 15 # pylint: disable=line-too-long 20 bazel-bin/tensorflow/tools/compatibility/update/generate_v2_renames_map 21 pyformat --in_place third_party/tensorflow/tools/compatibility/renames_v2.py 26 'tf.compat.v1.AUTO_REUSE', 28 'tf.compat.v1.AttrValue', 34 'tf.compat.v1.ConditionalAccumulator', 36 'tf.compat.v1.ConditionalAccumulatorBase', 38 'tf.compat.v1.ConfigProto', 40 'tf.compat.v1.Dimension', [all …]
|
| D | tf_upgrade_v2.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 30 # pylint: disable=g-explicit-bool-comparison,g-bool-id-comparison 50 compat_v1_import = VersionedTFImport("compat.v1") 51 compat_v2_import = VersionedTFImport("compat.v2") 60 ("tensorflow.compat.v1", "tf"): compat_v1_import, 61 ("tensorflow.compat.v2", "tf"): compat_v2_import, 66 """AST Visitor that replaces `import tensorflow.compat.v1 as tf`. 68 Converts `import tensorflow.compat.v1 as tf` to `import tensorflow as tf` 71 def visit_Import(self, node): # pylint: disable=invalid-name 79 if (import_alias.name == "tensorflow.compat.v1" and [all …]
|
| /external/libxkbcommon/test/ |
| D | rules-file.c | 27 #include "xkbcomp/xkbcomp-priv.h" 43 const char *compat; member 51 test_rules(struct xkb_context *ctx, struct test_data *data) in test_rules() argument 55 data->rules, data->model, data->layout, data->variant, data->options in test_rules() 59 fprintf(stderr, "\n\nChecking : %s\t%s\t%s\t%s\t%s\n", data->rules, in test_rules() 60 data->model, data->layout, data->variant, data->options); in test_rules() 62 if (data->should_fail) in test_rules() 66 data->keycodes, data->types, data->compat, data->symbols); in test_rules() 70 return data->should_fail; in test_rules() 74 kccgst.keycodes, kccgst.types, kccgst.compat, kccgst.symbols); in test_rules() [all …]
|
| D | rules-file-includes.c | 26 #include "test-config.h" 29 #include "xkbcomp/xkbcomp-priv.h" 45 const char *compat; member 53 test_rules(struct xkb_context *ctx, struct test_data *data) in test_rules() argument 57 data->rules, data->model, data->layout, data->variant, data->options in test_rules() 61 fprintf(stderr, "\n\nChecking : %s\t%s\t%s\t%s\t%s\n", data->rules, in test_rules() 62 data->model, data->layout, data->variant, data->options); in test_rules() 64 if (data->should_fail) in test_rules() 68 data->keycodes, data->types, data->compat, data->symbols); in test_rules() 72 return data->should_fail; in test_rules() [all …]
|
| /external/tensorflow/tensorflow/examples/speech_commands/ |
| D | train.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 17 This is a self-contained example script that will train a very basic audio 18 recognition model in TensorFlow. It downloads the necessary training data and 34 download over 1GB of open source training data, so you'll need enough free space 35 and a good internet connection. The default data is a collection of thousands of 36 one-second .wav files, each containing one spoken word. This data set is 44 If you want to train on your own data, you'll need to create .wavs with your 60 `--wanted_words` argument. In this case, 'up,down' might be what you want, and 65 bazel run tensorflow/examples/speech_commands:train -- \ 66 --data_dir=my_wavs --wanted_words=up,down [all …]
|
| D | input_data.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 33 from tensorflow.python.util import compat 35 tf.compat.v1.disable_eager_execution() 40 …ntend.python.ops import audio_microfrontend_op as frontend_op # pylint:disable=g-import-not-at-top 44 MAX_NUM_WAVS_PER_CLASS = 2**27 - 1 # ~134M 66 """Determines which data partition the file should belong to. 81 filename: File path of the data sample. 82 validation_percentage: How much of the data set to use for validation. 83 testing_percentage: How much of the data set to use for testing. 90 # deciding which set to put a wav in, so the data set creator has a way of [all …]
|
| D | test_streaming_accuracy.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 45 If you want to test natural data, you need to use a .wav with the same sample 47 sounds occur in time. Save this information out as a comma-separated text file, 53 bazel run tensorflow/examples/speech_commands:test_streaming_accuracy_py -- \ 54 --wav=/tmp/streaming_test_bg.wav \ 55 --ground-truth=/tmp/streaming_test_labels.txt --verbose \ 56 --model=/tmp/conv_frozen.pb \ 57 --labels=/tmp/speech_commands_train/conv_labels.txt \ 58 --clip_duration_ms=1000 --detection_threshold=0.70 --average_window_ms=500 \ 59 --suppression_ms=500 --time_tolerance_ms=1500 [all …]
|
| /external/libxkbcommon/test/data/ |
| D | sync.sh | 2 set -euo pipefail 7 if [ ! -d test/data ]; then 54 compat/complete \ 55 compat/lednum \ 56 compat/pc \ 57 compat/ledscroll \ 58 compat/basic \ 59 compat/misc \ 60 compat/iso9995 \ 61 compat/accessx \ [all …]
|
| /external/mbedtls/tests/scripts/ |
| D | test-ref-configs.pl | 3 # test-ref-configs.pl 6 # SPDX-License-Identifier: Apache-2.0 OR GPL-2.0-or-later 11 # configuration, run the test suites and compat.sh 13 # Usage: tests/scripts/test-ref-configs.pl [config-name [...]] 19 'config-ccm-psk-tls1_2.h' => { 20 'compat' => '-m tls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'', 23 'config-ccm-psk-dtls1_2.h' => { 24 'compat' => '-m dtls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'', 29 'config-no-entropy.h' => { 31 'config-suite-b.h' => { [all …]
|
| /external/openthread/third_party/mbedtls/repo/tests/scripts/ |
| D | test-ref-configs.pl | 3 # test-ref-configs.pl 6 # SPDX-License-Identifier: Apache-2.0 OR GPL-2.0-or-later 11 # configuration, run the test suites and compat.sh 13 # Usage: tests/scripts/test-ref-configs.pl [config-name [...]] 19 'config-ccm-psk-tls1_2.h' => { 20 'compat' => '-m tls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'', 23 'config-ccm-psk-dtls1_2.h' => { 24 'compat' => '-m dtls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'', 29 'config-mini-tls1_1.h' => { 30 'compat' => '-m tls1_1 -f \'^DES-CBC3-SHA$\|^TLS-RSA-WITH-3DES-EDE-CBC-SHA$\'', #', [all …]
|
| /external/federated-compute/fcp/artifact_building/ |
| D | graph_helpers_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 167 ds = tf.data.experimental.to_variant(tf.data.Dataset.range(3)) 172 ds2 = tf.data.experimental.from_variant( 176 with tf.compat.v1.Session() as sess: 186 x = tf.compat.v1.placeholder(dtype=tf.string) 190 with tf.compat.v1.Session() as sess: 198 with tf.compat.v1.control_dependencies([y]): 202 x = tf.compat.v1.placeholder(dtype=tf.int64) 203 y = tf.compat.v1.placeholder(dtype=tf.int64) 207 with tf.compat.v1.Session() as sess: [all …]
|
| D | graph_helpers.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 43 """Generates list of tff.compat.v1.placeholders for each leaf in a type spec. 57 signature of the input client data. 62 A list of tf.compat.v2.placeholders. 72 # stream of `tf.Example`s is selected from the data store, which is why we 74 return [tf.compat.v1.placeholder(tf.string, shape=[], name=name_prefix)] 91 ) -> tuple[tf.Tensor, list[MaybeSplitOutputs], list[tf.Tensor]]: 92 """Embeds the data logic into the current TensorFlow graph. 95 which returns a placeholder token. The initialization op and data values are 99 client_data_type: The TFF type signature of the input client data. [all …]
|
| D | type_checks.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 21 def _format_name_for_error(name: Optional[Any]) -> str: 38 ) -> None: 58 def check_callable(obj: Any, name: Optional[str] = None) -> None: 79 tf.data.Dataset, tf.compat.v1.data.Dataset, tf.compat.v2.data.Dataset 82 ) -> None: 94 tf.data.Dataset, 95 tf.compat.v1.data.Dataset, 96 tf.compat.v2.data.Dataset,
|
| /external/tensorflow/tensorflow/python/summary/ |
| D | summary.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 16 """Operations for writing summary data, for use in analysis and visualization. 28 # pylint: disable=unused-import, g-importing-member 31 …re.framework.summary_pb2 import SummaryMetadata as _SummaryMetadata # pylint: enable=unused-import 35 # pylint: enable=unused-import 44 from tensorflow.python.ops import gen_summary_ops as _gen_summary_ops # pylint: disable=unused-imp… 49 # pylint: disable=unused-import 52 # pylint: enable=unused-import 55 from tensorflow.python.util import compat as _compat 85 and there is an associated non-empty value for `step` (see [all …]
|
| /external/icu/android_icu4j/libcore_bridge/src/java/com/android/i18n/system/ |
| D | ZygoteHooks.java | 8 * http://www.apache.org/licenses/LICENSE-2.0 19 import android.compat.Compatibility; 20 import android.compat.annotation.ChangeId; 21 import android.compat.annotation.EnabledAfter; 32 import dalvik.annotation.compat.VersionCodes; 48 * Called when the Zygote begins preloading classes and data. 54 // Pin ICU data in memory from this point that would normally be held by soft references. in onBeginPreload() 59 // Explicitly exercise code to cache data apps/framework are likely to need. in onBeginPreload() 66 // Framework's LocalLog is used during app start-up. It indirectly uses the current ICU time in onBeginPreload() 67 // zone. Pre-loading the current time zone in ICU improves app startup time. b/150605074 in onBeginPreload() [all …]
|
| /external/conscrypt/repackaged/common/src/main/java/com/android/org/conscrypt/ |
| D | NativeCrypto.java | 9 * http://www.apache.org/licenses/LICENSE-2.0 55 …* <a href="https://github.com/google/error-prone/blob/master/docs/bugpattern/UnsafeFinalization.md… 61 // --- OpenSSL library initialization -------------------------------------- 88 // --- DSA/RSA public/private key handling functions ----------------------- 90 @android.compat.annotation.UnsupportedAppUsage 100 @android.compat.annotation.UnsupportedAppUsage static native void EVP_PKEY_free(long pkey); in EVP_PKEY_free() 106 static native long EVP_parse_private_key(byte[] data) throws ParsingException; in EVP_parse_private_key() argument 110 static native byte[] EVP_raw_X25519_private_key(byte[] data) in EVP_raw_X25519_private_key() argument 113 static native long EVP_parse_public_key(byte[] data) throws ParsingException; in EVP_parse_public_key() argument 123 @android.compat.annotation.UnsupportedAppUsage [all …]
|
| /external/flatbuffers/lua/flatbuffers/ |
| D | binaryarray.lua | 1 local compat = require("flatbuffers.compat") 2 -- locals for slightly faster access 3 local string_pack = compat.string_pack 4 local string_unpack = compat.string_unpack 7 local m = {} -- the module table 9 local mt = {} -- the module metatable 11 -- given a binary array, set a metamethod to return its length 12 -- (e.g., #binaryArray, calls this) 17 -- Create a new binary array of an initial size 19 -- the array storage itself [all …]
|
| /external/tensorflow/tensorflow/python/data/kernel_tests/ |
| D | list_files_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 15 """Tests for `tf.data.Dataset.list_files()`.""" 23 from tensorflow.python.data.kernel_tests import test_base 24 from tensorflow.python.data.ops import dataset_ops 28 from tensorflow.python.util import compat 67 compat.as_bytes(path.join(self.tmp_dir, filename)) 82 compat.as_bytes(path.join(self.tmp_dir, filename)) 95 compat.as_bytes(path.join(self.tmp_dir, filename)) 139 compat.as_bytes(path.join(self.tmp_dir, filename)) 153 compat.as_bytes(path.join(self.tmp_dir, filename)) [all …]
|
| /external/sdv/vsomeip/interface/compat/vsomeip/ |
| D | payload.hpp | 1 // Copyright (C) 2014-2017 Bayerische Motoren Werke Aktiengesellschaft (BMW AG) 11 #include "../../compat/vsomeip/export.hpp" 12 #include "../../compat/vsomeip/internal/deserializable.hpp" 13 #include "../../compat/vsomeip/internal/serializable.hpp" 14 #include "../../compat/vsomeip/primitive_types.hpp" 54 * \brief Copies the given data array to the payload object. 56 * The current payload content is replaced by the data provided. 59 * \param _data Pointer to a data buffer. 60 * \param _length Length of the data buffer. 66 * \brief Copies the given data array to the payload object. [all …]
|
| /external/tensorflow/tensorflow/python/ops/ |
| D | session_ops.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 18 # pylint: disable=g-bad-name 28 from tensorflow.python.util import compat 49 dtype: The data type of the tensor represented by `handle`. 52 self._handle = compat.as_str_any(handle) 69 self._resource_handle.device = self._handle.split(";")[-1] 119 handle_str = compat.as_str_any(handle) 120 return pydev.canonical_name(handle_str.split(";")[-1]) 126 return handle_parts[0] + ";" + handle_parts[-1] 135 def get_session_handle(data, name=None): argument [all …]
|
| D | init_ops.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 67 """Returns the configuration of the initializer as a JSON-serializable dict. 70 A JSON-serializable Python dict. 81 initializer = RandomUniform(-1, 1) 102 `tf.compat.v1.zeros_initializer` is compatible with eager execution 106 argument in `tf.compat.v1.zeros_initializer.__init__()` does not exist in 115 initializer = tf.compat.v1.zeros_initializer(dtype=tf.float32) 129 | :------------------- | :--------------- | :------------------------- | 131 | `partition_info` | - | (`__call__` arg in TF1) Not supported | 138 >>> initializer = tf.compat.v1.zeros_initializer(dtype=tf.float32) [all …]
|
| /external/noto-fonts/emoji-compat/ |
| D | README.android | 8 Noto Color Emoji Compat font is generated using Noto Color Emoji font using createfont.py. The 9 compat font is under font/ directory. 11 While generating the compat font, Noto Color Emoji font and data files from Unicode are used. 13 data/emoji-metadata.txt is updated using the Noto Color Emoji font and data files from 16 supported-emojis/emojis.txt file contains list of emojis that are supported by the font. Main 19 Noto Color Emoji font is under the <android_source>/external/noto-fonts/emoji/ directory. Unicode
|
| /external/federated-compute/fcp/demo/ |
| D | server_test.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 49 'id1': [b'1-1', b'1-2', b'1-3'], 50 'id2': [b'2-1', b'2-2'], 54 def create_plan() -> plan_pb2.Plan: 55 """Creates a test plan that counts examples, with a per-client cap.""" 57 with tf.compat.v1.Graph().as_default() as client_graph: 58 dataset_token = tf.compat.v1.placeholder(tf.string, shape=()) 59 input_filepath = tf.compat.v1.placeholder(tf.string, shape=()) 60 output_filepath = tf.compat.v1.placeholder(tf.string, shape=()) 68 data=[count]) [all …]
|