Home
last modified time | relevance | path

Searched +full:compat +full:- +full:data (Results 1 – 25 of 1032) sorted by relevance

12345678910>>...42

/external/tensorflow/tensorflow/tools/compatibility/
Dall_renames_v2.py7 # http://www.apache.org/licenses/LICENSE-2.0
18 # pylint: disable=line-too-long
29 "tf.compat.v1.batch_gather",
35 "tf.compat.v1.estimator.inputs",
84 "tf.contrib.data.AUTOTUNE":
85 "tf.data.experimental.AUTOTUNE",
86 "tf.contrib.data.Counter":
87 "tf.data.experimental.Counter",
88 "tf.contrib.data.CheckpointInputPipelineHook":
89 "tf.data.experimental.CheckpointInputPipelineHook",
[all …]
Dtf_upgrade_v2_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
23 import tensorflow.compat.v1 as tf
88 if hasattr(tf.compat, "v2"):
98 visitor.private_map["tf.compat"] = ["v1", "v2"]
99 traverse.traverse(tf.compat.v2, visitor)
101 if hasattr(tf.compat, "v1"):
111 visitor.private_map["tf.compat"] = ["v1", "v2"]
112 traverse.traverse(tf.compat.v1, visitor)
143 self.assertNotEqual(report.find("Failed to parse"), -1)
163 if not hasattr(tf.compat, "v2"):
[all …]
Drenames_v2.py7 # http://www.apache.org/licenses/LICENSE-2.0
15 # pylint: disable=line-too-long
20 bazel-bin/tensorflow/tools/compatibility/update/generate_v2_renames_map
21 pyformat --in_place third_party/tensorflow/tools/compatibility/renames_v2.py
26 'tf.compat.v1.AUTO_REUSE',
28 'tf.compat.v1.AttrValue',
34 'tf.compat.v1.ConditionalAccumulator',
36 'tf.compat.v1.ConditionalAccumulatorBase',
38 'tf.compat.v1.ConfigProto',
40 'tf.compat.v1.Dimension',
[all …]
Dtf_upgrade_v2.py7 # http://www.apache.org/licenses/LICENSE-2.0
30 # pylint: disable=g-explicit-bool-comparison,g-bool-id-comparison
50 compat_v1_import = VersionedTFImport("compat.v1")
51 compat_v2_import = VersionedTFImport("compat.v2")
60 ("tensorflow.compat.v1", "tf"): compat_v1_import,
61 ("tensorflow.compat.v2", "tf"): compat_v2_import,
66 """AST Visitor that replaces `import tensorflow.compat.v1 as tf`.
68 Converts `import tensorflow.compat.v1 as tf` to `import tensorflow as tf`
71 def visit_Import(self, node): # pylint: disable=invalid-name
79 if (import_alias.name == "tensorflow.compat.v1" and
[all …]
/external/libxkbcommon/test/
Drules-file.c27 #include "xkbcomp/xkbcomp-priv.h"
43 const char *compat; member
51 test_rules(struct xkb_context *ctx, struct test_data *data) in test_rules() argument
55 data->rules, data->model, data->layout, data->variant, data->options in test_rules()
59 fprintf(stderr, "\n\nChecking : %s\t%s\t%s\t%s\t%s\n", data->rules, in test_rules()
60 data->model, data->layout, data->variant, data->options); in test_rules()
62 if (data->should_fail) in test_rules()
66 data->keycodes, data->types, data->compat, data->symbols); in test_rules()
70 return data->should_fail; in test_rules()
74 kccgst.keycodes, kccgst.types, kccgst.compat, kccgst.symbols); in test_rules()
[all …]
Drules-file-includes.c26 #include "test-config.h"
29 #include "xkbcomp/xkbcomp-priv.h"
45 const char *compat; member
53 test_rules(struct xkb_context *ctx, struct test_data *data) in test_rules() argument
57 data->rules, data->model, data->layout, data->variant, data->options in test_rules()
61 fprintf(stderr, "\n\nChecking : %s\t%s\t%s\t%s\t%s\n", data->rules, in test_rules()
62 data->model, data->layout, data->variant, data->options); in test_rules()
64 if (data->should_fail) in test_rules()
68 data->keycodes, data->types, data->compat, data->symbols); in test_rules()
72 return data->should_fail; in test_rules()
[all …]
/external/tensorflow/tensorflow/examples/speech_commands/
Dtrain.py7 # http://www.apache.org/licenses/LICENSE-2.0
17 This is a self-contained example script that will train a very basic audio
18 recognition model in TensorFlow. It downloads the necessary training data and
34 download over 1GB of open source training data, so you'll need enough free space
35 and a good internet connection. The default data is a collection of thousands of
36 one-second .wav files, each containing one spoken word. This data set is
44 If you want to train on your own data, you'll need to create .wavs with your
60 `--wanted_words` argument. In this case, 'up,down' might be what you want, and
65 bazel run tensorflow/examples/speech_commands:train -- \
66 --data_dir=my_wavs --wanted_words=up,down
[all …]
Dinput_data.py7 # http://www.apache.org/licenses/LICENSE-2.0
33 from tensorflow.python.util import compat
35 tf.compat.v1.disable_eager_execution()
40 …ntend.python.ops import audio_microfrontend_op as frontend_op # pylint:disable=g-import-not-at-top
44 MAX_NUM_WAVS_PER_CLASS = 2**27 - 1 # ~134M
66 """Determines which data partition the file should belong to.
81 filename: File path of the data sample.
82 validation_percentage: How much of the data set to use for validation.
83 testing_percentage: How much of the data set to use for testing.
90 # deciding which set to put a wav in, so the data set creator has a way of
[all …]
Dtest_streaming_accuracy.py7 # http://www.apache.org/licenses/LICENSE-2.0
45 If you want to test natural data, you need to use a .wav with the same sample
47 sounds occur in time. Save this information out as a comma-separated text file,
53 bazel run tensorflow/examples/speech_commands:test_streaming_accuracy_py -- \
54 --wav=/tmp/streaming_test_bg.wav \
55 --ground-truth=/tmp/streaming_test_labels.txt --verbose \
56 --model=/tmp/conv_frozen.pb \
57 --labels=/tmp/speech_commands_train/conv_labels.txt \
58 --clip_duration_ms=1000 --detection_threshold=0.70 --average_window_ms=500 \
59 --suppression_ms=500 --time_tolerance_ms=1500
[all …]
/external/libxkbcommon/test/data/
Dsync.sh2 set -euo pipefail
7 if [ ! -d test/data ]; then
54 compat/complete \
55 compat/lednum \
56 compat/pc \
57 compat/ledscroll \
58 compat/basic \
59 compat/misc \
60 compat/iso9995 \
61 compat/accessx \
[all …]
/external/mbedtls/tests/scripts/
Dtest-ref-configs.pl3 # test-ref-configs.pl
6 # SPDX-License-Identifier: Apache-2.0 OR GPL-2.0-or-later
11 # configuration, run the test suites and compat.sh
13 # Usage: tests/scripts/test-ref-configs.pl [config-name [...]]
19 'config-ccm-psk-tls1_2.h' => {
20 'compat' => '-m tls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'',
23 'config-ccm-psk-dtls1_2.h' => {
24 'compat' => '-m dtls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'',
29 'config-no-entropy.h' => {
31 'config-suite-b.h' => {
[all …]
/external/openthread/third_party/mbedtls/repo/tests/scripts/
Dtest-ref-configs.pl3 # test-ref-configs.pl
6 # SPDX-License-Identifier: Apache-2.0 OR GPL-2.0-or-later
11 # configuration, run the test suites and compat.sh
13 # Usage: tests/scripts/test-ref-configs.pl [config-name [...]]
19 'config-ccm-psk-tls1_2.h' => {
20 'compat' => '-m tls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'',
23 'config-ccm-psk-dtls1_2.h' => {
24 'compat' => '-m dtls12 -f \'^TLS-PSK-WITH-AES-...-CCM-8\'',
29 'config-mini-tls1_1.h' => {
30 'compat' => '-m tls1_1 -f \'^DES-CBC3-SHA$\|^TLS-RSA-WITH-3DES-EDE-CBC-SHA$\'', #',
[all …]
/external/federated-compute/fcp/artifact_building/
Dgraph_helpers_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
167 ds = tf.data.experimental.to_variant(tf.data.Dataset.range(3))
172 ds2 = tf.data.experimental.from_variant(
176 with tf.compat.v1.Session() as sess:
186 x = tf.compat.v1.placeholder(dtype=tf.string)
190 with tf.compat.v1.Session() as sess:
198 with tf.compat.v1.control_dependencies([y]):
202 x = tf.compat.v1.placeholder(dtype=tf.int64)
203 y = tf.compat.v1.placeholder(dtype=tf.int64)
207 with tf.compat.v1.Session() as sess:
[all …]
Dgraph_helpers.py7 # http://www.apache.org/licenses/LICENSE-2.0
43 """Generates list of tff.compat.v1.placeholders for each leaf in a type spec.
57 signature of the input client data.
62 A list of tf.compat.v2.placeholders.
72 # stream of `tf.Example`s is selected from the data store, which is why we
74 return [tf.compat.v1.placeholder(tf.string, shape=[], name=name_prefix)]
91 ) -> tuple[tf.Tensor, list[MaybeSplitOutputs], list[tf.Tensor]]:
92 """Embeds the data logic into the current TensorFlow graph.
95 which returns a placeholder token. The initialization op and data values are
99 client_data_type: The TFF type signature of the input client data.
[all …]
Dtype_checks.py7 # http://www.apache.org/licenses/LICENSE-2.0
21 def _format_name_for_error(name: Optional[Any]) -> str:
38 ) -> None:
58 def check_callable(obj: Any, name: Optional[str] = None) -> None:
79 tf.data.Dataset, tf.compat.v1.data.Dataset, tf.compat.v2.data.Dataset
82 ) -> None:
94 tf.data.Dataset,
95 tf.compat.v1.data.Dataset,
96 tf.compat.v2.data.Dataset,
/external/tensorflow/tensorflow/python/summary/
Dsummary.py7 # http://www.apache.org/licenses/LICENSE-2.0
16 """Operations for writing summary data, for use in analysis and visualization.
28 # pylint: disable=unused-import, g-importing-member
31 …re.framework.summary_pb2 import SummaryMetadata as _SummaryMetadata # pylint: enable=unused-import
35 # pylint: enable=unused-import
44 from tensorflow.python.ops import gen_summary_ops as _gen_summary_ops # pylint: disable=unused-imp…
49 # pylint: disable=unused-import
52 # pylint: enable=unused-import
55 from tensorflow.python.util import compat as _compat
85 and there is an associated non-empty value for `step` (see
[all …]
/external/icu/android_icu4j/libcore_bridge/src/java/com/android/i18n/system/
DZygoteHooks.java8 * http://www.apache.org/licenses/LICENSE-2.0
19 import android.compat.Compatibility;
20 import android.compat.annotation.ChangeId;
21 import android.compat.annotation.EnabledAfter;
32 import dalvik.annotation.compat.VersionCodes;
48 * Called when the Zygote begins preloading classes and data.
54 // Pin ICU data in memory from this point that would normally be held by soft references. in onBeginPreload()
59 // Explicitly exercise code to cache data apps/framework are likely to need. in onBeginPreload()
66 // Framework's LocalLog is used during app start-up. It indirectly uses the current ICU time in onBeginPreload()
67 // zone. Pre-loading the current time zone in ICU improves app startup time. b/150605074 in onBeginPreload()
[all …]
/external/conscrypt/repackaged/common/src/main/java/com/android/org/conscrypt/
DNativeCrypto.java9 * http://www.apache.org/licenses/LICENSE-2.0
55 …* <a href="https://github.com/google/error-prone/blob/master/docs/bugpattern/UnsafeFinalization.md…
61 // --- OpenSSL library initialization --------------------------------------
88 // --- DSA/RSA public/private key handling functions -----------------------
90 @android.compat.annotation.UnsupportedAppUsage
100 @android.compat.annotation.UnsupportedAppUsage static native void EVP_PKEY_free(long pkey); in EVP_PKEY_free()
106 static native long EVP_parse_private_key(byte[] data) throws ParsingException; in EVP_parse_private_key() argument
110 static native byte[] EVP_raw_X25519_private_key(byte[] data) in EVP_raw_X25519_private_key() argument
113 static native long EVP_parse_public_key(byte[] data) throws ParsingException; in EVP_parse_public_key() argument
123 @android.compat.annotation.UnsupportedAppUsage
[all …]
/external/flatbuffers/lua/flatbuffers/
Dbinaryarray.lua1 local compat = require("flatbuffers.compat")
2 -- locals for slightly faster access
3 local string_pack = compat.string_pack
4 local string_unpack = compat.string_unpack
7 local m = {} -- the module table
9 local mt = {} -- the module metatable
11 -- given a binary array, set a metamethod to return its length
12 -- (e.g., #binaryArray, calls this)
17 -- Create a new binary array of an initial size
19 -- the array storage itself
[all …]
/external/tensorflow/tensorflow/python/data/kernel_tests/
Dlist_files_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
15 """Tests for `tf.data.Dataset.list_files()`."""
23 from tensorflow.python.data.kernel_tests import test_base
24 from tensorflow.python.data.ops import dataset_ops
28 from tensorflow.python.util import compat
67 compat.as_bytes(path.join(self.tmp_dir, filename))
82 compat.as_bytes(path.join(self.tmp_dir, filename))
95 compat.as_bytes(path.join(self.tmp_dir, filename))
139 compat.as_bytes(path.join(self.tmp_dir, filename))
153 compat.as_bytes(path.join(self.tmp_dir, filename))
[all …]
/external/sdv/vsomeip/interface/compat/vsomeip/
Dpayload.hpp1 // Copyright (C) 2014-2017 Bayerische Motoren Werke Aktiengesellschaft (BMW AG)
11 #include "../../compat/vsomeip/export.hpp"
12 #include "../../compat/vsomeip/internal/deserializable.hpp"
13 #include "../../compat/vsomeip/internal/serializable.hpp"
14 #include "../../compat/vsomeip/primitive_types.hpp"
54 * \brief Copies the given data array to the payload object.
56 * The current payload content is replaced by the data provided.
59 * \param _data Pointer to a data buffer.
60 * \param _length Length of the data buffer.
66 * \brief Copies the given data array to the payload object.
[all …]
/external/tensorflow/tensorflow/python/ops/
Dsession_ops.py7 # http://www.apache.org/licenses/LICENSE-2.0
18 # pylint: disable=g-bad-name
28 from tensorflow.python.util import compat
49 dtype: The data type of the tensor represented by `handle`.
52 self._handle = compat.as_str_any(handle)
69 self._resource_handle.device = self._handle.split(";")[-1]
119 handle_str = compat.as_str_any(handle)
120 return pydev.canonical_name(handle_str.split(";")[-1])
126 return handle_parts[0] + ";" + handle_parts[-1]
135 def get_session_handle(data, name=None): argument
[all …]
Dinit_ops.py7 # http://www.apache.org/licenses/LICENSE-2.0
67 """Returns the configuration of the initializer as a JSON-serializable dict.
70 A JSON-serializable Python dict.
81 initializer = RandomUniform(-1, 1)
102 `tf.compat.v1.zeros_initializer` is compatible with eager execution
106 argument in `tf.compat.v1.zeros_initializer.__init__()` does not exist in
115 initializer = tf.compat.v1.zeros_initializer(dtype=tf.float32)
129 | :------------------- | :--------------- | :------------------------- |
131 | `partition_info` | - | (`__call__` arg in TF1) Not supported |
138 >>> initializer = tf.compat.v1.zeros_initializer(dtype=tf.float32)
[all …]
/external/noto-fonts/emoji-compat/
DREADME.android8 Noto Color Emoji Compat font is generated using Noto Color Emoji font using createfont.py. The
9 compat font is under font/ directory.
11 While generating the compat font, Noto Color Emoji font and data files from Unicode are used.
13 data/emoji-metadata.txt is updated using the Noto Color Emoji font and data files from
16 supported-emojis/emojis.txt file contains list of emojis that are supported by the font. Main
19 Noto Color Emoji font is under the <android_source>/external/noto-fonts/emoji/ directory. Unicode
/external/federated-compute/fcp/demo/
Dserver_test.py7 # http://www.apache.org/licenses/LICENSE-2.0
49 'id1': [b'1-1', b'1-2', b'1-3'],
50 'id2': [b'2-1', b'2-2'],
54 def create_plan() -> plan_pb2.Plan:
55 """Creates a test plan that counts examples, with a per-client cap."""
57 with tf.compat.v1.Graph().as_default() as client_graph:
58 dataset_token = tf.compat.v1.placeholder(tf.string, shape=())
59 input_filepath = tf.compat.v1.placeholder(tf.string, shape=())
60 output_filepath = tf.compat.v1.placeholder(tf.string, shape=())
68 data=[count])
[all …]

12345678910>>...42