| /external/pytorch/tools/stats/ |
| D | README.md | 1 # PyTorch CI Stats 3 We track various stats about each CI job. 5 1. Jobs upload their artifacts to an intermediate data store (either GitHub 7 …/pytorch/blob/a9f6a35a33308f3be2413cc5c866baec5cfe3ba1/.github/workflows/_linux-build.yml#L144-L151 9 …`upload-test-stats.yml`](https://github.com/pytorch/pytorch/blob/d9fca126fca7d7780ae44170d30bda901… 10 3. `upload-test-stats` downloads the raw stats from the intermediate data store 15 J1[Job with AWS creds<br>e.g. linux, win] --raw stats--> S3[(AWS S3)] 16 J2[Job w/o AWS creds<br>e.g. mac] --raw stats--> GHA[(GH artifacts)] 18 S3 --> uts[upload-test-stats.yml] 19 GHA --> uts [all …]
|
| D | upload_test_stats_intermediate.py | 4 from tools.stats.test_dashboard import upload_additional_info 5 from tools.stats.upload_test_stats import get_tests 9 parser = argparse.ArgumentParser(description="Upload test stats to Rockset") 11 "--workflow-run-id", 16 "--workflow-run-attempt", 27 # Flush stdout so that any errors in Rockset upload show up last in the logs.
|
| D | upload_dynamo_perf_stats.py | 13 from tools.stats.upload_stats_lib import ( 22 "test-reports", 25 r"test-reports-test-(?P<name>[\w\-]+)-\d+-\d+-(?P<runner>[\w\.-]+)_(?P<job>\d+).zip" 35 ) -> list[dict[str, Any]]: 47 # Unzip to get perf stats csv files 51 print(f"Test report {path} has an invalid name. Skipping") 73 "workflow_id": workflow_run_id, # type: ignore[dict-item] 74 "run_attempt": workflow_run_attempt, # type: ignore[dict-item] 90 def generate_partition_key(repo: str, doc: Dict[str, Any]) -> str: 99 hash_content = hashlib.md5(json.dumps(doc).encode("utf-8")).hexdigest() [all …]
|
| D | upload_test_stats.py | 12 from tools.stats.test_dashboard import upload_additional_info 13 from tools.stats.upload_stats_lib import ( 26 ) -> list[dict[str, Any]]: 27 """Convert a test report xml file into a JSON-serializable list of test cases.""" 28 print(f"Parsing {tag}s for test report: {report}") 43 # The name of the file that the test is located in is not necessarily 44 # the same as the name of the file that invoked the test. 45 # For example, `test_jit.py` calls into multiple other test files (e.g. 46 # jit/test_dce.py). For sharding/test selection purposes, we want to 47 # record the file that invoked the test. [all …]
|
| D | upload_test_stat_aggregates.py | 13 from tools.stats.upload_stats_lib import upload_to_s3 16 def get_oncall_from_testfile(testfile: str) -> list[str] | None: 17 path = f"test/{testfile}" 20 # get oncall on test file 42 def get_test_stat_aggregates(date: datetime.date) -> Any: 70 description="Upload test stat aggregates to Rockset." 73 "--date", 75 help="Date to upload test stat aggregates for (YYYY-MM-DD). Must be in the last 30 days", 79 if args.date < datetime.datetime.now().date() - datetime.timedelta(days=30): 83 bucket_name="torchci-aggregated-stats",
|
| D | upload_sccache_stats.py | 10 from tools.stats.upload_stats_lib import ( 18 ) -> list[dict[str, Any]]: 24 download_s3_artifacts("sccache-stats", workflow_run_id, workflow_run_attempt) 34 parser = argparse.ArgumentParser(description="Upload test stats to Rockset") 36 "--workflow-run-id", 42 "--workflow-run-attempt", 48 stats = get_sccache_stats(args.workflow_run_id, args.workflow_run_attempt) variable 50 args.workflow_run_id, args.workflow_run_attempt, "sccache_stats", stats
|
| D | upload_artifacts.py | 6 from tools.stats.upload_stats_lib import download_gha_artifacts, upload_file_to_s3 10 "sccache-stats", 11 "test-jsons", 12 "test-reports", 13 "usage-log", 15 BUCKET_NAME = "gha-artifacts" 16 FILENAME_REGEX = r"-runattempt\d+" 19 def get_artifacts(repo: str, workflow_run_id: int, workflow_run_attempt: int) -> None: 30 … # GHA artifact is named as follows: NAME-runattempt${{ github.run_attempt }}-SUFFIX.zip 32 # pytorch/pytorch/WORKFLOW_ID/RUN_ATTEMPT/artifact/NAME-SUFFIX.zip [all …]
|
| D | check_disabled_tests.py | 11 from tools.stats.upload_stats_lib import ( 17 from tools.stats.upload_test_stats import process_xml_element 26 ) -> dict[str, dict[str, int]]: 28 Return a list of disabled tests that should be re-enabled and those that are still 35 # * Success test should be re-enable if it's green after rerunning in all platforms 37 # * Failures from pytest because pytest-flakefinder is used to run the same test 41 # We want to keep track of how many times the test fails (num_red) or passes (num_green) 47 # Under --rerun-disabled-tests mode, a test is skipped when: 49 # * it's skipped because it's a normal enabled test 72 # Check if the test is a failure [all …]
|
| /external/cronet/tot/net/reporting/ |
| D | reporting_delivery_agent_unittest.cc | 2 // Use of this source code is governed by a BSD-style license that can be 17 #include "base/test/metrics/histogram_tester.h" 18 #include "base/test/scoped_feature_list.h" 19 #include "base/test/simple_test_tick_clock.h" 20 #include "base/test/values_test_util.h" 49 // This is a private API of the reporting service, so no need to test the in ReportingDeliveryAgentTest() 50 // case kPartitionConnectionsByNetworkIsolationKey is disabled - the in ReportingDeliveryAgentTest() 60 policy.endpoint_backoff_policy.maximum_backoff_ms = -1; in ReportingDeliveryAgentTest() 72 cache()->AddReport(reporting_source, network_anonymization_key, url, in AddReport() 74 /*depth=*/0, /*queued=*/tick_clock()->NowTicks(), in AddReport() [all …]
|
| /external/cronet/stable/net/reporting/ |
| D | reporting_delivery_agent_unittest.cc | 2 // Use of this source code is governed by a BSD-style license that can be 17 #include "base/test/metrics/histogram_tester.h" 18 #include "base/test/scoped_feature_list.h" 19 #include "base/test/simple_test_tick_clock.h" 20 #include "base/test/values_test_util.h" 49 // This is a private API of the reporting service, so no need to test the in ReportingDeliveryAgentTest() 50 // case kPartitionConnectionsByNetworkIsolationKey is disabled - the in ReportingDeliveryAgentTest() 60 policy.endpoint_backoff_policy.maximum_backoff_ms = -1; in ReportingDeliveryAgentTest() 72 cache()->AddReport(reporting_source, network_anonymization_key, url, in AddReport() 74 /*depth=*/0, /*queued=*/tick_clock()->NowTicks(), in AddReport() [all …]
|
| /external/pytorch/.github/workflows/ |
| D | upload-test-stats.yml | 1 name: Upload test stats 5 …trunk, periodic, inductor, unstable, slow, unstable-periodic, inductor-periodic, rocm, inductor-mi… 7 - completed 11 …ion adapted from https://github.com/community/community/discussions/21090#discussioncomment-3226271 14 runs-on: ubuntu-latest 18 - name: Get workflow run conclusion 19 uses: octokit/request-action@v2.1.0 26 upload-test-stats: 32 runs-on: ubuntu-22.04 33 environment: upload-stats [all …]
|
| D | upload-torch-dynamo-perf-stats.yml | 1 name: Upload torch dynamo performance stats 5 …workflows: [inductor-A100-perf-nightly, inductor-perf-nightly-A10g, inductor-perf-nightly-aarch64,… 7 - completed 10 get-conclusion: 11 runs-on: ubuntu-latest 13 conclusion: ${{ fromJson(steps.get-conclusion.outputs.data).conclusion }} 15 - name: Get workflow run conclusion 16 uses: octokit/request-action@v2.1.0 17 id: get-conclusion 23 upload-perf-stats: [all …]
|
| D | upload_test_stats_intermediate.yml | 1 name: Upload test stats intermediate 15 name: Intermediate upload test stats for ${{ inputs.workflow_id }} 16 runs-on: ubuntu-22.04 17 environment: upload-stats 19 - name: Checkout PyTorch 20 uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.4 22 fetch-depth: 1 25 - uses: actions/setup-python@v4 27 python-version: '3.11' 30 - run: | [all …]
|
| D | _win-build.yml | 1 name: windows-build 6 build-environment: 9 description: Top-level label for what's being built/tested. 10 cuda-version: 14 build-with-debug: 19 sync-tag: 25 job with the same `sync-tag` is identical. 26 test-matrix: 30 An option JSON description of what test configs to run later on. This 31 is moved here from the Linux test workflow so that we can apply filter [all …]
|
| D | _mac-build.yml | 1 name: mac-build 6 build-environment: 9 description: Top-level label for what's being built/tested. 10 runner-type: 13 description: Name of the GitHub-managed runner type to use for the build. 14 build-generates-artifacts: 17 description: If set, upload generated build artifacts. 18 xcode-version: 23 sync-tag: 29 job with the same `sync-tag` is identical. [all …]
|
| D | upload-alerts.yml | 1 # upload alerts every 10 minutes 3 name: Upload Alerts to AWS/Rockset 7 - cron: '*/10 * * * *' 10 - 'tools/alerts/create_alerts.py' 11 - '.github/workflows/upload-alerts.yml' 14 upload-alerts: 16 runs-on: ubuntu-22.04 17 environment: upload-stats 19 - name: Checkout repo 22 fetch-depth: 1 [all …]
|
| D | _linux-build.yml | 1 name: linux-build 6 build-environment: 9 description: Top-level label for what's being built/tested. 10 docker-image-name: 14 build-generates-artifacts: 18 description: If set, upload generated build artifacts. 19 build-with-debug: 24 sync-tag: 30 job with the same `sync-tag` is identical. 31 cuda-arch-list: [all …]
|
| /external/pytorch/.github/scripts/ |
| D | rockset_mocks.json.gz | |
| /external/deqp/modules/gles3/performance/ |
| D | es3pBufferDataUploadTests.cpp | 1 /*------------------------------------------------------------------------- 3 * ------------------------------------------------- 11 * http://www.apache.org/licenses/LICENSE-2.0 21 * \brief Buffer data upload performance tests. 22 *//*--------------------------------------------------------------------*/ 302 SingleOperationStatistics upload; member 321 SingleOperationStatistics upload; member 745 return endTime - startTime; in medianTimeMemcpy() 765 sectionLens[sectionNdx] = sectionStarts[sectionNdx + 1] - sectionStarts[sectionNdx]; in medianTimeMemcpy() 786 sectionTimes[sectionNdx] = endTime - startTime; in medianTimeMemcpy() [all …]
|
| /external/pytorch/benchmarks/ |
| D | upload_scribe.py | 3 Currently supports data in pytest-benchmark format but can be extended. 45 def upload(self, messages): member in ScribeUploader 113 test = b["name"].split("[")[0] 115 benchmark_name = f"{test}[{net_name}]" 139 "stddev": b["stats"]["stddev"], 140 "rounds": b["stats"]["rounds"], 141 "min": b["stats"]["min"], 142 "median": b["stats"]["median"], 143 "max": b["stats"]["max"], 144 "mean": b["stats"]["mean"], [all …]
|
| /external/googleapis/google/ads/googleads/v14/resources/ |
| D | customer.proto | 7 // http://www.apache.org/licenses/LICENSE-2.0 69 // Optional, non-unique descriptive name of the customer. 88 // Whether auto-tagging is enabled for the customer. 101 // Output only. Whether the customer is a test account. 118 // This field is read-only. 129 // for all manager customers, and for unscored non-manager customers. 132 // https://support.google.com/google-ads/answer/9061546. 134 // This field is read-only. 141 // scores across multiple non-manager customers. The aggregate optimization 144 // field is 0 for all manager customers, and for unscored non-manager [all …]
|
| /external/curl/tests/http/ |
| D | test_07_upload.py | 2 # -*- coding: utf-8 -*- 23 # SPDX-License-Identifier: curl 47 env.make_data_file(indir=env.gen_dir, fname="data-10k", fsize=10*1024) 48 env.make_data_file(indir=env.gen_dir, fname="data-63k", fsize=63*1024) 49 env.make_data_file(indir=env.gen_dir, fname="data-64k", fsize=64*1024) 50 env.make_data_file(indir=env.gen_dir, fname="data-100k", fsize=100*1024) 51 env.make_data_file(indir=env.gen_dir, fname="data-1m+", fsize=(1024*1024)+1) 52 env.make_data_file(indir=env.gen_dir, fname="data-10m", fsize=10*1024*1024) 56 # upload small data, check that this is what was echoed 65 url = f'https://{env.authority_for(env.domain1, proto)}/curltest/echo?id=[0-0]' [all …]
|
| /external/mobile-data-download/javatests/com/google/android/libraries/mobiledatadownload/file/transforms/ |
| D | CompressTransformTest.java | 8 * http://www.apache.org/licenses/LICENSE-2.0 35 import org.junit.Test; 57 @Test 62 assertThat(byteCounter.stats()).isEqualTo(new long[] {0, 20}); in counterWithCompress() 67 assertThat(byteCounter.stats()).isEqualTo(new long[] {20, 20}); in counterWithCompress() 71 assertThat(byteCounter.stats()).isEqualTo(new long[] {20, 44}); in counterWithCompress() 72 ratio = text.length() / (44 - 20); in counterWithCompress() 77 assertThat(byteCounter.stats()).isEqualTo(new long[] {20, 76}); in counterWithCompress() 78 ratio = text.length() / (76 - 44); in counterWithCompress() 82 @Test [all …]
|
| /external/grpc-grpc-java/interop-testing/src/main/proto/grpc/testing/ |
| D | test.proto | 2 // Copyright 2015-2016 gRPC authors. 8 // http://www.apache.org/licenses/LICENSE-2.0 16 // An integration test service that covers all the method signature permutations 28 // A simple service to test the various types of RPCs and experiment with 47 // A sequence of requests followed by one response (streamed upload). 65 // The test server will not implement this method. It will be used 66 // to test the behavior when clients call unimplemented methods. 70 // A simple service NOT implemented at servers so clients can test for 83 // A service used to obtain stats for verifying LB behavior. 85 // Gets the backend distribution for RPCs sent by a test client. [all …]
|
| /external/rust/android-crates-io/crates/grpcio-sys/grpc/src/proto/grpc/testing/ |
| D | test.proto | 2 // Copyright 2015-2016 gRPC authors. 8 // http://www.apache.org/licenses/LICENSE-2.0 16 // An integration test service that covers all the method signature permutations 26 // A simple service to test the various types of RPCs and experiment with 45 // A sequence of requests followed by one response (streamed upload). 63 // The test server will not implement this method. It will be used 64 // to test the behavior when clients call unimplemented methods. 68 // A simple service NOT implemented at servers so clients can test for 81 // A service used to obtain stats for verifying LB behavior. 83 // Gets the backend distribution for RPCs sent by a test client. [all …]
|