Home
last modified time | relevance | path

Searched +full:upload +full:- +full:test +full:- +full:stats (Results 1 – 25 of 162) sorted by relevance

1234567

/external/pytorch/tools/stats/
DREADME.md1 # PyTorch CI Stats
3 We track various stats about each CI job.
5 1. Jobs upload their artifacts to an intermediate data store (either GitHub
7 …/pytorch/blob/a9f6a35a33308f3be2413cc5c866baec5cfe3ba1/.github/workflows/_linux-build.yml#L144-L151
9 …`upload-test-stats.yml`](https://github.com/pytorch/pytorch/blob/d9fca126fca7d7780ae44170d30bda901…
10 3. `upload-test-stats` downloads the raw stats from the intermediate data store
15 J1[Job with AWS creds<br>e.g. linux, win] --raw stats--> S3[(AWS S3)]
16 J2[Job w/o AWS creds<br>e.g. mac] --raw stats--> GHA[(GH artifacts)]
18 S3 --> uts[upload-test-stats.yml]
19 GHA --> uts
[all …]
Dupload_test_stats_intermediate.py4 from tools.stats.test_dashboard import upload_additional_info
5 from tools.stats.upload_test_stats import get_tests
9 parser = argparse.ArgumentParser(description="Upload test stats to Rockset")
11 "--workflow-run-id",
16 "--workflow-run-attempt",
27 # Flush stdout so that any errors in Rockset upload show up last in the logs.
Dupload_dynamo_perf_stats.py13 from tools.stats.upload_stats_lib import (
22 "test-reports",
25 r"test-reports-test-(?P<name>[\w\-]+)-\d+-\d+-(?P<runner>[\w\.-]+)_(?P<job>\d+).zip"
35 ) -> list[dict[str, Any]]:
47 # Unzip to get perf stats csv files
51 print(f"Test report {path} has an invalid name. Skipping")
73 "workflow_id": workflow_run_id, # type: ignore[dict-item]
74 "run_attempt": workflow_run_attempt, # type: ignore[dict-item]
90 def generate_partition_key(repo: str, doc: Dict[str, Any]) -> str:
99 hash_content = hashlib.md5(json.dumps(doc).encode("utf-8")).hexdigest()
[all …]
Dupload_test_stats.py12 from tools.stats.test_dashboard import upload_additional_info
13 from tools.stats.upload_stats_lib import (
26 ) -> list[dict[str, Any]]:
27 """Convert a test report xml file into a JSON-serializable list of test cases."""
28 print(f"Parsing {tag}s for test report: {report}")
43 # The name of the file that the test is located in is not necessarily
44 # the same as the name of the file that invoked the test.
45 # For example, `test_jit.py` calls into multiple other test files (e.g.
46 # jit/test_dce.py). For sharding/test selection purposes, we want to
47 # record the file that invoked the test.
[all …]
Dupload_test_stat_aggregates.py13 from tools.stats.upload_stats_lib import upload_to_s3
16 def get_oncall_from_testfile(testfile: str) -> list[str] | None:
17 path = f"test/{testfile}"
20 # get oncall on test file
42 def get_test_stat_aggregates(date: datetime.date) -> Any:
70 description="Upload test stat aggregates to Rockset."
73 "--date",
75 help="Date to upload test stat aggregates for (YYYY-MM-DD). Must be in the last 30 days",
79 if args.date < datetime.datetime.now().date() - datetime.timedelta(days=30):
83 bucket_name="torchci-aggregated-stats",
Dupload_sccache_stats.py10 from tools.stats.upload_stats_lib import (
18 ) -> list[dict[str, Any]]:
24 download_s3_artifacts("sccache-stats", workflow_run_id, workflow_run_attempt)
34 parser = argparse.ArgumentParser(description="Upload test stats to Rockset")
36 "--workflow-run-id",
42 "--workflow-run-attempt",
48 stats = get_sccache_stats(args.workflow_run_id, args.workflow_run_attempt) variable
50 args.workflow_run_id, args.workflow_run_attempt, "sccache_stats", stats
Dupload_artifacts.py6 from tools.stats.upload_stats_lib import download_gha_artifacts, upload_file_to_s3
10 "sccache-stats",
11 "test-jsons",
12 "test-reports",
13 "usage-log",
15 BUCKET_NAME = "gha-artifacts"
16 FILENAME_REGEX = r"-runattempt\d+"
19 def get_artifacts(repo: str, workflow_run_id: int, workflow_run_attempt: int) -> None:
30 … # GHA artifact is named as follows: NAME-runattempt${{ github.run_attempt }}-SUFFIX.zip
32 # pytorch/pytorch/WORKFLOW_ID/RUN_ATTEMPT/artifact/NAME-SUFFIX.zip
[all …]
Dcheck_disabled_tests.py11 from tools.stats.upload_stats_lib import (
17 from tools.stats.upload_test_stats import process_xml_element
26 ) -> dict[str, dict[str, int]]:
28 Return a list of disabled tests that should be re-enabled and those that are still
35 # * Success test should be re-enable if it's green after rerunning in all platforms
37 # * Failures from pytest because pytest-flakefinder is used to run the same test
41 # We want to keep track of how many times the test fails (num_red) or passes (num_green)
47 # Under --rerun-disabled-tests mode, a test is skipped when:
49 # * it's skipped because it's a normal enabled test
72 # Check if the test is a failure
[all …]
/external/cronet/tot/net/reporting/
Dreporting_delivery_agent_unittest.cc2 // Use of this source code is governed by a BSD-style license that can be
17 #include "base/test/metrics/histogram_tester.h"
18 #include "base/test/scoped_feature_list.h"
19 #include "base/test/simple_test_tick_clock.h"
20 #include "base/test/values_test_util.h"
49 // This is a private API of the reporting service, so no need to test the in ReportingDeliveryAgentTest()
50 // case kPartitionConnectionsByNetworkIsolationKey is disabled - the in ReportingDeliveryAgentTest()
60 policy.endpoint_backoff_policy.maximum_backoff_ms = -1; in ReportingDeliveryAgentTest()
72 cache()->AddReport(reporting_source, network_anonymization_key, url, in AddReport()
74 /*depth=*/0, /*queued=*/tick_clock()->NowTicks(), in AddReport()
[all …]
/external/cronet/stable/net/reporting/
Dreporting_delivery_agent_unittest.cc2 // Use of this source code is governed by a BSD-style license that can be
17 #include "base/test/metrics/histogram_tester.h"
18 #include "base/test/scoped_feature_list.h"
19 #include "base/test/simple_test_tick_clock.h"
20 #include "base/test/values_test_util.h"
49 // This is a private API of the reporting service, so no need to test the in ReportingDeliveryAgentTest()
50 // case kPartitionConnectionsByNetworkIsolationKey is disabled - the in ReportingDeliveryAgentTest()
60 policy.endpoint_backoff_policy.maximum_backoff_ms = -1; in ReportingDeliveryAgentTest()
72 cache()->AddReport(reporting_source, network_anonymization_key, url, in AddReport()
74 /*depth=*/0, /*queued=*/tick_clock()->NowTicks(), in AddReport()
[all …]
/external/pytorch/.github/workflows/
Dupload-test-stats.yml1 name: Upload test stats
5 …trunk, periodic, inductor, unstable, slow, unstable-periodic, inductor-periodic, rocm, inductor-mi…
7 - completed
11 …ion adapted from https://github.com/community/community/discussions/21090#discussioncomment-3226271
14 runs-on: ubuntu-latest
18 - name: Get workflow run conclusion
19 uses: octokit/request-action@v2.1.0
26 upload-test-stats:
32 runs-on: ubuntu-22.04
33 environment: upload-stats
[all …]
Dupload-torch-dynamo-perf-stats.yml1 name: Upload torch dynamo performance stats
5 …workflows: [inductor-A100-perf-nightly, inductor-perf-nightly-A10g, inductor-perf-nightly-aarch64,…
7 - completed
10 get-conclusion:
11 runs-on: ubuntu-latest
13 conclusion: ${{ fromJson(steps.get-conclusion.outputs.data).conclusion }}
15 - name: Get workflow run conclusion
16 uses: octokit/request-action@v2.1.0
17 id: get-conclusion
23 upload-perf-stats:
[all …]
Dupload_test_stats_intermediate.yml1 name: Upload test stats intermediate
15 name: Intermediate upload test stats for ${{ inputs.workflow_id }}
16 runs-on: ubuntu-22.04
17 environment: upload-stats
19 - name: Checkout PyTorch
20 uses: pytorch/pytorch/.github/actions/checkout-pytorch@release/2.4
22 fetch-depth: 1
25 - uses: actions/setup-python@v4
27 python-version: '3.11'
30 - run: |
[all …]
D_win-build.yml1 name: windows-build
6 build-environment:
9 description: Top-level label for what's being built/tested.
10 cuda-version:
14 build-with-debug:
19 sync-tag:
25 job with the same `sync-tag` is identical.
26 test-matrix:
30 An option JSON description of what test configs to run later on. This
31 is moved here from the Linux test workflow so that we can apply filter
[all …]
D_mac-build.yml1 name: mac-build
6 build-environment:
9 description: Top-level label for what's being built/tested.
10 runner-type:
13 description: Name of the GitHub-managed runner type to use for the build.
14 build-generates-artifacts:
17 description: If set, upload generated build artifacts.
18 xcode-version:
23 sync-tag:
29 job with the same `sync-tag` is identical.
[all …]
Dupload-alerts.yml1 # upload alerts every 10 minutes
3 name: Upload Alerts to AWS/Rockset
7 - cron: '*/10 * * * *'
10 - 'tools/alerts/create_alerts.py'
11 - '.github/workflows/upload-alerts.yml'
14 upload-alerts:
16 runs-on: ubuntu-22.04
17 environment: upload-stats
19 - name: Checkout repo
22 fetch-depth: 1
[all …]
D_linux-build.yml1 name: linux-build
6 build-environment:
9 description: Top-level label for what's being built/tested.
10 docker-image-name:
14 build-generates-artifacts:
18 description: If set, upload generated build artifacts.
19 build-with-debug:
24 sync-tag:
30 job with the same `sync-tag` is identical.
31 cuda-arch-list:
[all …]
/external/pytorch/.github/scripts/
Drockset_mocks.json.gz
/external/deqp/modules/gles3/performance/
Des3pBufferDataUploadTests.cpp1 /*-------------------------------------------------------------------------
3 * -------------------------------------------------
11 * http://www.apache.org/licenses/LICENSE-2.0
21 * \brief Buffer data upload performance tests.
22 *//*--------------------------------------------------------------------*/
302 SingleOperationStatistics upload; member
321 SingleOperationStatistics upload; member
745 return endTime - startTime; in medianTimeMemcpy()
765 sectionLens[sectionNdx] = sectionStarts[sectionNdx + 1] - sectionStarts[sectionNdx]; in medianTimeMemcpy()
786 sectionTimes[sectionNdx] = endTime - startTime; in medianTimeMemcpy()
[all …]
/external/pytorch/benchmarks/
Dupload_scribe.py3 Currently supports data in pytest-benchmark format but can be extended.
45 def upload(self, messages): member in ScribeUploader
113 test = b["name"].split("[")[0]
115 benchmark_name = f"{test}[{net_name}]"
139 "stddev": b["stats"]["stddev"],
140 "rounds": b["stats"]["rounds"],
141 "min": b["stats"]["min"],
142 "median": b["stats"]["median"],
143 "max": b["stats"]["max"],
144 "mean": b["stats"]["mean"],
[all …]
/external/googleapis/google/ads/googleads/v14/resources/
Dcustomer.proto7 // http://www.apache.org/licenses/LICENSE-2.0
69 // Optional, non-unique descriptive name of the customer.
88 // Whether auto-tagging is enabled for the customer.
101 // Output only. Whether the customer is a test account.
118 // This field is read-only.
129 // for all manager customers, and for unscored non-manager customers.
132 // https://support.google.com/google-ads/answer/9061546.
134 // This field is read-only.
141 // scores across multiple non-manager customers. The aggregate optimization
144 // field is 0 for all manager customers, and for unscored non-manager
[all …]
/external/curl/tests/http/
Dtest_07_upload.py2 # -*- coding: utf-8 -*-
23 # SPDX-License-Identifier: curl
47 env.make_data_file(indir=env.gen_dir, fname="data-10k", fsize=10*1024)
48 env.make_data_file(indir=env.gen_dir, fname="data-63k", fsize=63*1024)
49 env.make_data_file(indir=env.gen_dir, fname="data-64k", fsize=64*1024)
50 env.make_data_file(indir=env.gen_dir, fname="data-100k", fsize=100*1024)
51 env.make_data_file(indir=env.gen_dir, fname="data-1m+", fsize=(1024*1024)+1)
52 env.make_data_file(indir=env.gen_dir, fname="data-10m", fsize=10*1024*1024)
56 # upload small data, check that this is what was echoed
65 url = f'https://{env.authority_for(env.domain1, proto)}/curltest/echo?id=[0-0]'
[all …]
/external/mobile-data-download/javatests/com/google/android/libraries/mobiledatadownload/file/transforms/
DCompressTransformTest.java8 * http://www.apache.org/licenses/LICENSE-2.0
35 import org.junit.Test;
57 @Test
62 assertThat(byteCounter.stats()).isEqualTo(new long[] {0, 20}); in counterWithCompress()
67 assertThat(byteCounter.stats()).isEqualTo(new long[] {20, 20}); in counterWithCompress()
71 assertThat(byteCounter.stats()).isEqualTo(new long[] {20, 44}); in counterWithCompress()
72 ratio = text.length() / (44 - 20); in counterWithCompress()
77 assertThat(byteCounter.stats()).isEqualTo(new long[] {20, 76}); in counterWithCompress()
78 ratio = text.length() / (76 - 44); in counterWithCompress()
82 @Test
[all …]
/external/grpc-grpc-java/interop-testing/src/main/proto/grpc/testing/
Dtest.proto2 // Copyright 2015-2016 gRPC authors.
8 // http://www.apache.org/licenses/LICENSE-2.0
16 // An integration test service that covers all the method signature permutations
28 // A simple service to test the various types of RPCs and experiment with
47 // A sequence of requests followed by one response (streamed upload).
65 // The test server will not implement this method. It will be used
66 // to test the behavior when clients call unimplemented methods.
70 // A simple service NOT implemented at servers so clients can test for
83 // A service used to obtain stats for verifying LB behavior.
85 // Gets the backend distribution for RPCs sent by a test client.
[all …]
/external/rust/android-crates-io/crates/grpcio-sys/grpc/src/proto/grpc/testing/
Dtest.proto2 // Copyright 2015-2016 gRPC authors.
8 // http://www.apache.org/licenses/LICENSE-2.0
16 // An integration test service that covers all the method signature permutations
26 // A simple service to test the various types of RPCs and experiment with
45 // A sequence of requests followed by one response (streamed upload).
63 // The test server will not implement this method. It will be used
64 // to test the behavior when clients call unimplemented methods.
68 // A simple service NOT implemented at servers so clients can test for
81 // A service used to obtain stats for verifying LB behavior.
83 // Gets the backend distribution for RPCs sent by a test client.
[all …]

1234567