Home
last modified time | relevance | path

Searched full:rockset (Results 1 – 25 of 37) sorted by relevance

12

/external/pytorch/tools/stats/
Dupload_test_stat_aggregates.py11 import rockset # type: ignore[import]
43 # Initialize the Rockset client with your API key
45 rockset_api_server = "api.rs2.usw2.rockset.com"
47 rs = rockset.RocksetClient(host="api.usw2a1.rockset.com", api_key=rockset_api_key)
49 # Define the name of the Rockset collection and lambda function
53 rockset.models.QueryParameter(name="startTime", type="string", value=iso_date)
70 description="Upload test stat aggregates to Rockset."
Dupload_dynamo_perf_stats.py105 description="Upload dynamo perf stats from S3 to Rockset"
132 "--rockset-collection",
135 help="the name of the Rockset collection to store the stats",
138 "--rockset-workspace",
141 help="the name of the Rockset workspace to store the stats",
163 # TODO (huydhn): Write to both Rockset and DynamoDB, an one-off script to copy
164 # data from Rockset to DynamoDB is the next step before uploading to Rockset
DREADME.md11 and uploads them as JSON to Rockset, our metrics backend.
21 uts --json--> R[(Rockset)]
24 Why this weird indirection? Because writing to Rockset requires special
Dupload_stats_lib.py14 import rockset # type: ignore[import]
23 # NB: Rockset has an upper limit of 5000 documents in one request
125 client = rockset.RocksetClient(
126 host="api.usw2a1.rockset.com", api_key=os.environ["ROCKSET_API_KEY"]
133 print(f"Writing {to_index - from_index} documents to Rockset")
157 # This is to move away the _event_time field from Rockset, which we cannot use when
Dupload_test_stats_intermediate.py9 parser = argparse.ArgumentParser(description="Upload test stats to Rockset")
27 # Flush stdout so that any errors in Rockset upload show up last in the logs.
Dupload_test_stats.py70 # possible to make aggregation possible in Rockset.
220 parser = argparse.ArgumentParser(description="Upload test stats to Rockset")
258 # Flush stdout so that any errors in Rockset upload show up last in the logs.
262 # volume of writes we do to Rockset.
Dupload_sccache_stats.py34 parser = argparse.ArgumentParser(description="Upload test stats to Rockset")
/external/pytorch/.github/workflows/
Dnightly-rockset-uploads.yml1 name: Nightly Upload to rockset
11 - '.github/workflows/nightly-rockset-uploads.yml'
19 upload-stats-to-rockset:
35 pip3 install requests==2.32.2 rockset==1.0.3 boto3==1.19.12
Dupload-torch-dynamo-perf-stats.yml52 pip3 install requests==2.32.2 rockset==1.0.3 boto3==1.19.12
67 - name: Upload torch dynamo performance stats to Rockset
76 …{REPO_FULLNAME}" --head-branch "${HEAD_BRANCH}" --rockset-collection torch_dynamo_perf_stats --roc…
Dupload-test-stats.yml50 pip3 install requests==2.32.2 rockset==1.0.3 boto3==1.19.12
98 - name: Upload gpt-fast benchmark results to Rockset
109 …o "${REPO_FULLNAME}" --head-branch "${HEAD_BRANCH}" --rockset-collection oss_ci_benchmark --rockse…
Dclose-nonexistent-disable-issues.yml9 environment: rockset-read-only
22 pip3 install rockset==1.0.3
Dupload-alerts.yml3 name: Upload Alerts to AWS/Rockset
31 pip3 install rockset==1.0.3 boto3==1.19.12 requests==2.32.2
Dupdate-viablestrict.yml25 rockset-api-key: ${{ secrets.ROCKSET_API_KEY }}
Dupload_test_stats_intermediate.yml31 pip3 install requests==2.32.2 rockset==1.0.3 boto3==1.19.12
Dcherry-pick.yml29 - run: pip install pyyaml==6.0 rockset==1.0.3
Dcheck_mergeability_ghstack.yml58 - run: pip install pyyaml==6.0 rockset==1.0.3
/external/pytorch/.github/scripts/
Dclose_nonexistent_disable_issues.py10 import rockset # type: ignore[import]
53 "I cannot find any mention of this test in rockset for the past 7 days "
79 res = rockset.RocksetClient(
80 host="api.rs2.usw2.rockset.com", api_key=os.environ["ROCKSET_API_KEY"]
140 # Query rockset to see if the test is there
146 return True, "found in rockset"
Dtrymerge.py1182 # Finally, upload the record to Rockset. The list of pending and failed
1204 print("Missing comment ID or PR number, couldn't upload to Rockset")
1472 # Rockset merge records
1531 # Prepare the record to be written into Rockset
1553 # in rockset. Any unique string would work
1582 import rockset # type: ignore[import]
1584 res = rockset.RocksetClient(
1585 host="api.usw2a1.rockset.com", api_key=os.environ["ROCKSET_API_KEY"]
1589 print("Could not use RockSet as rocket dependency is missing")
2030 …egorization is used to keep track of all ignorable failures when saving the merge record on Rockset
[all …]
/external/pytorch/benchmarks/dynamo/ci_expected_accuracy/
Dupdate_expected.py34 # Note: the public query url targets this rockset lambda:
35 # https://console.rockset.com/lambdas/details/commons.artifacts
36 ARTIFACTS_QUERY_URL = "https://api.usw2a1.rockset.com/v1/public/shared_lambdas/4ca0033e-0117-41f5-b…
/external/pytorch/tools/testing/
Dupdate_slow_tests.py9 import rockset # type: ignore[import]
172 rs_client = rockset.RocksetClient(
173 host="api.usw2a1.rockset.com", api_key=os.environ["ROCKSET_API_KEY"]
/external/executorch/.github/scripts/
Dtrymerge.py1167 # Finally, upload the record to Rockset. The list of pending and failed
1192 print("Missing comment ID or PR number, couldn't upload to Rockset")
1460 # Rockset merge records
1519 This saves the merge records into Rockset, so we can query them (for fun and profit)
1522 # Decide not to save the record to Rockset if dry-run is set to not pollute
1527 import rockset # type: ignore[import]
1529 # Prepare the record to be written into Rockset
1553 client = rockset.RocksetClient(
1554 host="api.usw2a1.rockset.com", api_key=os.environ["ROCKSET_API_KEY"]
1563 print("Rockset is missing, no record will be saved")
[all …]
/external/pytorch/.github/
Drequirements-gha-cache.txt15 rockset==1.0.3
/external/pytorch/.github/requirements/
Dpip-requirements-macOS.txt27 rockset==1.0.3
/external/executorch/.github/workflows/
Dcheck-labels.yml47 - run: pip install pyyaml==6.0 rockset==1.0.3
Dcherry-pick.yml30 - run: pip install pyyaml==6.0 rockset==1.0.3

12