Searched +full:build +full:- +full:docker +full:- +full:cpu (Results 1 – 25 of 161) sorted by relevance
1234567
/external/pytorch/.github/workflows/ |
D | build-manywheel-images.yml | 1 name: Build manywheel docker images 7 - main 8 - release/* 10 … # NOTE: Binary build pipelines should only get triggered on release candidate or nightly builds 11 # Release candidate tags look like: v1.11.0-rc1 12 - v[0-9]+.[0-9]+.[0-9]+-rc[0-9]+ 14 - '.ci/docker/manywheel/*' 15 - '.ci/docker/manywheel/build_scripts/*' 16 - '.ci/docker/common/*' 17 - .github/workflows/build-manywheel-images.yml [all …]
|
D | build-conda-images.yml | 1 name: Build conda docker images 7 - main 8 - release/* 10 … # NOTE: Binary build pipelines should only get triggered on release candidate or nightly builds 11 # Release candidate tags look like: v1.11.0-rc1 12 - v[0-9]+.[0-9]+.[0-9]+-rc[0-9]+ 14 - '.ci/docker/conda/*' 15 - '.ci/docker/common/*' 16 - .github/workflows/build-conda-images.yml 19 - '.ci/docker/conda/*' [all …]
|
D | build-libtorch-images.yml | 1 name: Build libtorch docker images 6 - main 7 - release/* 9 … # NOTE: Binary build pipelines should only get triggered on release candidate or nightly builds 10 # Release candidate tags look like: v1.11.0-rc1 11 - v[0-9]+.[0-9]+.[0-9]+-rc[0-9]+ 13 - '.ci/docker/libtorch/*' 14 - '.ci/docker/common/*' 15 - .github/workflows/build-libtorch-images.yml 18 - '.ci/docker/libtorch/*' [all …]
|
D | trunk.yml | 6 - main 7 - release/* 8 - landchecks/* 10 - ciflow/trunk/* 13 - cron: 29 8 * * * # about 1:29am PDT 16 …-${{ github.event.pull_request.number || github.ref_name }}-${{ github.ref_type == 'branch' && git… 17 cancel-in-progress: true 19 permissions: read-all 22 llm-td: 23 name: before-test [all …]
|
D | generated-linux-s390x-binary-manywheel-nightly.yml | 5 name: linux-s390x-binary-manywheel 12 - nightly 14 # NOTE: Binary build pipelines should only get triggered on release candidate builds 15 # Release candidate tags look like: v1.11.0-rc1 16 - v[0-9]+.[0-9]+.[0-9]+-rc[0-9]+ 17 - 'ciflow/binaries/*' 18 - 'ciflow/binaries_wheel/*' 23 ALPINE_IMAGE: "docker.io/s390x/alpine" 25 AWS_DEFAULT_REGION: us-east-1 27 BUILD_ENVIRONMENT: linux-s390x-binary-manywheel [all …]
|
D | inductor-perf-test-nightly-x86.yml | 1 name: inductor-perf-nightly-x86 5 # - cron: 0 7 * * 1-6 6 # - cron: 0 7 * * 0 7 # Does not perform max_autotune on CPU, so skip the weekly run setup 8 - cron: 0 7 * * * 13 # CPU for training is not typical, but leave the option open here 45 …-${{ github.event.pull_request.number || github.ref_name }}-${{ github.ref_type == 'branch' && git… 46 cancel-in-progress: true 48 permissions: read-all 51 linux-jammy-cpu-py3_9-gcc11-inductor-build: [all …]
|
D | pull.yml | 5 branches-ignore: 6 - nightly 9 - main 10 - release/* 11 - landchecks/* 14 - cron: 29 8 * * * # about 1:29am PDT 17 …p: ${{ github.workflow }}-${{ github.event.pull_request.number || github.sha }}-${{ github.event_n… 18 cancel-in-progress: true 20 permissions: read-all 23 llm-td: [all …]
|
D | inductor.yml | 6 - main 7 - release/* 9 - ciflow/inductor/* 11 - cron: 29 8 * * * # about 1:29am PDT, for mem leak check and rerun disabled tests 15 …${{ github.workflow }}-${{ github.event.pull_request.number || github.ref_name }}-${{ github.ref_t… 16 cancel-in-progress: true 18 permissions: read-all 21 get-label-type: 22 name: get-label-type 23 uses: ./.github/workflows/_runner-determinator.yml [all …]
|
D | inductor-micro-benchmark-x86.yml | 1 name: inductor-micro-benchmark-x86 5 - cron: 0 7 * * * 8 - ciflow/inductor-micro-benchmark-cpu-x86/* 13 …-${{ github.event.pull_request.number || github.ref_name }}-${{ github.ref_type == 'branch' && git… 14 cancel-in-progress: true 16 permissions: read-all 19 linux-jammy-cpu-py3_9-gcc11-inductor-build: 20 name: linux-jammy-cpu-py3.9-gcc11-inductor 21 uses: ./.github/workflows/_linux-build.yml 23 build-environment: linux-jammy-py3.9-gcc11 [all …]
|
/external/pytorch/.ci/docker/manywheel/ |
D | build.sh | 4 set -eou pipefail 6 TOPDIR=$(git rev-parse --show-toplevel) 11 if [ -z "${image}" ]; then 18 DOCKER_REGISTRY="${DOCKER_REGISTRY:-docker.io}" 20 GPU_ARCH_TYPE=${GPU_ARCH_TYPE:-cpu} 21 GPU_ARCH_VERSION=${GPU_ARCH_VERSION:-} 22 MANY_LINUX_VERSION=${MANY_LINUX_VERSION:-} 23 DOCKERFILE_SUFFIX=${DOCKERFILE_SUFFIX:-} 24 WITH_PUSH=${WITH_PUSH:-} 27 cpu) [all …]
|
/external/pytorch/.ci/docker/libtorch/ |
D | build.sh | 4 set -eou pipefail 9 if [ -z "${image}" ]; then 16 TOPDIR=$(git rev-parse --show-toplevel) 18 GPU_ARCH_TYPE=${GPU_ARCH_TYPE:-cpu} 19 GPU_ARCH_VERSION=${GPU_ARCH_VERSION:-} 21 WITH_PUSH=${WITH_PUSH:-} 23 DOCKER=${DOCKER:-docker} 26 cpu) 27 BASE_TARGET=cpu 28 DOCKER_TAG=cpu [all …]
|
/external/tensorflow/tensorflow/tools/ci_build/ |
D | ci_build.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 17 # Usage: ci_build.sh <CONTAINER_TYPE> [--dockerfile <DOCKERFILE_PATH>] 20 # CONTAINER_TYPE: Type of the docker container used the run the build: 21 # e.g., (cpu | gpu | rocm | android | tensorboard) 23 # DOCKERFILE_PATH: (Optional) Path to the Dockerfile used for docker build. 25 # --dockerfile flag), default Dockerfiles in the same 28 # COMMAND: Command to be executed in the docker container, e.g., 29 # tensorflow/tools/ci_build/builds/pip.sh gpu -c opt --config=cuda 38 # Dockerfile to be used in docker build 42 if [[ "$1" == "--dockerfile" ]]; then [all …]
|
/external/pytorch/.github/scripts/s390x-ci/ |
D | README.md | 6 $ sudo dnf install podman podman-docker jq 12 $ sudo cp self-hosted-builder/*.service /etc/systemd/system/ 13 $ sudo systemctl daemon-reload 16 ## Download qemu-user-static image 19 # sudo docker pull docker.io/iiilinuxibmcom/qemu-user-static:6.1.0-1 25 $ sudo systemctl enable --now qemu-user-static 30 First build s390x builder image `docker.io/pytorch/manylinuxs390x-builder`, 37 $ git submodule update --init --recursive 38 $ GPU_ARCH_TYPE=cpu-s390x "$(pwd)/.ci/docker/manywheel/build.sh" manylinuxs390x-builder 39 $ docker image tag localhost/pytorch/manylinuxs390x-builder docker.io/pytorch/manylinuxs390x-builde… [all …]
|
/external/skia/infra/bots/recipe_modules/build/ |
D | canvaskit.py | 2 # Use of this source code is governed by a BSD-style license that can be 5 DOCKER_IMAGE = 'gcr.io/skia-public/canvaskit-emsdk:3.1.26_v2' 11 out_dir = api.vars.cache_dir.joinpath('docker', 'canvaskit') 15 # We want to make sure the directories exist and were created by chrome-bot, 16 # because if that isn't the case, docker will make them and they will be 19 # param. Instead, we use a "canvaskit" subdirectory in the "docker" named_cache. 22 # Download the emsdk binaries (we won't actually use the ones on the Docker 25 api.run(api.step, 'activate-emsdk', 26 cmd=['python3', skia_dir.joinpath('bin', 'activate-emsdk')], 29 # This uses the emscripten sdk docker image and says "run the [all …]
|
/external/pytorch/ |
D | docker.Makefile | 1 DOCKER_REGISTRY ?= docker.io 2 DOCKER_ORG ?= $(shell docker info 2>/dev/null | sed '/Username:/!d;s/.* //') 7 $(warning WARNING: No docker user found using results from whoami) 15 BASE_DEVEL = nvidia/cuda:$(CUDA_VERSION)-devel-ubuntu22.04 23 CUDA_PATH ?= cpu 24 ifneq ("$(CUDA_VERSION_SHORT)","cpu") 30 PYTORCH_VERSION ?= $(shell git describe --tags --always --match "v[1-9]*.*") 36 BUILD_ARGS = --build-arg BASE_IMAGE=$(BASE_IMAGE) \ 37 --build-arg PYTHON_VERSION=$(PYTHON_VERSION) \ 38 --build-arg CUDA_VERSION=$(CUDA_VERSION) \ [all …]
|
/external/tensorflow/tensorflow/tools/ci_build/builds/ |
D | docker_test.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 17 # Build and test TensorFlow docker images. 18 # The tests include Python unit tests-on-install and tutorial tests. 22 # IMAGE_TYPE : Type of the image: (CPU|GPU|ROCM) 23 # TAG : Docker image tag 24 # WHL_PATH : Path to the whl file to be installed inside the docker image 26 # e.g.: docker_test.sh CPU someone/tensorflow:0.8.0 pip_test/whl/tensorflow-0.8.0-cp27-none-linux… 45 test -e "$1" && echo "$PWD" && return || \ 59 if [[ "${IMAGE_TYPE}" == "cpu" ]]; then 60 DOCKERFILE="tensorflow/tools/docker/Dockerfile" [all …]
|
/external/pytorch/.ci/docker/conda/ |
D | build.sh | 4 set -eou pipefail 9 if [ -z "${image}" ]; then 18 TOPDIR=$(git rev-parse --show-toplevel) 20 CUDA_VERSION=${CUDA_VERSION:-12.1} 23 cpu) 25 DOCKER_TAG=cpu 39 set -x 40 docker build \ 41 --target final \ 42 --progress plain \ [all …]
|
/external/tensorflow/tensorflow/tools/dockerfiles/ |
D | README.md | 5 [TensorFlow's Docker Hub](https://hub.docker.com/r/tensorflow/tensorflow). 14 The Dockerfiles in the `dockerfiles` directory must have their build context set 18 $ docker build -f ./dockerfiles/cpu.Dockerfile -t tf . 21 Each Dockerfile has its own set of available `--build-arg`s which are documented 26 After building the image with the tag `tf` (for example), use `docker run` to 29 Note for new Docker users: the `-v` and `-u` flags share directories and 30 permissions between the Docker container and your machine. Without `-v`, your 31 work will be wiped once the container quits, and without `-u`, files created by 34 [Docker run documentation](https://docs.docker.com/engine/reference/run/) for 38 # Volume mount (-v) is optional but highly recommended, especially for Jupyter. [all …]
|
/external/perfetto/docs/deployment/ |
D | deploying-bigtrace-on-kubernetes.md | 3 … of Bigtrace services NOT Bigtrace users. This is also designed for non-Googlers - Googlers should… 11  14 There are three clients to interact with Bigtrace: a Python API, clickhouse-client and Apache Super… 15 - The Python API exists in the Perfetto python library and can be used similar to the TraceProcesso… 16 - Clickhouse is a data warehousing solution which gives a SQL based interface for the user to write… 17 - Superset is a GUI for Clickhouse which offers an SQLLab to run queries offering support for moder… 37 - A GCP Project 38 - GCS 39 - GKE 40 - gcloud (https://cloud.google.com/sdk/gcloud) [all …]
|
/external/skia/infra/bots/recipes/test_pathkit.expected/ |
D | Test-Debian10-EMCC-GCE-CPU-AVX2-asmjs-Debug-All-PathKit.json | 4 "name": "Docker setup" 9 "RECIPE_MODULE[skia::docker]/resources/get_uid_gid.py" 11 "name": "Docker setup.Get uid and gid", 19 "-u", 21 "--json-output", 23 "ensure-directory", 24 "--mode", 29 "name": "Docker setup.mkdirs out_dir", 41 "name": "Docker setup.chmod 777 [START_DIR]/[SWARM_OUT_DIR]", 53 "name": "Docker setup.chmod 755 [START_DIR]", [all …]
|
D | pathkit_trybot.json | 4 "name": "Docker setup" 9 "RECIPE_MODULE[skia::docker]/resources/get_uid_gid.py" 11 "name": "Docker setup.Get uid and gid", 19 "-u", 21 "--json-output", 23 "ensure-directory", 24 "--mode", 29 "name": "Docker setup.mkdirs out_dir", 41 "name": "Docker setup.chmod 777 [START_DIR]/[SWARM_OUT_DIR]", 53 "name": "Docker setup.chmod 755 [START_DIR]", [all …]
|
D | Test-Debian10-EMCC-GCE-CPU-AVX2-asmjs-Release-All-PathKit.json | 4 "name": "Docker setup" 9 "RECIPE_MODULE[skia::docker]/resources/get_uid_gid.py" 11 "name": "Docker setup.Get uid and gid", 19 "-u", 21 "--json-output", 23 "ensure-directory", 24 "--mode", 29 "name": "Docker setup.mkdirs out_dir", 41 "name": "Docker setup.chmod 777 [START_DIR]/[SWARM_OUT_DIR]", 53 "name": "Docker setup.chmod 755 [START_DIR]", [all …]
|
D | Test-Debian10-EMCC-GCE-CPU-AVX2-wasm-Debug-All-PathKit.json | 4 "name": "Docker setup" 9 "RECIPE_MODULE[skia::docker]/resources/get_uid_gid.py" 11 "name": "Docker setup.Get uid and gid", 19 "-u", 21 "--json-output", 23 "ensure-directory", 24 "--mode", 29 "name": "Docker setup.mkdirs out_dir", 41 "name": "Docker setup.chmod 777 [START_DIR]/[SWARM_OUT_DIR]", 53 "name": "Docker setup.chmod 755 [START_DIR]", [all …]
|
/external/pytorch/.ci/docker/ |
D | README.md | 1 # Docker images for GitHub CI and CD 3 This directory contains everything needed to build the Docker images 7 conditionally run build stages depending on build arguments passed to 8 `docker build`. This lets us use only a few Dockerfiles for many 13 See `build.sh` for valid build environments (it's the giant switch). 15 ## Docker CI builds 17 * `build.sh` -- dispatch script to launch all builds 18 * `common` -- scripts used to execute individual Docker build stages 19 * `ubuntu` -- Dockerfile for Ubuntu image for CPU build and test jobs 20 * `ubuntu-cuda` -- Dockerfile for Ubuntu image with CUDA support for nvidia-docker [all …]
|
/external/skia/infra/bots/recipes/test_canvaskit.expected/ |
D | canvaskit_trybot.json | 5 "-u", 7 "--json-output", 9 "ensure-directory", 10 "--mode", 12 "[START_DIR]/skia/modules/canvaskit/build" 19 "name": "Docker setup" 24 "RECIPE_MODULE[skia::docker]/resources/get_uid_gid.py" 26 "name": "Docker setup.Get uid and gid", 34 "-u", 36 "--json-output", [all …]
|
1234567