| /third_party/grpc/examples/python/ |
| D | Makefile | 7 # http://www.apache.org/licenses/LICENSE-2.0 17 ARTIFACTS= macro 19 ARTIFACTS += helloworld/helloworld_pb2.py 20 ARTIFACTS += helloworld/helloworld_pb2_grpc.py 21 ARTIFACTS += helloworld/helloworld_pb2.pyi 23 ARTIFACTS += timeout/helloworld_pb2.py 24 ARTIFACTS += timeout/helloworld_pb2_grpc.py 25 ARTIFACTS += timeout/helloworld_pb2.pyi 27 ARTIFACTS += keep_alive/helloworld_pb2.py 28 ARTIFACTS += keep_alive/helloworld_pb2_grpc.py [all …]
|
| /third_party/rust/crates/bindgen/.github/workflows/ |
| D | release.yml | 1 # Copyright 2022-2023, axodotdev 2 # SPDX-License-Identifier: MIT or Apache-2.0 7 # * builds artifacts with cargo-dist (archives, installers, hashes) 8 # * uploads those artifacts to temporary workflow zip 9 # * on success, uploads the artifacts to a Github Release 20 # like "1.0.0", "v0.1.0-prerelease.1", "my-app/0.1.0", "releases/v1.0.0", etc. 23 # must be a Cargo-style SemVer Version (must have at least major.minor.patch). 26 # package (erroring out if it doesn't have the given version or isn't cargo-dist-able). 29 # (cargo-dist-able) packages in the workspace with that version (this mode is 30 # intended for workspaces with only one dist-able package, or with all dist-able [all …]
|
| /third_party/mesa3d/.gitlab-ci/ |
| D | prepare-artifacts-python.sh | 6 . "${SCRIPTS_DIR}/setup-test-env.sh" 8 section_switch prepare-artifacts "artifacts: prepare" 10 set -e 11 set -o xtrace 13 mkdir -p artifacts/ 17 echo "$(cat VERSION) (git-$(git rev-parse HEAD | cut -b -10))" > artifacts/VERSION 18 cp -Rp .gitlab-ci/report-flakes.py artifacts/ 19 cp -Rp .gitlab-ci/setup-test-env.sh artifacts/ 20 cp -Rp .gitlab-ci/common artifacts/ci-common 21 cp -Rp .gitlab-ci/b2c artifacts/ [all …]
|
| D | prepare-artifacts.sh | 6 . "${SCRIPTS_DIR}/setup-test-env.sh" 8 section_switch prepare-artifacts "artifacts: prepare" 10 set -e 11 set -o xtrace 13 CROSS_FILE=/cross_file-"$CROSS".txt 15 # Delete unused bin and includes from artifacts to save space. 16 rm -rf install/bin install/include 17 rm -f install/lib/*.a 19 # Strip the drivers in the artifacts to cut 80% of the artifacts size. 20 if [ -n "$CROSS" ]; then [all …]
|
| /third_party/mesa3d/.gitlab-ci/lava/ |
| D | lava-gitlab-ci.yml | 2 …REGISTRY_IMAGE}/alpine/x86_64_lava_ssh_client:${ALPINE_X86_64_LAVA_SSH_TAG}--${MESA_TEMPLATES_COMM… 5 .lava-test: 10 FDO_CI_CONCURRENT: 6 # should be replaced by per-machine definitions 12 LAVA_HTTP_CACHE_URI: "http://caching-proxy/cache/?uri=" 17 # per-job build artifacts 18 JOB_ROOTFS_OVERLAY_PATH: "${JOB_ARTIFACTS_BASE}/job-rootfs-overlay.tar.gz" 20 LAVA_S3_ARTIFACT_NAME: "mesa-${ARCH}-default-debugoptimized" 21 S3_ARTIFACT_NAME: "mesa-python-ci-artifacts" 28 - !reference [.download_s3, before_script] 30 - . artifacts/setup-test-env.sh [all …]
|
| /third_party/grpc/tools/release/ |
| D | verify_python_release.py | 9 # http://www.apache.org/licenses/LICENSE-2.0 16 """Verifies that all gRPC Python artifacts have been successfully published. 18 This script is intended to be run from a directory containing the artifacts 19 that have been uploaded and only the artifacts that have been uploaded. We use 39 "grpcio-tools", 40 "grpcio-status", 41 "grpcio-health-checking", 42 "grpcio-reflection", 43 "grpcio-channelz", 44 "grpcio-testing", [all …]
|
| /third_party/grpc/tools/internal_ci/linux/ |
| D | grpc_distribtests_python.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 26 # some distribtests use a pre-registered binfmt_misc hook 34 # Build all python linux artifacts (this step actually builds all the binary wheels and source arch… 35 tools/run_tests/task_runner.py -f artifact linux python ${TASK_RUNNER_EXTRA_FILTERS} -j 12 -x build… 37 # the next step expects to find the artifacts from the previous step in the "input_artifacts" folde… 38 rm -rf input_artifacts 39 mkdir -p input_artifacts 40 cp -r artifacts/* input_artifacts/ || true 42 # This step simply collects python artifacts from subdirectories of input_artifacts/ and copies the… [all …]
|
| D | grpc_distribtests_php.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 26 # Build all PHP linux artifacts 27 tools/run_tests/task_runner.py -f artifact linux php ${TASK_RUNNER_EXTRA_FILTERS} -j 2 -x build_art… 29 # the next step expects to find the artifacts from the previous step in the "input_artifacts" folde… 30 rm -rf input_artifacts 31 mkdir -p input_artifacts 32 cp -r artifacts/* input_artifacts/ || true 34 # This step simply collects php artifacts from subdirectories of input_artifacts/ and copies them t… 35 tools/run_tests/task_runner.py -f package linux php -x build_packages/sponge_log.xml || FAILED="tru… [all …]
|
| D | grpc_distribtests_gcp_python.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 26 # some distribtests use a pre-registered binfmt_misc hook 34 # Build all python linux artifacts (this step actually builds all the binary wheels and source arch… 35 tools/run_tests/task_runner.py -f artifact linux python latest-manylinux ${TASK_RUNNER_EXTRA_FILTER… 37 # the next step expects to find the artifacts from the previous step in the "input_artifacts" folde… 38 rm -rf input_artifacts 39 mkdir -p input_artifacts 40 cp -r artifacts/* input_artifacts/ || true 41 rm -rf artifacts_from_build_artifacts_step [all …]
|
| D | grpc_distribtests_csharp.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 26 # some distribtests use a pre-registered binfmt_misc hook 34 # Build all protoc linux artifacts 35 tools/run_tests/task_runner.py -f artifact linux protoc ${TASK_RUNNER_EXTRA_FILTERS} -j 4 --inner_j… 37 # the next step expects to find the artifacts from the previous step in the "input_artifacts" folde… 38 rm -rf input_artifacts 39 mkdir -p input_artifacts 40 cp -r artifacts/* input_artifacts/ || true 44 # (this is required as we only have the linux artifacts at hand) [all …]
|
| D | grpc_distribtests_ruby.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 32 # Build all ruby linux artifacts (this step actually builds all the native and source gems) 33 tools/run_tests/task_runner.py -f artifact linux ruby ${TASK_RUNNER_EXTRA_FILTERS} -j 6 --inner_job… 37 # Note that by skipping the "build_package" step, we are also skipping the build of "grpc-tools" gem 40 # The next step expects to find the artifacts from the previous step in the "input_artifacts" folde… 41 # in addition to that, preserve the contents of "artifacts" directory since we want kokoro 42 # to upload its contents as job output artifacts. 43 rm -rf input_artifacts 44 mkdir -p input_artifacts [all …]
|
| /third_party/grpc/tools/internal_ci/macos/ |
| D | grpc_distribtests_csharp.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 27 # Build all protoc macos artifacts 28 tools/run_tests/task_runner.py -f artifact macos protoc ${TASK_RUNNER_EXTRA_FILTERS} -j 2 --inner_j… 30 # the next step expects to find the artifacts from the previous step in the "input_artifacts" folde… 31 rm -rf input_artifacts 32 mkdir -p input_artifacts 33 cp -r artifacts/* input_artifacts/ || true 37 # (this is required as we only have the macos artifacts at hand) 38 …LD_SINGLE_PLATFORM_NUGET=1 tools/run_tests/task_runner.py -f package macos csharp nuget -j 2 -x bu… [all …]
|
| D | grpc_distribtests_python.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 29 python3.8 -m pip install -U 'cython<4.0.0rc1' setuptools==65.4.1 wheel --user 30 python3.9 -m pip install -U 'cython<4.0.0rc1' setuptools==65.4.1 wheel --user 31 python3.10 -m pip install -U 'cython<4.0.0rc1' setuptools==65.4.1 wheel --user 32 python3.11 -m pip install -U 'cython<4.0.0rc1' setuptools==65.4.1 wheel --user 33 python3.12 -m pip install -U 'cython<4.0.0rc1' setuptools==65.4.1 wheel --user 34 python3.13 -m pip install -U 'cython<4.0.0rc1' setuptools==65.4.1 wheel --user 36 # Build all python macos artifacts (this step actually builds all the binary wheels and source arch… 37 tools/run_tests/task_runner.py -f artifact macos python ${TASK_RUNNER_EXTRA_FILTERS} -j 2 -x build_… [all …]
|
| D | grpc_distribtests_php.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 27 # Build all PHP macos artifacts 28 tools/run_tests/task_runner.py -f artifact macos php ${TASK_RUNNER_EXTRA_FILTERS} -j 4 -x build_art… 30 # PHP's "build_package" step is basically just a passthough, so the build artifacts can be used 33 # the next step expects to find the artifacts from the previous step in the "input_artifacts" folde… 34 # in addition to that, preserve the contents of "artifacts" directory since we want kokoro 35 # to upload its contents as job output artifacts 36 rm -rf input_artifacts 37 mkdir -p input_artifacts [all …]
|
| /third_party/grpc/tools/internal_ci/windows/ |
| D | grpc_distribtests_python.bat | 7 @rem http://www.apache.org/licenses/LICENSE-2.0 31 @rem Build all python windows artifacts 32 python tools/run_tests/task_runner.py -f artifact windows python %TASK_RUNNER_EXTRA_FILTERS% -j 4 -… 34 @rem the next step expects to find the artifacts from the previous step in the "input_artifacts" fo… 35 bash -c "rm -rf input_artifacts; mkdir -p input_artifacts; cp -r artifacts/* input_artifacts/ || tr… 37 @rem Collect the python artifact from subdirectories of input_artifacts/ to artifacts/ 38 @rem TODO(jtattermusch): when collecting the artifacts that will later be uploaded as kokoro job ar… 39 @rem potentially skip some file names that would clash with linux-created artifacts. 40 bash -c "cp -r input_artifacts/python_*/* artifacts/ || true" 45 @rem This step checks if any of the artifacts exceeds a per-file size limit.
|
| D | grpc_distribtests_csharp.bat | 7 @rem http://www.apache.org/licenses/LICENSE-2.0 33 @rem Install Msys2 zip to avoid crash when using cygwin's zip on grpc-win2016 kokoro workers. 37 curl -sSL --fail -o C:\zip\zip.exe https://storage.googleapis.com/grpc-build-helper/zip-3.0-1-x86_6… 39 zip --version 41 @rem Build all protoc windows artifacts 42 python tools/run_tests/task_runner.py -f artifact windows protoc %TASK_RUNNER_EXTRA_FILTERS% -j 4 -… 44 @rem the next step expects to find the artifacts from the previous step in the "input_artifacts" fo… 45 bash -c "rm -rf input_artifacts; mkdir -p input_artifacts; cp -r artifacts/* input_artifacts/ || tr… 49 @rem (this is required as we only have the windows artifacts at hand) 51 python tools/run_tests/task_runner.py -f package windows csharp nuget -j 2 -x build_packages/sponge… [all …]
|
| /third_party/rust/rust/src/tools/rust-analyzer/editors/code/src/ |
| D | toolchain.ts | 17 filter?: (artifacts: CompilationArtifact[]) => CompilationArtifact[]; 29 const cargoArgs = [...args, "--message-format=json"]; 32 // see crates\rust-analyzer\src\main_loop\handlers.rs, handle_code_lens 38 if (!cargoArgs.includes("--no-run")) { 39 cargoArgs.push("--no-run"); 47 // for instance, `crates\rust-analyzer\tests\heavy_tests\main.rs` tests 48 // produce 2 artifacts: {"kind": "bin"} and {"kind": "test"} 49 result.filter = (artifacts) => artifacts.filter((it) => it.isTest); 56 const artifacts: CompilationArtifact[] = []; constant 62 if (message.reason === "compiler-artifact" && message.executable) { [all …]
|
| /third_party/grpc/tools/bazelify_tests/ |
| D | grpc_build_artifact_task.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -e 26 tar -xopf ${ARCHIVE_WITH_SUBMODULES} 29 # Extract all input archives with artifacts into input_artifacts directory 31 mkdir -p input_artifacts 33 # all remaining args are .tar.gz archives with input artifacts 36 # extract the .tar.gz with artifacts into a directory named after a basename 38 # Note that input artifacts from different dependencies can have files 42 mkdir -p "${archive_extract_dir}" 44 tar --strip-components=1 -xopf ../../../${input_artifact_archive} [all …]
|
| /third_party/mesa3d/.gitlab-ci/test/ |
| D | gitlab-ci.yml | 7 - !reference [default, before_script] 9 - rm -rf install 10 - tar -xf artifacts/install.tar 11 - section_start ldd_section "Checking ldd on driver build" 12 - LD_LIBRARY_PATH=install/lib find install/lib -name "*.so" -print -exec ldd {} \; 13 - section_end ldd_section 14 artifacts: 18 - results/ 20 - !reference [.common-rules, rules] 21 - !reference [.never-post-merge-rules, rules] [all …]
|
| /third_party/grpc/tools/run_tests/artifacts/ |
| D | artifact_targets.py | 8 # http://www.apache.org/licenses/LICENSE-2.0 15 """Definition of targets to build artifacts.""" 41 environ["ARTIFACTS_OUT"] = "artifacts/%s" % name 45 docker_args += ["-e", "%s=%s" % (k, v)] 50 "OUTPUT_DIR": "artifacts", 83 environ["ARTIFACTS_OUT"] = os.path.join("..", "artifacts", name) 86 "tools/run_tests/artifacts/run_in_workspace.sh", 89 environ["ARTIFACTS_OUT"] = os.path.join("artifacts", name) 105 _MACOS_COMPAT_FLAG = "-mmacosx-version-min=10.10" 107 _ARCH_FLAG_MAP = {"x86": "-m32", "x64": "-m64"} [all …]
|
| /third_party/grpc/tools/internal_ci/linux/release/ |
| D | grpc_collect_all_packages.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 26 # Move artifacts generated by the previous step in the build chain to a place 29 mv ${KOKORO_GFILE_DIR}/github/grpc/artifacts input_artifacts || true 31 ls -R input_artifacts || true 33 # Collect the packages from the "build artifacts" phase and simply copy 35 # artifacts from the "build artifacts" phase are actually combined 36 # into a single multi-platform package file: 37 # - C#'s Grpc.Tools package requires protoc plugins built on multiple platforms 38 # - Ruby's grpc.tools gem also requires protoc plugins build on multiple platforms [all …]
|
| /third_party/grpc/src/csharp/ |
| D | build_nuget.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 16 set -ex 20 mkdir -p ../../artifacts 22 # Collect protoc artifacts built by the previous build step 23 mkdir -p protoc_plugins 24 cp -r "${EXTERNAL_GIT_ROOT}"/input_artifacts/protoc_* protoc_plugins || true 33 # we provide a way of building "partial" nugets that only include artifacts 35 # contain placeholders (empty files) for artifacts that normally need 43 # create placeholders for artifacts that can't be built 50 sed -ibak "s/<\/GrpcCsharpVersion>/-singleplatform<\/GrpcCsharpVersion>/" build/dependencies.props [all …]
|
| /third_party/grpc/tools/internal_ci/helper_scripts/ |
| D | store_artifacts_from_moved_src_tree.sh | 8 # http://www.apache.org/licenses/LICENSE-2.0 17 # and the script produces artifacts to be stored by kokoro, run this 20 set -ex 33 # we need to copy the artifacts produced by the build to a location that's stored by kokoro. 35 # artifacts in this directory will be stored by kokoro 36 mkdir -p "${GRPC_TEST_REPORT_BASE_DIR}/artifacts" 37 time cp -r artifacts/* "${GRPC_TEST_REPORT_BASE_DIR}/artifacts" || true
|
| /third_party/mbedtls/ |
| D | .gitignore | 6 # CMake build artifacts: 12 # CMake generates *.dir/ folders for in-tree builds (used by MSVC projects), ignore all of those: 19 # Test coverage build artifacts: 23 coverage-summary.txt 26 massif-* 33 # Unix-like build artifacts: 36 # MSVC build artifacts: 42 # Python build artifacts: 45 # CMake generates *.dir/ folders for in-tree builds (used by MSVC projects), ignore all of those: 55 /psa-arch-tests
|
| /third_party/ncurses/Ada95/src/ |
| D | library.gpr.in | 1 ------------------------------------------------------------------------------ 2 -- Copyright 2019,2020 Thomas E. Dickey -- 3 -- -- 4 -- Permission is hereby granted, free of charge, to any person obtaining a -- 5 -- copy of this software and associated documentation files (the -- 6 -- "Software"), to deal in the Software without restriction, including -- 7 -- without limitation the rights to use, copy, modify, merge, publish, -- 8 -- distribute, distribute with modifications, sublicense, and/or sell -- 9 -- copies of the Software, and to permit persons to whom the Software is -- 10 -- furnished to do so, subject to the following conditions: -- [all …]
|