| /external/mesa3d/.gitlab-ci/lava/ |
| D | lava-submit.sh | 1 #!/usr/bin/env bash 4 set -ex 6 # If we run in the fork (not from mesa or Marge-bot), reuse mainline kernel and rootfs, if exist. 9 if ! curl -s -X HEAD -L --retry 4 -f --retry-delay 60 \ 18 rm -rf results 19 mkdir -p results/job-rootfs-overlay/ 21 cp artifacts/ci-common/capture-devcoredump.sh results/job-rootfs-overlay/ 22 cp artifacts/ci-common/init-*.sh results/job-rootfs-overlay/ 23 cp artifacts/ci-common/intel-gpu-freq.sh results/job-rootfs-overlay/ 24 cp artifacts/ci-common/kdl.sh results/job-rootfs-overlay/ [all …]
|
| /external/mesa3d/.gitlab-ci/bare-metal/ |
| D | rootfs-setup.sh | 1 #!/usr/bin/env bash 6 mkdir -p $rootfs_dst/results 8 # Set up the init script that brings up the system. 9 cp $BM/bm-init.sh $rootfs_dst/init 14 # Make JWT token available as file in the bare-metal storage to enable access 20 cp $CI_COMMON/capture-devcoredump.sh $rootfs_dst/ 21 cp $CI_COMMON/intel-gpu-freq.sh $rootfs_dst/ 23 cp "$SCRIPTS_DIR/setup-test-env.sh" "$rootfs_dst/" 25 set +x 27 # Pass through relevant env vars from the gitlab job to the baremetal init script [all …]
|
| /external/mesa3d/.gitlab-ci/lava/utils/ |
| D | ssh_job_definition.py | 4 job is retried, causing delays in the overall pipeline executing, ultimately 11 via SSH pseudo terminal, instead of relying in the error-prone UART. 13 In more detail, we aim to use "export -p" to share the initial boot environment 14 with SSH LAVA test-cases. 15 The "init-stage1.sh" script handles tasks such as system mounting and network 16 setup, which are necessary for allocating a pseudo-terminal under "/dev/pts". 21 Therefore, we have divided the job definition into four parts: 24 2. [DUT] Running the "init-stage1.sh" script for the first SSH test case. 25 3. [DUT] Export the first boot environment to `/dut-env-vars.sh` file. 26 4. [SSH] Enabling the pseudo-terminal for colors and running the "init-stage2.sh" [all …]
|
| D | lava_job_definition.py | 30 job. 33 def __init__(self, job_submitter: "LAVAJobSubmitter") -> None: 36 def has_ssh_support(self) -> bool: 41 # which is required to follow the job in a SSH section 46 def generate_lava_yaml_payload(self) -> dict[str, Any]: 48 Generates a YAML payload for submitting a LAVA job, based on the provided arguments. 55 actions for the LAVA job submission. 60 "url": f"{args.rootfs_url_prefix}/lava-rootfs.tar.zst", 79 wrap_final_deploy_action(deploy_actions[-1]) 99 def generate_lava_job_definition(self) -> str: [all …]
|
| /external/mesa3d/.gitlab-ci/tests/data/ |
| D | UBOOT_force_uart=False_job_definition.yaml | 1 job_name: 'test-project: my_pipeline_info' 5 - my_visibility_group 10 job: 13 depthcharge-retry: 15 depthcharge-start: 17 depthcharge-action: 20 - deploy: 28 url: None/lava-rootfs.tar.zst 34 - boot: 36 method: u-boot [all …]
|
| D | FASTBOOT_force_uart=False_job_definition.yaml | 1 job_name: 'test-project: my_pipeline_info' 5 - my_visibility_group 10 job: 13 depthcharge-retry: 15 depthcharge-start: 17 depthcharge-action: 20 - deploy: 25 url: None/lava-rootfs.tar.zst 28 - deploy: 40 image: registry.gitlab.collabora.com/lava/health-check-docker [all …]
|
| D | UBOOT_force_uart=True_job_definition.yaml | 1 job_name: 'test-project: my_pipeline_info' 5 - my_visibility_group 10 job: 13 depthcharge-retry: 15 depthcharge-start: 17 depthcharge-action: 20 - deploy: 28 url: None/lava-rootfs.tar.zst 32 - boot: 34 method: u-boot [all …]
|
| D | FASTBOOT_force_uart=True_job_definition.yaml | 1 job_name: 'test-project: my_pipeline_info' 5 - my_visibility_group 10 job: 13 depthcharge-retry: 15 depthcharge-start: 17 depthcharge-action: 20 - deploy: 25 url: None/lava-rootfs.tar.zst 27 - deploy: 39 image: registry.gitlab.collabora.com/lava/health-check-docker [all …]
|
| /external/mesa3d/.gitlab-ci/test/ |
| D | gitlab-ci.yml | 2 # Cancel job if a newer commit is pushed to the same branch 7 - !reference [default, before_script] 9 - rm -rf install 10 - tar -xf artifacts/install.tar 11 - section_start ldd_section "Checking ldd on driver build" 12 - LD_LIBRARY_PATH=install/lib find install/lib -name "*.so" -print -exec ldd {} \; 13 - section_end ldd_section 18 - results/ 20 - !reference [.never-post-merge-rules, rules] 22 .formatting-check: [all …]
|
| /external/mesa3d/.gitlab-ci/common/ |
| D | init-stage2.sh | 7 # Second-stage init, used to set up devices and our job environment before 10 shopt -s extglob 20 set +x 32 kill -9 "$pid" 2>/dev/null || true 36 set -x 45 for path in '/dut-env-vars.sh' '/set-job-env-vars.sh' './set-job-env-vars.sh'; do 46 [ -f "$path" ] && source "$path" 48 . "$SCRIPTS_DIR"/setup-test-env.sh 50 set -ex 52 # Set up any devices required by the jobs [all …]
|
| /external/perfetto/infra/ci/worker/ |
| D | worker.py | 1 #!/usr/bin/env python3 8 # http://www.apache.org/licenses/LICENSE-2.0 37 WORKER_NAME = '%s-%s' % (os.getenv( 38 'WORKER_HOST', 'local').split('-')[-1], socket.gethostname()) 43 ''' Transactionally acquire the given job. 45 Returns the job JSON object if it managed to acquire and put it into the 48 logging.debug('Trying to acquire job %s', job_id) 51 job, etag = req('GET', uri, req_etag=True) 52 if job['status'] != 'QUEUED': 55 job['status'] = 'STARTED' [all …]
|
| /external/python/cpython3/.github/workflows/ |
| D | build.yml | 3 # gh-84728: "paths-ignore" is not used to skip documentation-only PRs, because 4 # it prevents to mark a job as mandatory. A PR cannot be merged if a job is 5 # mandatory but not scheduled because of "paths-ignore". 10 - 'main' 11 - '3.11' 12 - '3.10' 13 - '3.9' 14 - '3.8' 15 - '3.7' 18 - 'main' [all …]
|
| /external/lz4/.circleci/ |
| D | config.yml | 15 # In the meantime, set it to 1. 17 shell: /bin/bash --login 19 …n't or see https://circleci.com/docs/2.0/env-vars/#interpolating-environment-variables-to-set-othe… 21 CIRCLE_ARTIFACTS: /tmp/circleci-artifacts 22 CIRCLE_TEST_REPORTS: /tmp/circleci-test-results 23 …# In CircleCI 1.0 we used a pre-configured image with a large number of languages and other packag… 24 # In CircleCI 2.0 you can now specify your own image, or use one of our pre-configured images. 25 …ation line tells CircleCI to use the specified docker image as the runtime environment for you job. 26 # We have selected a pre-built image that mirrors the build environment we use on 28 # of each job. For more information on choosing an image (or alternatively using a [all …]
|
| /external/angle/doc/ |
| D | CaptureAndReplay.md | 9 * Mid-execution capture is supported with the Vulkan back-end. 10 * Mid-execution capture has many unimplemented features. 23 files. To enable capture, set the `ANGLE_CAPTURE_FRAME_START` and `ANGLE_CAPTURE_FRAME_END` 37 * Set to `0` to disable capture entirely. Default is `1`. 39 * Set to `0` to disable capture compression. Default is `1`. 45 * Uses mid-execution capture to write "Setup" functions that starts a Context at frame `n`. 65 * Set to `1` to enable GL state serialization. Default is `0`. 71 …E_END=4 ANGLE_CAPTURE_OUT_DIR=samples/capture_replay out/Debug/simple_texture_2d --use-angle=vulkan 89 $ autoninja -C out/Debug capture_replay_sample 103 src/tests/restricted_traces$ ls -1 desktop_test/ [all …]
|
| /external/mesa3d/docs/relnotes/ |
| D | 20.1.3.rst | 1 Mesa 20.1.3 Release Notes / 2020-07-08 20 --------------- 24 9872b8d46bee822177ffbe4292addff7bdb84cefc1fe776b8e6b2881a8362bf1 mesa-20.1.3.tar.xz 27 ------------ 29 - None 32 --------- 34 - vkGetSemaphoreCounterValue doesn't update without vkWaitSemaphores 36 - [v3d] corruption when GS omits some vertices 37 - [RADV] Non-precise occlusion queries return non-zero when all 39 - [DXVK] Project Cars rendering problems [all …]
|
| D | 23.1.0.rst | 1 Mesa 23.1.0 Release Notes / 2023-05-10 20 --------------- 24 a9dde3c76571c4806245a05bda1cceee347c3267127e9e549e4f4e225d92e992 mesa-23.1.0.tar.xz 28 ------------ 30 - VK_EXT_pipeline_library_group_handles on RADV 31 - VK_EXT_image_sliced_view_of_3d on RADV/GFX10+ 32 - VK_KHR_map_memory2 on ANV and RADV 33 - fullyCoveredFragmentShaderInputVariable on RADV/GFX9+ 34 - VK_EXT_discard_rectangles version 2 on RADV 35 - VK_EXT_graphics_pipeline_library on RADV [all …]
|
| /external/autotest/client/ |
| D | setup_modules.py | 17 py_version_re = re.compile(r'--py_version=(\w+)\b') 21 if not arg.startswith('--py_version'): 26 raise ValueError('--py_version may only be specified once.') 42 If the PY_VERSION env var is set, just return that. This is the case 43 when autoserv kicks of autotest on the server side via a job.run(), or 46 Otherwise, parse & pop the sys.argv for the '--py_version' flag. If no 47 flag is set, default to python 3. 50 # Even if the arg is in the env vars, we will attempt to get it from the 65 # There are cases were this can be set (ie by test_that), but a subprocess 84 # out the original record causing the error so there is -some- idea [all …]
|
| /external/tensorflow/tensorflow/core/grappler/ |
| D | grappler_item_builder.cc | 7 http://www.apache.org/licenses/LICENSE-2.0 58 auto flat = tensor->flat<float>(); in InitializeTensor() 64 auto flat = tensor->flat<int64_t>(); in InitializeTensor() 72 // Allocator will run non-trivial constructor/destructor for a Tensor with in InitializeTensor() 74 memset(const_cast<char*>(tensor->tensor_data().data()), 0, in InitializeTensor() 75 tensor->tensor_data().size()); in InitializeTensor() 86 item->graph = std::move(pruned_graph); in PruneGraph() 99 dim_proto.size() == -1) { in ReplaceUnknownShapeDim() 101 shape_pb_out->add_dim()->set_size( in ReplaceUnknownShapeDim() 105 shape_pb_out->add_dim()->set_size(dim_proto.size()); in ReplaceUnknownShapeDim() [all …]
|
| /external/curl/tests/ |
| D | testcurl.pl | 1 #!/usr/bin/env perl 22 # SPDX-License-Identifier: curl 32 # This is written for the purpose of being run from a crontab job or similar 34 # curl-autocompile@haxx.se to be dealt with automatically (make sure the 40 # testcurl.pl [options] [curl-daily-name] > output 44 # --configure=[options] Configure options 45 # --crosscompile This is a crosscompile 46 # --desc=[desc] Description of your test system 47 # --email=[email] Set email address to report as 48 # --extvercmd=[command] Command to use for displaying version with cross compiles. [all …]
|
| /external/tensorflow/tensorflow/dtensor/python/ |
| D | mesh_util.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 34 device_type: str, mesh: layout.Mesh) -> None: 38 # pylint: disable=protected-access 43 # pylint: enable=protected-access 49 ) -> Tuple[List[tf_device.DeviceSpec], str]: 74 device_type: Optional[str] = None) -> layout.Mesh: 75 """Creates a single-client mesh. 83 batch-parallel dimension called 'x' using all devices. As a special case, 84 a single-element mesh_dims whose dim_size is -1 also uses all devices. 92 A single-client mesh created from specified or default arguments. [all …]
|
| /external/tensorflow/tensorflow/python/distribute/coordinator/ |
| D | cluster_coordinator.py | 7 # http://www.apache.org/licenses/LICENSE-2.0 15 """Module for `ClusterCoordinator` and relevant cluster-worker related library. 50 # Maximum size for queued closures, "infinite" if set to 0. 53 # Note that using an "infinite" queue size can take a non-trivial portion of 61 _RPC_ERROR_FROM_PS = "GRPC error information from remote target /job:ps" 64 _JOB_WORKER_STRING_IDENTIFIER = "/job:worker" 86 # Avoid doubly-nested errors 102 This wrapper is used when a dependent training closure fails to set errors on 110 # Avoid doubly-nested errors 130 error = val._get_error() # pylint: disable=protected-access [all …]
|
| /external/libcups/scheduler/ |
| D | cupsfilter.c | 4 * Copyright 2007-2016 by Apple Inc. 5 * Copyright 1997-2006 by Easy Software Products, all rights reserved. 14 #include <cups/cups-private.h> 15 #include <cups/file-private.h> 16 #include <cups/ppd-private.h> 76 static void get_job_file(const char *job); 85 * 'main()' - Main entry for the test program. 88 int /* O - Exit status */ 89 main(int argc, /* I - Number of command-line args */ in main() 90 char *argv[]) /* I - Command-line arguments */ in main() [all …]
|
| /external/mksh/src/ |
| D | exec.c | 3 /*- 11 * is granted to deal in this work without restriction, including un- 71 if ((flags&XFORK) && !(flags&XEXEC) && t->type != TPIPE) in execute() 72 /* run in sub-process */ in execute() 73 return (exchild(t, flags & ~XTIME, xerrok, -1)); in execute() 80 if (t->type == TCOM) { in execute() 88 current_lineno = t->lineno; in execute() 93 t->args[0] == NULL && in execute() 95 t->vars[0] != NULL && t->vars[1] == NULL && in execute() 97 t->ioact != NULL && t->ioact[0] != NULL && in execute() [all …]
|
| /external/toybox/toys/pending/ |
| D | sh.c | 1 /* sh.c - toybox shell 6 * http://man7.org/linux/man-pages/man1/bash.1.html 8 * The POSIX-2008/SUSv4 shell spec is at: 19 * unset local export readonly set : . let history declare ulimit type 25 * TODO: case, wildcard +(*|?), job control (find_plus_minus), ${x//}, $(()) 33 * TODO: set -e -o pipefail, shopt -s nullglob 47 USE_SH(NEWTOY(cd, ">1LP[-LP]", TOYFLAG_NOFORK)) 57 USE_SH(NEWTOY(set, 0, TOYFLAG_NOFORK)) 68 USE_SH(OLDTOY(-sh, sh, 0)) 69 USE_SH(OLDTOY(-toysh, sh, 0)) [all …]
|
| /external/libcups/cgi-bin/ |
| D | ipp-var.c | 2 * CGI <-> IPP variable routines for CUPS. 4 * Copyright 2007-2016 by Apple Inc. 5 * Copyright 1997-2007 by Easy Software Products. 14 #include "cgi-private.h" 18 * 'cgiGetAttributes()' - Get the list of attributes that are needed 23 cgiGetAttributes(ipp_t *request, /* I - IPP request */ in cgiGetAttributes() 24 const char *tmpl) /* I - Base filename */ in cgiGetAttributes() 100 else if (nameptr < (name + sizeof(name) - 1)) in cgiGetAttributes() 103 *nameptr++ = '-'; in cgiGetAttributes() 129 * If we have attributes, add a requested-attributes attribute to the in cgiGetAttributes() [all …]
|