Home
last modified time | relevance | path

Searched +full:set +full:- +full:job +full:- +full:env +full:- +full:vars (Results 1 – 25 of 71) sorted by relevance

123

/external/mesa3d/.gitlab-ci/lava/
Dlava-submit.sh1 #!/usr/bin/env bash
4 set -ex
6 # If we run in the fork (not from mesa or Marge-bot), reuse mainline kernel and rootfs, if exist.
9 if ! curl -s -X HEAD -L --retry 4 -f --retry-delay 60 \
18 rm -rf results
19 mkdir -p results/job-rootfs-overlay/
21 cp artifacts/ci-common/capture-devcoredump.sh results/job-rootfs-overlay/
22 cp artifacts/ci-common/init-*.sh results/job-rootfs-overlay/
23 cp artifacts/ci-common/intel-gpu-freq.sh results/job-rootfs-overlay/
24 cp artifacts/ci-common/kdl.sh results/job-rootfs-overlay/
[all …]
/external/mesa3d/.gitlab-ci/bare-metal/
Drootfs-setup.sh1 #!/usr/bin/env bash
6 mkdir -p $rootfs_dst/results
8 # Set up the init script that brings up the system.
9 cp $BM/bm-init.sh $rootfs_dst/init
14 # Make JWT token available as file in the bare-metal storage to enable access
20 cp $CI_COMMON/capture-devcoredump.sh $rootfs_dst/
21 cp $CI_COMMON/intel-gpu-freq.sh $rootfs_dst/
23 cp "$SCRIPTS_DIR/setup-test-env.sh" "$rootfs_dst/"
25 set +x
27 # Pass through relevant env vars from the gitlab job to the baremetal init script
[all …]
/external/mesa3d/.gitlab-ci/lava/utils/
Dssh_job_definition.py4 job is retried, causing delays in the overall pipeline executing, ultimately
11 via SSH pseudo terminal, instead of relying in the error-prone UART.
13 In more detail, we aim to use "export -p" to share the initial boot environment
14 with SSH LAVA test-cases.
15 The "init-stage1.sh" script handles tasks such as system mounting and network
16 setup, which are necessary for allocating a pseudo-terminal under "/dev/pts".
21 Therefore, we have divided the job definition into four parts:
24 2. [DUT] Running the "init-stage1.sh" script for the first SSH test case.
25 3. [DUT] Export the first boot environment to `/dut-env-vars.sh` file.
26 4. [SSH] Enabling the pseudo-terminal for colors and running the "init-stage2.sh"
[all …]
Dlava_job_definition.py30 job.
33 def __init__(self, job_submitter: "LAVAJobSubmitter") -> None:
36 def has_ssh_support(self) -> bool:
41 # which is required to follow the job in a SSH section
46 def generate_lava_yaml_payload(self) -> dict[str, Any]:
48 Generates a YAML payload for submitting a LAVA job, based on the provided arguments.
55 actions for the LAVA job submission.
60 "url": f"{args.rootfs_url_prefix}/lava-rootfs.tar.zst",
79 wrap_final_deploy_action(deploy_actions[-1])
99 def generate_lava_job_definition(self) -> str:
[all …]
/external/mesa3d/.gitlab-ci/tests/data/
DUBOOT_force_uart=False_job_definition.yaml1 job_name: 'test-project: my_pipeline_info'
5 - my_visibility_group
10 job:
13 depthcharge-retry:
15 depthcharge-start:
17 depthcharge-action:
20 - deploy:
28 url: None/lava-rootfs.tar.zst
34 - boot:
36 method: u-boot
[all …]
DFASTBOOT_force_uart=False_job_definition.yaml1 job_name: 'test-project: my_pipeline_info'
5 - my_visibility_group
10 job:
13 depthcharge-retry:
15 depthcharge-start:
17 depthcharge-action:
20 - deploy:
25 url: None/lava-rootfs.tar.zst
28 - deploy:
40 image: registry.gitlab.collabora.com/lava/health-check-docker
[all …]
DUBOOT_force_uart=True_job_definition.yaml1 job_name: 'test-project: my_pipeline_info'
5 - my_visibility_group
10 job:
13 depthcharge-retry:
15 depthcharge-start:
17 depthcharge-action:
20 - deploy:
28 url: None/lava-rootfs.tar.zst
32 - boot:
34 method: u-boot
[all …]
DFASTBOOT_force_uart=True_job_definition.yaml1 job_name: 'test-project: my_pipeline_info'
5 - my_visibility_group
10 job:
13 depthcharge-retry:
15 depthcharge-start:
17 depthcharge-action:
20 - deploy:
25 url: None/lava-rootfs.tar.zst
27 - deploy:
39 image: registry.gitlab.collabora.com/lava/health-check-docker
[all …]
/external/mesa3d/.gitlab-ci/test/
Dgitlab-ci.yml2 # Cancel job if a newer commit is pushed to the same branch
7 - !reference [default, before_script]
9 - rm -rf install
10 - tar -xf artifacts/install.tar
11 - section_start ldd_section "Checking ldd on driver build"
12 - LD_LIBRARY_PATH=install/lib find install/lib -name "*.so" -print -exec ldd {} \;
13 - section_end ldd_section
18 - results/
20 - !reference [.never-post-merge-rules, rules]
22 .formatting-check:
[all …]
/external/mesa3d/.gitlab-ci/common/
Dinit-stage2.sh7 # Second-stage init, used to set up devices and our job environment before
10 shopt -s extglob
20 set +x
32 kill -9 "$pid" 2>/dev/null || true
36 set -x
45 for path in '/dut-env-vars.sh' '/set-job-env-vars.sh' './set-job-env-vars.sh'; do
46 [ -f "$path" ] && source "$path"
48 . "$SCRIPTS_DIR"/setup-test-env.sh
50 set -ex
52 # Set up any devices required by the jobs
[all …]
/external/perfetto/infra/ci/worker/
Dworker.py1 #!/usr/bin/env python3
8 # http://www.apache.org/licenses/LICENSE-2.0
37 WORKER_NAME = '%s-%s' % (os.getenv(
38 'WORKER_HOST', 'local').split('-')[-1], socket.gethostname())
43 ''' Transactionally acquire the given job.
45 Returns the job JSON object if it managed to acquire and put it into the
48 logging.debug('Trying to acquire job %s', job_id)
51 job, etag = req('GET', uri, req_etag=True)
52 if job['status'] != 'QUEUED':
55 job['status'] = 'STARTED'
[all …]
/external/python/cpython3/.github/workflows/
Dbuild.yml3 # gh-84728: "paths-ignore" is not used to skip documentation-only PRs, because
4 # it prevents to mark a job as mandatory. A PR cannot be merged if a job is
5 # mandatory but not scheduled because of "paths-ignore".
10 - 'main'
11 - '3.11'
12 - '3.10'
13 - '3.9'
14 - '3.8'
15 - '3.7'
18 - 'main'
[all …]
/external/lz4/.circleci/
Dconfig.yml15 # In the meantime, set it to 1.
17 shell: /bin/bash --login
19 …n't or see https://circleci.com/docs/2.0/env-vars/#interpolating-environment-variables-to-set-othe…
21 CIRCLE_ARTIFACTS: /tmp/circleci-artifacts
22 CIRCLE_TEST_REPORTS: /tmp/circleci-test-results
23 …# In CircleCI 1.0 we used a pre-configured image with a large number of languages and other packag…
24 # In CircleCI 2.0 you can now specify your own image, or use one of our pre-configured images.
25 …ation line tells CircleCI to use the specified docker image as the runtime environment for you job.
26 # We have selected a pre-built image that mirrors the build environment we use on
28 # of each job. For more information on choosing an image (or alternatively using a
[all …]
/external/angle/doc/
DCaptureAndReplay.md9 * Mid-execution capture is supported with the Vulkan back-end.
10 * Mid-execution capture has many unimplemented features.
23 files. To enable capture, set the `ANGLE_CAPTURE_FRAME_START` and `ANGLE_CAPTURE_FRAME_END`
37 * Set to `0` to disable capture entirely. Default is `1`.
39 * Set to `0` to disable capture compression. Default is `1`.
45 * Uses mid-execution capture to write "Setup" functions that starts a Context at frame `n`.
65 * Set to `1` to enable GL state serialization. Default is `0`.
71 …E_END=4 ANGLE_CAPTURE_OUT_DIR=samples/capture_replay out/Debug/simple_texture_2d --use-angle=vulkan
89 $ autoninja -C out/Debug capture_replay_sample
103 src/tests/restricted_traces$ ls -1 desktop_test/
[all …]
/external/mesa3d/docs/relnotes/
D20.1.3.rst1 Mesa 20.1.3 Release Notes / 2020-07-08
20 ---------------
24 9872b8d46bee822177ffbe4292addff7bdb84cefc1fe776b8e6b2881a8362bf1 mesa-20.1.3.tar.xz
27 ------------
29 - None
32 ---------
34 - vkGetSemaphoreCounterValue doesn't update without vkWaitSemaphores
36 - [v3d] corruption when GS omits some vertices
37 - [RADV] Non-precise occlusion queries return non-zero when all
39 - [DXVK] Project Cars rendering problems
[all …]
D23.1.0.rst1 Mesa 23.1.0 Release Notes / 2023-05-10
20 ---------------
24 a9dde3c76571c4806245a05bda1cceee347c3267127e9e549e4f4e225d92e992 mesa-23.1.0.tar.xz
28 ------------
30 - VK_EXT_pipeline_library_group_handles on RADV
31 - VK_EXT_image_sliced_view_of_3d on RADV/GFX10+
32 - VK_KHR_map_memory2 on ANV and RADV
33 - fullyCoveredFragmentShaderInputVariable on RADV/GFX9+
34 - VK_EXT_discard_rectangles version 2 on RADV
35 - VK_EXT_graphics_pipeline_library on RADV
[all …]
/external/autotest/client/
Dsetup_modules.py17 py_version_re = re.compile(r'--py_version=(\w+)\b')
21 if not arg.startswith('--py_version'):
26 raise ValueError('--py_version may only be specified once.')
42 If the PY_VERSION env var is set, just return that. This is the case
43 when autoserv kicks of autotest on the server side via a job.run(), or
46 Otherwise, parse & pop the sys.argv for the '--py_version' flag. If no
47 flag is set, default to python 3.
50 # Even if the arg is in the env vars, we will attempt to get it from the
65 # There are cases were this can be set (ie by test_that), but a subprocess
84 # out the original record causing the error so there is -some- idea
[all …]
/external/tensorflow/tensorflow/core/grappler/
Dgrappler_item_builder.cc7 http://www.apache.org/licenses/LICENSE-2.0
58 auto flat = tensor->flat<float>(); in InitializeTensor()
64 auto flat = tensor->flat<int64_t>(); in InitializeTensor()
72 // Allocator will run non-trivial constructor/destructor for a Tensor with in InitializeTensor()
74 memset(const_cast<char*>(tensor->tensor_data().data()), 0, in InitializeTensor()
75 tensor->tensor_data().size()); in InitializeTensor()
86 item->graph = std::move(pruned_graph); in PruneGraph()
99 dim_proto.size() == -1) { in ReplaceUnknownShapeDim()
101 shape_pb_out->add_dim()->set_size( in ReplaceUnknownShapeDim()
105 shape_pb_out->add_dim()->set_size(dim_proto.size()); in ReplaceUnknownShapeDim()
[all …]
/external/curl/tests/
Dtestcurl.pl1 #!/usr/bin/env perl
22 # SPDX-License-Identifier: curl
32 # This is written for the purpose of being run from a crontab job or similar
34 # curl-autocompile@haxx.se to be dealt with automatically (make sure the
40 # testcurl.pl [options] [curl-daily-name] > output
44 # --configure=[options] Configure options
45 # --crosscompile This is a crosscompile
46 # --desc=[desc] Description of your test system
47 # --email=[email] Set email address to report as
48 # --extvercmd=[command] Command to use for displaying version with cross compiles.
[all …]
/external/tensorflow/tensorflow/dtensor/python/
Dmesh_util.py7 # http://www.apache.org/licenses/LICENSE-2.0
34 device_type: str, mesh: layout.Mesh) -> None:
38 # pylint: disable=protected-access
43 # pylint: enable=protected-access
49 ) -> Tuple[List[tf_device.DeviceSpec], str]:
74 device_type: Optional[str] = None) -> layout.Mesh:
75 """Creates a single-client mesh.
83 batch-parallel dimension called 'x' using all devices. As a special case,
84 a single-element mesh_dims whose dim_size is -1 also uses all devices.
92 A single-client mesh created from specified or default arguments.
[all …]
/external/tensorflow/tensorflow/python/distribute/coordinator/
Dcluster_coordinator.py7 # http://www.apache.org/licenses/LICENSE-2.0
15 """Module for `ClusterCoordinator` and relevant cluster-worker related library.
50 # Maximum size for queued closures, "infinite" if set to 0.
53 # Note that using an "infinite" queue size can take a non-trivial portion of
61 _RPC_ERROR_FROM_PS = "GRPC error information from remote target /job:ps"
64 _JOB_WORKER_STRING_IDENTIFIER = "/job:worker"
86 # Avoid doubly-nested errors
102 This wrapper is used when a dependent training closure fails to set errors on
110 # Avoid doubly-nested errors
130 error = val._get_error() # pylint: disable=protected-access
[all …]
/external/libcups/scheduler/
Dcupsfilter.c4 * Copyright 2007-2016 by Apple Inc.
5 * Copyright 1997-2006 by Easy Software Products, all rights reserved.
14 #include <cups/cups-private.h>
15 #include <cups/file-private.h>
16 #include <cups/ppd-private.h>
76 static void get_job_file(const char *job);
85 * 'main()' - Main entry for the test program.
88 int /* O - Exit status */
89 main(int argc, /* I - Number of command-line args */ in main()
90 char *argv[]) /* I - Command-line arguments */ in main()
[all …]
/external/mksh/src/
Dexec.c3 /*-
11 * is granted to deal in this work without restriction, including un-
71 if ((flags&XFORK) && !(flags&XEXEC) && t->type != TPIPE) in execute()
72 /* run in sub-process */ in execute()
73 return (exchild(t, flags & ~XTIME, xerrok, -1)); in execute()
80 if (t->type == TCOM) { in execute()
88 current_lineno = t->lineno; in execute()
93 t->args[0] == NULL && in execute()
95 t->vars[0] != NULL && t->vars[1] == NULL && in execute()
97 t->ioact != NULL && t->ioact[0] != NULL && in execute()
[all …]
/external/toybox/toys/pending/
Dsh.c1 /* sh.c - toybox shell
6 * http://man7.org/linux/man-pages/man1/bash.1.html
8 * The POSIX-2008/SUSv4 shell spec is at:
19 * unset local export readonly set : . let history declare ulimit type
25 * TODO: case, wildcard +(*|?), job control (find_plus_minus), ${x//}, $(())
33 * TODO: set -e -o pipefail, shopt -s nullglob
47 USE_SH(NEWTOY(cd, ">1LP[-LP]", TOYFLAG_NOFORK))
57 USE_SH(NEWTOY(set, 0, TOYFLAG_NOFORK))
68 USE_SH(OLDTOY(-sh, sh, 0))
69 USE_SH(OLDTOY(-toysh, sh, 0))
[all …]
/external/libcups/cgi-bin/
Dipp-var.c2 * CGI <-> IPP variable routines for CUPS.
4 * Copyright 2007-2016 by Apple Inc.
5 * Copyright 1997-2007 by Easy Software Products.
14 #include "cgi-private.h"
18 * 'cgiGetAttributes()' - Get the list of attributes that are needed
23 cgiGetAttributes(ipp_t *request, /* I - IPP request */ in cgiGetAttributes()
24 const char *tmpl) /* I - Base filename */ in cgiGetAttributes()
100 else if (nameptr < (name + sizeof(name) - 1)) in cgiGetAttributes()
103 *nameptr++ = '-'; in cgiGetAttributes()
129 * If we have attributes, add a requested-attributes attribute to the in cgiGetAttributes()
[all …]

123