Home
last modified time | relevance | path

Searched full:jobs (Results 1 – 25 of 596) sorted by relevance

12345678910>>...24

/third_party/mesa3d/src/panfrost/lib/
Dpan_scoreboard.h37 /* The number of jobs in the primary batch, essentially */
49 /* The dependency for tiler jobs (i.e. the index of the last emitted
59 * There are various types of Mali jobs:
71 * of Mali jobs together forming a linked list. Within the job chain, each Mali
72 * job can set (up to) two dependencies on other earlier jobs in the chain.
81 * job, and tiler jobs must depend on it.
83 * - Vertex jobs and tiler jobs are independent.
89 * - Tiler jobs must depend on the write value job (chained or otherwise).
91 * - Tiler jobs must be strictly ordered. So each tiler job must depend on the
94 * - Jobs linking via next_job has no bearing on order of execution, rather it
[all …]
Dpan_indirect_draw.c103 struct jobs_data jobs; member
126 * the DDK seems to put all min/max search jobs at the beginning of the job chain
128 * the draw patching jobs which have the suppress_prefetch flag set. Merging the
312 builder->jobs.first_vertex_sysval = get_input_field(b, first_vertex_sysval); in extract_inputs()
313 builder->jobs.base_vertex_sysval = get_input_field(b, base_vertex_sysval); in extract_inputs()
314 builder->jobs.base_instance_sysval = get_input_field(b, base_instance_sysval); in extract_inputs()
315 builder->jobs.vertex_job = get_input_field(b, vertex_job); in extract_inputs()
316 builder->jobs.tiler_job = get_input_field(b, tiler_job); in extract_inputs()
385 builder->jobs.offset_start); in update_dcd()
396 builder->jobs.vertex_job : builder->jobs.tiler_job; in update_job()
[all …]
/third_party/skia/site/docs/dev/testing/
Dautomated_testing.md24 may automatically retry tasks within its set limits. Jobs are not retried.
25 Multiple jobs may share the same task, for example, tests on two different
28 Each Skia repository has an `infra/bots/tasks.json` file which defines the jobs
29 and tasks for the repo. Most jobs will run at every commit, but it is possible
30 to specify nightly and weekly jobs as well. For convenience, most repos also
45 Try Jobs
49 repo. You need to have permission to trigger try jobs; if you need permission,
56 or using `bin/try`, a small wrapper for `git cl try` which helps to choose try jobs.
100 <a name="adding-new-jobs"></a>
101 Adding new jobs
[all …]
/third_party/mesa3d/src/util/
Du_queue.c295 job = queue->jobs[queue->read_idx]; in util_queue_thread_func()
296 memset(&queue->jobs[queue->read_idx], 0, sizeof(struct util_queue_job)); in util_queue_thread_func()
314 /* signal remaining jobs if all threads are being terminated */ in util_queue_thread_func()
319 if (queue->jobs[i].job) { in util_queue_thread_func()
320 if (queue->jobs[i].fence) in util_queue_thread_func()
321 util_queue_fence_signal(queue->jobs[i].fence); in util_queue_thread_func()
322 queue->jobs[i].job = NULL; in util_queue_thread_func()
448 queue->jobs = (struct util_queue_job*) in util_queue_init()
450 if (!queue->jobs) in util_queue_init()
477 if (queue->jobs) { in util_queue_init()
[all …]
/third_party/mesa3d/src/intel/vulkan/tests/
Dstate_pool_no_free.c38 } jobs[NUM_THREADS]; variable
71 jobs[i].pool = &state_pool; in run_test()
72 jobs[i].id = i; in run_test()
73 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); in run_test()
77 pthread_join(jobs[i].thread, NULL); in run_test()
92 if (thread_max < jobs[i].offsets[next[i]]) { in run_test()
93 thread_max = jobs[i].offsets[next[i]]; in run_test()
105 ASSERT(jobs[max_thread_idx].offsets[next[max_thread_idx]] > highest); in run_test()
107 highest = jobs[max_thread_idx].offsets[next[max_thread_idx]]; in run_test()
Dblock_pool_no_free.c39 } jobs[NUM_THREADS]; variable
45 uint32_t job_id = job - jobs; in alloc_blocks()
126 jobs[i].pool = &pool; in run_test()
127 jobs[i].id = i; in run_test()
128 pthread_create(&jobs[i].thread, NULL, alloc_blocks, &jobs[i]); in run_test()
132 pthread_join(jobs[i].thread, NULL); in run_test()
137 block_ptrs[i] = jobs[i].blocks; in run_test()
142 block_ptrs[i] = jobs[i].back_blocks; in run_test()
Dstate_pool_test_helper.h30 } jobs[NUM_THREADS]; variable
64 jobs[i].pool = state_pool; in run_state_pool_test()
65 jobs[i].id = i; in run_state_pool_test()
66 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); in run_state_pool_test()
70 pthread_join(jobs[i].thread, NULL); in run_state_pool_test()
/third_party/mbedtls/tests/scripts/
Ddepends.py14 Then, test domains (groups of jobs, tests) are built based on predefined data
42 Lastly, the collected jobs are executed and (optionally) tested, with
313 """Initialize the jobs container"""
314 self.jobs = []
322 The domain contains a set of jobs that enable one of the elements
325 If exclude is a regular expression, skip generated jobs whose description
340 self.jobs.append(job)
346 If exclude is a regular expression, skip generated jobs whose description
360 self.jobs.append(job)
365 each call adds respective jobs. The job array initialization is done once in
[all …]
/third_party/vixl/tools/
Dtest.py150 '--jobs', '-j', metavar='N', type=int, nargs='?',
153 help='''Runs the tests using N jobs. If the option is set but no value is
154 provided, the script will use as many jobs as it thinks useful.''')
260 def RunLinter(jobs): argument
263 jobs = args.jobs, progress_prefix = 'cpp lint: ')
266 def RunClangFormat(clang_path, jobs): argument
269 jobs = jobs,
272 def RunClangTidy(clang_path, jobs): argument
275 jobs = jobs,
282 def BuildAll(build_options, jobs, environment_options): argument
[all …]
Dlint.py62 parser.add_argument('--jobs', '-j', metavar='N', type=int, nargs='?',
65 help='''Runs the tests using N jobs. If the option is set
66 but no value is provided, the script will use as many jobs
132 jobs = 1, argument
159 pool = multiprocessing.Pool(jobs)
229 def RunLinter(files, jobs=1, progress_prefix='', cached=True): argument
233 jobs=jobs,
248 retcode = RunLinter(files, jobs=args.jobs, cached=cached)
/third_party/mesa3d/.gitlab-ci/bin/
Dci_run_n_monitor.py11 Helper script to restrict running only required CI jobs
46 # dependency graph to see which jobs the target jobs need
125 """Monitors pipeline and delegate canceling jobs"""
138 for job in pipeline.jobs.list(all=True, sort="desc"):
139 # target jobs
154 # all jobs
193 pjob = project.jobs.get(job.id, lazy=True)
204 pjob = project.jobs.get(job.id, lazy=True)
210 """Cancel unwanted GitLab jobs"""
223 job = project.jobs.get(job_id)
[all …]
/third_party/node/deps/npm/node_modules/node-gyp/lib/
Dbuild.js26 const jobs = gyp.opts.jobs || process.env.JOBS
151 if (jobs) {
152 j = parseInt(jobs, 10)
155 } else if (jobs.toUpperCase() === 'MAX') {
164 if (jobs) {
165 j = parseInt(jobs, 10)
167 argv.push('--jobs')
169 } else if (jobs.toUpperCase() === 'MAX') {
170 argv.push('--jobs')
/third_party/libdrm/tests/exynos/
Dexynos_fimg2d_event.c103 * We need to wait until all G2D jobs are finished, otherwise we
110 static void wait_all_jobs(struct g2d_job* jobs, unsigned num_jobs) in wait_all_jobs() argument
115 while (jobs[i].busy) in wait_all_jobs()
121 static struct g2d_job* free_job(struct g2d_job* jobs, unsigned num_jobs) in free_job() argument
126 if (jobs[i].busy == 0) in free_job()
127 return &jobs[i]; in free_job()
136 struct g2d_job *jobs = calloc(num_jobs, sizeof(struct g2d_job)); in g2d_work() local
142 jobs[i].id = i; in g2d_work()
150 j = free_job(jobs, num_jobs); in g2d_work()
189 wait_all_jobs(jobs, num_jobs); in g2d_work()
[all …]
/third_party/cups-filters/scripting/php/
DREADME74 Gets a list of jobs:
78 The "dest" string can be blank for jobs on all destinations.
79 Pass TRUE for "myjobs" to only get jobs for the current user.
80 The "completed" argument can be 0 for pending jobs, 1 for
81 completed jobs, and -1 for all jobs.
100 $jobs = cups_get_jobs("", FALSE, -1);
/third_party/mesa3d/docs/ci/
DLAVA.rst17 taking in jobs for the hardware that the LAVA lab contains. The
19 connects to the LAVA lab using a predefined token to submit jobs under
22 The LAVA instance manages scheduling those jobs to the boards present.
30 some boards booting using test jobs. Start with the stock QEMU
37 that device type and the Mesa jobs will be scheduled to any of them.
50 to restrict the jobs it takes or it will grab random jobs from tasks
81 Now it's time to define your test jobs in the driver-specific
/third_party/node/deps/npm/node_modules/tar/lib/
Dpack.js40 const JOBS = Symbol('jobs') constant
118 this[JOBS] = 0
119 this.jobs = +opt.jobs || 4
164 this[JOBS] += 1
179 this[JOBS] += 1
183 this[JOBS] -= 1
206 this[JOBS] += 1
209 this[JOBS] -= 1
230 w !== null && this[JOBS] < this.jobs;
242 if (this[ENDED] && !this[QUEUE].length && this[JOBS] === 0) {
[all …]
/third_party/skia/infra/bots/
DREADME.md7 Tasks and Jobs
15 Jobs are collections of related tasks which help define sub-sections of the DAG,
16 for example, to be used as try jobs. Each job is defined as an entry point into
19 The tasks.json file in this directory is the list of tasks and jobs for
23 reads the tasks.json file at each commit to determine which jobs to run. For
30 * jobs.json - The list of all jobs to run. Edit this to add or remove
/third_party/node/deps/v8/src/compiler-dispatcher/
Dlazy-compile-dispatcher.h70 // LazyCompileDispatcher::DoIdleWork tries to advance as many jobs out of jobs_
74 // LazyCompileDispatcher::DoBackgroundWork advances one of the pending jobs,
100 // Aborts all jobs, blocking until all jobs are aborted.
206 // The set of jobs that can be run on a background thread.
209 // The set of jobs that can be finalized on the main thread.
212 // The total number of jobs ready to execute on background, both those pending
217 // The set of all allocated jobs, used for verification of the various queues
222 // A queue of jobs to delete on the background thread(s). Jobs in this queue
/third_party/mesa3d/src/panfrost/perf/
DT72x.xml32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple…
33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) …
34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
DT82x.xml32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple…
33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) …
34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
DT83x.xml32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple…
33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) …
34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
DT86x.xml32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple…
33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) …
34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
/third_party/node/deps/openssl/openssl/crypto/async/
Dasync.c112 * no max size and no pre-created jobs in async_get_pool_job()
119 job = sk_ASYNC_JOB_pop(pool->jobs); in async_get_pool_job()
147 sk_ASYNC_JOB_push(pool->jobs, job); in async_release_job()
324 if (pool == NULL || pool->jobs == NULL) in async_empty_pool()
328 job = sk_ASYNC_JOB_pop(pool->jobs); in async_empty_pool()
374 pool->jobs = sk_ASYNC_JOB_new_reserve(NULL, init_size); in ASYNC_init_thread()
375 if (pool->jobs == NULL) { in ASYNC_init_thread()
383 /* Pre-create jobs as required */ in ASYNC_init_thread()
390 * skip creation of any more jobs in ASYNC_init_thread()
396 sk_ASYNC_JOB_push(pool->jobs, job); /* Cannot fail due to reserve */ in ASYNC_init_thread()
[all …]
/third_party/openssl/crypto/async/
Dasync.c112 * no max size and no pre-created jobs in async_get_pool_job()
119 job = sk_ASYNC_JOB_pop(pool->jobs); in async_get_pool_job()
147 sk_ASYNC_JOB_push(pool->jobs, job); in async_release_job()
324 if (pool == NULL || pool->jobs == NULL) in async_empty_pool()
328 job = sk_ASYNC_JOB_pop(pool->jobs); in async_empty_pool()
374 pool->jobs = sk_ASYNC_JOB_new_reserve(NULL, init_size); in ASYNC_init_thread()
375 if (pool->jobs == NULL) { in ASYNC_init_thread()
383 /* Pre-create jobs as required */ in ASYNC_init_thread()
390 * skip creation of any more jobs in ASYNC_init_thread()
396 sk_ASYNC_JOB_push(pool->jobs, job); /* Cannot fail due to reserve */ in ASYNC_init_thread()
[all …]
/third_party/skia/tools/skqp/
Dfind_commit_with_best_gold_results.py23 path = skia_directory + '/infra/bots/jobs.json'
36 jobs = json.load(f)
37 for job in jobs:
62 def get_results_for_commit(commit, jobs): argument
81 for job in jobs for config in CONFIGS]
94 jobs = [j for j in get_jobs()]
98 results.append((commit_hash, get_results_for_commit(commit_hash, jobs)))

12345678910>>...24