| /third_party/mesa3d/src/panfrost/lib/ |
| D | pan_scoreboard.h | 37 /* The number of jobs in the primary batch, essentially */ 49 /* The dependency for tiler jobs (i.e. the index of the last emitted 59 * There are various types of Mali jobs: 71 * of Mali jobs together forming a linked list. Within the job chain, each Mali 72 * job can set (up to) two dependencies on other earlier jobs in the chain. 81 * job, and tiler jobs must depend on it. 83 * - Vertex jobs and tiler jobs are independent. 89 * - Tiler jobs must depend on the write value job (chained or otherwise). 91 * - Tiler jobs must be strictly ordered. So each tiler job must depend on the 94 * - Jobs linking via next_job has no bearing on order of execution, rather it [all …]
|
| D | pan_indirect_draw.c | 103 struct jobs_data jobs; member 126 * the DDK seems to put all min/max search jobs at the beginning of the job chain 128 * the draw patching jobs which have the suppress_prefetch flag set. Merging the 312 builder->jobs.first_vertex_sysval = get_input_field(b, first_vertex_sysval); in extract_inputs() 313 builder->jobs.base_vertex_sysval = get_input_field(b, base_vertex_sysval); in extract_inputs() 314 builder->jobs.base_instance_sysval = get_input_field(b, base_instance_sysval); in extract_inputs() 315 builder->jobs.vertex_job = get_input_field(b, vertex_job); in extract_inputs() 316 builder->jobs.tiler_job = get_input_field(b, tiler_job); in extract_inputs() 385 builder->jobs.offset_start); in update_dcd() 396 builder->jobs.vertex_job : builder->jobs.tiler_job; in update_job() [all …]
|
| /third_party/skia/site/docs/dev/testing/ |
| D | automated_testing.md | 24 may automatically retry tasks within its set limits. Jobs are not retried. 25 Multiple jobs may share the same task, for example, tests on two different 28 Each Skia repository has an `infra/bots/tasks.json` file which defines the jobs 29 and tasks for the repo. Most jobs will run at every commit, but it is possible 30 to specify nightly and weekly jobs as well. For convenience, most repos also 45 Try Jobs 49 repo. You need to have permission to trigger try jobs; if you need permission, 56 or using `bin/try`, a small wrapper for `git cl try` which helps to choose try jobs. 100 <a name="adding-new-jobs"></a> 101 Adding new jobs [all …]
|
| /third_party/mesa3d/src/util/ |
| D | u_queue.c | 295 job = queue->jobs[queue->read_idx]; in util_queue_thread_func() 296 memset(&queue->jobs[queue->read_idx], 0, sizeof(struct util_queue_job)); in util_queue_thread_func() 314 /* signal remaining jobs if all threads are being terminated */ in util_queue_thread_func() 319 if (queue->jobs[i].job) { in util_queue_thread_func() 320 if (queue->jobs[i].fence) in util_queue_thread_func() 321 util_queue_fence_signal(queue->jobs[i].fence); in util_queue_thread_func() 322 queue->jobs[i].job = NULL; in util_queue_thread_func() 448 queue->jobs = (struct util_queue_job*) in util_queue_init() 450 if (!queue->jobs) in util_queue_init() 477 if (queue->jobs) { in util_queue_init() [all …]
|
| /third_party/mesa3d/src/intel/vulkan/tests/ |
| D | state_pool_no_free.c | 38 } jobs[NUM_THREADS]; variable 71 jobs[i].pool = &state_pool; in run_test() 72 jobs[i].id = i; in run_test() 73 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); in run_test() 77 pthread_join(jobs[i].thread, NULL); in run_test() 92 if (thread_max < jobs[i].offsets[next[i]]) { in run_test() 93 thread_max = jobs[i].offsets[next[i]]; in run_test() 105 ASSERT(jobs[max_thread_idx].offsets[next[max_thread_idx]] > highest); in run_test() 107 highest = jobs[max_thread_idx].offsets[next[max_thread_idx]]; in run_test()
|
| D | block_pool_no_free.c | 39 } jobs[NUM_THREADS]; variable 45 uint32_t job_id = job - jobs; in alloc_blocks() 126 jobs[i].pool = &pool; in run_test() 127 jobs[i].id = i; in run_test() 128 pthread_create(&jobs[i].thread, NULL, alloc_blocks, &jobs[i]); in run_test() 132 pthread_join(jobs[i].thread, NULL); in run_test() 137 block_ptrs[i] = jobs[i].blocks; in run_test() 142 block_ptrs[i] = jobs[i].back_blocks; in run_test()
|
| D | state_pool_test_helper.h | 30 } jobs[NUM_THREADS]; variable 64 jobs[i].pool = state_pool; in run_state_pool_test() 65 jobs[i].id = i; in run_state_pool_test() 66 pthread_create(&jobs[i].thread, NULL, alloc_states, &jobs[i]); in run_state_pool_test() 70 pthread_join(jobs[i].thread, NULL); in run_state_pool_test()
|
| /third_party/mbedtls/tests/scripts/ |
| D | depends.py | 14 Then, test domains (groups of jobs, tests) are built based on predefined data 42 Lastly, the collected jobs are executed and (optionally) tested, with 313 """Initialize the jobs container""" 314 self.jobs = [] 322 The domain contains a set of jobs that enable one of the elements 325 If exclude is a regular expression, skip generated jobs whose description 340 self.jobs.append(job) 346 If exclude is a regular expression, skip generated jobs whose description 360 self.jobs.append(job) 365 each call adds respective jobs. The job array initialization is done once in [all …]
|
| /third_party/vixl/tools/ |
| D | test.py | 150 '--jobs', '-j', metavar='N', type=int, nargs='?', 153 help='''Runs the tests using N jobs. If the option is set but no value is 154 provided, the script will use as many jobs as it thinks useful.''') 260 def RunLinter(jobs): argument 263 jobs = args.jobs, progress_prefix = 'cpp lint: ') 266 def RunClangFormat(clang_path, jobs): argument 269 jobs = jobs, 272 def RunClangTidy(clang_path, jobs): argument 275 jobs = jobs, 282 def BuildAll(build_options, jobs, environment_options): argument [all …]
|
| D | lint.py | 62 parser.add_argument('--jobs', '-j', metavar='N', type=int, nargs='?', 65 help='''Runs the tests using N jobs. If the option is set 66 but no value is provided, the script will use as many jobs 132 jobs = 1, argument 159 pool = multiprocessing.Pool(jobs) 229 def RunLinter(files, jobs=1, progress_prefix='', cached=True): argument 233 jobs=jobs, 248 retcode = RunLinter(files, jobs=args.jobs, cached=cached)
|
| /third_party/mesa3d/.gitlab-ci/bin/ |
| D | ci_run_n_monitor.py | 11 Helper script to restrict running only required CI jobs 46 # dependency graph to see which jobs the target jobs need 125 """Monitors pipeline and delegate canceling jobs""" 138 for job in pipeline.jobs.list(all=True, sort="desc"): 139 # target jobs 154 # all jobs 193 pjob = project.jobs.get(job.id, lazy=True) 204 pjob = project.jobs.get(job.id, lazy=True) 210 """Cancel unwanted GitLab jobs""" 223 job = project.jobs.get(job_id) [all …]
|
| /third_party/node/deps/npm/node_modules/node-gyp/lib/ |
| D | build.js | 26 const jobs = gyp.opts.jobs || process.env.JOBS 151 if (jobs) { 152 j = parseInt(jobs, 10) 155 } else if (jobs.toUpperCase() === 'MAX') { 164 if (jobs) { 165 j = parseInt(jobs, 10) 167 argv.push('--jobs') 169 } else if (jobs.toUpperCase() === 'MAX') { 170 argv.push('--jobs')
|
| /third_party/libdrm/tests/exynos/ |
| D | exynos_fimg2d_event.c | 103 * We need to wait until all G2D jobs are finished, otherwise we 110 static void wait_all_jobs(struct g2d_job* jobs, unsigned num_jobs) in wait_all_jobs() argument 115 while (jobs[i].busy) in wait_all_jobs() 121 static struct g2d_job* free_job(struct g2d_job* jobs, unsigned num_jobs) in free_job() argument 126 if (jobs[i].busy == 0) in free_job() 127 return &jobs[i]; in free_job() 136 struct g2d_job *jobs = calloc(num_jobs, sizeof(struct g2d_job)); in g2d_work() local 142 jobs[i].id = i; in g2d_work() 150 j = free_job(jobs, num_jobs); in g2d_work() 189 wait_all_jobs(jobs, num_jobs); in g2d_work() [all …]
|
| /third_party/cups-filters/scripting/php/ |
| D | README | 74 Gets a list of jobs: 78 The "dest" string can be blank for jobs on all destinations. 79 Pass TRUE for "myjobs" to only get jobs for the current user. 80 The "completed" argument can be 0 for pending jobs, 1 for 81 completed jobs, and -1 for all jobs. 100 $jobs = cups_get_jobs("", FALSE, -1);
|
| /third_party/mesa3d/docs/ci/ |
| D | LAVA.rst | 17 taking in jobs for the hardware that the LAVA lab contains. The 19 connects to the LAVA lab using a predefined token to submit jobs under 22 The LAVA instance manages scheduling those jobs to the boards present. 30 some boards booting using test jobs. Start with the stock QEMU 37 that device type and the Mesa jobs will be scheduled to any of them. 50 to restrict the jobs it takes or it will grab random jobs from tasks 81 Now it's time to define your test jobs in the driver-specific
|
| /third_party/node/deps/npm/node_modules/tar/lib/ |
| D | pack.js | 40 const JOBS = Symbol('jobs') constant 118 this[JOBS] = 0 119 this.jobs = +opt.jobs || 4 164 this[JOBS] += 1 179 this[JOBS] += 1 183 this[JOBS] -= 1 206 this[JOBS] += 1 209 this[JOBS] -= 1 230 w !== null && this[JOBS] < this.jobs; 242 if (this[ENDED] && !this[QUEUE].length && this[JOBS] === 0) { [all …]
|
| /third_party/skia/infra/bots/ |
| D | README.md | 7 Tasks and Jobs 15 Jobs are collections of related tasks which help define sub-sections of the DAG, 16 for example, to be used as try jobs. Each job is defined as an entry point into 19 The tasks.json file in this directory is the list of tasks and jobs for 23 reads the tasks.json file at each commit to determine which jobs to run. For 30 * jobs.json - The list of all jobs to run. Edit this to add or remove
|
| /third_party/node/deps/v8/src/compiler-dispatcher/ |
| D | lazy-compile-dispatcher.h | 70 // LazyCompileDispatcher::DoIdleWork tries to advance as many jobs out of jobs_ 74 // LazyCompileDispatcher::DoBackgroundWork advances one of the pending jobs, 100 // Aborts all jobs, blocking until all jobs are aborted. 206 // The set of jobs that can be run on a background thread. 209 // The set of jobs that can be finalized on the main thread. 212 // The total number of jobs ready to execute on background, both those pending 217 // The set of all allocated jobs, used for verification of the various queues 222 // A queue of jobs to delete on the background thread(s). Jobs in this queue
|
| /third_party/mesa3d/src/panfrost/perf/ |
| D | T72x.xml | 32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple… 33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) … 34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
|
| D | T82x.xml | 32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple… 33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) … 34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
|
| D | T83x.xml | 32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple… 33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) … 34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
|
| D | T86x.xml | 32 …="JS0_JOBS" title="Job Manager Jobs" name="JS0 jobs" description="Number of Jobs (fragment) comple… 33 …JOBS" title="Job Manager Jobs" name="JS1 jobs" description="Number of Jobs (vertex/tiler/compute) … 34 …S2_JOBS" title="Job Manager Jobs" name="JS2 jobs" description="Number of Jobs (vertex/compute) com…
|
| /third_party/node/deps/openssl/openssl/crypto/async/ |
| D | async.c | 112 * no max size and no pre-created jobs in async_get_pool_job() 119 job = sk_ASYNC_JOB_pop(pool->jobs); in async_get_pool_job() 147 sk_ASYNC_JOB_push(pool->jobs, job); in async_release_job() 324 if (pool == NULL || pool->jobs == NULL) in async_empty_pool() 328 job = sk_ASYNC_JOB_pop(pool->jobs); in async_empty_pool() 374 pool->jobs = sk_ASYNC_JOB_new_reserve(NULL, init_size); in ASYNC_init_thread() 375 if (pool->jobs == NULL) { in ASYNC_init_thread() 383 /* Pre-create jobs as required */ in ASYNC_init_thread() 390 * skip creation of any more jobs in ASYNC_init_thread() 396 sk_ASYNC_JOB_push(pool->jobs, job); /* Cannot fail due to reserve */ in ASYNC_init_thread() [all …]
|
| /third_party/openssl/crypto/async/ |
| D | async.c | 112 * no max size and no pre-created jobs in async_get_pool_job() 119 job = sk_ASYNC_JOB_pop(pool->jobs); in async_get_pool_job() 147 sk_ASYNC_JOB_push(pool->jobs, job); in async_release_job() 324 if (pool == NULL || pool->jobs == NULL) in async_empty_pool() 328 job = sk_ASYNC_JOB_pop(pool->jobs); in async_empty_pool() 374 pool->jobs = sk_ASYNC_JOB_new_reserve(NULL, init_size); in ASYNC_init_thread() 375 if (pool->jobs == NULL) { in ASYNC_init_thread() 383 /* Pre-create jobs as required */ in ASYNC_init_thread() 390 * skip creation of any more jobs in ASYNC_init_thread() 396 sk_ASYNC_JOB_push(pool->jobs, job); /* Cannot fail due to reserve */ in ASYNC_init_thread() [all …]
|
| /third_party/skia/tools/skqp/ |
| D | find_commit_with_best_gold_results.py | 23 path = skia_directory + '/infra/bots/jobs.json' 36 jobs = json.load(f) 37 for job in jobs: 62 def get_results_for_commit(commit, jobs): argument 81 for job in jobs for config in CONFIGS] 94 jobs = [j for j in get_jobs()] 98 results.append((commit_hash, get_results_for_commit(commit_hash, jobs)))
|