/third_party/node/test/sequential/ |
D | test-inspector-port-cluster.js | 23 workers: [{ expectedPort: 9230 }] property 28 workers: [ property 40 workers: [ property 51 workers: [{ expectedPort: port + 1 }] property 58 workers: [{ expectedPort: port + 1 }] property 65 workers: [{ expectedPort: port + 1, expectedHost: '0.0.0.0' }] property 72 workers: [{ expectedPort: port + 1, expectedHost: '127.0.0.1' }] property 80 workers: [{ expectedPort: port + 1, expectedHost: '::' }] property 87 workers: [{ expectedPort: port + 1, expectedHost: '::1' }] property 99 workers: [{ expectedPort: port + 2 }] property [all …]
|
/third_party/node/benchmark/cluster/ |
D | echo.js | 7 workers: [1], property 16 workers, property 21 const expectedPerBroadcast = sendsPerBroadcast * workers; 40 for (let i = 0; i < workers; ++i) 44 if (++readies === workers) { 53 for (const id in cluster.workers) 54 cluster.workers[id].disconnect(); 57 for (const id in cluster.workers) { 58 const worker = cluster.workers[id];
|
/third_party/ltp/testcases/kernel/io/direct_io/ |
D | dma_thread_diotest.c | 134 static int workers; variable 231 workers = sysconf(_SC_NPROCESSORS_ONLN); in main() 232 if (workers > MAX_WORKERS) in main() 233 workers = MAX_WORKERS; in main() 260 for (j = 0; j < workers; j++) { in dma_thread_diotest_verify() 270 for (j = 0; j < workers; j++) { in dma_thread_diotest_verify() 277 worker[workers - 1].length = in dma_thread_diotest_verify() 278 READSIZE - PAGE_SIZE * (workers - 1); in dma_thread_diotest_verify() 288 for (j = 0; j < workers; j++) { in dma_thread_diotest_verify() 298 for (j = 0; j < workers; j++) { in dma_thread_diotest_verify() [all …]
|
/third_party/ltp/testcases/kernel/fs/fs_fill/ |
D | fs_fill.c | 25 static struct worker *workers; variable 71 SAFE_PTHREAD_CREATE(&threads[i], NULL, worker, &workers[i]); in testrun() 95 workers = SAFE_MALLOC(sizeof(struct worker) * nthreads); in setup() 98 snprintf(workers[i].dir, sizeof(workers[i].dir), in setup() 100 SAFE_MKDIR(workers[i].dir, 0700); in setup() 108 free(workers); in cleanup()
|
/third_party/node/lib/internal/cluster/ |
D | shared_handle.js | 11 this.workers = new SafeMap(); 28 assert(!this.workers.has(worker.id)); 29 this.workers.set(worker.id, worker); 34 if (!this.workers.has(worker.id)) 37 this.workers.delete(worker.id); 39 if (this.workers.size !== 0)
|
D | primary.js | 39 cluster.workers = {}; 99 for (const worker of ObjectValues(cluster.workers)) { 144 delete cluster.workers[worker.id]; 146 if (ObjectKeys(cluster.workers).length === 0) { 215 cluster.workers[worker.id] = worker; 224 const workers = ObjectKeys(cluster.workers); 226 if (workers.length === 0) { 229 for (const worker of ObjectValues(cluster.workers)) {
|
/third_party/libabigail/tests/ |
D | test-read-write.cc | 40 using abigail::workers::queue; 41 using abigail::workers::task; 42 using abigail::workers::task_sptr; 43 using abigail::workers::get_number_of_threads; 250 struct test_task : public abigail::workers::task 352 using abigail::workers::queue; in main() 353 using abigail::workers::task; in main() 354 using abigail::workers::task_sptr; in main() 355 using abigail::workers::get_number_of_threads; in main()
|
D | test-types-stability.cc | 61 struct test_task : public abigail::workers::task 112 using abigail::workers::queue; in main() 113 using abigail::workers::task; in main() 114 using abigail::workers::task_sptr; in main() 115 using abigail::workers::get_number_of_threads; in main()
|
D | test-diff-filter.cc | 833 struct test_task : public abigail::workers::task 926 using abigail::workers::queue; in main() 927 using abigail::workers::task; in main() 928 using abigail::workers::task_sptr; in main() 929 using abigail::workers::get_number_of_threads; in main()
|
/third_party/ltp/testcases/kernel/fs/read_all/ |
D | read_all.c | 81 static struct worker *workers; variable 272 struct worker *wa = workers; in spawn_workers() 274 memset(workers, 0, worker_count * sizeof(*workers)); in spawn_workers() 302 ret = queue_push(workers[i].q, buf); in work_push_retry() 320 if (!workers) in stop_workers() 324 if (workers[i].q) in stop_workers() 329 if (workers[i].q) { in stop_workers() 330 queue_destroy(workers[i].q, 0); in stop_workers() 331 workers[i].q = 0; in stop_workers() 370 workers = SAFE_MALLOC(worker_count * sizeof(*workers)); in setup() [all …]
|
/third_party/libabigail/src/ |
D | abg-workers.cc | 34 namespace workers namespace 137 std::vector<worker> workers; member 170 workers.push_back(w); in create_workers() 188 if (workers.empty() || !t) in schedule_task() 231 if (workers.empty()) in do_bring_workers_down() 246 for (std::vector<worker>::const_iterator i = workers.begin(); in do_bring_workers_down() 247 i != workers.end(); in do_bring_workers_down() 250 workers.clear(); in do_bring_workers_down()
|
/third_party/node/test/report/ |
D | test-report-worker.js | 22 assert.strictEqual(report.workers.length, 1); 23 helper.validateContent(report.workers[0]); 24 assert.strictEqual(report.workers[0].header.threadId, w.threadId); 40 assert.strictEqual(report.workers.length, 1); 41 helper.validateContent(report.workers[0]);
|
/third_party/node/benchmark/worker/ |
D | echo.js | 7 workers: [1], property 15 function main({ n, workers, sendsPerBroadcast: sends, payload: payloadType }) { property 16 const expectedPerBroadcast = sends * workers; 35 for (let i = 0; i < workers; ++i) { 43 if (++readies === workers) {
|
/third_party/node/test/internet/ |
D | test-dgram-broadcast-multi-process.js | 59 const workers = {}; constant 73 killSubprocesses(workers); 82 workers[worker.pid] = worker; 107 killSubprocesses(workers); 137 Object.keys(workers).forEach((pid) => { 138 const worker = workers[pid]; 160 killSubprocesses(workers);
|
D | test-dgram-multicast-multi-process.js | 40 const workers = {}; constant 47 workers[worker.pid] = worker; 97 Object.keys(workers).forEach(function(pid) { 98 const worker = workers[pid]; 119 killSubprocesses(workers); 144 killSubprocesses(workers);
|
D | test-dgram-multicast-ssm-multi-process.js | 19 const workers = {}; constant 41 workers[worker.pid] = worker; 91 Object.keys(workers).forEach(function(pid) { 92 const worker = workers[pid]; 113 killChildren(workers); 138 killChildren(workers);
|
D | test-dgram-multicast-ssmv6-multi-process.js | 19 const workers = {}; constant 41 workers[worker.pid] = worker; 91 Object.keys(workers).forEach(function(pid) { 92 const worker = workers[pid]; 113 killChildren(workers); 138 killChildren(workers);
|
D | test-dgram-multicast-set-interface-lo.js | 79 const workers = {}; constant 98 killSubprocesses(workers); 118 workers[worker.pid] = worker; 141 killSubprocesses(workers); 171 Object.keys(workers).forEach((pid) => { 172 const worker = workers[pid]; 196 killSubprocesses(workers);
|
/third_party/skia/third_party/externals/abseil-cpp/absl/synchronization/ |
D | blocking_counter_test.cc | 42 std::vector<std::thread> workers; in TEST() local 47 workers.reserve(num_workers); in TEST() 49 workers.emplace_back( in TEST() 61 for (std::thread& w : workers) { in TEST()
|
/third_party/ffmpeg/libavutil/ |
D | slicethread.c | 36 WorkerContext *workers; member 119 if (nb_workers && !(ctx->workers = av_calloc(nb_workers, sizeof(*ctx->workers)))) { in avpriv_slicethread_create() 139 WorkerContext *w = &ctx->workers[i]; in avpriv_slicethread_create() 178 WorkerContext *w = &ctx->workers[i]; in avpriv_slicethread_execute() 214 WorkerContext *w = &ctx->workers[i]; in avpriv_slicethread_free() 222 WorkerContext *w = &ctx->workers[i]; in avpriv_slicethread_free() 230 av_freep(&ctx->workers); in avpriv_slicethread_free()
|
/third_party/node/test/parallel/ |
D | test-cluster-concurrent-disconnect.js | 12 const workers = []; variable 17 workers[i] = worker; 20 for (const worker of workers) 37 for (const worker of workers)
|
/third_party/skia/third_party/externals/spirv-cross/include/spirv_cross/ |
D | thread_group.hpp | 34 workers[i].start(&impl[i]); in ThreadGroup() 39 for (auto &worker : workers) in run() 45 for (auto &worker : workers) in wait() 110 Thread workers[Size]; member in spirv_cross::ThreadGroup
|
/third_party/python/Lib/ |
D | compileall.py | 49 rx=None, quiet=0, legacy=False, optimize=-1, workers=1, argument 84 if workers < 0: 86 if workers != 1: 92 workers = 1 99 if workers != 1 and ProcessPoolExecutor is not None: 101 workers = workers or None 102 with ProcessPoolExecutor(max_workers=workers) as executor: 441 args.legacy, workers=args.workers,
|
/third_party/node/doc/api/ |
D | cluster.md | 28 // Fork workers. 57 // Fork workers. 77 Running Node.js will now share port 8000 between the workers: 103 port, accepts new connections and distributes them across the workers 108 socket and sends it to interested workers. The workers then accept 138 Because workers are all separate processes, they can be killed or 140 workers. As long as there are some workers still alive, the server will 141 continue to accept connections. If no workers are alive, existing connections 143 automatically manage the number of workers, however. It is the application's 158 In the primary it can be obtained using `cluster.workers`. In a worker [all …]
|
/third_party/node/benchmark/http/ |
D | cluster.js | 22 let workers = 0; 27 workers++; 28 if (workers < 2)
|