/external/python/cpython2/Doc/includes/ |
D | mp_synchronize.py | 11 import multiprocessing # may get overwritten 21 print '\n\t\t\t' + str(multiprocessing.current_process()) + ' has finished' 27 running = multiprocessing.Value('i', TASKS) 28 mutex = multiprocessing.Lock() 31 p = multiprocessing.Process(target=value_func, args=(running, mutex)) 54 q = multiprocessing.Queue() 56 p = multiprocessing.Process(target=queue_func, args=(q,)) 83 cond = multiprocessing.Condition() 85 p = multiprocessing.Process(target=condition_func, args=(cond,)) 123 print '%s has finished' % multiprocessing.current_process() [all …]
|
D | mp_benchmarks.py | 8 import time, sys, multiprocessing, threading, Queue, gc 72 c, d = multiprocessing.Pipe() 73 cond = multiprocessing.Condition() 80 p = multiprocessing.Process(target=pipe_func, 183 manager = multiprocessing.Manager() 191 test_queuespeed(multiprocessing.Process, multiprocessing.Queue(), 192 multiprocessing.Condition()) 194 test_queuespeed(multiprocessing.Process, manager.Queue(), 206 test_seqspeed(multiprocessing.Array('i', range(10), lock=False)) 208 test_seqspeed(multiprocessing.Array('i', range(10), lock=True)) [all …]
|
D | mp_pool.py | 8 import multiprocessing 20 multiprocessing.current_process().name, 49 print 'cpu_count() = %d\n' % multiprocessing.cpu_count() 57 pool = multiprocessing.Pool(PROCESSES) 196 except multiprocessing.TimeoutError: 209 except multiprocessing.TimeoutError: 266 pool = multiprocessing.Pool(2) 284 pool = multiprocessing.Pool(2) 301 multiprocessing.freeze_support() 309 import multiprocessing.dummy as multiprocessing namespace
|
/external/python/cpython2/Lib/multiprocessing/ |
D | __init__.py | 64 from multiprocessing.process import Process, current_process, active_children 65 from multiprocessing.util import SUBDEBUG, SUBWARNING 97 from multiprocessing.managers import SyncManager 106 from multiprocessing.connection import Pipe 144 from multiprocessing.forking import freeze_support 151 from multiprocessing.util import get_logger 158 from multiprocessing.util import log_to_stderr 165 from multiprocessing import reduction 175 from multiprocessing.synchronize import Lock 182 from multiprocessing.synchronize import RLock [all …]
|
/external/grpc-grpc/tools/run_tests/ |
D | run_microbenchmark.py | 17 import multiprocessing 87 '%d' % multiprocessing.cpu_count() 117 if len(benchmarks) >= min(16, multiprocessing.cpu_count()): 122 multiprocessing.cpu_count() / 2)) 123 jobset.run(profile_analysis, maxjobs=multiprocessing.cpu_count()) 124 jobset.run(cleanup, maxjobs=multiprocessing.cpu_count()) 130 jobset.run(benchmarks, maxjobs=max(1, multiprocessing.cpu_count() / 2)) 131 jobset.run(profile_analysis, maxjobs=multiprocessing.cpu_count()) 132 jobset.run(cleanup, maxjobs=multiprocessing.cpu_count()) 140 '%d' % multiprocessing.cpu_count() [all …]
|
/external/rust/crates/grpcio-sys/grpc/tools/run_tests/ |
D | run_microbenchmark.py | 18 import multiprocessing 124 if len(benchmarks) >= min(16, multiprocessing.cpu_count()): 129 multiprocessing.cpu_count() / 2)) 130 jobset.run(profile_analysis, maxjobs=multiprocessing.cpu_count()) 131 jobset.run(cleanup, maxjobs=multiprocessing.cpu_count()) 137 jobset.run(benchmarks, maxjobs=max(1, multiprocessing.cpu_count() / 2)) 138 jobset.run(profile_analysis, maxjobs=multiprocessing.cpu_count()) 139 jobset.run(cleanup, maxjobs=multiprocessing.cpu_count()) 181 jobset.run(profile_analysis, maxjobs=multiprocessing.cpu_count()) 182 jobset.run(cleanup, maxjobs=multiprocessing.cpu_count()) [all …]
|
/external/toolchain-utils/bestflags/ |
D | pipeline_process.py | 13 import multiprocessing 19 class PipelineProcess(multiprocessing.Process): 61 multiprocessing.Process.__init__(self) 75 manager = multiprocessing.Manager() 89 work_pool = multiprocessing.Pool(self._num_processes) 92 helper_process = multiprocessing.Process(
|
D | steering_test.py | 11 import multiprocessing 107 manager = multiprocessing.Manager() 111 steering_process = multiprocessing.Process( 154 manager = multiprocessing.Manager() 158 steering_process = multiprocessing.Process(
|
/external/python/cpython3/Lib/test/ |
D | _test_multiprocessing.py | 43 import multiprocessing.connection 44 import multiprocessing.dummy 45 import multiprocessing.heap 46 import multiprocessing.managers 47 import multiprocessing.pool 48 import multiprocessing.queues 50 from multiprocessing import util 53 from multiprocessing import reduction 59 from multiprocessing.sharedctypes import Value, copy 65 from multiprocessing import shared_memory [all …]
|
D | mp_fork_bomb.py | 1 import multiprocessing, sys 11 multiprocessing.set_start_method(sys.argv[1]) 13 multiprocessing.set_start_method('spawn') 15 p = multiprocessing.Process(target=foo)
|
D | mp_preload.py | 1 import multiprocessing 3 multiprocessing.Lock() 11 ctx = multiprocessing.get_context("forkserver")
|
/external/python/cpython3/Doc/includes/ |
D | mp_pool.py | 1 import multiprocessing 13 multiprocessing.current_process().name, 45 with multiprocessing.Pool(PROCESSES) as pool: 132 except multiprocessing.TimeoutError: 145 except multiprocessing.TimeoutError: 152 multiprocessing.freeze_support()
|
/external/tensorflow/tensorflow/python/distribute/ |
D | multi_process_lib.py | 20 import multiprocessing 63 multiprocessing.context.ForkServerProcess): 69 class AbslForkServerContext(multiprocessing.context.ForkServerContext): 73 multiprocessing = AbslForkServerContext() variable 74 Process = multiprocessing.Process 129 multiprocessing.get_context().set_executable(sys.argv[0])
|
/external/python/cpython2/Lib/test/ |
D | test_multiprocessing.py | 30 import multiprocessing.dummy 31 import multiprocessing.connection 32 import multiprocessing.managers 33 import multiprocessing.heap 34 import multiprocessing.pool 36 from multiprocessing import util 39 from multiprocessing import reduction 45 from multiprocessing.sharedctypes import Value, copy 283 cpus = multiprocessing.cpu_count() 304 from multiprocessing import forking [all …]
|
D | mp_fork_bomb.py | 1 import multiprocessing 10 r, w = multiprocessing.Pipe(False) 11 p = multiprocessing.Process(target=foo, args=(w,))
|
/external/rust/crates/grpcio-sys/grpc/tools/distrib/python/grpcio_tools/ |
D | _parallel_compile_patch.py | 27 import multiprocessing 28 BUILD_EXT_COMPILER_JOBS = multiprocessing.cpu_count() 55 import multiprocessing.pool 56 multiprocessing.pool.ThreadPool(BUILD_EXT_COMPILER_JOBS).map(
|
/external/chromium-trace/catapult/common/py_utils/py_utils/ |
D | lock_unittest.py | 9 import multiprocessing 64 p = multiprocessing.Process( 95 p = multiprocessing.Process( 116 p = multiprocessing.Process( 134 p = multiprocessing.Process( 153 p = multiprocessing.Process( 162 p = multiprocessing.Process(
|
/external/python/cpython2/Doc/library/ |
D | multiprocessing.rst | 1 :mod:`multiprocessing` --- Process-based "threading" interface 4 .. module:: multiprocessing 13 :mod:`multiprocessing` is a package that supports spawning processes using an 14 API similar to the :mod:`threading` module. The :mod:`multiprocessing` package 17 to this, the :mod:`multiprocessing` module allows the programmer to fully 21 The :mod:`multiprocessing` module also introduces APIs which do not have 30 from multiprocessing import Pool 47 In :mod:`multiprocessing`, processes are spawned by creating a :class:`Process` 52 from multiprocessing import Process 64 from multiprocessing import Process [all …]
|
/external/vixl/tools/ |
D | threaded_tests.py | 28 import multiprocessing 42 n_tests_passed = multiprocessing.Value('i', 0) 43 n_tests_failed = multiprocessing.Value('i', 0) 44 n_tests_skipped = multiprocessing.Value('i', 0) 45 manager = multiprocessing.Manager()
|
D | thread_pool.py | 29 import multiprocessing 41 __run_tests_lock__ = multiprocessing.Lock() 49 pool = multiprocessing.Pool(num_threads)
|
/external/python/cpython3/Doc/library/ |
D | multiprocessing.rst | 1 :mod:`multiprocessing` --- Process-based parallelism 4 .. module:: multiprocessing 7 **Source code:** :source:`Lib/multiprocessing/` 14 :mod:`multiprocessing` is a package that supports spawning processes using an 15 API similar to the :mod:`threading` module. The :mod:`multiprocessing` package 19 to this, the :mod:`multiprocessing` module allows the programmer to fully 23 The :mod:`multiprocessing` module also introduces APIs which do not have 25 :class:`~multiprocessing.pool.Pool` object which offers a convenient means of 30 of data parallelism using :class:`~multiprocessing.pool.Pool`, :: 32 from multiprocessing import Pool [all …]
|
/external/autotest/site_utils/tester_feedback/ |
D | request_multiplexer.py | 7 from multiprocessing import reduction 10 import multiprocessing 34 self._request_queue = multiprocessing.Queue() 134 self._request_handling_process = multiprocessing.Process( 164 reply_pipe_send, reply_pipe_recv = multiprocessing.Pipe()
|
/external/python/cpython2/Doc/tools/ |
D | susp-ignored.csv | 80 library/multiprocessing,,:queue,">>> QueueManager.register('get_queue', callable=lambda:queue)" 81 library/multiprocessing,,`,">>> l._callmethod('__getitem__', (20,)) # equiv to `l[20]`" 82 library/multiprocessing,,`,">>> l._callmethod('__getslice__', (2, 7)) # equiv to `l[2:7]`" 83 library/multiprocessing,,`,# `BaseManager`. 84 library/multiprocessing,,`,# `Pool.imap()` (which will save on the amount of code needed anyway). 85 library/multiprocessing,,`,# A test file for the `multiprocessing` package 86 library/multiprocessing,,`,# A test of `multiprocessing.Pool` class 87 library/multiprocessing,,`,# Add more tasks using `put()` 88 library/multiprocessing,,`,# in the original order then consider using `Pool.map()` or 89 library/multiprocessing,,`,# Not sure if we should synchronize access to `socket.accept()` method by [all …]
|
/external/autotest/site_utils/ |
D | rpc_flight_recorder.py | 12 import multiprocessing 97 pool = multiprocessing.Pool(processes=multiprocessing.cpu_count() * 4) 124 self._manager = multiprocessing.Manager() 136 self._stop_event = multiprocessing.Event() 143 self._update_shards_ps = multiprocessing.Process( 155 self._poll_rpc_server_ps = multiprocessing.Process(
|
/external/llvm/utils/lit/lit/ |
D | run.py | 16 import multiprocessing 18 multiprocessing = None variable 118 self.queue = multiprocessing.Queue() 215 if jobs != 1 and use_processes and multiprocessing: 217 task_impl = multiprocessing.Process 218 queue_impl = multiprocessing.Queue 219 canceled_flag = multiprocessing.Value('i', 0)
|