/external/autotest/server/samples/ |
D | profilertest.srv | 3 ssh_hosts = [hosts.create_host(m) for m in machines] 8 machines, name): 18 at.host.hostname, "PROF_MASTER", str(machines))) 27 def wait_for_profilers(machines, timeout = 180): 31 sync_bar.rendezvous_servers("PROF_MASTER", *machines) 34 def start_profilers(machines, timeout = 180): 38 start_bar.rendezvous_servers("PROF_MASTER", *machines) 41 def stop_profilers(machines, timeout = 120): 43 stop_bar.rendezvous_servers("PROF_MASTER", *machines) 47 wait_for_profilers(machines) [all …]
|
D | run_test.srv | 8 print "usage: -t <test name> -m <machines> -l <log dir>" 38 print "Going to launch %s on %r with log dir of %s." % (test, machines, logdir) 39 parallel_simple(run, machines)
|
D | autoserv_console.srv | 11 if machines: 12 host = hosts.create_host(machines[0])
|
/external/autotest/server/ |
D | standalone_profiler.py | 37 def generate_test(machines, hostname, profilers, timeout_start, timeout_stop, argument 65 hostname, _PROF_MASTER, machines)) 74 def wait_for_profilers(machines, timeout=300): argument 77 sb.rendezvous_servers(_PROF_MASTER, *machines) 80 def start_profilers(machines, timeout=120): argument 83 sb.rendezvous_servers(_PROF_MASTER, *machines) 86 def stop_profilers(machines, timeout=120): argument 89 sb.rendezvous_servers(_PROF_MASTER, *machines) 92 def finish_profilers(machines, timeout=120): argument 95 sb.rendezvous_servers(_PROF_MASTER, *machines)
|
D | autoserv | 89 if parser.options.machines: 90 machines = parser.options.machines.replace(',', ' ').strip().split() 92 machines = [] 95 machines = [] 100 machines.append(m) 102 logging.debug('Machines: %s', ','.join(machines)) 104 if machines: 105 for machine in machines: 108 machines = list(set(machines)) 109 machines.sort() [all …]
|
D | autoserv.py | 89 if parser.options.machines: 90 machines = parser.options.machines.replace(',', ' ').strip().split() 92 machines = [] 95 machines = [] 100 machines.append(m) 102 logging.debug('Machines: %s', ','.join(machines)) 104 if machines: 105 for machine in machines: 108 machines = list(set(machines)) 109 machines.sort() [all …]
|
D | base_utils.py | 197 def default_mappings(machines): argument 205 mach = machines[0] 207 if len(machines) > 1: 208 machines = machines[1:] 209 for machine in machines: 215 def form_ntuples_from_machines(machines, n=2, mapping_func=default_mappings): argument 220 (mappings, failures) = mapping_func(machines)
|
D | server_job.py | 245 def __init__(self, control, args, resultdir, label, user, machines, argument 314 self.machines = machines 340 'hostname' : ','.join(machines), 487 if not self.machines: 503 if not self.machines: 519 if not self.machines: 535 if not self.machines: 585 def _make_parallel_wrapper(self, function, machines, log): argument 594 if (machines and isinstance(machines, list) 595 and isinstance(machines[0], dict)): [all …]
|
D | autoserv_utils.py | 20 def autoserv_run_job_command(autoserv_directory, machines, argument 96 if machines: 97 command += ['-m', machines]
|
D | base_utils_unittest.py | 14 self.machines = ['mach1', 'mach2', 'mach3', 'mach4', 'mach5', 24 (ntuples, failures) = utils.form_ntuples_from_machines(self.machines)
|
/external/autotest/server/site_tests/p2p_EndToEndTest/ |
D | control | 26 The test runs over a set of N machines generating a random file in one of 27 them (called the "main") and sharing it with the rest of the machines. The 28 success condition of this test occurs when all the N machines have the same 43 peers=machines, 46 if len(machines) < 2: 47 raise error.TestError('At least two machines are needed for this test') 49 # The file ID shared among all test machines. 53 barrier = queue_barrier.QueueBarrier(len(machines)-1) 56 main = machines[0] 58 job.parallel_simple(run, machines)
|
/external/mesa3d/src/gallium/drivers/softpipe/ |
D | sp_compute.c | 112 struct tgsi_exec_machine **machines) in run_workgroup() argument 120 grp_hit_barrier |= cs_run(cs, g_w, g_h, g_d, machines[i], restart_threads); in run_workgroup() 174 struct tgsi_exec_machine **machines; in softpipe_launch_grid() local 193 machines = CALLOC(sizeof(struct tgsi_exec_machine *), num_threads_in_group); in softpipe_launch_grid() 194 if (!machines) { in softpipe_launch_grid() 204 machines[idx] = tgsi_exec_machine_create(PIPE_SHADER_COMPUTE); in softpipe_launch_grid() 206 machines[idx]->LocalMem = local_mem; in softpipe_launch_grid() 207 machines[idx]->LocalMemSize = cs->shader.req_local_mem; in softpipe_launch_grid() 208 cs_prepare(cs, machines[idx], in softpipe_launch_grid() 215 tgsi_exec_set_constant_buffers(machines[idx], PIPE_MAX_CONSTANT_BUFFERS, in softpipe_launch_grid() [all …]
|
/external/toolchain-utils/cros_utils/ |
D | locks.py | 18 def AcquireLock(machines, chromeos_root, timeout=1200): argument 25 lock_machine.LockManager(machines, False, 33 repr(machines), timeout, str(e))) 39 def ReleaseLock(machines, chromeos_root): argument 43 lock_machine.LockManager(machines, False, 48 'Could not unlock %s. %s' % (repr(machines), str(e)))
|
/external/squashfs-tools/RELEASE-READMEs/ |
D | README-2.0-AMD64 | 5 filesystems on amd64 machines. These filesystems work correctly on amd64 6 machines, but cannot be mounted on non-amd64 machines. Likewise, filesystems 7 generated on non amd64 machines could not be mounted on amd64 machines. 12 generated by previous releases will not be mountable on amd64 machines
|
/external/toolchain-utils/ |
D | lock_machine.py | 22 from cros_utils import machines 97 self.machines = list(set(remotes)) or [] 100 if not self.machines: 101 self.machines = self.toolchain_lab_machines 117 if not machines.MachineIsPingable(machine, logging_level='none'): 119 if not machines.MachineIsPingable(cros_machine, logging_level='none'): 267 for m in self.machines: 284 self.machines = updated_machines 300 self.machines = [ 301 m for m in self.machines if m not in (cros_machine, machine) [all …]
|
/external/elfutils/tests/ |
D | saridx.c | 30 static const char *machines[] = variable 216 (ehdr.e_machine >= (sizeof (machines) in main() 217 / sizeof (machines[0])) in main() 218 || machines[ehdr.e_machine] == NULL) in main() 220 : machines[ehdr.e_machine]); in main()
|
/external/autotest/utils/ |
D | site_check_dut_usage.py | 88 machines = set() 92 machines.add(machine) 94 num_machines = len(machines) 107 for machine in machines:
|
/external/elfutils/libebl/ |
D | eblopenbackend.c | 73 } machines[] = variable 155 #define nmachines (sizeof (machines) / sizeof (machines[0])) 299 if ((emulation != NULL && strcmp (emulation, machines[cnt].emulation) == 0) in openbackend() 300 || (emulation == NULL && machines[cnt].em == machine)) in openbackend() 303 result->emulation = machines[cnt].emulation; in openbackend() 317 result->machine = machines[cnt].em; in openbackend() 318 result->class = machines[cnt].class; in openbackend() 319 result->data = machines[cnt].data; in openbackend() 328 if (machines[cnt].init && in openbackend() 329 machines[cnt].init (elf, machine, result)) in openbackend()
|
/external/ltp/testcases/network/nfsv4/locks/ |
D | deploy_info | 6 Setting up lock test with several test machines. 19 Running lock test over several test machines. 21 Run test on already configured test machines.
|
/external/autotest/server/tests/sleeptest/ |
D | control | 9 runs sleep for one second on the list of machines. 16 job.parallel_simple(run, machines)
|
/external/autotest/server/site_tests/hardware_StorageStress/ |
D | control.stress | 49 job.parallel_simple(run_hardware_storage_stress_suspend, machines) 50 job.parallel_simple(run_hardware_storage_stress, machines) 51 job.parallel_simple(run_hardware_storage_stress_reboot, machines)
|
/external/autotest/client/site_tests/dummy_Pass/ |
D | wifichaos.star | 9 chaos_dut machines. 15 purpose = "To re-image chaos_dut machines nightly.",
|
D | control.wifichaos | 7 PURPOSE = "To re-image chaos_dut machines nightly." 18 chaos_dut machines
|
/external/grpc-grpc/tools/run_tests/performance/ |
D | README.md | 14 ### On remote machines, to start the driver and workers manually: 16 be used with remote machines, but for e.g., profiling the server, 19 1. You'll need a "driver" and separate "worker" machines. 21 GCE "worker" machines that are in the same zone. 81 machines. Note that the driver will start the "benchmark server" on the first 129 Set this to a comma separated list of QPS worker processes/machines.
|
/external/rust/crates/grpcio-sys/grpc/tools/run_tests/performance/ |
D | README.md | 14 ### On remote machines, to start the driver and workers manually: 16 be used with remote machines, but for e.g., profiling the server, 19 1. You'll need a "driver" and separate "worker" machines. 21 GCE "worker" machines that are in the same zone. 81 machines. Note that the driver will start the "benchmark server" on the first 129 Set this to a comma separated list of QPS worker processes/machines.
|