/test/vti/test_serving/gae/webapp/src/endpoint/ |
D | job_queue.py | 23 from webapp.src.proto import model 29 JOB_QUEUE_RESOURCE = endpoints.ResourceContainer(model.JobMessage) 41 model.JobLeaseResponse, 47 job_query = model.JobModel.query( 48 model.JobModel.hostname == request.hostname, 49 model.JobModel.status == Status.JOB_STATUS_DICT["ready"]) 61 job_message = model.JobMessage() 62 common_attributes = self.GetCommonAttributes(job, model.JobMessage) 66 device_query = model.DeviceModel.query( 67 model.DeviceModel.serial.IN(job.serial)) [all …]
|
D | schedule_info.py | 23 from webapp.src.proto import model 26 SCHEDULE_INFO_RESOURCE = endpoints.ResourceContainer(model.ScheduleInfoMessage) 28 model.ScheduleSuspendMessage) 37 model.DefaultResponse, 43 schedule_query = model.ScheduleModel.query( 44 model.ScheduleModel.schedule_type != "green") 48 return model.DefaultResponse( 49 return_code=model.ReturnCodeMessage.SUCCESS) 53 model.DefaultResponse, 59 exist_on_both = self.GetCommonAttributes(request, model.ScheduleModel) [all …]
|
D | lab_info.py | 25 from webapp.src.proto import model 27 LAB_INFO_RESOURCE = endpoints.ResourceContainer(model.LabInfoMessage) 28 LAB_HOST_INFO_RESOURCE = endpoints.ResourceContainer(model.LabHostInfoMessage) 37 model.DefaultResponse, 43 lab_query = model.LabModel.query() 47 return model.DefaultResponse( 48 return_code=model.ReturnCodeMessage.SUCCESS) 52 model.DefaultResponse, 61 duplicate_query = model.LabModel.query( 62 model.LabModel.name == request.name, [all …]
|
D | build_info.py | 21 from webapp.src.proto import model 23 BUILD_INFO_RESOURCE = endpoints.ResourceContainer(model.BuildInfoMessage) 32 model.DefaultResponse, 38 build_query = model.BuildModel.query( 39 model.BuildModel.build_id == request.build_id, 40 model.BuildModel.build_target == request.build_target, 41 model.BuildModel.build_type == request.build_type, 42 model.BuildModel.artifact_type == request.artifact_type) 58 build = model.BuildModel() 60 model.BuildModel) [all …]
|
D | host_info.py | 25 from webapp.src.proto import model 27 HOST_INFO_RESOURCE = endpoints.ResourceContainer(model.HostInfoMessage) 40 device_query = model.DeviceModel.query( 41 model.DeviceModel.hostname == hostname, 42 model.DeviceModel.product == _NULL_DEVICE_PRODUCT_TYPE 50 device = model.DeviceModel() 69 model.DefaultResponse, 82 device_query = model.DeviceModel.query( 83 model.DeviceModel.serial == request_device.serial 89 device = model.DeviceModel() [all …]
|
D | endpoint_base_test.py | 29 from webapp.src.proto import model 47 job_message = model.JobMessage() 55 job = model.JobModel() 77 job_message = model.JobMessage() 100 job = model.JobModel() 112 model.GetRequestMessage).combined_message_class( 121 metaclass=model.DeviceModel, 122 message=model.DeviceInfoMessage) 133 model.GetRequestMessage).combined_message_class( 142 metaclass=model.DeviceModel, [all …]
|
D | build_info_test.py | 26 from webapp.src.proto import model 39 builds = model.BuildModel.query().fetch() 52 self.assertEqual(response.return_code, model.ReturnCodeMessage.SUCCESS) 53 builds = model.BuildModel.query().fetch() 64 builds = model.BuildModel.query().fetch() 77 self.assertEqual(response.return_code, model.ReturnCodeMessage.SUCCESS) 78 builds = model.BuildModel.query().fetch() 91 self.assertEqual(response.return_code, model.ReturnCodeMessage.SUCCESS) 92 builds = model.BuildModel.query().fetch() 103 builds = model.BuildModel.query().fetch() [all …]
|
D | job_queue_test.py | 28 from webapp.src.proto import model 83 job = model.JobModel() 95 model.ReturnCodeMessage.SUCCESS) 106 devices = model.DeviceModel.query().fetch() 123 model.ReturnCodeMessage.SUCCESS) 125 jobs = model.JobModel.query().fetch() 143 model.ReturnCodeMessage.SUCCESS) 145 jobs = model.JobModel.query().fetch() 151 devices = model.DeviceModel.query().fetch()
|
D | host_info_test.py | 27 from webapp.src.proto import model 57 devices = model.DeviceModel.query().fetch() 74 devices = model.DeviceModel.query().fetch() 86 devices = model.DeviceModel.query().fetch()
|
D | lab_info_test.py | 26 from webapp.src.proto import model 70 devices = model.DeviceModel.query().fetch() 80 devices = model.DeviceModel.query().fetch() 125 devices = model.DeviceModel.query().fetch()
|
/test/vti/test_serving/gae/webapp/src/tasks/ |
D | indexing.py | 21 from webapp.src.proto import model 30 "build": model.BuildModel, 31 "device": model.DeviceModel, 32 "lab": model.LabModel, 33 "job": model.JobModel, 34 "schedule": model.ScheduleModel 79 reload(model) 111 parent_schedule_query = model.ScheduleModel.query( 112 model.ScheduleModel.priority == entity.priority, 113 model.ScheduleModel.test_name == entity.test_name, [all …]
|
D | indexing_test.py | 25 from webapp.src.proto import model 52 schedules = model.ScheduleModel.query().fetch() 57 job = model.JobModel() 78 jobs = model.JobModel.query().fetch() 82 jobs = model.JobModel.query().fetch() 90 jobs = model.JobModel.query().fetch() 124 schedules = model.ScheduleModel.query().fetch() 130 job = model.JobModel() 151 jobs = model.JobModel.query().fetch() 155 jobs = model.JobModel.query().fetch() [all …]
|
D | removing_outdated_devices.py | 21 from webapp.src.proto import model 33 device_query = model.DeviceModel.query( 34 model.DeviceModel.timestamp < datetime.datetime.now() -
|
D | removing_outdated_devices_test.py | 26 from webapp.src.proto import model 76 devices = model.DeviceModel.query().fetch() 90 devices = model.DeviceModel.query().fetch() 104 devices = model.DeviceModel.query().fetch()
|
/test/vti/test_serving/gae/webapp/src/utils/ |
D | model_util_test.py | 27 from webapp.src.proto import model 67 jobs = model.JobModel.query().fetch() 74 parent_from_db = model.ScheduleModel.query().fetch()[0] 88 jobs = model.JobModel.query().fetch() 94 parent_from_db = model.ScheduleModel.query().fetch()[0] 99 device = model.DeviceModel.query().fetch()[0] 107 jobs = model.JobModel.query().fetch() 110 ready_jobs = model.JobModel.query( 111 model.JobModel.status == Status.JOB_STATUS_DICT[ 118 parent_from_db = model.ScheduleModel.query().fetch()[0] [all …]
|
/test/vti/test_serving/gae/webapp/src/scheduler/ |
D | schedule_worker_test.py | 27 from webapp.src.proto import model 70 self.assertEqual(1, len(model.JobModel.query().fetch())) 73 device_query = model.DeviceModel.query( 74 model.DeviceModel.serial == device.serial) 112 schedules = model.ScheduleModel.query().fetch() 159 jobs = model.JobModel.query().fetch() 171 jobs = model.JobModel.query().fetch() 189 jobs = model.JobModel.query().fetch() 254 jobs = model.JobModel.query().fetch() 291 jobs = model.JobModel.query().fetch() [all …]
|
D | schedule_worker.py | 26 from webapp.src.proto import model 131 device_query = model.DeviceModel.query( 132 model.DeviceModel.serial.IN(target_device_serials)) 164 build_query = model.BuildModel.query( 165 model.BuildModel.artifact_type == artifact_type, 166 model.BuildModel.manifest_branch == manifest_branch, 167 model.BuildModel.build_target == build_target, 168 model.BuildModel.build_type == build_type) 196 schedule_query = model.ScheduleModel.query( 197 model.ScheduleModel.suspended != True) [all …]
|
D | device_heartbeat.py | 25 from webapp.src.proto import model 46 device_query = model.DeviceModel.query( 47 model.DeviceModel.status != 64 lab_query = model.LabModel.query( 65 model.LabModel.hostname == device.hostname)
|
D | job_heartbeat.py | 25 from webapp.src.proto import model 47 job_query = model.JobModel.query( 48 model.JobModel.status == Status.JOB_STATUS_DICT["leased"]) 72 device_query = model.DeviceModel.query( 73 model.DeviceModel.serial.IN(job.serial))
|
D | job_heartbeat_test.py | 27 from webapp.src.proto import model 85 jobs = model.JobModel.query().fetch() 110 jobs = model.JobModel.query().fetch() 118 devices = model.DeviceModel.query( 119 model.DeviceModel.serial.IN(serials)).fetch()
|
/test/vti/test_serving/gae/webapp/src/testing/ |
D | unittest_base.py | 31 from webapp.src.proto import model 85 lab = model.LabModel() 107 device = model.DeviceModel() 192 schedule = model.ScheduleModel() 263 build = model.BuildModel() 297 build_query = model.BuildModel.query( 298 model.BuildModel.artifact_type == build.artifact_type, 299 model.BuildModel.build_target == build.build_target, 300 model.BuildModel.signed == build.signed, 301 model.BuildModel.manifest_branch == build.manifest_branch) [all …]
|
/test/mlts/models/ |
D | README.txt | 22 MobileNet tensorflow lite model based on: 34 8bit quantized MobileNet tensorflow lite model based on: 46 MobileNet v2 tensorflow lite model based on: 58 8bit quantized MobileNet v2 tensorflow lite model based on: 70 Float version of MobileNet SSD tensorflow model based on: 83 8bit quantized MobileNet SSD tensorflow lite model based on: 96 TTS tensorflow lite model based on: 102 Note that the tensorflow lite model is the acoustic model in the paper. It is used because it is 103 much heavier than the duration model. 106 ASR tensorflow lite model based on the ASR acoustic model in: [all …]
|
/test/mlts/benchmark/jni/ |
D | random_graph_test_jni.cpp | 190 GeneratedModel model; in runRandomModel() local 191 createModel(*testModel, &model); in runRandomModel() 192 if (!model.isValid()) { in runRandomModel() 196 auto modelFinishResult = model.finish(); in runRandomModel() 212 model.getHandle(), &mDevice, 1, opsSupportedFlags.get()); in runRandomModel() 247 auto [compilationResult, compilation] = CreateCompilation(model, devices); in runRandomModel() 404 android::nn::generated_tests::GeneratedModel model; in OperationsSupportedByDevice() local 405 createModel(*testModel, &model); in OperationsSupportedByDevice() 406 if (!model.isValid()) { in OperationsSupportedByDevice() 410 auto modelFinishResult = model.finish(); in OperationsSupportedByDevice() [all …]
|
/test/mlts/benchmark/src/com/android/nn/crashtest/app/ |
D | AcceleratorSpecificTestSupport.java | 46 for (TestModels.TestModelEntry model : TestModels.modelsList()) { in findTestModelRunningOnAccelerator() 47 if (Processor.isTestModelSupportedByAccelerator(context, model, acceleratorName)) { in findTestModelRunningOnAccelerator() 48 return Optional.of(model); in findTestModelRunningOnAccelerator() 57 for (TestModels.TestModelEntry model : TestModels.modelsList()) { in findAllTestModelsRunningOnAccelerator() 58 if (Processor.isTestModelSupportedByAccelerator(context, model, acceleratorName)) { in findAllTestModelsRunningOnAccelerator() 59 result.add(model); in findAllTestModelsRunningOnAccelerator()
|
/test/mlts/benchmark/ |
D | README.txt | 44 partition the model and assign the best available one(s) by using the 62 on a single model in multiple processes and threads with different probabilities in client process 65 * multi-process-model-load-stress: this extends the `parallel-inference-stress` running inference 66 on a single model in multiple processes and threads with different probabilities in client process 69 * memory-mapped-model-load-stress: runs a series of parallel model compilation with memory mapped 72 * model-load-random-stress: test compiling a large set of randomly generated models
|