Searched refs:inference_preference (Results 1 – 7 of 7) sorted by relevance
32 jboolean quantized_models_allowed, jint inference_preference) { in Java_org_tensorflow_lite_gpu_GpuDelegate_createDelegate() argument44 options.inference_preference = static_cast<int32_t>(inference_preference); in Java_org_tensorflow_lite_gpu_GpuDelegate_createDelegate()
73 EXPECT_EQ(best_options.inference_preference, in TEST_F()74 default_options.inference_preference); in TEST_F()
73 int32_t inference_preference; member
310 options.usage = ToUsage(delegate_options.inference_preference); in InitializeOpenClApi()328 options.usage = ToUsage(delegate_options.inference_preference); in InitializeOpenGlApi()446 .inference_preference = in TfLiteGpuDelegateOptionsV2Default()
130 options.inference_preference = in CreateGPUDelegate()
130 options.inference_preference = in main()
118 gpu_opts.inference_preference = in CreateGPUDelegate()