/external/apache-commons-math/src/main/java/org/apache/commons/math3/stat/inference/ |
D | TestUtils.java | 17 package org.apache.commons.math3.stat.inference; 36 * A collection of static methods to create inference test instances or to 37 * perform inference tests. 68 * @see org.apache.commons.math3.stat.inference.TTest#homoscedasticT(double[], double[]) 76 …* @see org.apache.commons.math3.stat.inference.TTest#homoscedasticT(org.apache.commons.math3.stat.… 85 …* @see org.apache.commons.math3.stat.inference.TTest#homoscedasticTTest(double[], double[], double) 95 * @see org.apache.commons.math3.stat.inference.TTest#homoscedasticTTest(double[], double[]) 103 …* @see org.apache.commons.math3.stat.inference.TTest#homoscedasticTTest(org.apache.commons.math3.s… 112 * @see org.apache.commons.math3.stat.inference.TTest#pairedT(double[], double[]) 121 * @see org.apache.commons.math3.stat.inference.TTest#pairedTTest(double[], double[], double) [all …]
|
/external/aws-sdk-java-v2/services/elasticinference/src/main/resources/codegen-resources/ |
D | service-2.json | 5 "endpointPrefix":"api.elastic-inference", 8 "serviceAbbreviation":"Amazon Elastic Inference", 9 "serviceFullName":"Amazon Elastic Inference", 10 "serviceId":"Elastic Inference", 12 "signingName":"elastic-inference", 13 "uid":"elastic-inference-2017-07-25" 29 …ng April 15, 2023, AWS will not onboard new customers to Amazon Elastic Inference (EI), and will h… 42 …ng April 15, 2023, AWS will not onboard new customers to Amazon Elastic Inference (EI), and will h… 57 …ng April 15, 2023, AWS will not onboard new customers to Amazon Elastic Inference (EI), and will h… 72 …lastic Inference Accelerator. </p> <p> February 15, 2023: Starting April 15, 2023, AWS will not on… [all …]
|
D | endpoint-tests.json | 7 "url": "https://api.elastic-inference.ap-northeast-1.amazonaws.com" 20 "url": "https://api.elastic-inference.ap-northeast-2.amazonaws.com" 33 "url": "https://api.elastic-inference.eu-west-1.amazonaws.com" 46 "url": "https://api.elastic-inference.us-east-1.amazonaws.com" 59 "url": "https://api.elastic-inference.us-east-2.amazonaws.com" 72 "url": "https://api.elastic-inference.us-west-2.amazonaws.com" 85 "url": "https://api.elastic-inference-fips.us-east-1.api.aws" 98 "url": "https://api.elastic-inference-fips.us-east-1.amazonaws.com" 111 "url": "https://api.elastic-inference.us-east-1.api.aws" 124 … "url": "https://api.elastic-inference-fips.cn-north-1.api.amazonwebservices.com.cn" [all …]
|
/external/apache-commons-math/src/main/java/org/apache/commons/math/stat/inference/ |
D | TestUtils.java | 17 package org.apache.commons.math.stat.inference; 24 * A collection of static methods to create inference test instances or to 25 * perform inference tests. 156 * @see org.apache.commons.math.stat.inference.TTest#homoscedasticT(double[], double[]) 164 …* @see org.apache.commons.math.stat.inference.TTest#homoscedasticT(org.apache.commons.math.stat.de… 173 … * @see org.apache.commons.math.stat.inference.TTest#homoscedasticTTest(double[], double[], double) 182 * @see org.apache.commons.math.stat.inference.TTest#homoscedasticTTest(double[], double[]) 190 …* @see org.apache.commons.math.stat.inference.TTest#homoscedasticTTest(org.apache.commons.math.sta… 199 * @see org.apache.commons.math.stat.inference.TTest#pairedT(double[], double[]) 207 * @see org.apache.commons.math.stat.inference.TTest#pairedTTest(double[], double[], double) [all …]
|
/external/armnn/samples/ObjectDetection/include/ |
D | ObjectDetectionPipeline.hpp | 18 …eric object detection pipeline with 3 steps: data pre-processing, inference execution and inference 27 * @param executor - unique pointer to inference runner 28 * @param decoder - unique pointer to inference results decoder 44 * @brief Executes inference 46 * Calls inference runner provided during instance construction. 48 * @param[in] processed - input inference data. Data type should be aligned with input tensor. 49 * @param[out] result - raw floating point inference results. 51 virtual void Inference(const cv::Mat& processed, common::InferenceResults<float>& result); 54 * @brief Standard inference results post-processing implementation. 56 * Decodes inference results using decoder provided during construction. [all …]
|
/external/armnn/samples/KeywordSpotting/include/ |
D | KeywordSpottingPipeline.hpp | 16 …eric Keyword Spotting pipeline with 3 steps: data pre-processing, inference execution and inference 26 * @param executor - unique pointer to inference runner 27 * @param decoder - unique pointer to inference results decoder 36 * Preprocesses and prepares the data for inference by 45 * @brief Executes inference 47 * Calls inference runner provided during instance construction. 49 …* @param[in] preprocessedData - input inference data. Data type should be aligned with input tenso… 50 * @param[out] result - raw inference results. 52 …void Inference(const std::vector<int8_t>& preprocessedData, common::InferenceResults<int8_t>& resu… 55 * @brief Standard inference results post-processing implementation. [all …]
|
/external/pytorch/docs/cpp/source/notes/ |
D | inference_mode.rst | 1 Inference Mode 12 all newly allocated (non-view) tensors are marked as inference tensors. Inference tensors: 21 A non-view tensor is an inference tensor if and only if it was allocated inside ``InferenceMode``. 22 A view tensor is an inference tensor if and only if the tensor it is a view of is an inference tens… 27 This applies to both inference tensors and normal tensors. 28 - View operations on inference tensors do not do view tracking. View and non-view inference tensors… 30 - Inplace operations on inference tensors are guaranteed not to do a version bump. 37 In production use of PyTorch for inference workload, we have seen a proliferation 40 current colloquial of this guard for inference workload is unsafe: it's possible to 49 1. Users trying to run workload in inference only mode (like loading a pretrained JIT model and [all …]
|
/external/armnn/samples/SpeechRecognition/include/ |
D | SpeechRecognitionPipeline.hpp | 16 …ic Speech Recognition pipeline with 3 steps: data pre-processing, inference execution and inference 26 * @param executor - unique pointer to inference runner 27 * @param decoder - unique pointer to inference results decoder 35 * Preprocesses and prepares the data for inference by 51 * @brief Executes inference 53 * Calls inference runner provided during instance construction. 55 …* @param[in] preprocessedData - input inference data. Data type should be aligned with input tenso… 56 * @param[out] result - raw inference results. 59 … void Inference(const std::vector<T>& preprocessedData, common::InferenceResults<int8_t>& result) in Inference() function in asr::ASRPipeline 66 * @brief Standard inference results post-processing implementation. [all …]
|
/external/aws-sdk-java-v2/services/sagemakerruntime/src/main/resources/codegen-resources/ |
D | service-2.json | 47 …Inference requests sent to this API are enqueued for asynchronous processing. The processing of th… 65 …inference response as a stream. The inference stream provides the response payload incrementally a… 172 …"documentation":"<p>The desired MIME type of the inference response from the model container.</p>", 178 …"documentation":"<p>Provides additional information about a request for an inference submitted to … 184 …"documentation":"<p>The identifier for the inference request. Amazon SageMaker will generate an id… 186 "locationName":"X-Amzn-SageMaker-Inference-Id" 190 "documentation":"<p>The Amazon S3 URI where the inference request payload is stored.</p>", 213 …"documentation":"<p>Identifier for an inference request. This will be the same as the <code>Infere… 217 … "documentation":"<p>The Amazon S3 URI where the inference response payload is stored.</p>", 223 …"documentation":"<p>The Amazon S3 URI where the inference failure response payload is stored.</p>", [all …]
|
/external/tensorflow/tensorflow/lite/g3doc/guide/ |
D | inference.md | 1 # TensorFlow Lite inference 3 The term *inference* refers to the process of executing a TensorFlow Lite model 5 inference with a TensorFlow Lite model, you must run it through an 11 an inference using C++, Java, and Python, plus links to other resources for each 18 TensorFlow Lite inference typically follows the following steps: 31 1. **Running inference** 39 When you receive results from the model inference, you must interpret the 48 TensorFlow inference APIs are provided for most common mobile/embedded platforms 53 use. TensorFlow Lite is designed for fast inference on small devices, so it 59 inputs, and retrieve inference outputs. [all …]
|
/external/armnn/python/pyarmnn/examples/speech_recognition/ |
D | audio_utils.py | 10 """Decodes the integer encoded results from inference into a string. 13 model_output: Results from running inference. 27 results: List of top 1 results from inference. 57 …es the current right context, and updates it for each inference. Will get used after last inferenc… 60 is_first_window: Boolean to show if it is the first window we are running inference on 62 output_result: the output from the inference 74 # Since it's the first inference, keep the left context, and inner context, and decode 80 # Store the right context, we will need it after the last inference
|
/external/armnn/ |
D | InstallationViaAptRepository.md | 153 libarmnn-cpuref-backend23 - Arm NN is an inference engine for CPUs, GPUs and NPUs 154 libarmnn-cpuref-backend24 - Arm NN is an inference engine for CPUs, GPUs and NPUs 155 libarmnn-dev - Arm NN is an inference engine for CPUs, GPUs and NPUs 156 …libarmnntfliteparser-dev - Arm NN is an inference engine for CPUs, GPUs and NPUs # Note: removal o… 157 libarmnn-tfliteparser23 - Arm NN is an inference engine for CPUs, GPUs and NPUs 158 …libarmnntfliteparser24 - Arm NN is an inference engine for CPUs, GPUs and NPUs # Note: removal of … 159 …libarmnntfliteparser24.5 - Arm NN is an inference engine for CPUs, GPUs and NPUs # Note: removal o… 160 libarmnn23 - Arm NN is an inference engine for CPUs, GPUs and NPUs 161 libarmnn24 - Arm NN is an inference engine for CPUs, GPUs and NPUs 162 libarmnn25 - Arm NN is an inference engine for CPUs, GPUs and NPUs [all …]
|
/external/tensorflow/tensorflow/core/framework/ |
D | full_type_inference_util.h | 33 // inference functions. 40 // same can be said about the shape inference function. 42 // Note: Unlike type constructors, which describe op definitions, type inference 46 // Helper for a no-op type inference function that indicates type inference 48 // This is the same as not defining a type inference function at all, but 52 // Helper for a type inference function which has the same type as the i'th 59 // Helper for a type inference function which has the same type as a variadic 74 // Helper for the type inference counterpart of Unary, that is (U -> 116 // Auxiliary constructs to help creation of type inference functions. 117 // TODO(mdan): define these as type inference functions as well. [all …]
|
D | op_def_builder.h | 17 // inference function for Op registration. 42 // A type inference function, called for each node during type inference 100 // Forward type inference function. This callable infers the return type of an 103 // Note that the type constructor and forward inference functions need not be 107 // forward inference function. 117 // These type inference functions are intermediate solutions as well: once the 119 // a solver-based type inference, it will replace these functions. 121 // TODO(mdan): Merge with shape inference. 122 // TODO(mdan): Replace with a union-based type inference algorithm. 125 // Reverse type inference function. This callable infers some input types [all …]
|
/external/aws-sdk-java-v2/services/bedrockruntime/src/main/resources/codegen-resources/ |
D | service-2.json | 35 …inference using the input provided in the request body. You use InvokeModel to run inference for t… 58 …inference using the input provided. Return the response in a stream.</p> <p>For more information, … 105 …://docs.aws.amazon.com/bedrock/latest/userguide/model-parameters.html\">Inference parameters</a>.<… 115 …"documentation":"<p>The desired MIME type of the inference body in the response. The default value… 137 …Inference response from the model in the format specified in the content-type header field. To see… 141 "documentation":"<p>The MIME type of the inference result.</p>", 157 …Inference input in the format specified by the content-type. To see the format and content of this… 167 …"documentation":"<p>The desired MIME type of the inference body in the response. The default value… 189 …Inference response from the model in the format specified by Content-Type. To see the format and c… 193 "documentation":"<p>The MIME type of the inference result.</p>", [all …]
|
/external/aws-sdk-java-v2/services/lookoutequipment/src/main/resources/codegen-resources/ |
D | service-2.json | 51 …mentation":"<p> Creates a scheduled inference. Scheduling an inference is setting up a continuous … 107 …"documentation":"<p>Creates a machine learning model for data inference. </p> <p>A machine-learnin… 142 …ataset and associated artifacts. The operation will check to see if any inference scheduler or dat… 159 …"documentation":"<p>Deletes an inference scheduler that has been set up. Prior inference results w… 210 …zon Lookout for Equipment. This will prevent it from being used with an inference scheduler, even … 295 …"documentation":"<p> Specifies information about the inference scheduler being used, including nam… 484 …"documentation":"<p> Lists all inference events that have been found for the specified inference s… 501 …"documentation":"<p> Lists all inference executions that have been performed by the specified infe… 517 …"documentation":"<p>Retrieves a list of all inference schedulers currently available for your acco… 688 "documentation":"<p>Starts an inference scheduler. </p>" [all …]
|
/external/armnn/samples/SpeechRecognition/ |
D | Readme.md | 16 sample rate. Top level inference API is provided by Arm NN library. 104 3. Executing Inference 106 5. Decoding and Processing Inference Output 163 Using the `Optimize()` function we optimize the graph for inference and load the optimized network … 195 …on pipeline has 3 steps to perform, data pre-processing, run inference and decode inference results 202 …ositioned window of data, sized appropriately for the given model, to pre-process before inference. 212 After all the MFCCs needed for an inference have been extracted from the audio data, we convolve th… 215 #### Executing Inference 219 asrPipeline->Inference<int8_t>(preprocessedData, results); 221 Inference step will call `ArmnnNetworkExecutor::Run` method that will prepare input tensors and exe… [all …]
|
/external/pytorch/docs/source/ |
D | torch.compiler_aot_inductor.rst | 16 which are frequently employed for inference deployments on the server side. 40 Otherwise, the compiled artifact will run on CPU. For better performance during CPU inference, 84 Inference in C++ 87 Next, we use the following C++ file ``inference.cpp`` to load the shared library generated in the 113 std::cout << "Result from the first inference:"<< std::endl; 116 // The second inference uses a different batch size and it works because we 118 std::cout << "Result from the second inference:"<< std::endl; 127 ``inference.cpp`` into an executable binary named ``aoti_example``. 136 add_executable(aoti_example inference.cpp model.so) 157 inference.cpp [all …]
|
/external/armnn/samples/KeywordSpotting/ |
D | Readme.md | 17 …data from file, and to re-sample to the expected sample rate. Top level inference API is provided … 123 3. Executing Inference 125 5. Decoding and Processing Inference Output 188 Using the `Optimize()` function we optimize the graph for inference and load the optimized network … 224 …ng pipeline has 3 steps to perform: data pre-processing, run inference and decode inference result… 231 …ositioned window of data, sized appropriately for the given model, to pre-process before inference. 241 After all the MFCCs needed for an inference have been extracted from the audio data they are concat… 243 #### Executing Inference 248 kwsPipeline->Inference(preprocessedData, results); 251 Inference step will call `ArmnnNetworkExecutor::Run` method that will prepare input tensors and exe… [all …]
|
/external/pytorch/test/cpp/api/ |
D | inference_mode.cpp | 25 - Autograd=false, ADInplaceOrView=false (inference tensor) 26 Tensors created in InferenceMode are mostly inference tensors. The only 64 // New tensor created through constructors are inference tensors. in TEST() 69 // requires_grad doesn't change inference tensor behavior inside in TEST() 165 "Inplace update to inference tensor outside InferenceMode is not allowed"); in TEST() 308 "A view was created in inference mode and is being modified inplace") in TEST() 328 // add(Tensor, Tensor) is safe with inference tensor since it doesn't save in TEST() 335 // leaf inference tensor with requires_grad=true can still have gradient. in TEST() 344 c.mul(s), "Inference tensors cannot be saved for backward."); in TEST() 346 // Inference tensor in TensorList input in TEST() [all …]
|
/external/tensorflow/tensorflow/lite/delegates/xnnpack/ |
D | README.md | 3 XNNPACK is a highly optimized library of neural network inference operators for 6 library as an inference engine for TensorFlow Lite. 12 for floating-point inference. 85 inference by default.** 129 // Run inference using XNNPACK 178 // and inference. 197 The weights cache has to be finalized before any inference, it will be an error 421 XNNPACK supports half-precision (using IEEE FP16 format) inference for a subset 423 inference when the following conditions are met: 431 * IEEE FP16 inference is supported for every floating-point operator in the [all …]
|
/external/armnn/python/pyarmnn/src/pyarmnn/_utilities/ |
D | profiling_helper.py | 10 ProfilerData.__doc__ = """Container to hold the profiling inference data, and the profiling data pe… 13 inference_data (dict): holds end-to-end inference performance data. Keys: 15 … 'execution_time' - list of total inference execution times for each inference run. 19 … 'execution_time' - list of total execution times for each inference run. 56 …#Get the inference measurements dict, this will be just one value for key starting with "inference… 72 # This is the total inference time map
|
/external/pytorch/torch/onnx/_internal/exporter/ |
D | _onnx_program.py | 34 """Initialize an ONNX Runtime inference session with the specified model.""" 106 logger.debug("Running the inference session with %s arguments.", len(ort_input)) 108 logger.debug("Inference session run completed.") 137 during inference. 145 inference and want to minimize the size of the saved model. 208 """Initialize the ONNX Runtime inference session. 211 initializer: The function to initialize the ONNX Runtime inference 215 # TODO(justinchuby): Allow different inference options 216 logger.debug("Initializing the inference session.") 230 logger.debug("Inference session initialized.") [all …]
|
/external/tensorflow/tensorflow/lite/g3doc/inference_with_metadata/task_library/ |
D | bert_nl_classifier.md | 29 ## Run inference in Java 60 ### Step 2: Run inference using the API 71 // Run inference 79 ## Run inference in Swift 92 ### Step 2: Run inference using the API 99 // Run inference 107 ## Run inference in C++ 115 // Run inference with your input, `input_text`.
|
/external/javaparser/javaparser-symbol-solver-core/src/main/java/com/github/javaparser/symbolsolver/resolution/typeinference/ |
D | TypeInference.java | 65 // throw new IllegalArgumentException("Type inference unnecessary as type arguments have… in instantiationInference() 71 …// - Where P1, ..., Pp (p ≥ 1) are the type parameters of m, let α1, ..., αp be inference variable… in instantiationInference() 134 // inference variables in B2 succeeds (§18.4). in instantiationInference() 184 …:=α1, ..., Pp:=αp] defined in §18.5.1 to replace the type parameters of m with inference variables. in invocationTypeInferenceBoundsSetB3() 186 … in §18.5.1. (While it was necessary in §18.5.1 to demonstrate that the inference variables in B2 … in invocationTypeInferenceBoundsSetB3() 197 …// for fresh inference variables β1, ..., βn, the constraint formula ‹G<β1, ..., βn> → T› is r… in invocationTypeInferenceBoundsSetB3() 200 // - Otherwise, if R θ is an inference variable α, and one of the following is true: in invocationTypeInferenceBoundsSetB3() 256 …inference variable α can influence an inference variable β if α depends on the resolution of β (§1… in invocationTypeInference() 272 … //Finally, if B4 does not contain the bound false, the inference variables in B4 are resolved. in invocationTypeInference() 274 …//If resolution succeeds with instantiations T1, ..., Tp for inference variables α1, ..., αp, let … in invocationTypeInference() [all …]
|