Home
last modified time | relevance | path

Searched refs:optimization (Results 1 – 25 of 1482) sorted by relevance

12345678910>>...60

/external/mesa3d/src/compiler/glsl/
Dtest_optpass.cpp57 do_optimization(struct exec_list *ir, const char *optimization, in do_optimization() argument
66 if (sscanf(optimization, "do_common_optimization ( %d ) ", &int_0) == 1) { in do_optimization()
68 } else if (strcmp(optimization, "do_algebraic") == 0) { in do_optimization()
70 } else if (strcmp(optimization, "do_constant_folding") == 0) { in do_optimization()
72 } else if (strcmp(optimization, "do_constant_variable") == 0) { in do_optimization()
74 } else if (strcmp(optimization, "do_constant_variable_unlinked") == 0) { in do_optimization()
76 } else if (strcmp(optimization, "do_copy_propagation_elements") == 0) { in do_optimization()
78 } else if (strcmp(optimization, "do_constant_propagation") == 0) { in do_optimization()
80 } else if (strcmp(optimization, "do_dead_code") == 0) { in do_optimization()
82 } else if (strcmp(optimization, "do_dead_code_local") == 0) { in do_optimization()
[all …]
/external/apache-commons-math/src/main/java/org/apache/commons/math/optimization/general/
DAbstractScalarDifferentiableOptimizer.java18 package org.apache.commons.math.optimization.general;
25 import org.apache.commons.math.optimization.DifferentiableMultivariateRealOptimizer;
26 import org.apache.commons.math.optimization.GoalType;
27 import org.apache.commons.math.optimization.OptimizationException;
28 import org.apache.commons.math.optimization.RealConvergenceChecker;
29 import org.apache.commons.math.optimization.RealPointValuePair;
30 import org.apache.commons.math.optimization.SimpleScalarValueChecker;
/external/tensorflow/tensorflow/core/common_runtime/
Doptimization_registry.h137 #define REGISTER_OPTIMIZATION(grouping, phase, optimization) \ argument
138 REGISTER_OPTIMIZATION_UNIQ_HELPER(__COUNTER__, grouping, phase, optimization)
140 #define REGISTER_OPTIMIZATION_UNIQ_HELPER(ctr, grouping, phase, optimization) \ argument
141 REGISTER_OPTIMIZATION_UNIQ(ctr, grouping, phase, optimization)
143 #define REGISTER_OPTIMIZATION_UNIQ(ctr, grouping, phase, optimization) \ argument
148 new optimization()), \
149 #optimization)
/external/apache-commons-math/src/main/java/org/apache/commons/math/optimization/fitting/
DGaussianFitter.java18 package org.apache.commons.math.optimization.fitting;
21 import org.apache.commons.math.optimization.DifferentiableMultivariateVectorialOptimizer;
22 import org.apache.commons.math.optimization.OptimizationException;
23 import org.apache.commons.math.optimization.fitting.CurveFitter;
24 import org.apache.commons.math.optimization.fitting.WeightedObservedPoint;
/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/optimization/
Dchoose_fastest_branch_dataset_test.py22 from tensorflow.python.data.experimental.ops import optimization
44 choose_fastest = optimization._ChooseFastestBranchDataset(
65 choose_fastest = optimization._ChooseFastestBranchDataset(
81 choose_fastest = optimization._ChooseFastestBranchDataset(
98 choose_fastest = optimization._ChooseFastestBranchDataset(
116 choose_fastest = optimization._ChooseFastestBranchDataset(
129 choose_fastest = optimization._ChooseFastestBranchDataset(
145 return optimization._ChooseFastestBranchDataset(
169 choose_fastest = optimization._ChooseFastestBranchDataset(
Dchoose_fastest_dataset_test.py22 from tensorflow.python.data.experimental.ops import optimization
37 merge = optimization._ChooseFastestDataset([dataset, dataset])
46 merge = optimization._ChooseFastestDataset([dataset for _ in range(5)])
58 merge = optimization._ChooseFastestDataset([dataset_a, dataset_b])
93 merge = optimization._ChooseFastestDataset([dataset_a, dataset_b])
95 merge = optimization._ChooseFastestDataset([dataset_a, dataset_b])
/external/tensorflow/tensorflow/python/data/experimental/ops/
Doptimization_options.py299 for optimization in optimizations_to_disable:
300 if getattr(self, optimization) is None:
301 result.default.append(optimization)
305 for optimization in all_optimizations:
306 if getattr(self, optimization) is True: # pylint: disable=g-bool-id-comparison
307 result.enabled.append(optimization)
308 elif getattr(self, optimization) is False: # pylint: disable=g-bool-id-comparison
309 result.disabled.append(optimization)
340 for optimization in autotune_only_optimizations:
341 graph_rewrite_configs.append(optimization + ":autotune:false")
[all …]
/external/apache-commons-math/src/main/java/org/apache/commons/math/optimization/linear/
DLinearOptimizer.java18 package org.apache.commons.math.optimization.linear;
22 import org.apache.commons.math.optimization.GoalType;
23 import org.apache.commons.math.optimization.OptimizationException;
24 import org.apache.commons.math.optimization.RealPointValuePair;
DAbstractLinearOptimizer.java18 package org.apache.commons.math.optimization.linear;
23 import org.apache.commons.math.optimization.GoalType;
24 import org.apache.commons.math.optimization.OptimizationException;
25 import org.apache.commons.math.optimization.RealPointValuePair;
/external/llvm-project/clang/test/CodeGenObjC/
Ddisable-tail-call-escaping-block.m43 BlockTy b0 = ^{ (void)a; }; // disable tail-call optimization.
45 callee0(^{ (void)a; }); // enable tail-call optimization.
46 callee1(^{ (void)a; }); // disable tail-call optimization.
48 BlockTy b1 = ^{ (void)a; }; // disable tail-call optimization.
50 [c0 m0:^{ (void)a; }]; // enable tail-call optimization.
51 [c0 m1:^{ (void)a; }]; // disable tail-call optimization.
53 noescapefunc(^{ (void)a; }); // enable tail-call optimization.
/external/tensorflow/tensorflow/core/api_def/base_api/
Dapi_def_RetrieveTPUEmbeddingAdadeltaParametersGradAccumDebug.pbtxt7 Parameter parameters updated by the Adadelta optimization algorithm.
13 Parameter accumulators updated by the Adadelta optimization algorithm.
19 Parameter updates updated by the Adadelta optimization algorithm.
25 Parameter gradient_accumulators updated by the Adadelta optimization algorithm.
30 An op that retrieves optimization parameters from embedding to host
Dapi_def_RetrieveTPUEmbeddingRMSPropParametersGradAccumDebug.pbtxt7 Parameter parameters updated by the RMSProp optimization algorithm.
13 Parameter ms updated by the RMSProp optimization algorithm.
19 Parameter mom updated by the RMSProp optimization algorithm.
25 Parameter gradient_accumulators updated by the RMSProp optimization algorithm.
30 An op that retrieves optimization parameters from embedding to host
Dapi_def_RetrieveTPUEmbeddingMDLAdagradLightParameters.pbtxt7 Parameter parameters updated by the MDL Adagrad Light optimization algorithm.
13 Parameter accumulators updated by the MDL Adagrad Light optimization algorithm.
19 Parameter weights updated by the MDL Adagrad Light optimization algorithm.
25 Parameter benefits updated by the MDL Adagrad Light optimization algorithm.
30 An op that retrieves optimization parameters from embedding to host
Dapi_def_RetrieveTPUEmbeddingFTRLParametersGradAccumDebug.pbtxt7 Parameter parameters updated by the FTRL optimization algorithm.
13 Parameter accumulators updated by the FTRL optimization algorithm.
19 Parameter linears updated by the FTRL optimization algorithm.
25 Parameter gradient_accumulators updated by the FTRL optimization algorithm.
30 An op that retrieves optimization parameters from embedding to host
Dapi_def_RetrieveTPUEmbeddingCenteredRMSPropParameters.pbtxt7 Parameter parameters updated by the centered RMSProp optimization algorithm.
13 Parameter ms updated by the centered RMSProp optimization algorithm.
19 Parameter mom updated by the centered RMSProp optimization algorithm.
25 Parameter mg updated by the centered RMSProp optimization algorithm.
30 An op that retrieves optimization parameters from embedding to host
Dapi_def_RetrieveTPUEmbeddingADAMParametersGradAccumDebug.pbtxt7 Parameter parameters updated by the ADAM optimization algorithm.
13 Parameter momenta updated by the ADAM optimization algorithm.
19 Parameter velocities updated by the ADAM optimization algorithm.
25 Parameter gradient_accumulators updated by the ADAM optimization algorithm.
30 An op that retrieves optimization parameters from embedding to host
Dapi_def_LoadTPUEmbeddingRMSPropParametersGradAccumDebug.pbtxt7 Value of parameters used in the RMSProp optimization algorithm.
13 Value of ms used in the RMSProp optimization algorithm.
19 Value of mom used in the RMSProp optimization algorithm.
25 Value of gradient_accumulators used in the RMSProp optimization algorithm.
30 An op that loads optimization parameters into HBM for embedding. Must be
Dapi_def_LoadTPUEmbeddingAdadeltaParametersGradAccumDebug.pbtxt7 Value of parameters used in the Adadelta optimization algorithm.
13 Value of accumulators used in the Adadelta optimization algorithm.
19 Value of updates used in the Adadelta optimization algorithm.
25 Value of gradient_accumulators used in the Adadelta optimization algorithm.
30 An op that loads optimization parameters into HBM for embedding. Must be
Dapi_def_LoadTPUEmbeddingCenteredRMSPropParameters.pbtxt7 Value of parameters used in the centered RMSProp optimization algorithm.
13 Value of ms used in the centered RMSProp optimization algorithm.
19 Value of mom used in the centered RMSProp optimization algorithm.
25 Value of mg used in the centered RMSProp optimization algorithm.
30 An op that loads optimization parameters into HBM for embedding. Must be
Dapi_def_LoadTPUEmbeddingADAMParametersGradAccumDebug.pbtxt7 Value of parameters used in the ADAM optimization algorithm.
13 Value of momenta used in the ADAM optimization algorithm.
19 Value of velocities used in the ADAM optimization algorithm.
25 Value of gradient_accumulators used in the ADAM optimization algorithm.
30 An op that loads optimization parameters into HBM for embedding. Must be
Dapi_def_LoadTPUEmbeddingFTRLParametersGradAccumDebug.pbtxt7 Value of parameters used in the FTRL optimization algorithm.
13 Value of accumulators used in the FTRL optimization algorithm.
19 Value of linears used in the FTRL optimization algorithm.
25 Value of gradient_accumulators used in the FTRL optimization algorithm.
30 An op that loads optimization parameters into HBM for embedding. Must be
Dapi_def_LoadTPUEmbeddingMDLAdagradLightParameters.pbtxt7 Value of parameters used in the MDL Adagrad Light optimization algorithm.
13 Value of accumulators used in the MDL Adagrad Light optimization algorithm.
19 Value of weights used in the MDL Adagrad Light optimization algorithm.
25 Value of benefits used in the MDL Adagrad Light optimization algorithm.
30 An op that loads optimization parameters into HBM for embedding. Must be
Dapi_def_RetrieveTPUEmbeddingADAMParameters.pbtxt7 Parameter parameters updated by the ADAM optimization algorithm.
13 Parameter momenta updated by the ADAM optimization algorithm.
19 Parameter velocities updated by the ADAM optimization algorithm.
24 An op that retrieves optimization parameters from embedding to host
Dapi_def_RetrieveTPUEmbeddingRMSPropParameters.pbtxt7 Parameter parameters updated by the RMSProp optimization algorithm.
13 Parameter ms updated by the RMSProp optimization algorithm.
19 Parameter mom updated by the RMSProp optimization algorithm.
24 An op that retrieves optimization parameters from embedding to host
/external/apache-commons-math/src/main/java/org/apache/commons/math/optimization/direct/
DPowellOptimizer.java18 package org.apache.commons.math.optimization.direct;
23 import org.apache.commons.math.optimization.GoalType;
24 import org.apache.commons.math.optimization.OptimizationException;
25 import org.apache.commons.math.optimization.RealPointValuePair;
26 import org.apache.commons.math.optimization.general.AbstractScalarDifferentiableOptimizer;
27 import org.apache.commons.math.optimization.univariate.AbstractUnivariateRealOptimizer;
28 import org.apache.commons.math.optimization.univariate.BracketFinder;
29 import org.apache.commons.math.optimization.univariate.BrentOptimizer;

12345678910>>...60