/external/tensorflow/tensorflow/cc/gradients/ |
D | grad_testutil.cc | 26 ops::GradFunc grad_fn; in CallGradFunction() local 28 op.node()->type_string(), &grad_fn)); in CallGradFunction() 29 TF_RETURN_IF_ERROR(grad_fn(scope, op, grad_inputs, grad_outputs)); in CallGradFunction()
|
/external/tensorflow/tensorflow/python/ops/ |
D | custom_gradient.py | 330 result, grad_fn = f(*args) 379 grad_argspec = tf_inspect.getfullargspec(grad_fn) 398 input_grads, variable_grads = grad_fn(*result_grads, variables=variables) 403 input_grads = grad_fn(*result_grads) 438 result, grad_fn = f(*args, **kwargs) 448 grad_argspec = tf_inspect.getfullargspec(grad_fn) 468 input_grads, variable_grads = grad_fn(*result_grads, variables=variables) 473 input_grads = grad_fn(*result_grads)
|
D | gradients_util.py | 322 def _MaybeCompile(scope, op, func, grad_fn): argument 340 return grad_fn() # Exit early 356 return grad_fn() 598 grad_fn = None 608 grad_fn = ops.get_gradient_function(op) 634 grad_fn = func_call.python_grad_func 656 if (grad_fn or is_func_call) and has_out_grads: 662 (not grad_fn and is_func_call) 680 if grad_fn: 684 lambda: grad_fn(op, *out_grads))
|
D | gradient_checker_v2.py | 169 grad_fn = _prepare(lambda dy, *xs: grad_fn_unprep(*xs, dy=dy), 174 grad = _to_numpy(grad_fn(dy_data, *xs)[0]) 190 grad = _to_numpy(grad_fn(dy_data, *xs)[0])
|
D | gradient_checker_v2_test.py | 224 def grad_fn(dy): function 229 return y, grad_fn 246 def grad_fn(dy): function 251 return y, grad_fn
|
/external/tensorflow/tensorflow/python/eager/ |
D | backprop.py | 148 grad_fn = ops._gradient_registry.lookup(op_name) # pylint: disable=protected-access 149 if grad_fn is None: 159 return grad_fn(mock_op, *out_grads) 161 return grad_fn(mock_op, *out_grads) 228 def grad_fn(*args, **kwds): function 256 return grad_fn 301 def grad_fn(*args, **kwds): function 305 return grad_fn
|
D | function_gradients_test.py | 429 grad_fn = backprop.implicit_grad(sum_gather) 430 gradient = grad_fn()
|
D | backprop_test.py | 226 grad_fn = backprop.gradients_function(f) 227 self.assertAllEqual(2., grad_fn(1., dy=2.)[0]) 1603 def grad_fn(x): function 1606 grad_ops = grad_fn.get_concrete_function(
|
/external/tensorflow/tensorflow/cc/framework/ |
D | gradients.cc | 375 ops::GradFunc grad_fn; in IsPrimitiveOpWithNoGrad() local 376 Status s = registry_->Lookup(opname, &grad_fn); in IsPrimitiveOpWithNoGrad() 377 return s.ok() && (grad_fn == nullptr); in IsPrimitiveOpWithNoGrad() 384 ops::GradFunc grad_fn; in CallGradFunction() local 385 TF_RETURN_IF_ERROR(registry_->Lookup(op.node()->type_string(), &grad_fn)); in CallGradFunction() 386 TF_RETURN_IF_ERROR(grad_fn(scope_, op, grad_inputs, grad_outputs)); in CallGradFunction()
|
/external/tensorflow/tensorflow/python/distribute/ |
D | strategy_test_lib.py | 155 grad_fn = backprop.implicit_grad(loss) 156 grad_fn = optimizer.get_filtered_grad_fn(grad_fn) 166 g_v = d.extended.call_for_each_replica(grad_fn, args=(one,)) 213 grad_fn = backprop.implicit_grad(loss) 223 g_v = d.extended.call_for_each_replica(grad_fn, args=(one,))
|
D | collective_all_reduce_strategy_test.py | 134 def grad_fn(x): function 151 g_v = d.extended.call_for_each_replica(grad_fn, args=[one])
|
D | parameter_server_strategy_test.py | 463 def grad_fn(x): function 480 g_v = d.extended.call_for_each_replica(grad_fn, args=(one,))
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | cholesky_op_test.py | 357 def _BenchmarkGrad(grad_fn, name, device): argument 366 grad = grad_fn(l, grad_matrix)
|
D | while_v2_test.py | 117 …def grad_fn(dy, variables=None): # pylint: disable=invalid-name, unused-argument, redefined-outer… function 120 return v * v * m, grad_fn
|
D | tensor_array_ops_test.py | 1015 grad_fn = backprop.gradients_function(func) 1016 v0_grad, state0_grad, var_grad = grad_fn(v0, state0, var, dy=grad_val)
|
/external/tensorflow/tensorflow/python/training/ |
D | optimizer.py | 49 def get_filtered_grad_fn(grad_fn): argument 60 return [(g, v) for g, v in grad_fn(*args, **kwargs) if g is not None]
|
/external/tensorflow/tensorflow/compiler/tests/ |
D | eager_test.py | 163 grad_fn = backprop.gradients_function(f) 164 self.assertAllEqual(2., grad_fn(1., dy=2.)[0])
|