Searched refs:grad_grad (Results 1 – 6 of 6) sorted by relevance
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | softplus_op_test.py | 118 (grad_grad,) = gradients_impl.gradients(grad, x) 124 x, [2, 5], grad_grad, [2, 5], x_init_value=x_init)
|
D | while_v2_test.py | 174 grad_grad = gradients_impl.gradients(grad, [x]) # 12x**2 177 self.assertSequenceEqual(self.evaluate(grad_grad), [48.]) 186 grad_grad = gradients_impl.gradients(grad, [x]) # 12x**2 190 self.assertSequenceEqual(self.evaluate(grad_grad), [48.])
|
D | control_flow_ops_py_test.py | 986 grad_grad = gradients_impl.gradients(grad, x)[0] 993 self.assertIsNone(grad_grad) 996 self.assertEqual(sess.run(grad_grad, {pred: True, x: 1.0, y: 2.0}), 0.0) 997 self.assertEqual(sess.run(grad_grad, {pred: False, x: 1.0, y: 2.0}), 0.0) 2255 grad_grad = gradients_impl.gradients(grad, x_init)[0] 2263 self.assertEqual(sess.run(grad_grad, {pred: True}), 0.0) 2264 self.assertEqual(sess.run(grad_grad, {pred: False}), 0.0)
|
/external/tensorflow/tensorflow/cc/gradients/ |
D | nn_grad.cc | 91 auto grad_grad = grad_inputs[1]; in SoftmaxCrossEntropyWithLogitsGrad() local 94 if (!IsZero(scope, grad_grad)) { in SoftmaxCrossEntropyWithLogitsGrad() 98 auto grad_grad_expand = ExpandDims(scope, grad_grad, 1); in SoftmaxCrossEntropyWithLogitsGrad() 104 auto subtraction_result = Subtract(scope, grad_grad, squeeze_result); in SoftmaxCrossEntropyWithLogitsGrad()
|
/external/tensorflow/tensorflow/python/ops/ |
D | nn_grad.py | 512 def _SoftmaxCrossEntropyWithLogitsGrad(op, grad_loss, grad_grad): argument 533 if grad_grad is not None and not IsZero(grad_grad): 536 grad += ((grad_grad - array_ops.squeeze( 538 array_ops.expand_dims(grad_grad, 1),
|
/external/tensorflow/tensorflow/python/ops/parallel_for/ |
D | control_flow_ops_test.py | 314 grad_grad = g.gradient(grad, ones) 315 return output, grad, grad_grad 335 grad_grad = g.gradient(grad, ones) 336 return output, grad, grad_grad
|