/external/tensorflow/tensorflow/python/eager/ |
D | tape.py | 40 def __init__(self, tape): argument 53 def push_tape(tape): argument 58 def watch(tape, tensor): argument 95 def watch_variable(tape, variable): argument 149 def pop_tape(tape): argument
|
D | imperative_grad.py | 33 def imperative_grad(tape, argument
|
D | pywrap_tfe_src.cc | 1572 GradientTape* tape; member 1579 static void TFE_Py_Tape_Delete(PyObject* tape) { in TFE_Py_Tape_Delete() 1869 TFE_Py_Tape* tape = PyObject_NEW(TFE_Py_Tape, &TFE_Py_Tape_Type); in TFE_Py_TapeSetNew() local 1878 void TFE_Py_TapeSetAdd(PyObject* tape) { in TFE_Py_TapeSetAdd() 1896 void TFE_Py_TapeSetRemove(PyObject* tape) { in TFE_Py_TapeSetRemove() 2060 void TFE_Py_TapeWatch(PyObject* tape, PyObject* tensor) { 2206 void TFE_Py_TapeWatchVariable(PyObject* tape, PyObject* variable) { 2213 PyObject* TFE_Py_TapeWatchedVariables(PyObject* tape) { 2718 PyObject* TFE_Py_TapeGradient(PyObject* tape, PyObject* target,
|
/external/llvm-project/clang/test/SemaCXX/ |
D | constexpr-turing-cxx2a.cpp | 8 bool tape; member 37 Tape *tape = new Tape; in run() local
|
D | constexpr-turing.cpp | 10 bool tape; member 35 constexpr unsigned run(const State *tm, const Tape &tape, unsigned state) { in run()
|
/external/tensorflow/tensorflow/c/eager/ |
D | mnist_gradients_testutil.cc | 49 auto tape = new Tape(/*persistent=*/false); in AddGradModel() local 73 auto tape = new Tape(/*persistent=*/false); in MatMulGradModel() local 171 auto tape = new Tape(/*persistent=*/true); in MNISTGradModel() local
|
D | gradients_test.cc | 78 auto tape = std::make_unique<Tape>(/*persistent=*/false); in IdentityNGradModel() local 196 auto tape = std::make_unique<Tape>(/*persistent=*/false); in TEST_P() local 210 Tape tape(/*persistent=*/false); in RecordOperationWithNullGradientFunctionModel() local
|
D | gradients.cc | 468 ForwardOperation* forward_op_, Tape* tape, in Execute()
|
/external/clang/test/SemaCXX/ |
D | constexpr-turing.cpp | 10 bool tape; member 35 constexpr unsigned run(const State *tm, const Tape &tape, unsigned state) { in run()
|
/external/tensorflow/tensorflow/c/experimental/gradients/ |
D | nn_grad_test.cc | 44 Tape tape(/*persistent=*/false); in ReluGradModel() local 84 Tape tape(/*persistent=*/false); in SparseSoftmaxCrossEntropyWithLogitsGradModel() local 114 Tape tape(/*persistent=*/false); in BiasAddGradModel() local
|
D | custom_gradient_test.cc | 71 Tape tape(/*persistent=*/false); in ExpWithPassThroughGrad() local
|
D | grad_test_helper.cc | 115 Tape tape(/*persistent=*/false); in BuildGradModel() local
|
/external/tensorflow/tensorflow/c/experimental/gradients/tape/ |
D | tape_context.cc | 22 TapeContext::TapeContext(AbstractContext* c, Tape* tape, in TapeContext()
|
D | tape_operation.cc | 22 TapeOperation::TapeOperation(AbstractOperation* parent_op, Tape* tape, in TapeOperation()
|
/external/tensorflow/tensorflow/python/framework/experimental/ |
D | tape.cc | 70 [](AbstractContext* ctx, Tape* tape, GradientRegistry* registry) { in PYBIND11_MODULE()
|
/external/tensorflow/tensorflow/python/keras/optimizer_v2/ |
D | optimizer_v2.py | 474 def _get_gradients(self, tape, loss, var_list, grad_loss=None): argument 500 def minimize(self, loss, var_list, grad_loss=None, name=None, tape=None): argument 535 def _compute_gradients(self, loss, var_list, grad_loss=None, tape=None): argument
|
/external/tensorflow/tensorflow/python/ |
D | tfe_wrapper.cc | 886 [](const py::handle& tape) { TFE_Py_TapeSetAdd(tape.ptr()); }); in PYBIND11_MODULE() 888 [](const py::handle& tape) { TFE_Py_TapeSetRemove(tape.ptr()); }); in PYBIND11_MODULE() 947 [](const py::handle& tape, const py::handle& tensor) { in PYBIND11_MODULE() 951 [](const py::handle& tape, const py::handle& variable) { in PYBIND11_MODULE() 954 m.def("TFE_Py_TapeWatchedVariables", [](const py::handle& tape) { in PYBIND11_MODULE()
|
/external/tensorflow/tensorflow/python/keras/ |
D | optimizer_v1.py | 766 def minimize(self, loss, var_list, grad_loss=None, tape=None): argument
|
/external/tensorflow/tensorflow/python/keras/mixed_precision/ |
D | loss_scale_optimizer.py | 677 def _compute_gradients(self, loss, var_list, grad_loss=None, tape=None): argument
|