/external/tensorflow/tensorflow/python/keras/wrappers/ |
D | scikit_learn_test.py | 39 model.add(keras.layers.Activation('relu')) 41 model.add(keras.layers.Activation('relu')) 43 model.add(keras.layers.Activation('softmax')) 75 model.add(keras.layers.Activation('relu')) 77 model.add(keras.layers.Activation('relu')) 79 model.add(keras.layers.Activation('linear'))
|
/external/tensorflow/tensorflow/tools/api/golden/v2/ |
D | tensorflow.keras.layers.-activation.pbtxt | 1 path: "tensorflow.keras.layers.Activation" 3 is_instance: "<class \'tensorflow.python.keras.layers.core.Activation\'>"
|
D | tensorflow.keras.layers.pbtxt | 8 name: "Activation"
|
/external/tensorflow/tensorflow/tools/api/golden/v1/ |
D | tensorflow.keras.layers.-activation.pbtxt | 1 path: "tensorflow.keras.layers.Activation" 3 is_instance: "<class \'tensorflow.python.keras.layers.core.Activation\'>"
|
D | tensorflow.keras.layers.pbtxt | 8 name: "Activation"
|
/external/tensorflow/tensorflow/core/kernels/ |
D | conv_ops_fused_impl.h | 157 template <typename T, typename Activation = Identity> 175 output = Activation::template apply<decltype(expr)>(expr); in operator() 185 template <typename T, typename Activation = Identity> 217 output = Activation::template apply<decltype(shifted)>(shifted); in operator()
|
/external/tensorflow/tensorflow/python/keras/layers/ |
D | core_test.py | 248 keras.layers.Activation, 254 keras.layers.Activation,
|
D | core.py | 337 class Activation(Layer): class 354 super(Activation, self).__init__(**kwargs) 366 base_config = super(Activation, self).get_config()
|
D | wrappers_test.py | 124 model.add(keras.layers.Activation('relu')) 139 model.add(keras.layers.Activation('relu')) 155 model.add(keras.layers.Activation('relu'))
|
D | __init__.py | 72 from tensorflow.python.keras.layers.core import Activation
|
/external/tensorflow/tensorflow/contrib/fused_conv/ |
D | BUILD | 2 # A Fused Conv Bias Activation operator wrapper.
|
/external/tensorflow/tensorflow/contrib/eager/python/examples/revnet/ |
D | blocks.py | 433 self.activation = tf.keras.layers.Activation("relu") 491 self.activation = tf.keras.layers.Activation("relu")
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | summary_ops_test.py | 43 from tensorflow.python.keras.layers.core import Activation 963 Activation('relu', name='my_relu')]) 972 Activation('relu', name='my_relu')]) 989 self.activation = Activation('relu', name='my_relu')
|
/external/tensorflow/tensorflow/contrib/keras/api/keras/layers/ |
D | __init__.py | 70 from tensorflow.python.keras.layers.core import Activation
|
/external/v8/src/wasm/ |
D | wasm-interpreter.cc | 1078 struct Activation { struct in v8::internal::wasm::__anon49171a620111::ThreadImpl 1081 Activation(uint32_t fp, sp_t sp) : fp(fp), sp(sp) {} in Activation() argument 1145 Activation act = current_activation(); in GetReturnValue() 1215 Activation& act = activations_.back(); in HandleException() 1260 ZoneVector<Activation> activations_; 2919 inline Activation current_activation() { in current_activation() 2920 return activations_.empty() ? Activation(0, 0) : activations_.back(); in current_activation()
|
/external/tensorflow/tensorflow/python/keras/saving/ |
D | saved_model_test.py | 425 outputs = keras.layers.Activation(relu6)(inputs)
|
/external/python/cpython3/Doc/library/ |
D | site.rst | 141 Activation of rlcompleter and history was made automatic.
|
/external/tensorflow/tensorflow/python/keras/engine/ |
D | base_layer_test.py | 615 (keras.layers.Activation, (2, 2),
|
D | training_test.py | 1280 model.add(keras.layers.Activation('softmax')) 1368 model.add(keras.layers.Activation('softmax'))
|
/external/tensorflow/tensorflow/lite/g3doc/convert/ |
D | cmdline_examples.md | 341 * Activation arrays are gray:
|
/external/webrtc/webrtc/modules/audio_device/ios/ |
D | audio_device_ios.mm | 178 // Activate the audio session. Activation can fail if another active audio
|
/external/u-boot/doc/driver-model/ |
D | README.txt | 647 2. Activation/probe
|
/external/tensorflow/tensorflow/stream_executor/cuda/ |
D | cudnn_6_0.inc | 1222 … - Per-Activation Normalization : tensors are expected to have dims of 1xCxHxW
|
D | cudnn_7_0.inc | 1334 … - Per-Activation Normalization : tensors are expected to have dims of 1xCxHxW
|
/external/freetype/ |
D | ChangeLog.27 | 647 Activation of the code follows in another commit.
|