/third_party/mindspore/tests/ut/python/communication/ |
D | test_comm.py | 152 optimizer = Momentum(filter(lambda x: x.requires_grad, network.get_parameters()), 176 optimizer = Momentum(filter(lambda x: x.requires_grad, network.get_parameters()), 190 optimizer = Momentum(filter(lambda x: x.requires_grad, network.get_parameters()), 205 optimizer = Momentum(filter(lambda x: x.requires_grad, network.get_parameters()), 226 optimizer = Momentum(filter(lambda x: x.requires_grad, network.get_parameters()), 241 optimizer = Momentum(filter(lambda x: x.requires_grad, network.get_parameters()),
|
/third_party/mindspore/tests/ut/python/nn/optim/ |
D | test_rmsprop.py | 60 RMSProp(net.get_parameters(), momentum=-0.1, learning_rate=0.1, weight_decay=0.9) 63 RMSProp(net.get_parameters(), momentum=1, learning_rate=0.1, weight_decay=0.9)
|
D | test_adam.py | 209 AdamWeightDecay(net.get_parameters(), beta1=1.0, learning_rate=0.1) 215 AdamWeightDecay(net.get_parameters(), beta2=1.0, learning_rate=0.1) 221 AdamWeightDecay(net.get_parameters(), eps=-0.1, learning_rate=0.1) 227 AdamWeightDecay(net.get_parameters())
|
D | test_lr_schedule.py | 57 self.weights = ParameterTuple(network.get_parameters())
|
/third_party/mindspore/tests/st/pynative/ |
D | test_pynative_temporary_cell.py | 73 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9) 80 …compare_optimizer = Momentum(filter(lambda x: x.requires_grad, compare_net.get_parameters()), 0.1,…
|
D | test_pynative_lenet.py | 124 self.weights = ParameterTuple(filter(lambda x: x.requires_grad, network.get_parameters())) 147 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9) 188 weights = ParameterTuple(filter(lambda x: x.requires_grad, net.get_parameters()))
|
D | test_pynative_hook.py | 111 self.weights = ParameterTuple(filter(lambda x: x.requires_grad, network.get_parameters())) 173 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9)
|
/third_party/mindspore/tests/ut/python/nn/ |
D | test_cell_wrapper.py | 44 …optimizer = Momentum(net.get_parameters(), Tensor(np.array([0.1, 0.01, 0.001]), mstype.float32), 0… 75 optimizer = Momentum(net.get_parameters(), 0.01, 0.001)
|
/third_party/mindspore/tests/ut/python/parallel/ |
D | test_auto_parallel_resnet.py | 302 opt = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.01, 0.9) 678 opt = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.01, 0.9) 706 opt = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.01, 0.9) 732 opt = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.01, 0.9) 758 opt = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.01, 0.9)
|
D | test_loss_scale.py | 187 opt = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.01, 0.9) 200 opt = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.01, 0.9)
|
/third_party/mindspore/tests/st/pynative/ms_function/ |
D | test_pynative_ms_function.py | 83 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9) 140 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9) 195 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9)
|
D | test_pynative_lenet_ms_function.py | 157 self.weights = ParameterTuple(filter(lambda x: x.requires_grad, network.get_parameters())) 180 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9)
|
/third_party/mindspore/tests/st/nontask_sink/ |
D | test_lenet.py | 123 self.weights = ParameterTuple(filter(lambda x: x.requires_grad, network.get_parameters())) 138 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9)
|
/third_party/mindspore/tests/ut/cpp/python_input/gtest_input/pipeline/parse/ |
D | parse_compile.py | 52 opt = Momentum(net.get_parameters(), learning_rate=0.1, momentum=0.9)
|
/third_party/mindspore/tests/ut/python/pynative_mode/ |
D | test_hook.py | 134 self.weights = ParameterTuple(filter(lambda x: x.requires_grad, network.get_parameters())) 143 optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), 0.1, 0.9)
|
D | test_ms_function.py | 62 … self.optimizer = Momentum(filter(lambda x: x.requires_grad, self.net.get_parameters()), 0.1, 0.9)
|
/third_party/mindspore/tests/st/quantization/ops/ |
D | test_Conv2dBnFoldQuant.py | 41 …opt = nn.Momentum(filter(lambda x: x.requires_grad, network.get_parameters()), learning_rate=0.1, …
|
/third_party/mindspore/tests/st/ops/cpu/ |
D | test_momentum_op.py | 54 …optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), learning_rate, momen…
|
D | test_adam_op.py | 51 net.get_parameters()), learning_rate=0.01)
|
D | test_adam_weight_decay_op.py | 51 …optimizer = AdamWeightDecayOp(filter(lambda x: x.requires_grad, net.get_parameters()), learning_ra…
|
D | test_sgd_op.py | 56 …optimizer = SGD(filter(lambda x: x.requires_grad, net.get_parameters()), learning_rate, momentum, …
|
/third_party/mindspore/tests/st/networks/models/resnet50/src_thor/ |
D | thor.py | 161 params = filter(lambda x: x.requires_grad, net.get_parameters()) 171 … self.matrix_A_cov = ParameterTuple(filter(lambda x: 'matrix_a' in x.name, net.get_parameters())) 172 … self.matrix_G_cov = ParameterTuple(filter(lambda x: 'matrix_g' in x.name, net.get_parameters())) 173 …elf.A_normalizer = ParameterTuple(filter(lambda x: 'a_normalizer' in x.name, net.get_parameters())) 174 …elf.G_normalizer = ParameterTuple(filter(lambda x: 'g_normalizer' in x.name, net.get_parameters()))
|
/third_party/mindspore/tests/st/ps/part_ps/ |
D | test_ps_embedding_heterogeneous_conv2d_adam.py | 147 opt = Adam(params=filter(lambda x: x.requires_grad, net.get_parameters())) 164 opt = Adam(params=filter(lambda x: x.requires_grad, net.get_parameters()))
|
/third_party/mindspore/tests/st/ops/gpu/ |
D | test_momentum_op.py | 54 …optimizer = Momentum(filter(lambda x: x.requires_grad, net.get_parameters()), learning_rate, momen…
|
D | test_sgd_op.py | 56 …optimizer = SGD(filter(lambda x: x.requires_grad, net.get_parameters()), learning_rate, momentum, …
|