• Home
  • Raw
  • Download

Lines Matching full:optimizer

33   OptimizerClass optimizer(model->parameters(), options);  in test_optimizer_xor()  local
47 auto step = [&](OptimizerClass& optimizer, in test_optimizer_xor()
52 optimizer.zero_grad(); in test_optimizer_xor()
58 return optimizer.step(closure); in test_optimizer_xor()
61 torch::Tensor loss = step(optimizer, model, inputs, labels); in test_optimizer_xor()
122 auto optimizer = OptimizerClass(parameters.values(), options); in check_exact_values() local
128 optimizer.zero_grad(); in check_exact_values()
134 optimizer.step(closure); in check_exact_values()
164 auto optimizer = Adagrad(params, options); in TEST() local
166 auto& options_ = static_cast<AdagradOptions&>(optimizer.defaults()); in TEST()
169 auto& params_groups = optimizer.param_groups(); in TEST()
180 optimizer.add_param_group(OptimizerParamGroup(params)), in TEST()
185 *(optimizer.state()[params_1[0].unsafeGetTensorImpl()])); in TEST()
213 struct MyOptimizer : Optimizer { in TEST()
214 using Optimizer::Optimizer; in TEST()
222 Optimizer( in TEST()
229 MyOptimizer optimizer(parameters); in TEST() local
232 OLD_INTERFACE_WARNING_CHECK(size = optimizer.size()); in TEST()
237 MyOptimizer optimizer(params); in TEST() local
241 OLD_INTERFACE_WARNING_CHECK(size = optimizer.size()); in TEST()
244 OLD_INTERFACE_WARNING_CHECK(optimizer.add_parameters(parameters)); in TEST()
246 OLD_INTERFACE_WARNING_CHECK(size = optimizer.size()); in TEST()
250 OLD_INTERFACE_WARNING_CHECK(params_ = optimizer.parameters()); in TEST()
257 MyOptimizer optimizer(linear->parameters()); in TEST() local
261 OLD_INTERFACE_WARNING_CHECK(size = optimizer.size()); in TEST()
419 SGD optimizer(model->parameters(), 0.1); in TEST() local
434 optimizer.zero_grad(); in TEST()
454 SGD optimizer(parameters, 1.0); in TEST() local
456 optimizer.step(); in TEST()
474 LBFGS optimizer(std::vector<torch::Tensor>{}, 1.0); in TEST() local
475 OLD_INTERFACE_WARNING_CHECK(optimizer.add_parameters(parameters)); in TEST()
477 optimizer.step([]() { return torch::tensor(1); }); in TEST()
482 // Check whether the learning rate of the parameter groups in the optimizer are
485 Optimizer& optimizer, in check_lr_change() argument
505 optimizer.param_groups()[0].options().get_lr()) < in check_lr_change()
508 optimizer.step(); in check_lr_change()
517 Optimizer& optimizer, in check_lr_change_for_reduce_on_plateau() argument
537 optimizer.param_groups()[0].options().get_lr()) < in check_lr_change_for_reduce_on_plateau()
540 optimizer.step(); in check_lr_change_for_reduce_on_plateau()
547 auto optimizer = Adam({parameters}, AdamOptions().lr(1e-3)); in TEST() local
551 StepLR step_lr_scheduler(optimizer, step_size, gamma); in TEST()
556 check_lr_change(optimizer, step_lr_scheduler, expected_epoch_lrs); in TEST()
561 auto optimizer = Adam({parameters}, AdamOptions().lr(1e-3)); in TEST() local
565 optimizer, in TEST()
574 optimizer, reduce_lr_on_plateau_scheduler, expected_epoch_lrs); in TEST()