1 /** 2 * Copyright 2021 Huawei Technologies Co., Ltd 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 #ifndef MINDSPORE_INCLUDE_API_CALLBACK_LR_SCHEDULER_H 17 #define MINDSPORE_INCLUDE_API_CALLBACK_LR_SCHEDULER_H 18 19 #include <cstddef> 20 #include <string> 21 #include <vector> 22 #include <memory> 23 #include "include/api/callback/callback.h" 24 25 namespace mindspore { 26 27 constexpr int DONT_UPDATE_LR = 0; 28 constexpr int UPDATE_LR = 1; 29 30 using LR_Lambda = std::function<int(float *lr, int epoch, void *cb_data)>; 31 32 /// \brief Multiply the LR by a factor of gamma every epoch 33 int MultiplicativeLRLambda(float *lr, int epoch, void *multiplication); 34 35 /// \brief Multiply the LR by a factor of gamma every step_size 36 int StepLRLambda(float *lr, int epoch, void *step_size); 37 struct StepLRLambda { 38 StepLRLambda(int step, float g) : step_size(step), gamma(g) {} 39 40 int step_size; // period of LR decay 41 float gamma; // LR decay factor 42 }; 43 44 class LRScheduler: public TrainCallBack { 45 public: 46 explicit LRScheduler(LR_Lambda lambda_func, void *lr_cb_data = nullptr, int step = 1); 47 virtual ~LRScheduler(); 48 }; 49 50 } // namespace mindspore 51 #endif // MINDSPORE_INCLUDE_API_CALLBACK_LR_SCHEDULER_H 52