/third_party/mindspore/mindspore/ccsrc/ps/ |
D | optimizer_info_builder.h | 33 explicit OptimizerInfoBuilder(size_t worker_num) : worker_num_(worker_num) {} in OptimizerInfoBuilder() argument 37 … const Values &values, const Lengths &lens, const InputsShapePtr &inputs_shape, size_t worker_num, 41 … const Lengths &lens, const InputsShapePtr &inputs_shape, size_t worker_num, 44 … void BuildWorkspaces(OptimizerInfo *info, const std::vector<size_t> &ws_sizes, size_t worker_num); 45 virtual void BuildOutputs(OptimizerInfo *info, size_t worker_num) {} in BuildOutputs() argument 57 explicit MomentumOptimInfoBuilder(size_t worker_num) : OptimizerInfoBuilder(worker_num) {} in MomentumOptimInfoBuilder() argument 60 const InputsShapePtr &inputs_shape, size_t worker_num, 66 explicit SparseAdamOptimInfoBuilder(size_t worker_num) : OptimizerInfoBuilder(worker_num) {} in SparseAdamOptimInfoBuilder() argument 69 const InputsShapePtr &inputs_shape, size_t worker_num, 75 explicit SparseFtrlOptimInfoBuilder(size_t worker_num) : OptimizerInfoBuilder(worker_num) {} in SparseFtrlOptimInfoBuilder() argument [all …]
|
/third_party/mindspore/tests/st/fl/cross_silo_faster_rcnn/ |
D | run_cross_silo_fasterrcnn_worker.py | 36 worker_num = args.worker_num variable 48 local_worker_num = worker_num 50 assert local_worker_num <= worker_num, "The local worker number should not be bigger than total wor… 61 cmd_worker += " --worker_num=" + str(worker_num)
|
D | run_cross_silo_fasterrcnn_sched.py | 33 worker_num = args.worker_num variable 48 cmd_sched += " --worker_num=" + str(worker_num)
|
D | run_cross_silo_fasterrcnn_server.py | 45 worker_num = args.worker_num variable 79 cmd_server += " --worker_num=" + str(worker_num)
|
/third_party/mindspore/tests/st/fl/hybrid_lenet/ |
D | run_hybrid_train_worker.py | 39 worker_num = args.worker_num variable 52 local_worker_num = worker_num 54 assert local_worker_num <= worker_num, "The local worker number should not be bigger than total wor… 66 cmd_worker += " --worker_num=" + str(worker_num)
|
D | run_hybrid_train_sched.py | 37 worker_num = args.worker_num variable 55 cmd_sched += " --worker_num=" + str(worker_num)
|
D | run_smlt.sh | 19 worker_num=$2 23 for((i=0;i<worker_num;i++));
|
/third_party/mindspore/tests/st/fl/cross_silo_lenet/ |
D | run_cross_silo_lenet_worker.py | 37 worker_num = args.worker_num variable 50 local_worker_num = worker_num 52 assert local_worker_num <= worker_num, "The local worker number should not be bigger than total wor… 64 cmd_worker += " --worker_num=" + str(worker_num)
|
D | run_cross_silo_lenet_sched.py | 32 worker_num = args.worker_num variable 46 cmd_sched += " --worker_num=" + str(worker_num)
|
D | run_cross_silo_lenet_server.py | 43 worker_num = args.worker_num variable 75 cmd_server += " --worker_num=" + str(worker_num)
|
/third_party/mindspore/tests/st/fl/cross_silo_femnist/ |
D | run_cross_silo_femnist_worker.py | 38 worker_num = args.worker_num variable 52 local_worker_num = worker_num 54 assert local_worker_num <= worker_num, "The local worker number should not be bigger than total wor… 65 cmd_worker += " --worker_num=" + str(worker_num)
|
D | run_cross_silo_femnist_sched.py | 33 worker_num = args.worker_num variable 48 cmd_sched += " --worker_num=" + str(worker_num)
|
D | run_cross_silo_femnist_server.py | 44 worker_num = args.worker_num variable 77 cmd_server += " --worker_num=" + str(worker_num)
|
/third_party/mindspore/mindspore/ccsrc/ps/core/ |
D | node_recovery.cc | 29 …int32_t worker_num = std::strtol(recovery_storage_->Get(kRecoveryWorkerNum, "").c_str(), nullptr, … in Recover() local 30 node_->set_worker_num(worker_num); in Recover() 58 …MS_LOG(INFO) << "The worker num:" << node_->worker_num() << ", the server num:" << node_->server_n… in Recover()
|
D | cluster_metadata.h | 34 …ClusterMetadata(const uint32_t &worker, const uint32_t &server) : worker_num(worker), server_num(s… in ClusterMetadata() 36 uint32_t worker_num; member
|
D | node_manager.cc | 111 if (rank_id >= meta_data_->worker_num) { in NextRankId() 112 … MS_LOG(WARNING) << "The rank id is greater than the number of workers:" << meta_data_->worker_num; in NextRankId() 304 …oid NodeManager::set_worker_num(const int32_t &worker_num) { meta_data_->worker_num = IntToUint(wo… in set_worker_num() argument 308 int32_t NodeManager::worker_num() const { return UintToInt(meta_data_->worker_num); } in worker_num() function in mindspore::ps::core::NodeManager
|
D | cluster_config.h | 34 …explicit ClusterConfig(const uint32_t &worker_num, const uint32_t &server_num, std::string host, c… in ClusterConfig() 35 : initial_worker_num(worker_num), in ClusterConfig()
|
/third_party/mindspore/tests/st/fl/albert/ |
D | run_hybrid_train_sched.py | 36 worker_num = args.worker_num variable 53 cmd_sched += " --worker_num=" + str(worker_num)
|
/third_party/mindspore/tests/st/fl/mobile/ |
D | run_mobile_sched.py | 39 worker_num = args.worker_num variable 58 cmd_sched += " --worker_num=" + str(worker_num)
|
D | run_smlt.sh | 19 worker_num=$2 23 for((i=0;i<worker_num;i++));
|
/third_party/mindspore/tests/ut/cpp/ps/core/ |
D | cluster_metadata_test.cc | 37 std::string worker_num = "1"; in TEST_F() local 41 common::SetEnv(kEnvWorkerNum, worker_num.c_str()); in TEST_F()
|
D | cluster_available_timeout_test.cc | 35 std::string worker_num = "1"; in TEST_F() local 39 common::SetEnv(kEnvWorkerNum, worker_num.c_str()); in TEST_F()
|
/third_party/mindspore/mindspore/ccsrc/backend/kernel_compiler/cpu/ps/ |
D | apply_momentum_ps_kernel.h | 30 ApplyMomentumPSKernel(size_t rank_id, size_t pserver_num, size_t worker_num) in ApplyMomentumPSKernel() argument 31 : PServerKernel(rank_id, pserver_num, worker_num) {} in ApplyMomentumPSKernel()
|
D | sparse_apply_adam_ps_kernel.h | 31 SparseApplyAdamPSKernel(size_t rank_id, size_t pserver_num, size_t worker_num) in SparseApplyAdamPSKernel() argument 32 : PServerKernel(rank_id, pserver_num, worker_num) {} in SparseApplyAdamPSKernel()
|
D | sparse_apply_lazy_adam_ps_kernel.h | 30 SparseApplyLazyAdamPSKernel(size_t rank_id, size_t pserver_num, size_t worker_num) in SparseApplyLazyAdamPSKernel() argument 31 : PServerKernel(rank_id, pserver_num, worker_num) {} in SparseApplyLazyAdamPSKernel()
|