/third_party/mindspore/mindspore/lite/test/st/scripts/ |
D | run_benchmark_x86.sh | 24 model_name=`echo ${parallel_split_line_info}|awk -F ' ' '{print $1}'` 25 echo ${model_name} >> "${run_converter_log_file}" 26 echo 'convert mode name: '${model_name}' begin.' 27 … --fmk=TFLITE --modelFile='${models_path}'/'${model_name}' --outputFile='${ms_models_path}'/'${mo… 28 …er_lite --fmk=TFLITE --modelFile=$models_path/${model_name} --outputFile=${ms_models_path}/${mode… 30 …converter_result='converter 1_1_parallel_split '${model_name}' pass';echo ${converter_result} >> $… 32 …converter_result='converter 1_1_parallel_split '${model_name}' failed';echo ${converter_result} >>… 38 … --fmk=TFLITE --modelFile='${models_path}'/'${model_name}' --outputFile='${ms_models_path}'/'${mo… 39 …er_lite --fmk=TFLITE --modelFile=$models_path/${model_name} --outputFile=${ms_models_path}/${mode… 41 …converter_result='converter 1_2_parallel_split '${model_name}' pass';echo ${converter_result} >> $… [all …]
|
D | run_benchmark_codegen.sh | 32 model_name=${line} 33 if [[ $model_name == \#* || $model_name == "" ]]; then 46 echo ${model_name} >> "$4" 47 …${CODEGEN_PATH}/codegen --codePath=$1 --modelPath=$2/${model_name}.ms --supportParallel=${support_… 49 mkdir -p $1/${model_name}/build && cd $1/${model_name}/build || exit 1 50 cmake -DPKG_PATH=${x86_path}/mindspore-lite-${version}-linux-x64 $1/${model_name} >> $4 53 echo "net file: $1/${model_name}/src/net.bin" >> $4 54 …_path}/input_output/input/${model_name}.ms.bin $1/${model_name}/src/net.bin 1 ${models_path}/input… 55 …_path}/input_output/input/${model_name}.ms.bin $1/${model_name}/src/net.bin 1 ${models_path}/input… 57 run_result='x86_codegen'${suffix}': '${model_name}' pass'; echo ${run_result} >> $5 [all …]
|
D | run_net_train.sh | 15 model_name=${line_array[0]} 16 if [[ $model_name == \#* ]]; then 19 echo ${model_name}'_train_export.py' >> "${export_log_file}" 20 echo 'exporting' ${model_name} 22 …t/share --privileged=true '${docker_image}' python '${models_path}'/'${model_name}'_train_export.p… 23 …pt/share --privileged=true "${docker_image}" python ${models_path}'/'${model_name}_train_export.py… 25 …echo 'CLOUD_MODEL_ZOO=${CLOUD_MODEL_ZOO} python '${models_path}'/'${model_name}'_train_export.py' … 26 …CLOUD_MODEL_ZOO=${CLOUD_MODEL_ZOO} python ${models_path}'/'${model_name}_train_export.py "${epoch_… 29 …export_result='export mindspore '${model_name}'_train_export pass';echo ${export_result} >> ${expo… 31 …export_result='export mindspore '${model_name}'_train_export failed';echo ${export_result} >> ${ex… [all …]
|
D | base_functions.sh | 6 …local cfg_file_list model_info model_name extra_info model_type cfg_file_name model_file weight_fi… 15 model_name=${model_info%%;*} 17 model_type=${model_name##*.} 39 model_file=$2"/"${model_name} 45 output_file=$3"/"${model_name} 63 config_file="${quant_config_path}/${model_name}_posttraining.config" 69 echo ${model_name} >> "$4" 77 …converter_result='converter '${model_type}''${quant_type}' '${model_name}' pass';echo ${converter_… 79 …converter_result='converter '${model_type}''${quant_type}' '${model_name}' failed';echo ${converte… 128 …local cfg_file_list cfg_file_name line_info model_info spec_acc_limit model_name input_num input_s… [all …]
|
D | run_benchmark_gpu.sh | 36 model_name=${line} 37 if [[ $model_name == \#* ]]; then 41 model_name=`echo ${line} | awk -F ';' '{print $1}'` 46 output=${data_path}'output/'${model_name}'.ms.out' 47 input=${data_path}'input/'${model_name}'.ms.bin' 48 model=${model_name}'.ms' 49 echo ${model_name} >> "${run_parallel_log_file}" 50 echo "run mindrt parallel test : ${model_name}" 61 …run_result='mindrt_parallel_CPU_CPU: '${model_name}' pass'; echo ${run_result} >> ${run_benchmark_… 63 …run_result='mindrt_parallel_CPU_CPU: '${model_name}' failed'; echo ${run_result} >> ${run_benchmar… [all …]
|
D | run_benchmark_cropping_size.sh | 53 …pping_cpu.cfg --modelFile=${ms_models_path}/${model_name}.ms --outputFile=./libmindspore-lite-${mo… 54 …pping_cpu.cfg --modelFile=${ms_models_path}/${model_name}.ms --outputFile=./libmindspore-lite-${mo… 62 ls -l libmindspore-lite-${model_name}.a || exit 1 63 ls -l -h libmindspore-lite-${model_name}.a || exit 1 65 cp libmindspore-lite-${model_name}.a mm/ || exit 1 67 ar -x libmindspore-lite-${model_name}.a || exit 1 68 ar -d libmindspore-lite-${model_name}.a *.S.o || exit 1 70 echo "after ar -d libmindspore-lite-${model_name}.a *.S.o" 71 ls -l mm/libmindspore-lite-${model_name}.a || exit 1 72 ls -l -h mm/libmindspore-lite-${model_name}.a || exit 1 [all …]
|
D | run_cropper.sh | 45 model_name=`echo ${line}|awk -F ' ' '{print $1}'` 47 …pping_npu.cfg --modelFile=${ms_models_path}/${model_name}.ms --outputFile=./libmindspore-lite-${mo… 48 …pping_npu.cfg --modelFile=${ms_models_path}/${model_name}.ms --outputFile=./libmindspore-lite-${mo… 56 …ns/llvm/prebuilt/linux-x86_64/bin/clang++ -Wl,--whole-archive ./libmindspore-lite-${model_name}.a \ 79 echo "mindspore run cropper: ${model_name}, accuracy limit:4" >> "${run_cropper_log_file}" 81 …./benchmark --device=GPU --modelFile=/data/local/tmp/benchmark_test/'${model_name}'.ms --loopCount… 82 …./benchmark --device=GPU --modelFile=/data/local/tmp/benchmark_test/'${model_name}'.ms --loopCount… 85 … run_result='run_benchmark: '${model_name}' pass'; echo ${run_result} >> "${run_cropper_result}" 87 …run_result='run_benchmark: '${model_name}' failed'; echo ${run_result} >> "${run_cropper_result}";…
|
/third_party/mindspore/mindspore/lite/examples/export_models/ |
D | prepare.sh | 83 model_name=${line_array[0]} 84 if [[ $model_name == \#* ]]; then 87 echo 'exporting' ${model_name} 90 …L_ZOO=${CLOUD_MODEL_ZOO} PYTHONPATH=${CLOUD_MODEL_ZOO} python models/${model_name}${suffix}_export… 91 && chmod 444 mindir/${model_name}_head.mindir && chmod 444 mindir/${model_name}_bb.mindir" 93 …ATH=${CLOUD_MODEL_ZOO} python models/${model_name}${suffix}_export.py ${TRAIN_IO} && chmod 444 min… 96 PYTHONPATH=${CLOUD_MODEL_ZOO} python models/${model_name}${suffix}_export.py ${TRAIN_IO} 99 …export_result='export mindspore '${model_name}${suffix}'_export pass';echo ${export_result} >> ${e… 101 …export_result='export mindspore '${model_name}${suffix}'_export failed';echo ${export_result} >> $…
|
/third_party/mindspore/mindspore/lite/test/st/scripts/nnie/ |
D | run_benchmark_nnie.sh | 19 model_name=${model_info%%;*} 29 … input_files=${input_files}${basepath}'/../input_output/input/'${model_name}'.ms.bin_'$i',' 36 …input_files=${input_files}${basepath}'/../input_output/input/'${model_name}'.ms.bin_'$cur_input_nu… 40 input_files=${basepath}/../input_output/input/${model_name}.ms.bin 45 …${model_name}'.ms --inDataFile='${input_files}' --inputShapes='${input_shapes}' --benchmarkDataFil… 46 …th}/${model_name}.ms --inDataFile=${input_files} --inputShapes=${input_shapes} --benchmarkDataFile… 48 … run_result='hi3516: '${model_name}' pass'; echo ${run_result} >> ${run_benchmark_result_file} 50 …run_result='hi3516: '${model_name}' failed'; echo ${run_result} >> ${run_benchmark_result_file}; r…
|
D | run_converter_nnie.sh | 37 model_name=${model_info%%;*} 38 cp ${models_path}/${model_location}/${model_name}.cfg ./ || exit 1 39 echo 'export NNIE_CONFIG_PATH=./'${model_name}'.cfg' >> "${run_converter_log_file}" 40 export NNIE_CONFIG_PATH=./${model_name}.cfg 41 echo ${model_name} >> "${run_converter_log_file}" 42 …model_name}'.prototxt --weightFile='${models_path}'/'${model_location}'/model/'${model_name}'.caff… 43 …model_name}.prototxt --weightFile=${models_path}/${model_location}/model/${model_name}.caffemodel … 45 …converter_result='converter CAFFE '${model_name}' pass';echo ${converter_result} >> ${run_converte… 47 …converter_result='converter CAFFE '${model_name}' failed';echo ${converter_result} >> ${run_conver… 71 run_result='hi3516: '${model_name}' pass'; echo ${run_result} >> ${run_benchmark_result_file}; [all …]
|
/third_party/mindspore/tests/st/model_zoo_tests/bgcf/ |
D | test_BGCF_amazon_beauty.py | 28 model_name = "bgcf" 29 utils.copy_files(model_path, cur_path, model_name) 30 cur_model_path = os.path.join(cur_path, model_name) 43 .format(model_name, data_path) 65 model_name = "bgcf" 66 utils.copy_files(model_path, cur_path, model_name) 67 cur_model_path = os.path.join(cur_path, model_name) 70 …shell = "cd {}; python export.py --ckpt_file={} --file_format=MINDIR".format(model_name, ckpt_path) 72 assert os.path.exists(os.path.join(cur_model_path, "{}.mindir".format(model_name)))
|
/third_party/mindspore/tests/st/model_zoo_tests/maskrcnn/ |
D | test_maskrcnn.py | 35 model_name = "maskrcnn" 36 utils.copy_files(model_path, cur_path, model_name) 37 cur_model_path = os.path.join(cur_path, model_name) 41 … exec_export_shell = "cd {}; python export.py --config_path default_config.yaml".format(model_name) 43 assert os.path.exists(os.path.join(cur_model_path, "{}.air".format(model_name)))
|
/third_party/mindspore/tests/st/model_zoo_tests/lenet/ |
D | test_lenet_MNIST.py | 28 model_name = "lenet" 29 utils.copy_files(model_path, cur_path, model_name) 30 cur_model_path = os.path.join(cur_path, model_name) 36 .format(model_name, dataset_path, train_log) 40 .format(model_name, dataset_path, ckpt_file, infer_log)
|
/third_party/mindspore/mindspore/lite/src/delegate/npu/ |
D | npu_manager.cc | 146 …::AddModel(std::shared_ptr<domi::ModelBufferData> model_buffer_data, const std::string &model_name, in AddModel() argument 148 auto model = std::make_shared<SubGraphModel>(subgraph_index_, model_name, model_buffer_data); in AddModel() 149 auto desc = std::make_shared<hiai::AiModelDescription>(model_name, frequency, 0, 0, 0); in AddModel() 151 models_.insert({model_name, model}); in AddModel() 229 std::shared_ptr<hiai::AiModelMngerClient> NPUManager::GetClient(const std::string &model_name) { in GetClient() argument 230 if (models_.find(model_name) == models_.end() || models_[model_name] == nullptr) { in GetClient() 233 return models_[model_name]->client_; in GetClient()
|
D | npu_manager.h | 37 …SubGraphModel(int index, std::string model_name, std::shared_ptr<domi::ModelBufferData> model_buff… in SubGraphModel() 38 : index_(index), model_name_(std::move(model_name)), model_buffer_data_(model_buffer_data) {} in SubGraphModel() 58 …shared_ptr<domi::ModelBufferData> model_buffer_data, const std::string &model_name, int frequency); 64 std::shared_ptr<hiai::AiModelMngerClient> GetClient(const std::string &model_name);
|
/third_party/mindspore/tests/st/model_zoo_tests/face_detection/ |
D | test_FaceDetection_WIDER.py | 28 model_name = "FaceDetection" 29 utils.copy_files(model_path, cur_path, model_name) 30 cur_model_path = os.path.join(cur_path, model_name) 37 .format(model_name, dataset_path, device_id)
|
/third_party/mindspore/mindspore/lite/tools/benchmark/ |
D | run_benchmark.cc | 67 auto model_name = flags.model_file_.substr(flags.model_file_.find_last_of(DELIM_SLASH) + 1); in RunBenchmark() local 71 BENCHMARK_LOG_ERROR("Run Benchmark " << model_name << " Failed : " << status); in RunBenchmark() 75 MS_LOG(INFO) << "Run Benchmark " << model_name << " Success."; in RunBenchmark() 76 std::cout << "Run Benchmark " << model_name << " Success." << std::endl; in RunBenchmark()
|
/third_party/mindspore/tests/st/model_zoo_tests/deeplabv3/ |
D | test_DeeplabV3_voc2007.py | 28 model_name = "deeplabv3" 29 utils.copy_files(model_path, cur_path, model_name) 30 cur_model_path = os.path.join(cur_path, model_name) 46 model_name, utils.rank_table_path)
|
/third_party/mindspore/tests/st/model_zoo_tests/ |
D | utils.py | 24 def copy_files(from_, to_, model_name): argument 25 if not os.path.exists(os.path.join(from_, model_name)): 26 raise ValueError("There is no file or path", os.path.join(from_, model_name)) 27 if os.path.exists(os.path.join(to_, model_name)): 28 shutil.rmtree(os.path.join(to_, model_name)) 29 return os.system("cp -r {0} {1}".format(os.path.join(from_, model_name), to_))
|
/third_party/wpa_supplicant/wpa_supplicant-2.9_standard/src/wps/ |
D | wps_dev_attr.c | 45 len = dev->model_name ? os_strlen(dev->model_name) : 0; in wps_build_model_name() 59 wpabuf_put_data(msg, dev->model_name, len); in wps_build_model_name() 289 os_free(dev->model_name); in wps_process_model_name() 290 dev->model_name = dup_binstr(str, str_len); in wps_process_model_name() 291 if (dev->model_name == NULL) in wps_process_model_name() 381 wps_process_model_name(dev, attr->model_name, in wps_process_device_attrs() 436 os_free(dev->model_name); in wps_device_data_free() 437 dev->model_name = NULL; in wps_device_data_free()
|
/third_party/wpa_supplicant/wpa_supplicant-2.9/src/wps/ |
D | wps_dev_attr.c | 45 len = dev->model_name ? os_strlen(dev->model_name) : 0; in wps_build_model_name() 59 wpabuf_put_data(msg, dev->model_name, len); in wps_build_model_name() 274 os_free(dev->model_name); in wps_process_model_name() 275 dev->model_name = dup_binstr(str, str_len); in wps_process_model_name() 276 if (dev->model_name == NULL) in wps_process_model_name() 366 wps_process_model_name(dev, attr->model_name, in wps_process_device_attrs() 421 os_free(dev->model_name); in wps_device_data_free() 422 dev->model_name = NULL; in wps_device_data_free()
|
/third_party/mindspore/tests/st/model_zoo_tests/resnet50/ |
D | test_resnet50_cifar10.py | 29 model_name = "resnet" 30 utils.copy_files(model_path, cur_path, model_name) 60 model_name = "resnet" 61 utils.copy_files(model_path, cur_path, model_name)
|
/third_party/cef/libcef/browser/media_router/ |
D | media_sink_impl.cc | 48 CefString(&device_info.model_name) = cast_data.model_name; in GetSinkInternalAndContinue() 55 CefString(&device_info.model_name) = dial_data.model_name; in GetSinkInternalAndContinue()
|
/third_party/mindspore/tests/st/model_zoo_tests/ssd_mobilenet_fpn/ |
D | test_ssd_modelnet_fpn.py | 29 model_name = "ssd" 30 utils.copy_files(model_path, cur_path, model_name) 31 cur_model_path = os.path.join(cur_path, model_name)
|
/third_party/mindspore/tests/st/model_zoo_tests/centernet/ |
D | test_centernet.py | 28 model_name = "centernet" 29 utils.copy_files(model_path, cur_path, model_name) 30 cur_model_path = os.path.join(cur_path, model_name)
|