diff --git a/tools/visual-pipeline-and-platform-evaluation-tool/models/model_manager.sh b/tools/visual-pipeline-and-platform-evaluation-tool/models/model_manager.sh index d2605b6e2..c6738be1a 100755 --- a/tools/visual-pipeline-and-platform-evaluation-tool/models/model_manager.sh +++ b/tools/visual-pipeline-and-platform-evaluation-tool/models/model_manager.sh @@ -116,6 +116,13 @@ download_omz_models() { cp "$proc_src" "$target_dir/vehicle-attributes-recognition-barrier-0039.json" echo "Copied model_proc file for $model." fi + elif [ "$model" == "human-pose-estimation-0001" ]; then + mv "$tmp_models_dir/intel/$model" "$target_dir" + local proc_src="/opt/intel/dlstreamer/samples/gstreamer/model_proc/intel/human-pose-estimation-0001.json" + if [ -f "$proc_src" ]; then + cp "$proc_src" "$target_dir/human-pose-estimation-0001.json" + echo "Copied model_proc file for $model." + fi elif [ "$model" == "mobilenet-v2-pytorch" ]; then mv "$tmp_models_dir/public/$model" "$target_dir" local proc_src="/opt/intel/dlstreamer/samples/gstreamer/model_proc/public/preproc-aspect-ratio.json" diff --git a/tools/visual-pipeline-and-platform-evaluation-tool/models/supported_models.lst b/tools/visual-pipeline-and-platform-evaluation-tool/models/supported_models.lst index 2001399be..13ec71cee 100644 --- a/tools/visual-pipeline-and-platform-evaluation-tool/models/supported_models.lst +++ b/tools/visual-pipeline-and-platform-evaluation-tool/models/supported_models.lst @@ -1,5 +1,6 @@ ch_PP-OCRv4_rec_infer|PaddleOCR (FP32)|public|classification|default efficientnet-b0_INT8|EfficientNet B0 (INT8)|pipeline-zoo-models|classification|default +human-pose-estimation-0001|Human Pose Estimation 0001 (FP16)|omz|classification|optional mobilenet-v2-pytorch|MobileNet V2 PyTorch (FP16)|omz|classification|optional resnet-50-tf_INT8|ResNet-50 TF (INT8)|pipeline-zoo-models|classification|default ssdlite_mobilenet_v2_INT8|SSDLite MobileNet V2 (INT8)|pipeline-zoo-models|detection|default diff --git a/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/simplevs/config.yaml b/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/simplevs/config.yaml index 572462b13..01ba632d0 100644 --- a/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/simplevs/config.yaml +++ b/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/simplevs/config.yaml @@ -33,4 +33,5 @@ parameters: - "ResNet-50 TF (INT8)" - "PaddleOCR (FP32)" - "Vehicle Attributes Recognition Barrier 0039 (FP16)" + - "Human Pose Estimation 0001 (FP16)" classification_model_default: "PaddleOCR (FP32)" diff --git a/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/smartnvr/config.yaml b/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/smartnvr/config.yaml index f40eba360..722513b42 100644 --- a/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/smartnvr/config.yaml +++ b/tools/visual-pipeline-and-platform-evaluation-tool/pipelines/smartnvr/config.yaml @@ -29,4 +29,5 @@ parameters: - "EfficientNet B0 (INT8)" - "MobileNet V2 PyTorch (FP16)" - "ResNet-50 TF (INT8)" + - "Human Pose Estimation 0001 (FP16)" classification_model_default: "ResNet-50 TF (INT8)" diff --git a/tools/visual-pipeline-and-platform-evaluation-tool/utils.py b/tools/visual-pipeline-and-platform-evaluation-tool/utils.py index 1c49ee99c..04c91c1cd 100644 --- a/tools/visual-pipeline-and-platform-evaluation-tool/utils.py +++ b/tools/visual-pipeline-and-platform-evaluation-tool/utils.py @@ -212,6 +212,13 @@ def prepare_video_and_constants( constants["OBJECT_CLASSIFICATION_MODEL_PROC"] = ( f"{MODELS_PATH}/omz/vehicle-attributes-recognition-barrier-0039/vehicle-attributes-recognition-barrier-0039.json" ) + case "Human Pose Estimation 0001 (FP16)": + constants["OBJECT_CLASSIFICATION_MODEL_PATH"] = ( + f"{MODELS_PATH}/omz/human-pose-estimation-0001/human-pose-estimation-0001/FP16/human-pose-estimation-0001.xml" + ) + constants["OBJECT_CLASSIFICATION_MODEL_PROC"] = ( + f"{MODELS_PATH}/omz/human-pose-estimation-0001/human-pose-estimation-0001.json" + ) case _: raise ValueError("Unrecognized Object Classification Model")