Skip to content

Latest commit

 

History

History
49 lines (39 loc) · 2.82 KB

File metadata and controls

49 lines (39 loc) · 2.82 KB

Single Human Pose Estimation Demo (top-down pipeline)

This demo showcases top-down pipeline for human pose estimation on video or image. The task is to predict bboxes for every person on frame and then to predict a pose for every detected person. The pose may contain up to 17 keypoints: ears, eyes, nose, shoulders, elbows, wrists, hips, knees, and ankles.

How It Works

On the start-up, the application reads command line parameters and loads detection person model and single human pose estimation model. Upon getting a frame from the OpenCV VideoCapture, the demo executes top-down pipeline for this frame and displays the results.

Running

Running the application with the -h option yields the following usage message:

usage: single_human_pose_estimation_demo.py [-h] -m_od MODEL_OD -m_hpe
                                            MODEL_HPE [-i INPUT [INPUT ...]]
                                            [-d DEVICE]
                                            [--person_label PERSON_LABEL]
                                            [--no_show]
                                            [-u UTILIZATION_MONITORS]

optional arguments:
  -h, --help            show this help message and exit
  -m_od MODEL_OD, --model_od MODEL_OD
                        path to model of object detector in xml format
  -m_hpe MODEL_HPE, --model_hpe MODEL_HPE
                        path to model of human pose estimator in xml format
  -i INPUT [INPUT ...], --input INPUT [INPUT ...]
                        path to video or image/images
  -d DEVICE, --device DEVICE
                        Specify the target to infer on CPU or GPU
  --person_label PERSON_LABEL
                        Label of class person for detector
  --no_show             Optional. Do not display output.
  -u UTILIZATION_MONITORS, --utilization_monitors UTILIZATION_MONITORS
                        Optional. List of monitors to show initially.

To run the demo, you can use public or pre-trained models. To download the pre-trained models, use the OpenVINO Model Downloader or go to https://download.01.org/opencv/.

For example, to do inference on a CPU, run the following command:

python single_human_pose_estimation_demo.py --model_od <path_to_dir__with_models>/mobilenet-ssd.xml --model_hpe <path_to_dir__with_models>/single-human-pose-estimation-0001.xml --input <path_to_video>/back-passengers.avi

The demo uses OpenCV to display the resulting frame with estimated poses and reports performance in the following format: summary inference FPS (single human pose inference FPS / detector inference FPS).

See Also