# TensorRT OpenPifPaf Pose Estimation
TensorRT OpenPifPaf Pose Estimation is a Jetson-friendly application that runs inference using a [TensorRT](https://developer.nvidia.com/tensorrt) engine to extract human poses. The provided TensorRT engine is generated from an ONNX model exported from [OpenPifPaf](https://github.com/vita-epfl/openpifpaf) version 0.10.0 using [ONNX-TensorRT](https://github.com/onnx/onnx-tensorrt) repo.
You can read [this article]() on our website to learn more about the TensorRT OpenPifPaf Pose Estimation application.
## Getting Started
The following instructions will help you get started.
### Prerequisites
**Hardware**
* [NVIDIA Jetson TX2](https://developer.nvidia.com/embedded/jetson-tx2)
**Software**
* You should have [Docker](https://docs.docker.com/get-docker/) on your device.
### Install
```bash
git clone https://github.com/neuralet/neuralet.git
cd neuralet/applications/pose-estimation-tensorrt/
```
### Usage
##### Run on Jetson TX2
* You need to have [JetPack 4.3](https://developer.nvidia.com/jetpack-43-archive) installed on your Jetson TX2.
```bash
# 1) Download TensorRT engine file built with JetPack 4.3:
./download_engine.sh
# 2) Download/Copy Sample image
./download_sample_image.sh
# 3) Build Docker image for Jetson TX2 (This step is optional, you can skip it if you want to pull the container from neuralet dockerhub)
docker build -f jetson-tx2-openpifpaf.Dockerfile -t "neuralet/applications-openpifpaf:latest-jetson-tx2" .
# 4) Run Docker container:
docker run --runtime nvidia --privileged -it -v $PWD:/repo neuralet/applications-openpifpaf:latest-jetson-tx2
```