Skip to content

Live video to 3D Mesh Generation using Lightweight Graph Transformer Network for Human Mesh Reconstruction from 2D Human Pose and Lightweight OpenPose for 2D pose estimation.

Notifications You must be signed in to change notification settings

foram-s1/ITCS6166-Video2Mesh

 
 

Repository files navigation

Final Project

ITCS 6166: Communication And Computer Networks

Team 4:

  • Foram Shah
  • Samarth Bhole
  • Parshwa Shah
  • Theodore Huang

Project Description:

This project implements A Lightweight Graph Transformer Network for Human Mesh Reconstruction from 2D Human Pose. The project is based on the paper "A Lightweight Graph Transformer Network for Human Mesh Reconstruction". The paper was published in 2021. The project is implemented in PyTorch. We connected the project to a light weight openpose implementation to get the 2D human pose and then used GTRS model to get the 3D human mesh.

Project Structure:

  • GTRS
    • Contrains the implementation of the GTRS model. From link
  • PoseDetector
    • Contains the implementation of the OpenPose model. From link
  • Tests
    • Contains the test data for the project
  • models
    • Contains the trained models for the project (GTRS and OpenPose) (Exported)
  • requirements.txt
    • Contains the required libraries for the project
  • liveWebCam.py
    • Contains the code to run the project on live webcam

How to run the project (Colab without exported models):

  • Open the colab notebook Colab Link
  • Run the cells in the notebook

How to run the project (Local with exported models):

  • Clone the repository
  • Create a virtual environment using the command conda create -n <env_name> python=3.9
  • Install the required libraries using the command pip install -r requirements.txt
  • Run the project using the command python liveWebCam.py

Training

  • Prepare data as described in GTRS and pose2mesh.
  • Activate the aforementioned virtual environment and cd GTRS
  • If you would like to re-train the PAM module, run
    python main/train.py --gpu 0,1,2,3 --cfg ./asset/yaml/pam_cocoJ_train_human36_coco_muco.yml
    
  • If you only want to train the MRM with pre-trained PAM, run
    python main/train.py --gpu 0,1,2,3 --cfg ./asset/yaml/gtrs_cocoJ_train_human36_coco_muco.yml
    

About

Live video to 3D Mesh Generation using Lightweight Graph Transformer Network for Human Mesh Reconstruction from 2D Human Pose and Lightweight OpenPose for 2D pose estimation.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%