Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Vision-Guided Action: Enhancing 3D Human Motion Prediction with Gaze-informed Affordance in 3D Scenes

Notifications You must be signed in to change notification settings

OpenMICG/GAP3DS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Vision-Guided Action: Enhancing 3D Human Motion Prediction with Gaze-informed Affordance in 3D Scenes

Ting Yu1  ·  Yi Lin1  ·  Jun Yu2  ·  Zhenyu Lou3  ·  Qiongjie Cui4

1Hangzhou Normal University
2Harbin Institute of Technology
3Zhejiang University
4Singapore University of Technology and Design

CVPR 2025

Setup


Step 1: Install Required Dependencies

Begin by installing the required Python packages listed in requirements.txt:

pip install -r requirements.txt

Step 2: Install PointNet++, SoftGroup, Affordancenet

Clone the PointNet++ repository and follow the instructions provided in this link:

git clone --recursive https://github.com/erikwijmans/Pointnet2_PyTorch
cd Pointnet2_PyTorch

Important: You need to modify the code in the repository to avoid issues with the build. Specifically:

  • Comment out lines 100-101 in sampling_gpu.cu:

    # https://github.com/erikwijmans/Pointnet2_PyTorch/blob/master/pointnet2_ops_lib/pointnet2_ops/_ext-src/src/sampling_gpu.cu#L100-L101
    
  • Edit lines 196-198 in pointnet2_modules.py (located in [PATH-TO-VENV]/lib64/python3.8/site-packages/pointnet2_ops/):

    interpolated_feats = known_feats.repeat(1, 1, unknown.shape[1])
    

Clone the Softgroup repository and follow the instructions provided in this link:

git clone --recursive https://github.com/thangvubk/SoftGroup.git
cd Softgroup

Softgroup provides the custom dataset guide, so you can process scene data youself, or use the results we processed.

Clone the Affordancenet repository and follow the instructions provided in this link:

cd model
git clone --recursive https://github.com/Gorilla-Lab-SCUT/AffordanceNet.git
cd AffordanceNet

you can easily train the affordancenet following the official implementation.

  1. download the affordance dataset

  2. train

    python train.py config/dgcnn/estimation_cfg.py --work_dir TPATH_TO_LOG_DIR --gpu 0,1
  3. test

    python test.py config/dgcnn/estimation_cfg.py --work_dir PATH_TO_LOG_DIR --gpu 0,1 --checkpoint PATH_TO_CHECKPOINT

After making the changes, run the following commands to install dependencies:

pip install -r requirements.txt

Step 3: Install Additional Dependencies

Download and install the following dependencies:

Dataset


The GAP3DS method utilizes a standard-processed dataset. However, due to confidentiality constraints, we are unable to release the processed version.

To obtain the raw dataset, please follow the instructions provided in the official GIMO repository.

After downloading and unzipping the raw dataset, the directory structure should look like the following:

--data_root
     |--bedroom0122
           |--2022-01-21-194925
                 |--eye_pc
                 |--PV
                 |--smplx_local
                 |--transform_info.json
                 ...
           |--2022-01-21-195107
           ...
     |--bedroom0123
     |--bedroom0210
     |--classroom0219
     ...

On the first run, our code will automatically preprocess the data. Ensure that the dataroot is correctly set before running the program. After preprocessing, the dataset will be stored in the same location as the raw dataset, and the folder structure will be as follows:

--data_root
      |--SLICES_8s
            |--train
                 |--gazes.pth
                 |--joints_input.pth
                 |--joints_label.pth
                 |--poses_input.pth
                 |--poses_label.pth
                 |--scene_points_<sample_points>.pth
            |--test
                 |--gazes.pth
                 |--joints_input.pth
                 |--joints_label.pth
                 |--poses_input.pth
                 |--poses_label.pth
                 |--scene_points_<sample_points>.pth
     |--bedroom0122
     |--bedroom0123
     |--bedroom0210
     |--classroom0219
     ...

Quickstart Guide


Evaluation

To evaluate the model, execute the following command:

bash scripts/eval.sh

the weight's results you can use

Training

To train the model, use the following command:

bash scripts/train.sh

About

Vision-Guided Action: Enhancing 3D Human Motion Prediction with Gaze-informed Affordance in 3D Scenes

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published