Revisiting Point Cloud Shape Classification with a Simple and Effective Baseline
Ankit Goyal, Hei Law, Bowei Liu, Alejandro Newell, Jia Deng
International Conference on Machine Learning (ICML), 2021
If you find our work useful in your research, please consider citing:
@article{goyal2021revisiting,
title={Revisiting Point Cloud Shape Classification with a Simple and Effective Baseline},
author={Goyal, Ankit and Law, Hei and Liu, Bowei and Newell, Alejandro and Deng, Jia},
journal={International Conference on Machine Learning},
year={2021}
}
Getting Started
First clone the repository. We would refer to the directory containing the code as SimpleView
.
git clone [email protected]:princeton-vl/SimpleView.git
Requirements
The code is tested on Linux OS with Python version 3.7.5, CUDA version 10.0, CuDNN version 7.6 and GCC version 5.4. We recommend using these versions especially for installing pointnet++ custom CUDA modules.
Install Libraries
We recommend you first install Anaconda and create a virtual environment.
conda create --name simpleview python=3.7.5
Activate the virtual environment and install the libraries. Make sure you are in SimpleView
.
conda activate simpleview
pip install -r requirements.txt
conda install sed # for downloading data and pretrained models
For PointNet++, we need to install custom CUDA modules. Make sure you have access to a GPU during this step. You might need to set the appropriate TORCH_CUDA_ARCH_LIST
environment variable depending on your GPU model. The following command should work for most cases export TORCH_CUDA_ARCH_LIST="6.0;6.1;6.2;7.0;7.5"
. However, if the install fails, check if TORCH_CUDA_ARCH_LIST
is correctly set. More details could be found here.
cd pointnet2_pyt && pip install -e . && cd ..
Download Datasets and Pre-trained Models
Make sure you are in SimpleView
. download.sh
script can be used for downloading all the data and the pretrained models. It also places them at the correct locations. First, use the following command to provide execute permission to the download.sh
script.
chmod +x download.sh
To download ModelNet40 execute the following command. This will download the ModelNet40 point cloud dataset released with pointnet++ as well as the validation splits used in our work.
./download.sh modelnet40
To download the pretrained models, execute the following command.
./download.sh pretrained
Code Organization
SimpleView/models
: Code for various models in PyTorch.SimpleView/configs
: Configuration files for various models.SimpleView/main.py
: Training and testing any models.SimpleView/configs.py
: Hyperparameters for different models and dataloader.SimpleView/dataloader.py
: Code for different variants of the dataloader.SimpleView/*_utils.py
: Code for various utility functions.
Running Experiments
Training and Config files
To train or test any model, we use the main.py
script. The format for running this script is as follows.
python main.py --exp-config <path to the config>
The config files are named as <protocol>_<model_name><_extra>_run_<seed>.yaml
(<protocol> ∈ [dgcnn, pointnet2, rscnn]
; <model_name> ∈ [dgcnn, pointnet2, rscnn, pointnet, simpleview]
; <_extra> ∈ ['',valid,0.5,0.25]
). For example, the config file to run an experiment for PointNet++ in DGCNN protocol with seed 1 dgcnn_pointnet2_run_1.yaml
. To run a new experiment with a different seed, you need to change the SEED
parameter in the config file. For all our experiments (including on the validation set) we do 4 runs with different seeds.
As discussed in the paper for the PointNet++ and SimpleView protocols, we need to first run an experiment to tune the number of epochs on the validation set. This could be done by first running the experiment <pointnet2/dgcnn>_<model_name>_valid_run_<seed>.yaml
and then running the experiment <pointnet2/dgcnn>_<model_name>_run_<seed>.yaml
. Based on the number of epochs achieving the best performance on the validation set, one could use the model trained on the complete training set to get the final test performance.
To train models on the partial training set (Table 7), use the configs named as dgcnn_<model_name>_valid_<0.25/0.5>_run_<seed>.yaml
and <dgcnn>_<model_name>_<0.25/0.5>_run_<seed>.yaml
.
Even with the same SEED the results could vary slightly because of the randomization introduced for faster cuDNN operations. More details could be found here
SimpleView Protocol
To run an experiment in the SimpleView protocol, there are two stages.
- First tune the number of epochs on the validation set. This is done using configs
dgcnn_<model_name>_valid_run_<seed>.yaml
. Find the best number of epochs on the validation set, evaluated at every 25th epoch. - Train the model on the complete training set using configs
dgcnn_<model_name>_run_<seed>.yaml
. Use the performance on the test set at the fine-tuned number of epochs as the final performance.
Evaluate a pretrained model
We provide pretrained models. They can be downloaded using the ./download pretrained
command and are stored in the SimpleView/pretrained
folder. To test a pretrained model, the command is of the following format.
python main.py --entry <test/rscnn_vote/pn2_vote> --model-path pretrained/<cfg_name>/<model_name>.pth --exp-config configs/<cfg_name>.yaml
We list the evaluation commands in the eval_models.sh
script. For example to evaluate models on the SimpleView protocol, use the commands here. Note that for the SimpleView and the Pointnet2 protocols, the model path has names in the format model_<epoch_id>.pth
. Here epoch_id
represents the number of epochs tuned on the validation set.
Performance of the released pretrained models on ModelNet40
Protocol → | DGCNN - Smooth | DCGNN - CE. | RSCNN - No Vote | PointNet - No Vote | SimpleView |
---|---|---|---|---|---|
Method↓ | (Tab. 2, Col. 7) | (Tab. 2, Col. 6) | (Tab. 2, Col. 5) | (Tab. 2, Col. 2) | (Tab. 4, Col. 2) |
SimpleView | 93.9 | 93.2 | 92.7 | 90.8 | 93.3 |
PointNet++ | 93.0 | 92.8 | 92.6 | 89.7 | 92.6 |
DGCNN | 92.6 | 91.8 | 92.2 | 89.5 | 92.0 |
RSCNN | 92.3 | 92.0 | 92.2 | 89.4 | 91.6 |
PointNet | 90.7 | 90.0 | 89.7 | 88.8 | 90.1 |
Acknowlegements
We would like to thank the authors of the following reposities for sharing their code.