| Model | Benchmark | Num GPUs | Val mIoU | Tensorboard | Exp Record |
|---|---|---|---|---|---|
| PointSSM | ScanNet | 2 | 78.1% | link | link |
| PointSSM | ScanNet200 | 2 | 35.7% | link | link |
| PointSSM | S3DIS (Area5) | 2 | 72.8% | link | link |
| Model | Benchmark | Num GPUs | Val mIoU | Tensorboard | Exp Record |
|---|---|---|---|---|---|
| PointSSM | nuScenes | 2 | 80.7% | link | link |
| PointSSM | SemanticKITTI | 2 | 70.8% | – | – |
| PointSSM | DALES | 2 | 82.3% | - | - |
- Scannet,S3DIS datasets can be downloaded following by (https://huggingface.co/Pointcept/datasets)
- DALES can be downloaded at (https://drive.google.com/file/d/1Ta5Hg7e1dyUCSaDRCfAt_sDyisBuoikw/view?usp=sharing)
Our database builds on Pointcept codebase.
- Ubuntu: 18.04 and above.
- CUDA: 11.8 and above.
- PyTorch: 1.10.0 and above.
conda create -n pointssm python=3.8 -y
conda activate pointssm
conda install ninja -y
# Choose version you want here: https://pytorch.org/get-started/previous-versions/
conda install pytorch==1.12.1 torchvision==0.13.1 torchaudio==0.12.1 cudatoolkit=11.3 -c pytorch -y
conda install h5py pyyaml -c anaconda -y
conda install sharedarray tensorboard tensorboardx yapf addict einops scipy plyfile termcolor timm -c conda-forge -y
conda install pytorch-cluster pytorch-scatter pytorch-sparse -c pyg -y
pip install torch-geometric
# spconv (SparseUNet)
# refer https://github.com/traveller59/spconv
pip install spconv-cu113
# PPT (clip)
pip install ftfy regex tqdm
pip install git+https://github.com/openai/CLIP.git
# PTv1 & PTv2 or precise eval
cd libs/pointops
# usual
python setup.py install
# docker & multi GPU arch
TORCH_CUDA_ARCH_LIST="ARCH LIST" python setup.py install
# e.g. 7.5: RTX 3000; 8.0: a100 More available in: https://developer.nvidia.com/cuda-gpus
TORCH_CUDA_ARCH_LIST="7.5 8.0" python setup.py install
cd ../..
# Open3D (visualization, optional)
pip install open3d
# Mamba-ssm
pip install mamba-ssm==1.0.1Please follow the Pointcept codebase.
We thank the authors of Point Transformer V3. Our implementation is heavily built upon their codes.
