There has been a lot of work involving using re-trained 2D vision models with 3D frameworks. We chose to build upon Lexicon3D Man et al. [2024], in which an ablation study demonstrated improved outcome in the semantic segmentation task by creating an ensembe of models involving LSeg, Stable Diffusion, and Swin3D. Weinvestigated feature based fusion strategies, including additive and interleaved approaches, to refine the Mixture of Features paradigm. By extending the concept to encompass semantic segmentation, our goal is to identify optimal fusion techniques that outperform existing benchmarks, leveraging diverse pre-trained embeddings to improve understanding in 3D spaces
The paper for the work done is: MoF-Paper.pdf
Please install the required packages and dependencies according to the requirements.txt
file.
In addition,
- in order to use the LSeg model, please follow this repo to install the necessary dependencies.
- in order to use the Swin3D model, please follow this repo and this repo to install the necessary dependencies.
Finally, please download the ScanNet dataset from the official website and follow the instructions here to preprocess the ScanNet dataset and get RGB video frames and point clouds for each scannet scene.
To extract features from the foundation models, please run the corresponding scripts in the lexicon3d
folder. For example, to extract features from the LSeg model, please run the following command:
python fusion_scannet_clip.py --data_dir dataset/ScanNet/openscene/ --output_dir dataset/lexicon3d/clip/ --split train --prefix clip
This script will extract features from the LSeg model for the ScanNet dataset. The extracted features will be saved in the output_dir
folder, containing the feature embeddings, points, and voxel grids.
This repo is built based on the fantastic work of Lexicon3D & OpenScene.