We provide converted datasets in our paper, You can use directly train on datasets from hugging face link
data
|____bear
|____lerf
| |____figurines
|____mipnerf360
| |____counter
For your custom dataset, you can follow this step to create masks for training. If you want to prepare masks on your own dataset, you will need DEVA python environment and checkpoints.
<location>
|---input
|---<image 0>
|---<image 1>
|---...
Firstly, convert initial camera pose and point cloud with colmap
python convert.py -s <location>
Then, convert SAM associated object masks. Note that the quality of converted-mask will largely affect the results of 3D segmentation and editing. And getting the mask is very fast. So it is best to adjust the parameters of anything segment first to make the mask as consistent and reasonable as possible from multiple views.
Example1. Bear dataset
bash script/prepare_pseudo_label.sh bear 1
Example2. figurines dataset
bash script/prepare_pseudo_label.sh lerf/figurines 1
Example3. counter dataset
bash script/prepare_pseudo_label.sh mipnerf360/counter 2
For Gaussian Grouping training and segmentation rendering of trained 3D Gaussian Grouping model:
Example1. Bear dataset
bash script/train.sh bear 1
Example2. figurines dataset
bash script/train_lerf.sh lerf/figurines 1
Example3. counter dataset
bash script/train.sh mipnerf360/counter 2