Static Segmentation by Tracking: A Frustratingly Label-Efficient Approach to Fine-Grained Segmentation
Zhenyang Feng, Zihe Wang, Saul Ibaven Bueno, Tomasz Frelek, Advikaa Ramesh, Jingyan Bai, Lemeng Wang, Zanming Huang, Jianyang Gu, Jinsu Yoo, Tai-Yu Pan, Arpita Chowdhury, Michelle Ramirez, Elizabeth G Campolongo, Matthew J Thompson, Christopher G. Lawrence, Sydne Record, Neil Rosser, Anuj Karpatne, Daniel Rubenstein, Hilmar Lapp, Charles V. Stewart, Tanya Berger-Wolf, Yu Su, Wei-Lun Chao
- Release inference code
- Release beetle part segmentation dataset
- Release online demo
- Release one-shot fine-tuning (OC-CCL) code
- Release trait retrieval code
- Release butterfly trait segmentation dataset
Set CUDA_HOME to your cuda path (this is for grounding DINO)
For example:
export CUDA_HOME=/usr/local/cuda
Then sync uv packages:
uv sync
Download weights into checkpoints folder:
For wget
cd checkpoints
wget https://dl.fbaipublicfiles.com/segment_anything_2/072824/sam2_hiera_large.pt
wget -q https://github.com/IDEA-Research/GroundingDINO/releases/download/v0.1.0-alpha/groundingdino_swint_ogc.pth
For curl:
cd checkpoints
curl https://dl.fbaipublicfiles.com/segment_anything_2/072824/sam2_hiera_large.pt --output sam2_hiera_large.pt
curl https://github.com/IDEA-Research/GroundingDINO/releases/download/v0.1.0-alpha/groundingdino_swint_ogc.pth --output groundingdino_swint_ogc.pth
Go to the SAM demo, upload a representative image (e.g., img001.png), click the portions to segment, and select "Cut out object" from the sidebar. Right click and save the extraction (img001_extracted.png).
See the two examples below:
img001.png |
img001_extracted.png |
|---|---|
![]() |
![]() |
Then run the following two commands to generate the mask (like a guide for the model in segmentation shape--note the final processed image will appear to be an all black image):
uv run python src/sst/get_mask_from_crop.py \
--image_path img001.png \
--image_crop_path img001_extracted.png \
--mask_image_path_out img001_extracted_processed.png
Example output:
img001_extracted_processed.png |
|---|
![]() |
uv run python src/sst/prepare_starter_mask.py \
--mask_image_path img001_extracted_processed.png \
--mask_image_path_out img001_extracted_processed.png
Example output (NOTE: the color is very faint):
img001_extracted_processed.png |
|---|
![]() |
Now that the mask has been generated, the following command can be run to segment your remaining images.
uv run python src/sst/segment_and_crop.py \
--support_image img001.png \
--support_mask img001_extracted_processed.png \
--query_images [PATH_TO_IMAGE_DIRECTORY] \
--output [PATH_TO_SEGMENTED_OUTPUT_DIRECTORY]
The above script is RAM intensive on large datasets. To process individually run the above with src/sst/segment_and_crop_individual.py
For one-shot trait/part segmentation, please run the following demo code:
python code/segment.py --support_image /path/to/sample/image.png \
--support_mask /path/to/greyscale_mask.png \
--query_images /path/to/query/images/folder \
--output /path/to/output/folder \
--output_format "png" # png or gif, optionalFor trait-based retrieval, please refer to the demo code below:
python code/trait_retrieval.py --support_image /path/to/sample/image.png \
--support_mask /path/to/greyscale_mask.png \
--trait_id 1 \ # target trait to retrieve, denote by the value in support mask \
--query_images /path/to/query/images/folder \
--output /path/to/output/folder \
--output_format "png" \ # png or gif, optional
--top_k 5 # n top retrievals to save as resultsBeetle part segmentation dataset is available here.
Butterfly trait segmentation dataset can be accessed here.
The instructions and appropriate citations for these datasets are provided in the Citation section of their respective READMEs.
This project makes use of the SAM2 and GroundingDINO codebases. We are grateful to the developers and maintainers of these projects for their contributions to the open-source community. We thank LoRA for their great work.
We also thank David Carlyn for his contributions to improving the repository’s ease of setup, workflows, and overall usability; and Sam Stevens for developing a nice interactive tool for mask generation, selection, and visualization.
If you find our work helpful for your research, please consider citing using the following BibTeX entry:
@misc{feng2025staticsegmentationtrackingfrustratingly,
title={Static Segmentation by Tracking: A Frustratingly Label-Efficient Approach to Fine-Grained Segmentation},
author={Zhenyang Feng and Zihe Wang and Saul Ibaven Bueno and Tomasz Frelek and Advikaa Ramesh and Jingyan Bai and Lemeng Wang and Zanming Huang and Jianyang Gu and Jinsu Yoo and Tai-Yu Pan and Arpita Chowdhury and Michelle Ramirez and Elizabeth G. Campolongo and Matthew J. Thompson and Christopher G. Lawrence and Sydne Record and Neil Rosser and Anuj Karpatne and Daniel Rubenstein and Hilmar Lapp and Charles V. Stewart and Tanya Berger-Wolf and Yu Su and Wei-Lun Chao},
year={2025},
eprint={2501.06749},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2501.06749},
}



