Zhe Zhu1, Le Wan2, Rui Xu3, Yiheng Zhang4, Honghua Chen5, Zhiyang Dou3, Cheng Lin6, Yuan Liu2†, Mingqiang Wei1†
† Corresponding authors
Affiliations: 1 Nanjing University of Aeronautics and Astronautics 2 Hong Kong University of Science and Technology 3 The University of Hong Kong 4 National University of Singapore 5 Lingnan University 6 Macau University of Science and Technology
Segmenting 3D objects into parts is a long-standing challenge in computer vision. To overcome taxonomy constraints and generalize to unseen 3D objects, recent works turn to open-world part segmentation. These approaches typically transfer supervision from 2D foundation models, such as SAM, by lifting multi-view masks into 3D. However, this indirect paradigm fails to capture intrinsic geometry, leading to surface-only understanding, uncontrolled decomposition, and limited generalization. We present PartSAM, the first promptable part segmentation model trained natively on large-scale 3D data. Following the design philosophy of SAM, PartSAM employs an encoder-decoder architecture in which a triplane-based dual-branch encoder produces spatially structured tokens for scalable part-aware representation learning. To enable large-scale supervision, we further introduce a model-in-the-loop annotation pipeline that curates over five million 3D shape-part pairs from online assets, providing diverse and fine-grained labels. This combination of scalable architecture and diverse 3D data yields emergent open-world capabilities: with a single prompt, PartSAM achieves highly accurate part identification, and in a Segment-Every-Part mode, it automatically decomposes shapes into both surface and internal structures. Extensive experiments show that PartSAM outperforms state-of-the-art methods by large margins across multiple benchmarks, marking a decisive step toward foundation models for 3D part understanding. Our code and model will be released soon.
- Release inference code of PartSAM
- Realse the pre-trained models
If you find this work useful, please cite our paper:
@article{zhu2025partsam,
title={PartSAM: A Scalable Promptable Part Segmentation Model Trained on Native 3D Data},
author={Zhe Zhu and Le Wan and Rui Xu and Yiheng Zhang and Honghua Chen and Zhiyang Dou and Cheng Lin and Yuan Liu and Mingqiang Wei},
journal={arXiv preprint arXiv:2509.21965},
year={2025}
}