Point Cloud Part Editing: Segmentation, Generation, Assembly, and Selection
DOI:
https://doi.org/10.1609/aaai.v38i7.28547Keywords:
CV: 3D Computer Vision, ML: Deep Generative Models & AutoencodersAbstract
Ideal part editing should guarantee the diversity of edited parts, the fidelity to the remaining parts, and the quality of the results. However, previous methods do not disentangle each part completely, which means the edited parts will affect the others, resulting in poor diversity and fidelity. In addition, some methods lack constraints between parts, which need manual selections of edited results to ensure quality. Therefore, we propose a four-stage process for point cloud part editing: Segmentation, Generation, Assembly, and Selection. Based on this process, we introduce SGAS, a model for part editing that employs two strategies: feature disentanglement and constraint. By independently fitting part-level feature distributions, we realize the feature disentanglement. By explicitly modeling the transformation from object-level distribution to part-level distributions, we realize the feature constraint. Considerable experiments on different datasets demonstrate the efficiency and effectiveness of SGAS on point cloud part editing. In addition, SGAS can be pruned to realize unsupervised part-aware point cloud generation and achieves state-of-the-art results.Downloads
Published
2024-03-24
How to Cite
Zhang, K., Chen, Y., Yang, X., Zhang, W., & Jin, C. (2024). Point Cloud Part Editing: Segmentation, Generation, Assembly, and Selection. Proceedings of the AAAI Conference on Artificial Intelligence, 38(7), 7187-7195. https://doi.org/10.1609/aaai.v38i7.28547
Issue
Section
AAAI Technical Track on Computer Vision VI