Sella_Blended_Point_Cloud_Diffusion_for_Localized_Text-guided_Shape_Editing@ICCV2025@CVF

Total: 1

#1 Blended Point Cloud Diffusion for Localized Text-guided Shape Editing [PDF] [Copy] [Kimi] [REL]

Authors: Etai Sella, Noam Atia, Ron Mokady, Hadar Averbuch-Elor

Natural language offers a highly intuitive interface for enabling localized fine-grained edits of 3D shapes. However, prior works face challenges in preserving global coherence while locally modifying the input 3D shape. In this work, we introduce an inpainting-based framework for editing shapes represented as point clouds. Our approach leverages foundation 3D diffusion models for achieving localized shape edits, adding structural guidance in the form of a partial conditional shape, ensuring that other regions correctly preserve the shape's identity. Furthermore, to encourage identity preservation also within the local edited region, we propose an inference-time coordinate blending algorithm which balances reconstruction of the full shape with inpainting at a progression of noise levels during the inference process. Our coordinate blending algorithm seamlessly blends the original shape with its edited version, enabling a fine-grained editing of 3D shapes, all while circumventing the need for computationally expensive and often innacurate inversion. Extensive experiments show that our method outperforms alternative techniques across a wide range of metrics that evaluate both fidelity to the original shape and also adherence to the textual description. We will release our code and trained models.

Subject: ICCV.2025 - Highlight