Drag Your Gaussian: Effective Drag-Based Editing with
Score Distillation for 3D Gaussian Splatting
Ministry of Education of China, Xiamen University.
2Baidu Inc.
Dragging results of DYG: The left video displays the original 3D scene and user input,
while the right video presents the results after the dragging edit.
TL;DR
Pipeline
The overall framework of DYG. Left: Given a 3D Gaussian scene, users provide 3D masks and several pairs of control points as input. Top-right: The Smooth Geometric Editing module predicts positional offsets for 3D Gaussians, resolving the issue of sparse distributions within the target region while ensuring seamless local editing. We adopt a two-stage training strategy: the first stage constructs the geometric scaffold of the edited Gaussians, and the second stage refines the texture details. Bottom-right: In the Score Distillation Guidance Module, to ensure stable optimization, 3D control points are projected onto 2D control points for a specified viewpoint. The RGB image and 2D mask, rendered from the mirrored initial 3D Gaussians, are encoded into point embeddings (P-Emb) and appearance embeddings (A-Emb), which act as conditions for the drag-based LDM. This process leverages our proposed Drag-SDS loss function to enable flexible and view-consistent 3D drag-based editing.
Video Comparison
Below shows the comparison of our results with other 3D Gaussian scene editing methods.
More Results
Multi-Round Edit
Our method supports multi-round dragging on different objects and the same objects.
Multi-round Dragging on Different Objects |
Multi-round Dragging on Same Objects |
Generalizability
Besides editing real scenes, our method can also edit recent text-to-3D generation results.
Acknowledgements
The website template was borrowed from Michaël Gharbi and MipNeRF360.