DreamCraft3D++: Efficient Hierarchical 3D Generation with Multi-Plane Reconstruction Model

submited by
Style Pass
2024-10-27 11:30:05

We present DreamCraft3D++, an enhanced iteration of DreamCraft3D, a multi-stage 3D generation framework that enables efficient, high-quality production of complex 3D assets. DreamCraft3D++ significantly improves upon its predecessor in both generation speed and quality, reducing creation time to 10 minutes — a 20-fold acceleration. While retaining the multi-stage generation process, DreamCraft3D++ introduces two key innovations: (1) A feed-forward multi-plane based reconstruction model replaces the time-consuming geometry sculpting optimization, achieving a 1000x speedup in this stage. (2) For texture refinement, we propose a novel training-free IP-Adapter module that dynamically selects embeddings based on camera position, enhancing texture and geometry consistency. This approach provides a 4x faster alternative to DreamCraft3D's DreamBooth fine-tuning. Compared to current LRM-based methods, DreamCraft3D++ achieves a substantial improvement in both texture and geometric quality. Extensive experiments across diverse datasets demonstrate DreamCraft3D++'s superior capability in generating creative 3D assets with intricate geometry and realistic 360° textures, outperforming state-of-the-art image-to-3D methods. To foster further advancements in 3D content creation, we will open-source the complete implementation of DreamCraft3D++.

Given a single input image, Dreamcraft3D++ processes it with multi-view diffusion models to generate orthogonal, consistent views and normal maps. A feed-forward sparse-view 3D reconstruction model then infers textured meshes from these multi-view images, using a convolutional U-Net to map the input to non-orthogonal planes, which are subsequently decoded into Flexicubes. Finally, a training-free object-aware diffusion prior enhances high-frequency geometry and texture details through score distillation.

Leave a Comment