Home/Explore/3D Creation/wavespeed-ai/hunyuan3d/v2-base

image-to-3d

wavespeed-ai/hunyuan3d/v2-base

Hunyuan3D-V2-Base is a state-of-the-art image-to-3D generative model developed by Tencent and now available on WaveSpeedAI.

Doc

Hint: You can drag and drop a file or click to upload

preview

Idle

Your request will cost $0.16 per run.

For $10 you can run this model approximately 62 times.

ExamplesView all

README

Hunyuan3D-V2-Mini

Hunyuan3D-2 is an open-source 3D generation model series launched by Tencent. As of March 2025, supporting the generation of high-fidelity 3D models with high-resolution texture maps through text, image, or sketch inputs.

Technical Highlights

The system adopts a separated process of geometry generation + texture synthesis:

-Geometry Generation (Hunyuan3D-DiT): Based on a flow diffusion model that generates untextured 3D geometric models, with 2.6B parameters, capable of precisely extracting geometric information from input images or text.

  • Texture Synthesis (Hunyuan3D-Paint): Adds high-resolution (4K) textures to geometric models, with 1.3B parameters, supporting multi-view diffusion generation technology to ensure realistic textures and consistent lighting.
  • By decoupling shape and texture generation, it effectively reduces complexity and improves generation quality.

Performance and Efficiency Optimization

  • Fast Generation: Completes model generation in as fast as 30 seconds, with the accelerated version (Hunyuan3D-DiT-v2-0-Fast) shortening inference time by 50% through guidance distillation techniques.
  • Multi-modal Input: Supports various input methods including text descriptions, images, and sketches, compatible with Blender plugins and Gradio applications, lowering the usage threshold.
  • Open Source Model Ecosystem The project has open-sourced 6 models (some simplified versions), covering different scenario needs: