GET3D (Nvidia)
GET3D is NVIDIA’s AI model that generates detailed, textured 3D meshes directly from 2D images. Ideal for gaming, animation, and virtual world creation—no 3D scanning required.
Pricing options
- Free
About GET3D
What is GET3D?
GET3D is an advanced generative model developed by NVIDIA that creates high-quality, textured 3D meshes directly from 2D image collections. Unlike traditional 3D modeling pipelines that require scans, sensors, or CAD tools, GET3D leverages deep learning to generate complex 3D objects—ready to use in animation, games, and virtual production.
A Leap in 3D Content Creation
Trained using adversarial learning and differentiable rendering, GET3D can produce diverse objects with realistic textures and geometry. It outputs meshes with high fidelity, arbitrary topology, and intricate material details, bridging the gap between AI and production-ready 3D assets.
How GET3D Works
Latent Space Representation
GET3D generates two distinct latent codes: one for shape (geometry) and another for texture. These are used to produce a signed distance field (SDF) and a texture field that define the 3D mesh and surface appearance.
Mesh Extraction & Texturing
Using DMTet (Deep Marching Tetrahedra), GET3D converts the SDF into a triangular mesh. Then, it queries the texture field to paint the mesh with detailed color and material features.
Training with 2D Discriminators
GET3D is trained using 2D images and silhouettes with adversarial losses. Differentiable rendering allows the model to backpropagate errors from image space into 3D space, enabling learning without explicit 3D supervision.
Key Capabilities of GET3D
High-Quality 3D Meshes
GET3D generates textured 3D objects with fine details such as headlights, seams, fur, and reflections—making it suitable for animation and simulation tasks.
Arbitrary Topology Support
Unlike many earlier models, GET3D can generate complex, non-rigid shapes across a wide range of categories including animals, vehicles, furniture, shoes, and human avatars.
Disentangled Control of Shape & Texture
GET3D separates geometry and texture into distinct latent codes. Users can independently manipulate either aspect to achieve greater control in asset generation.
Latent Code Interpolation
By interpolating between latent vectors, GET3D enables smooth transitions and morphing between shapes and textures. This feature is useful for animation, asset variation, and design iteration.
Text-Guided Generation
Incorporating CLIP-based directional loss (as seen in StyleGAN-NADA), GET3D supports text-guided shape generation. Users can fine-tune outputs using natural language prompts for creative control.
Material and Lighting Effects
When combined with DIBR++ (a hybrid renderer), GET3D can also simulate materials and lighting effects in an unsupervised fashion, enhancing realism in renders.
Applications of GET3D
Gaming & Interactive Media
Game developers can rapidly generate character models, props, and environments with consistent geometry and texture, significantly reducing manual modeling time.
Animation & Film Production
GET3D enables fast prototyping of stylized or photorealistic assets with flexible design variation and direct export into rendering pipelines.
Virtual Reality & Metaverse
Ideal for VR creators, GET3D provides a scalable way to populate virtual spaces with high-quality 3D content—without the need for traditional scanning or modeling.
3D E-Commerce & Digital Twins
Retailers and industrial designers can use GET3D to visualize products in 3D from catalog images, enhancing interactive shopping and simulation workflows.
Research Highlights
- Disentangled Geometry and Texture: Independent control of mesh shape and surface appearance.
- Adversarial Image-Based Training: No 3D labels or models required—just image collections.
- Latent Code Interpolation: Smooth transitions between different shapes and styles.
- High Compatibility: Outputs standard mesh formats compatible with Blender, Unity, Unreal, and other engines.
Project Origins & Contributors
GET3D is the result of collaborative research between NVIDIA, the University of Toronto, and the Vector Institute, presented at NeurIPS 2022. It builds on prior work like DMTet, EG3D, and DIBR++, further advancing 3D generative modeling.
Resources and Access
- GET3D GitHub & Codebase
- Research Paper PDF & arXiv
- Citation & BibTeX Information Available on Project Page
GET3D (Nvidia) Reviews
Share your experience of using GET3D (Nvidia) with us!
Pricing options
- Free