Meta 3D Gen
Meta introduces Meta 3D Gen (3DGen), a fast text-to-3D asset tool with high prompt fidelity and PBR support. It integrates AssetGen and TextureGen components, outperforming industry baselines in speed and quality.
Read original articleMeta has introduced Meta 3D Gen (3DGen), a fast pipeline for text-to-3D asset generation. This new tool offers high prompt fidelity and quality 3D shapes and textures in under a minute, supporting physically-based rendering (PBR) for real-world applications. 3DGen also allows generative retexturing of 3D shapes using textual inputs. It integrates Meta 3D AssetGen and Meta 3D TextureGen components, achieving a 68% win rate compared to single-stage models. The tool outperforms industry baselines in prompt fidelity and visual quality for complex textual prompts while being faster. Authors of related publications include Raphael Bensadoun, Tom Monnier, Yanir Kleiman, and others. Meta 3D AssetGen focuses on text-to-mesh generation, while Meta 3D TextureGen specializes in texture generation for 3D objects. These advancements showcase Meta's commitment to innovation in graphics and computer vision research.
Related
MeshAnything – Converts 3D representations into efficient 3D meshes
MeshAnything efficiently generates high-quality Artist-Created Meshes with optimized topology, fewer faces, and precise shapes. Its innovative approach enhances 3D industry applications by improving storage and rendering efficiencies.
HybridNeRF: Efficient Neural Rendering
HybridNeRF combines surface and volumetric representations for efficient neural rendering, achieving 15-30% error rate improvement over baselines. It enables real-time framerates of 36 FPS at 2K×2K resolutions, outperforming VR-NeRF in quality and speed on various datasets.
Figma AI: Empowering designers with intelligent tools
Figma AI enhances designers' workflow with AI-powered features like Visual Search, Asset Search, text tools, and content generation. It aims to streamline tasks, boost efficiency, and spark creativity while prioritizing data privacy.
Declaratively build your APIs in TypeScript and Python
Metatype is a declarative API development platform focusing on creating reliable, modular components. It offers Typegraphs for composing APIs, Typegate for GraphQL/REST, and Meta CLI for live reloading. Users can access documentation and a quick start guide on the Metatype GitHub repository.
Our guidance on using LLMs (for technical writing)
The Ritza Handbook advises on using GPT and GenAI models for writing, highlighting benefits like code samples and overcoming writer's block. However, caution is urged against using GPT-generated text in published articles.
3D has an extremely steep learning curve once you try to do anything non-trivial, especially in terms of asset creation for VR etc. but my real interest is where this leads in terms of real-world items. One of the major hurdles is that in the real-world we aren't as forgiving as we are in VR/games. I'm not entirely surprised to see that most of the outputs are "artistic" ones, but I'm really interested to see where this ends up when we can give AI combined inputs from text/photos/LIDAR etc and have it make the model for a physical item that can be 3D printed.
[1] https://www.technicalchops.com/articles/ai-inputs-and-output...
That being said, I wonder if the use of signed distance fields (SDFs) results in bad topology.
I saw a paper earlier this week that was recently released that seems to build "game-ready" topology --- stuff that might actually be riggable for animation. https://github.com/buaacyw/MeshAnything
When the person then emerges from this virtual world, it'll be like an egg hatching into a new birth, having learned the lessons in their virtual cocoon.
If you don't like this idea, it's an interesting thought experiment regardless as we can't verify, we're not already in a form of this.
I'll use it to upscale 8x all meshes and textures in the original Mafia and Unreal Tournament, write a good bye letter to my family and disappear.
I think the kids will understand when they grow up.
Or, just throw a PS1 filter on top and make some retro games
- Image Input to 3D model Output
- 3D model(format) as Input
Question: What is the current state of the art commercially available product in that niche?
Well played, Meta.
Related
MeshAnything – Converts 3D representations into efficient 3D meshes
MeshAnything efficiently generates high-quality Artist-Created Meshes with optimized topology, fewer faces, and precise shapes. Its innovative approach enhances 3D industry applications by improving storage and rendering efficiencies.
HybridNeRF: Efficient Neural Rendering
HybridNeRF combines surface and volumetric representations for efficient neural rendering, achieving 15-30% error rate improvement over baselines. It enables real-time framerates of 36 FPS at 2K×2K resolutions, outperforming VR-NeRF in quality and speed on various datasets.
Figma AI: Empowering designers with intelligent tools
Figma AI enhances designers' workflow with AI-powered features like Visual Search, Asset Search, text tools, and content generation. It aims to streamline tasks, boost efficiency, and spark creativity while prioritizing data privacy.
Declaratively build your APIs in TypeScript and Python
Metatype is a declarative API development platform focusing on creating reliable, modular components. It offers Typegraphs for composing APIs, Typegate for GraphQL/REST, and Meta CLI for live reloading. Users can access documentation and a quick start guide on the Metatype GitHub repository.
Our guidance on using LLMs (for technical writing)
The Ritza Handbook advises on using GPT and GenAI models for writing, highlighting benefits like code samples and overcoming writer's block. However, caution is urged against using GPT-generated text in published articles.