MeshAnything – Converts 3D representations into efficient 3D meshes
MeshAnything efficiently generates high-quality Artist-Created Meshes with optimized topology, fewer faces, and precise shapes. Its innovative approach enhances 3D industry applications by improving storage and rendering efficiencies.
Read original articleMeshAnything is a model designed to extract meshes from 3D representations, mimicking human artists' work. It can be integrated into 3D asset production pipelines to create Artist-Created Meshes efficiently. Compared to existing methods, MeshAnything generates meshes with significantly fewer faces, improving storage and rendering efficiencies while maintaining precision. The model consists of a VQ-VAE and a shape-conditioned decoder-only transformer, enabling shape-conditioned autoregressive mesh generation. By focusing on optimized topology rather than complex 3D shape distributions, MeshAnything reduces training complexity and enhances scalability. The approach produces meshes aligned with specified shapes, demonstrating better topology and fewer faces compared to ground truth. This indicates the model's ability to construct meshes efficiently without overfitting. MeshAnything's potential lies in enhancing 3D industry applications by providing high-quality, controllable Artist-Created Mesh generation.
Related
20x Faster Background Removal in the Browser Using ONNX Runtime with WebGPU
Using ONNX Runtime with WebGPU and WebAssembly in browsers achieves 20x speedup for background removal, reducing server load, enhancing scalability, and improving data security. ONNX models run efficiently with WebGPU support, offering near real-time performance. Leveraging modern technology, IMG.LY aims to enhance design tools' accessibility and efficiency.
Lessons About the Human Mind from Artificial Intelligence
In 2022, a Google engineer claimed AI chatbot LaMDA was self-aware, but further scrutiny revealed it mimicked human-like responses without true understanding. This incident underscores AI limitations in comprehension and originality.
Mip-Splatting: Alias-Free 3D Gaussian Splatting
The paper introduces Mip-Splatting, enhancing 3D Gaussian Splatting by addressing artifacts with a 3D smoothing filter and a 2D Mip filter, achieving alias-free renderings and improved image fidelity in 3D rendering applications.
I am using AI to drop hats outside my window onto New Yorkers
dropofahat.zone showcases a project by a New York City resident using AI to drop hats onto passersby from their window. Challenges include hat selection, window opening, and AI implementation. The goal is to redefine "Window Shopping."
Homegrown Rendering with Rust
Embark Studios develops a creative platform for user-generated content, emphasizing gameplay over graphics. They leverage Rust for 3D rendering, introducing the experimental "kajiya" renderer for learning purposes. The team aims to simplify rendering for user-generated content, utilizing Vulkan API and Rust's versatility for GPU programming. They seek to enhance Rust's ecosystem for GPU programming.
Weird custom non-commercial license unfortunately. Notes from the GitHub readme:
> It takes about 7GB and 30s to generate a mesh on an A6000 GPU
> trained on meshes with fewer than 800 faces and cannot generate meshes with more than 800 faces
Its an area where things can be improved a lot imho - I did some work a while back fitting flat planes to pointclouds, and ended up with mesh model anything from 40x to 100x smaller data than the ptcloud dataset. see quato.xyz for samples where you can compare the cloud, the mesh produced.. and view the 3D model in recent browsers.
My approach had some similarity to gaussian splats... but using only planar regions .. great for buildings made of flat slabs, less so for smooth curves and foliage.
Applying their MeshAnything algo to fine meshes from photogrammetry scans of buildings would be of great benefit - probably getting those meshes down to a size where they can be shared as 3D webgl/threejs pages.
Even deciding on triangle points to efficiently tesselate / cover a planar region with holes etc, is basically a knapsack problem, which heuristics, monte-carlo and ML can improve upon.
Still triangles rather than polygons, but we are getting closer.
The end goal should be:
1) Polygons, mostly 4 sided, rather than triangles.
2) Edge smoothness/creases to separate hard coders from soft corners. (Which when combined with polygons enables SubD support: https://graphics.pixar.com/opensubdiv/docs/subdivision_surfa...)
3) UV for textures that are aligned with the natural flow of textures on those components.
4) Repeating textures (although sometimes not) that work with the UVs and combine to create PBR textures. (Getting closer all the time: https://gvecchio.com/stablematerials/)
After the above works, I think people should move on to inferring proper CAD models from an image. Basically infer all the constraints and the various construction steps.
So the scale shown in this paper feels like toys! Not undermining the effort at all. We need to start somewhere anyway.
For the same reason, I feel puzzled looking at Industrial scenes in Video Games. They are like 3 order of magnitude simplified compared to a real plant.
The triangle topologies in this paper made don’t follow the logical loops that an artist would work as. Generally it’s rare an artist would work directly in triangles, versus quads. But that aside, you’d place the loops in more logical places along the surface.
The face and toilet really stand out to me as examples of meshes that look really off.
Anyway, I think this is a good attempt at a reasonable topology generation, but the tag line is a miss.
Beside good topology is dependent on the use case, it’s very different if you are doing animation, a 3D print, a game or just a render.
https://huggingface.co/spaces/Yiwen-ntu/MeshAnything
on the provided sample "hat". I tried with and without checking "Preprocess with marching cubes" and "Random Sample". Both outputs had holes in the output mesh where the original did not.
Am I doing this wrong, or is the algorithm buggy?
Sentiments aside, that's an impressive approach.
https://github.com/buaacyw/MeshAnything/blob/main/LICENSE.tx...
Related
20x Faster Background Removal in the Browser Using ONNX Runtime with WebGPU
Using ONNX Runtime with WebGPU and WebAssembly in browsers achieves 20x speedup for background removal, reducing server load, enhancing scalability, and improving data security. ONNX models run efficiently with WebGPU support, offering near real-time performance. Leveraging modern technology, IMG.LY aims to enhance design tools' accessibility and efficiency.
Lessons About the Human Mind from Artificial Intelligence
In 2022, a Google engineer claimed AI chatbot LaMDA was self-aware, but further scrutiny revealed it mimicked human-like responses without true understanding. This incident underscores AI limitations in comprehension and originality.
Mip-Splatting: Alias-Free 3D Gaussian Splatting
The paper introduces Mip-Splatting, enhancing 3D Gaussian Splatting by addressing artifacts with a 3D smoothing filter and a 2D Mip filter, achieving alias-free renderings and improved image fidelity in 3D rendering applications.
I am using AI to drop hats outside my window onto New Yorkers
dropofahat.zone showcases a project by a New York City resident using AI to drop hats onto passersby from their window. Challenges include hat selection, window opening, and AI implementation. The goal is to redefine "Window Shopping."
Homegrown Rendering with Rust
Embark Studios develops a creative platform for user-generated content, emphasizing gameplay over graphics. They leverage Rust for 3D rendering, introducing the experimental "kajiya" renderer for learning purposes. The team aims to simplify rendering for user-generated content, utilizing Vulkan API and Rust's versatility for GPU programming. They seek to enhance Rust's ecosystem for GPU programming.