August 21st, 2024

Degas: Detailed Expressions on Full-Body Gaussian Avatars

DEGAS introduces a method for creating full-body avatars with detailed facial expressions using 3D Gaussian Splatting, integrating body motion and expressions through a conditional variational autoencoder and 2D images.

Read original articleLink Icon
ExcitementCuriosityApprehension
Degas: Detailed Expressions on Full-Body Gaussian Avatars

DEGAS (Detailed Expressions on Full-Body Gaussian Avatars) is a novel modeling method that utilizes 3D Gaussian Splatting (3DGS) to create full-body avatars capable of exhibiting rich facial expressions. Despite advancements in neural rendering for lifelike avatars, the integration of detailed expressions into full-body models has been underexplored. DEGAS addresses this gap by employing a conditional variational autoencoder trained on multiview videos, which captures both body motion and facial expressions. Unlike traditional methods that rely on 3D Morphable Models (3DMMs), DEGAS uses an expression latent space derived from 2D portrait images, effectively linking 2D talking faces with 3D avatars. The resulting avatars can be reenacted to produce photorealistic images with nuanced facial expressions. The method's effectiveness is demonstrated through experiments on existing datasets and a newly introduced DREAMS Avatar Dataset, which includes multi-view captures of six subjects performing various expressions and motions. Additionally, an audio-driven extension of the method is proposed, leveraging 2D talking faces to enhance interactivity in AI agents.

- DEGAS is the first method to model full-body avatars with detailed facial expressions using 3D Gaussian Splatting.

- The approach integrates body motion and facial expressions through a conditional variational autoencoder.

- It utilizes an expression latent space based on 2D images, bridging 2D and 3D avatar technologies.

- The DREAMS Avatar Dataset features multi-view captures of subjects performing standard and freestyle motions.

- An audio-driven extension of DEGAS opens new avenues for interactive AI applications.

AI: What people are saying
The comments on the article about DEGAS's avatar creation technology reveal a mix of excitement and concern regarding its implications.
  • Many users express enthusiasm for the potential of creating realistic avatars for various applications, including video meetings and content creation.
  • There are concerns about the ethical implications of deepfake technology and the ease of impersonation it may enable.
  • Questions arise about the technical requirements for using the technology, such as data needs and real-time processing capabilities.
  • Some commenters note the impressive realism achieved, while also pointing out existing weaknesses, particularly in cloth simulation.
  • Overall, the discussion reflects a blend of fascination with the technology and caution regarding its misuse.
Link Icon 10 comments
By @olivierduval - 8 months
It's so amazing that it's frightening !!!!

With that kind of technology, what are the key problem still to be solved before being massively applied to deepfakes ? More specifically:

- how much datas (pictures or video) of the "target" is needed to use this ? Does it requires a specific lighting, a lot of different poses... or is it possible to just use some "online" videos (found on tiktok for example) or to record the "target" in the street with a phone ? How is it to create a "virtual doppelganger" ?

- when there is a "target" model, is it possible to use this in realtime ? How much power would it need ? A small laptop ? A big machine in the cloud ? Only a state-sponsored infrastructure ?

It looks like this technology has a real potential to "impersonate" anybody really efficiently

By @instagraham - 8 months
> see project > excited to try it > code (coming soon) > bookmark project > forget about project

Barring the obvious deepfake implications though, I'd be excited to see a new era of SFM-style content made with this

By @jimmySixDOF - 8 months
So much cool work bringing realism to Gaussian sources I think avatar mediated collaboration will get across the uncanny valley so where you are and how well you can communicate are not related anymore.

Also if you like Degas, this is another state of the art project in progress called VR-GS: A Physical Dynamics-Aware Interactive Gaussian Splatting System in Virtual Reality

https://yingjiang96.github.io/VR-GS/

By @SiempreViernes - 8 months
The second to last video shows a Gaussian splatting advantage I didn't think of: when gaussians clip into each other the failure is more gradual than when polygons do it.
By @unraveller - 8 months
Realism in more hands will make everyone a loveable artist. I look forward to being able to jump scare people with The Mummy face in video meetings until the last of their believability is shot.
By @MPSimmons - 8 months
Still some weaknesses with cloth but that's really impressive
By @jgord - 8 months
scary good demos.