New Meta FAIR Research and Models
Meta FAIR has released new research artifacts, including Meta Motivo for humanoid agents, Meta Video Seal for video watermarking, and frameworks like Flow Matching and Large Concept Model to enhance AI capabilities.
Read original articleMeta FAIR has announced the release of several new research artifacts aimed at advancing machine intelligence. Key innovations include Meta Motivo, a foundation model for controlling virtual agents, and Meta Video Seal, an open-source model for video watermarking. These releases are part of Meta's commitment to democratizing access to cutting-edge technologies and fostering collaboration within the research community. Meta Motivo utilizes unsupervised reinforcement learning to enable humanoid agents to perform complex tasks with human-like behaviors, demonstrating robustness to environmental changes. Meta Video Seal provides a framework for embedding imperceptible watermarks in videos, enhancing content traceability and security against manipulation. Additionally, Meta introduced Flow Matching, a generative AI framework that improves performance across various modalities, and the Large Concept Model (LCM), which decouples reasoning from language representation for better hierarchical thinking. Other notable releases include the Dynamic Byte Latent Transformer for tokenizer-free processing, Memory Layers for improved factuality in language models, and an evaluation toolbox for text-to-image generative models. These initiatives reflect Meta's ongoing efforts to promote responsible AI development and encourage community engagement in advancing AI technologies.
- Meta FAIR has released new models and datasets to enhance machine intelligence.
- Meta Motivo enables humanoid agents to perform tasks with human-like behaviors.
- Meta Video Seal offers a watermarking solution for video content security.
- Flow Matching and Large Concept Models improve generative AI capabilities.
- Meta emphasizes collaboration and responsible AI development in its research initiatives.
Related
Big tech wants to make AI cost nothing
Meta has open-sourced its Llama 3.1 language model for organizations with fewer than 700 million users, aiming to enhance its public image and increase product demand amid rising AI infrastructure costs.
Sam 2: The next generation of Meta Segment Anything Model
Meta has launched SAM 2, a real-time object segmentation model for images and videos, enhancing accuracy and reducing interaction time. It supports diverse applications and is available under an Apache 2.0 license.
Meta Movie Gen
Meta has launched Movie Gen, an AI model for creating and editing high-definition videos from text inputs, allowing personalized content generation and sound integration while emphasizing responsible AI development.
Meta Permits Its A.I. Models to Be Used for U.S. Military Purposes
Meta has shifted its policy to allow U.S. government and contractors to use its AI models for military purposes, emphasizing responsible use while collaborating with defense firms amid potential scrutiny.
Meta unveils a new, more efficient Llama model
Meta has launched the Llama 3.3 70B generative AI model, outperforming competitors while reducing costs. The company is investing $10 billion in AI infrastructure amid regulatory challenges in the EU.
Here they are presented as separate things, each of which apparently improves quality / efficiency. I wonder what the quality / efficiency increase is of all those methods put together? Maybe that’s what Llama 4 will be?
This looks like a lot of innovation is happening at Meta in those areas, really cool!
https://i.imgur.com/O5hGMo5.gif
Then the "Meta Explore Theory of Mind" is even more interesting. There was a thread about a month ago in which some of us were discussing some of the concepts here like "beliefs" and updating a model of the world accordingly. https://news.ycombinator.com/item?id=42035985
Interesting that it's a a hierarchical structure but only two levels of hierarchy. Stacking more levels seems like an obvious direction for further research.
He's recorded the full talk here now: https://www.youtube.com/watch?v=S5l5OvJ01ws
I had missed how much Meta have been doing on reasoning, ToM etc.
My initial thought is that they want to please/distract the doomers, but I’m prolly just self-centered!
Yet again interesting progress.
Also I like the idea of using the pose model to generate not a NPC but a avatar living in my phone or glas cube as a hologram. That would be quite scifi futuristic
Related
Big tech wants to make AI cost nothing
Meta has open-sourced its Llama 3.1 language model for organizations with fewer than 700 million users, aiming to enhance its public image and increase product demand amid rising AI infrastructure costs.
Sam 2: The next generation of Meta Segment Anything Model
Meta has launched SAM 2, a real-time object segmentation model for images and videos, enhancing accuracy and reducing interaction time. It supports diverse applications and is available under an Apache 2.0 license.
Meta Movie Gen
Meta has launched Movie Gen, an AI model for creating and editing high-definition videos from text inputs, allowing personalized content generation and sound integration while emphasizing responsible AI development.
Meta Permits Its A.I. Models to Be Used for U.S. Military Purposes
Meta has shifted its policy to allow U.S. government and contractors to use its AI models for military purposes, emphasizing responsible use while collaborating with defense firms amid potential scrutiny.
Meta unveils a new, more efficient Llama model
Meta has launched the Llama 3.3 70B generative AI model, outperforming competitors while reducing costs. The company is investing $10 billion in AI infrastructure amid regulatory challenges in the EU.