World Emulation via DNN
Ollin Boer Bohan developed a neural network-based "neural world" for exploring a forest trail online, enhancing image generation through user controls and recorded video, envisioning it as a unique creative medium.
Read original articleOllin Boer Bohan has developed a neural network-based "neural world" that allows users to explore a forest trail through their web browser. This innovative approach utilizes a neural network to generate images based on previous frames and user controls, without traditional game development elements like level geometry or scripted animations. The project builds on Bohan's earlier work, where he created a 2D game world by training a neural network on gameplay videos. The new neural world was created by recording approximately 15 minutes of video while walking through a forest, which was then processed into input-output pairs for training. Despite initial challenges in achieving a coherent output, Bohan made several upgrades to the training process, including enhancing control inputs, increasing memory capacity, and refining the network architecture. The final model, which consists of around 5 million parameters, was trained on a dataset of 22,814 frames. Bohan argues that while traditional game worlds are crafted like paintings, neural worlds are akin to photographs, capturing real-life details directly from recorded data. He envisions a future where neural worlds can achieve high fidelity and become a distinct creative medium, similar to how photography evolved. Bohan encourages programmers interested in world modeling to explore existing frameworks and expresses a desire to continue improving this technology.
- Bohan created a neural world that generates images based on recorded video and user controls.
- The project builds on previous work with 2D game emulation using neural networks.
- Significant upgrades to the training process improved the quality of the generated world.
- Bohan compares neural worlds to photographs, emphasizing their basis in real-world recordings.
- He envisions future advancements that could make neural worlds a unique creative medium.
Related
Open and remotely accessible Neuroplatform for research in wetware computing
An open Neuroplatform for wetware computing research combines electrophysiology and AI with living neurons. It enables long-term experiments on brain organoids remotely, supporting complex studies for energy-efficient computing advancements.
Diffusion Models Are Real-Time Game Engines
GameNGen, developed by Google and Tel Aviv University, simulates DOOM in real-time at over 20 frames per second using a two-phase training process, highlighting the potential of neural models in gaming.
WonderWorld: Interactive 3D Scene Generation from a Single Image
WonderWorld is a novel framework from Stanford and MIT that generates interactive 3D scenes from a single image in under 10 seconds, allowing user-defined content and real-time navigation.
The Matrix: Infinite-Horizon World Generation with Real-Time Interaction
The Matrix project aims to create an immersive digital universe with real-time interactions, featuring AAA-level graphics, frame-level precision, and open-sourced data to promote further research in world simulation technology.
Google is building its own 'world modeling' AI team for games and robot training
Google DeepMind is creating a team to develop AI "world models" for simulating environments, enhancing video games and robot training, while competing with OpenAI and Nvidia in the AGI race.
- Many users express admiration for the innovative approach and potential applications of the project.
- There are inquiries about the technical knowledge required to create similar models and the feasibility of open-sourcing the training code.
- Some comments draw comparisons to existing technologies and express a desire for further exploration in related areas.
- Users appreciate the transparency regarding the challenges faced during development.
- Questions arise about the computational resources needed for training the models.
I don't get this analogy at all. Instead of a human information flows through a neural network which alters the information.
> Every lifelike detail in the final world is only there because my phone recorded it.
I might be wrong here but I don't think this is true. It might also be there because the network inferred that it is there based on previous data.
Imo this just takes the human out of a artistic process - creating video game worlds and I'm not sure if this is worth archiving.
What could go wrong?
Jokes aside, this is insanely cool!
More interesting is that you made an easy to use environment authoring tool that (I haven’t tried it yet) seems really slick.
Both of those are impressive alone but together that’s very exciting.
I didn't see it in an obvious place on your github, do you have any plans to open source the training code?
Is OP the blog’s author? Because in the post the author said that the purpose of the project is to show why NN are truly special and I wanted a more articulate view of why he/she thinks that? Good work anyway!
Imagine a similar technique but with productivity software.
And a pre-trained network that adapts quickly.
edit: I see now that you mention a pricepoint of 100 GPU-hours/roughly 100$. My mistake.
Related
Open and remotely accessible Neuroplatform for research in wetware computing
An open Neuroplatform for wetware computing research combines electrophysiology and AI with living neurons. It enables long-term experiments on brain organoids remotely, supporting complex studies for energy-efficient computing advancements.
Diffusion Models Are Real-Time Game Engines
GameNGen, developed by Google and Tel Aviv University, simulates DOOM in real-time at over 20 frames per second using a two-phase training process, highlighting the potential of neural models in gaming.
WonderWorld: Interactive 3D Scene Generation from a Single Image
WonderWorld is a novel framework from Stanford and MIT that generates interactive 3D scenes from a single image in under 10 seconds, allowing user-defined content and real-time navigation.
The Matrix: Infinite-Horizon World Generation with Real-Time Interaction
The Matrix project aims to create an immersive digital universe with real-time interactions, featuring AAA-level graphics, frame-level precision, and open-sourced data to promote further research in world simulation technology.
Google is building its own 'world modeling' AI team for games and robot training
Google DeepMind is creating a team to develop AI "world models" for simulating environments, enhancing video games and robot training, while competing with OpenAI and Nvidia in the AGI race.