August 8th, 2024

GPUDrive: Data-driven, multi-agent driving simulation at 1M FPS

The paper introduces GPUDrive, a GPU-accelerated simulator that generates over a million experience steps per second, enhancing multi-agent planning and training reinforcement learning agents using the Waymo Motion dataset.

Read original articleLink Icon
GPUDrive: Data-driven, multi-agent driving simulation at 1M FPS

The paper titled "GPUDrive: Data-driven, multi-agent driving simulation at 1 million FPS" presents a new GPU-accelerated simulator designed to enhance multi-agent planning through rapid experience generation. The authors, Saman Kazemkhani and colleagues, highlight that traditional multi-agent learning algorithms require extensive experience, often in the billions of steps, which has limited their practical application. GPUDrive, built on the Madrona Game Engine, addresses this bottleneck by generating over a million steps of experience per second. The simulator allows for the definition of complex agent behaviors using C++, which are then optimized for performance with CUDA. The authors demonstrate the effectiveness of GPUDrive by training reinforcement learning agents on the Waymo Motion dataset, achieving rapid training times for both individual scenes and general capabilities. The trained agents are made available as part of the code base, promoting further research and development in multi-agent systems.

- GPUDrive can generate over a million steps of experience per second for multi-agent simulations.

- The simulator is built on the Madrona Game Engine and utilizes C++ and CUDA for performance optimization.

- It addresses the challenge of extensive experience requirements in traditional multi-agent learning algorithms.

- The authors successfully trained reinforcement learning agents using the Waymo Motion dataset.

- The trained agents are included in the code base for public access and further research.

Related

Show HN: UNet diffusion model in pure CUDA

Show HN: UNet diffusion model in pure CUDA

The GitHub content details optimizing a UNet diffusion model in C++/CUDA to match PyTorch's performance. It covers custom convolution kernels, forward pass improvements, backward pass challenges, and future optimization plans.

The History of Machine Learning in Trackmania

The History of Machine Learning in Trackmania

Trackmania, a racing game, inspires machine learning projects aiming for superhuman performance. Players like Rottaca and TMRL use supervised and unsupervised learning. Innovations include track lookahead and diverse learning algorithms. Programs show progress but not human-level skills. The goal is a top-performing AI on a desktop PC, blending gaming and machine learning for innovation.

Canada-based Artificial Agency raises $16M to alter NPC behaviour in Games

Canada-based Artificial Agency raises $16M to alter NPC behaviour in Games

Artificial Agency, a Canadian firm, raised $16 million to advance an AI behavior engine for video game NPCs. Led by ex-Google DeepMind experts, the company aims to innovate NPC interactions, diverging from traditional scripted approaches.

Four co's are hoarding billions worth of Nvidia GPU chips. Meta has 350K of them

Four co's are hoarding billions worth of Nvidia GPU chips. Meta has 350K of them

Meta has launched Llama 3.1, a large language model outperforming ChatGPT 4o on some benchmarks. The model's development involved significant investment in Nvidia GPUs, reflecting high demand for AI training resources.

The Path to StyleGan2 – Implementing the Progressive Growing GAN

The Path to StyleGan2 – Implementing the Progressive Growing GAN

The article outlines the implementation of Progressive Growing GAN (PGGAN) for generating high-resolution images, focusing on a two-phase approach and key concepts to enhance training efficiency and image quality.

Link Icon 4 comments
By @cs702 - 8 months
Unless I'm missing something big, this looks like a significant deal for independent developers of self-driving AI software: GPUDrive enables them to run driving simulations with hundreds of AI agents on consumer-grade GPUs at 1M FPS, and it comes with Python bindings, wrappers for Pytorch and Jax, and a friendly standard MIT license. Thank you for sharing this on HN!
By @foota - 8 months
Is this just the location data being trained on, or is there image and sensor input data too? It looks like it's just location, which seems like it limits the applicability, but I'm not sur

Edit: reading a bit more it's somewhere in between. Afaict no raw sensor data etc.,. but different "parsed" sensor inputs are supported. I'm not sure whether this is synthetic or not? E.g., is the LIDAR view real LIDAR data from some system or a processed result of what the system thinks LIDAR would be able to see? I can't tell.

By @BetterWhisper - 8 months
By @toppy - 8 months
I don't know this field of research thus my question: why such a high framerate is consider as a feature at all? Does it help with learning rate?