Launch HN: Silurian (YC S24) – Simulate the Earth
Silurian, a team developing advanced weather forecasting models, introduced the Generative Forecasting Transformer (GFT), predicting weather up to 14 days ahead and showing promise in forecasting hurricane tracks for 2024.
Silurian, a team comprising Jayesh, Cris, and Nikhil, is developing foundation models to simulate Earth, focusing initially on weather forecasting. They highlight that traditional forecasting systems have improved at a rate of one day per decade, but recent advancements in GPUs and deep learning have accelerated this process. Notably, companies like NVIDIA, Google DeepMind, Huawei, and Microsoft have been researching deep learning systems for weather prediction, which can outperform traditional physics-based simulations. The founders, having previously led significant projects at Microsoft, aim to enhance these models to their full potential, ultimately simulating various infrastructures affected by weather, such as energy grids and agriculture. This summer, they introduced their own model, the Generative Forecasting Transformer (GFT), which can predict global weather up to 14 days ahead with high resolution. Despite limited historical extreme weather data, GFT has shown promising results in forecasting hurricane tracks for 2024. The team encourages users to explore their hurricane forecasts through their dedicated platform.
- Silurian focuses on advanced weather forecasting using deep learning models.
- Traditional forecasting has improved slowly, but new technologies are accelerating accuracy.
- The Generative Forecasting Transformer (GFT) can predict weather up to 14 days ahead.
- GFT has demonstrated strong performance in predicting hurricane tracks for 2024.
- The team aims to simulate various infrastructures impacted by weather conditions.
Related
El Niño forecasts extended to 18 months with physics-based model
Researchers at the University of Hawai'i at Mānoa introduce the XRO model, extending El Niño forecasts to 18 months. This physics-based approach enhances predictability and understanding of ENSO events and other climate variabilities.
AI helps to produce breakthrough in weather and climate forecasting
Artificial intelligence, exemplified by Google's NeuralGCM model, enhances weather forecasting by combining AI with traditional physics models. This collaboration improves speed, accuracy, and efficiency, marking a significant advancement in climate prediction.
Neural general circulation models for weather and climate
Neural general circulation models (GCMs) integrate machine learning with atmospheric dynamics for accurate weather and climate predictions. They show competitive performance, combining traditional GCM strengths with machine learning efficiency for improved forecasting.
Fast, accurate climate modeling with NeuralGCM
NeuralGCM is a new machine learning model for simulating Earth's atmosphere, providing accurate 2-15 day weather forecasts, outperforming traditional models, and being computationally efficient for broader climate predictions.
Artificial intelligence gives weather forecasters a new edge
AI is improving weather forecasting speed and accuracy, notably for hurricanes. DeepMind's GraphCast outperforms traditional models, providing timely warnings and enhancing forecasting accuracy while emphasizing the need for human oversight.
- Many commenters express enthusiasm for the potential of GFT and its innovative approach to weather forecasting.
- There are discussions about the comparison of GFT with existing models, particularly regarding the integration of physics in forecasting.
- Some users raise questions about the accuracy and reliability of AI-driven models compared to traditional methods.
- Several comments suggest collaboration opportunities and the sharing of metrics for better comparison with other models.
- Concerns about the business viability and competition in the weather forecasting space are also noted.
One nit on your framing: NeuralGCM (https://www.nature.com/articles/s41586-024-07744-y), built by my team at Google, is currently at the top of the WeatherBench leaderboard and actually builds in lots of physics :).
We would love to metrics from your model in WeatherBench for comparison. When/if you have that, please do reach out.
Haha. The old NLP saying "every time I fire a linguist, my performance goes up", now applies to the physicists....
What else do you hope to simulate, if this becomes successful?
Shameless plug: recently we've built a demo that allows you to search for objects in San Francisco using natural language. You can look for things like Tesla cars, dry patches, boats, and more. Link: https://demo.bluesight.ai/
We've tried using Clay embeddings but we quickly found out that they perform poorly for similarity search compared to embeddings produced by CLIP fine tuned on OSM captions (SkyScript).
Disclosure: I work there.
…we’ve built our own foundation model, GFT (Generative Forecasting Transformer), a 1.5B parameter frontier model that simulates global weather…
I’m constantly scolding people for trying to use LLMs for non-linguistic tasks, and thus getting deceptively disappointing results. The quintessential example is arithmetic, which makes me immediately dubious of a transformer built to model physics. That said, you’ve obviously found great empirical success already, so something’s working. Can you share some of your philosophical underpinnings for this approach, if they exist beyond “it’s a natural evolution of other DL tech”? Does your transformer operate in the same rough way as LLMs, or have you radically changed the architecture to better approach this problem? Hence: simulate the Earth.
When I read “simulate”, I immediately think of physics simulations built around interpretable/symbolic systems of elements and forces, which I would usually put in basic opposition to unguided/connectionist ML models. Why choose the word “simulate”, given that your models are essentially black boxes? Again, a pretty philosophical question that you don’t necessarily have to have an answer to for YC reasons, lolBest of luck, and thanks for taking the leap! Humanity will surely thank you. Hopefully one day you can claim a bit of the NWS’ $1.2B annual budget, or the US Navy’s $infinity budget — if you haven’t, definitely reach out to NRL and see if they’ll buy what you’re selling!
Oh and C) reach out if you ever find the need to contract out a naive, cheap, and annoyingly-optimistic full stack engineer/philosopher ;)
What will your differentiators be?
Are you paying for weather data products?
Once upon a time I converted spectral-transform-shallow-water-model (STSWM or parallelized as PSTSWM) from FORTRAN to Verilog. I believe this is the spectral-transform method we have run for the last 30 years to do forecasting. The forecasting would be ~20% different results for 10-day predictions if we truncated each operation to FP64 instead of Intel's FP80.
It seems like this is another instance of The Bitter Lesson, no?
Have specific industries reached out to you for your commerical potential – natural resource exploration, for example?
As a fellow deep learning modeler of Earth systems, I can also say that what they're doing really is 100% top notch. Congrats to the team and YC.
Using the full expressive power of a programming language to model the real world and then execute AI algorithms on highly structured and highly understood data seems like the right way to go!
1. How will you handle one-off events like volcanic eruptions for instance? 2. Where do you start with this too? Do you pitch a meteorology team? Is it like a "compare and see for yourself"?
I had a web app online in 2020-22 called Skim Day that predicted skimboarding conditions on California beaches that was mostly powered by weather APIs. The tide predictions were solid, but the weather itself was almost never right, especially wind speed. Additionally there were some missing metrics like slope of beach which changes significantly throughout the year and is very important for skimboarding.
Basically, I needed AI. And this looks incredible. Love your website and even the name and concept of "Generative Forecasting Transformer (GFT)" - very cool. I imagine the likes of Surfline, The Weather Channel, and NOAA would be interested to say the least.
Related
El Niño forecasts extended to 18 months with physics-based model
Researchers at the University of Hawai'i at Mānoa introduce the XRO model, extending El Niño forecasts to 18 months. This physics-based approach enhances predictability and understanding of ENSO events and other climate variabilities.
AI helps to produce breakthrough in weather and climate forecasting
Artificial intelligence, exemplified by Google's NeuralGCM model, enhances weather forecasting by combining AI with traditional physics models. This collaboration improves speed, accuracy, and efficiency, marking a significant advancement in climate prediction.
Neural general circulation models for weather and climate
Neural general circulation models (GCMs) integrate machine learning with atmospheric dynamics for accurate weather and climate predictions. They show competitive performance, combining traditional GCM strengths with machine learning efficiency for improved forecasting.
Fast, accurate climate modeling with NeuralGCM
NeuralGCM is a new machine learning model for simulating Earth's atmosphere, providing accurate 2-15 day weather forecasts, outperforming traditional models, and being computationally efficient for broader climate predictions.
Artificial intelligence gives weather forecasters a new edge
AI is improving weather forecasting speed and accuracy, notably for hurricanes. DeepMind's GraphCast outperforms traditional models, providing timely warnings and enhancing forecasting accuracy while emphasizing the need for human oversight.