Essay by Eric Worrall
“… Google adds machine learning to climate models for ‘faster forecasts’ …”
The secret to better weather forecasts may be a dash of AI
Google adds machine learning to climate models for ‘faster forecasts’
Tobias Mann
Sat 27 Jul 2024 // 13:27 UTC…
In a paper published in the journal Nature this week, a team from Google and the European Centre for Medium-Range Weather Forecasts (ECMWF) detailed a novel approach that uses machine learning to overcome limitations in existing climate models and try to generate forecasts faster and more accurately than existing methods.
Dubbed NeuralGCM, the model was developed using historical weather data gathered by ECMWF, and uses neural networks to augment more traditional HPC-style physics simulations.
As Stephan Hoyer, one of the crew behind NeuralGCM wrote in a recent report, most climate models today make predictions by dividing up the globe into cubes 50-100 kilometers on each side and then simulating how air and moisture move within them based on known laws of physics.
NeuralGCM works in a similar fashion, but the added machine learning is used to track climate processes that aren’t necessarily as well understood or which take place at smaller scales.
Read more: https://www.theregister.com/2024/07/27/google_ai_weather/
…
The abstract of the study;
- Article
- Open access
- Published: 22 July 2024
Neural general circulation models for weather and climate
Dmitrii Kochkov, Janni Yuval, Ian Langmore, Peter Norgaard, Jamie Smith, Griffin Mooers, Milan Klöwer, James Lottes, Stephan Rasp, Peter Düben, Sam Hatfield, Peter Battaglia, Alvaro Sanchez-Gonzalez, Matthew Willson, Michael P. Brenner & Stephan Hoyer
Abstract
General circulation models (GCMs) are the foundation of weather and climate prediction1,2. GCMs are physics-based simulators that combine a numerical solver for large-scale dynamics with tuned representations for small-scale processes such as cloud formation. Recently, machine-learning models trained on reanalysis data have achieved comparable or better skill than GCMs for deterministic weather forecasting3,4. However, these models have not demonstrated improved ensemble forecasts, or shown sufficient stability for long-term weather and climate simulations. Here we present a GCM that combines a differentiable solver for atmospheric dynamics with machine-learning components and show that it can generate forecasts of deterministic weather, ensemble weather and climate on par with the best machine-learning and physics-based methods. NeuralGCM is competitive with machine-learning models for one- to ten-day forecasts, and with the European Centre for Medium-Range Weather Forecasts ensemble prediction for one- to fifteen-day forecasts. With prescribed sea surface temperature, NeuralGCM can accurately track climate metrics for multiple decades, and climate forecasts with 140-kilometre resolution show emergent phenomena such as realistic frequency and trajectories of tropical cyclones. For both weather and climate, our approach offers orders of magnitude computational savings over conventional GCMs, although our model does not extrapolate to substantially different future climates. Our results show that end-to-end deep learning is compatible with tasks performed by conventional GCMs and can enhance the large-scale physical simulations that are essential for understanding and predicting the Earth system.
Read more: https://www.nature.com/articles/s41586-024-07744-y
Reading the main study, they appear to be claiming adding neural net magic sauce produces better short term weather predictions and climate predictions.
The researchers attempted to test their model by holding back some of the training data, and using their trained neural network to product weather forecasts based on real world data which had never been seen before by the AI. They also discuss how their model predictions diverge from reality after 3 days – “At longer lead times, RMSE rapidly increases owing to chaotic divergence of nearby weather trajectories, making RMSE less informative for deterministic models”, but claim their approach still performs better than traditional approaches, once that chaotic divergence is taken into account.
I’m a bit dubious about putting faith in the predictive skill of neural net black boxes. History is littered with scientists who followed all the steps the authors described, only to see the neural net diverge wildly from expected behaviour on demonstration day. I would have preferred if they made more effort to reverse engineer their neural net, to tease out what it actually discovered, if anything, to see if it discovered new atmospheric physics which can be used to create better deterministic white box models.
In 2018 Amazon suffered a serious embarrassment when one of their neural nets went off the rails. They tried to use neural nets to filter tech candidates, but they discovered the neural net exhibited bias against women candidates. The neural net had noticed most of the candidates were men, and inferred it should discard applications from female candidates based on their gender.
Anyone who thinks there is any basis to that Amazon neural net gender bias needs to pay a visit to a tech shop in Asia. Somehow in the West we are convincing our girls from a young age they are not suited for tech jobs. The few women who make it through this Western cultural filter, women like Margaret Hamilton who led the team which programmed the Apollo guidance computer, more than demonstrate their ability. Hamilton is credited with inventing the term “software engineering”.
Neural Nets are incredibly useful, they can identify relationships which aren’t obvious. But that ability to see non-obvious patterns in data carries a severe risk of false positives, seeing patterns which don’t exist. Especially when analysing physical phenomena, if you don’t try to reverse engineer the inner workings of your magic black box neural network once it has allegedly demonstrated superior skill, you just don’t know whether what you are seeing is genuine skill or a subtle false positive.
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.
An issue is that the reporting stations have a large amount of Urban Heat Island contamination, so the input data is flawed.
in theory an AI should have a chance of figuring this out, but there are many more ways to be wrong than right. You have to crack open the black box to be sure when doing AI physics.
Given Large Language Models known tendency to “hallucinate” sources, I would doubt anything one came up with. Plus the biases induced by the training set.
The image Google came up with when asked to show “German soldiers in 1945” coming up with Black or Asian guys in SS uniforms shows the possibility.
Hallucinations are another manifestation of the same underlying issue.
With all the ghost stations, they won’t ever know the true temps. much less be able to predict the future temps.
GIGO… by a faster route.
Oh goodie !! 😉
You can bet they tell it that CO2 causes warming, and other similar garbage conjectures.
Or maybe think of it as Artificial Garbage
Yes. The machines are equally adept at producing artificial intelligence and artificial stupidity.
When the ‘machine learning’ finds patterns just as predictive, or better, than the current physics-based bottom-up models then they mainly just serve to illustrate how bad the current models are.
The initials are AI–actual idiocracy.
UHI is a real phenomenon. Stations reporting higher temperatures in urban environments is not a flaw; it’s reality.
Yes, but what it is “detecting” is something other than GHG caused global warming.
True, but irrelevant. The input data is valid regardless on what causes it to be what it is.
WRONG! Input data from surface site fabrications is TOTALLY BOGUS.
They are looking for “global” models.
UHI effect, although making up a large part of the surface data, is only 5% or so of the Earth’s surface.
Only if it is the data as recorded. Adjusted data leads to adjusted output. If the AI is truly smart it can find problems with the input data and determine if it is fit for purpose.
It’s reality in urban environments. Totally irrelevant in the (fictitious) “average global temperature.”
Too bad UHI is not an indicator of CO2 in the GHG theory. UHI may be proper to use in measuring heat, but not for determining CO2’s effects on the atmosphere.
Claim: Google NeuralGCM AI
Mightwill not Improve Climate ModelsIt is claimed that AI and a powerful personal computer does predict the weather slightly better than anything else currently being used.
🙂 for 3 days ahead … no better than physically looking at the cloud formations and air pressure. Still a lot better than many of the weather forecasters, the weather forecast on Microsoft systems is consistently 2C higher than actual and forever predicting a “near record” warm temperatures that do not eventuate. On Saturday here in Brisbane, Australia, was forecast to reach 26C and it barely made 23C … lovely winters day!
This looks like a sales job, gooogle trying to extract more of our money
More like trying desperately to justify the massive amounts of money they’ve poured into the sinkhole.
It seems to me that this study really exposes the limitations of the researchers who appear not to have any appreciable knowledge of the intricacies of AI learning. Am I wrong?
A faster “climate” model seems unnecessary.
A faster weather model won’t do much good
because it will be just as wrong as a slower one.
Their “black box” will need a little periscope so
it can look out to see if it is raining or not.
Unless AI can rediscover Lechâtelier’s Principle (LCP) that has not been included in climate models, it will stump AI attempts to improve the models significantly. LCP states that in an interactive system of components, any perturbing change in one (or more) of the components induces changes in the other components such as to resist the perturbing change, thereby resulting in a change much reduced from what was expected.
The “components” in the climate system include composition of the oceans, atmosphere and biosphere, temperatures, pressures, physical states of water (liquid, solid and gas) enthalpy, etc. An easy to understand example is if the bulk atmosphere is heated, it expands which is a cooling reaction to the heating, resulting in a much reduced final T° at equilibrium.
Another somewhat more complex example is, if the CO2 composition of the atmosphere is increased by burning fossil fuels or volcanic eruption, the biosphere reacts by increasing photosynthesis to expand the Greening of the planet, and increased growth of the existing plants. At the same time increased partial pressure of CO2 in the atmosphere induces increased solubility of CO2 in the oceans. This added CO2 to the oceans, induces photosynthesis in phytoplankton, seaweeds and shellfish calcium carbonate shells. The equilibrium result in the atmosphere is to remove half or more CO2 from the added amount. All the other components also change at least in small ways. Photosynthesis is also an endothermic (cooling) reaction on land and in the ocean!
It is totally unscientific to use ceteris paribus (all else remaining the same) calculations to determine what temperature will be. Simply, the other components do not remain the same! Physicists seem unaware of LCP. To get “The Physics” right, you had better get “The Chemistry” right!
Actually, Gary, physicists call the LCP ‘negative feedback’ or ‘entropy production optimisation’. But you are right, ceteris paribus rarely applies. Your chemistry reference is a good one, because chemical buffering of CO2 absorbed in ocean waters is an example of LCP.
Ceteris paribus is useful only to understand the effect of a single parameter or process, not to understand the response of a complex system, which our Earthian atmospheric climate certainly is.
But it makes great copy for warmunist ideologues and their handmaidens in the mainstream media who love nothing better than to predict gloom.
Saw a headline on CNN this morning about yet another warmunist “study” that said that increased heat and humidity will make some regions of the planet “uninhabitable”. Quick, somebody better tell all those millions of northerners moving to the US Gulf of Mexico region, particularly the two fastest growing states of Florida and Texas, with their extremely well known heat and humidity, that their new home is “uninhabitable”! If they only knew they were jumping into a frying pan, they’d all stay up north.
As Google’s search engines are now a product of over-commercialised drivel, it’s hardly wise to put global trust in their interest in anything but making money from ‘climate models’. They will lie, lie and lie again so long as they make money from their lying.
Climate models fail because they try and fit real data to hypothetical models not taking account of some of the most important features of climate (e.g. clouds).
It’s better to stop using their rubbish than to assign influence to Google, whose mantra should in fact be ‘if it makes us lots of money, then evil is absolutely OK’.
It is easy to beat the current IPCC climate model’s forecast.
Just forecast the next 30 years(climate) will be about the same as the last 30 years.
An inspirational article. I’m convinced that if I have an AI analyse bird entrails it will help improve my predictions of future events. All I need is a few million dollars to train the AI. Send cheques to…
You beat me to that one. Except I recommend tea leaves.
How absurd! Tea leaves, bird entrails? It’s settled science to use tortoise shells.
I find that AI (actual intelligence, AKA Joe Bastardi) is a terrific prognosticator.
When the model itself is rubbish no amount of ‘improving’ will make it better. Lipstick on a pig will not change the pig into a horse.
Might look more attractive to Brandon, though!
[BTW I don’t think you should be (quite) so disrespectful to Kamala.]
And who defines what a pattern is if it isn’t the programmer? Machines may be useful in context but they still cannot make something out of nothing and nor will they ever be able to.
No. These models are self-learning. They are great at finding patterns.
When they work, these AI systems can find innovative solutions.
When they don’t work, AI systems produce utter nonsense.
But the programmer had to define how to measurement success, which means defining the concept of “pattern”. Looking at AI from the outside, if it either finds innovative solutions or produces utter nonsense, then by definition it’s finding random “solutions”.
For this kind of problem success is easy to define as given the conditions today compare your prediction to the actual weather tomorrow. And that applies to all history. A well trained and not overfitted model can use its historic data in that way.
It’s the same idea when training-vision only AI to estimate 3D properties such as distance to an object in an image. Eg drive around taking pictures and lidar measurements then train the model by back propagating the actual distances to the objects given by lidar.
You can’t expect a programmer to be able to program intuition into an AI. It can be programmed to look for patterns that it can define with current mathematics, but doing a leap of knowledge through intuition will be beyond it. So much of this depends on proper measurements of all the physical parameters of the earth, sun, planets, etc. If basic temperature data must be “adjusted” then it is by implication not fit for purpose. Just looking at temperature is also way too limited.
“The neural net had noticed most of the candidates were men, and inferred it should discard applications from female candidates based on their gender.”
How stupid of it to only notice 2 genders. /s
How did the neural net know they were men or women? That’s not allowed on employment applications. I’m thinking the neural net found the “best” candidates in the input set and they just happened to be men but that didn’t meet the HR requirements and the politically correct philosophy so they invented another reason.
While there may be some advantage to having machines detect patterns in the data that humans are not predisposed to take note of, it is still the case of GIGO – bad or incomplete or biased or manipulated data will skew any result that AI produces from reviewing and analyzing such data. And as pointed out in this post by Eric with the example of Amazon’s job screening AI, even perfectly reliable data can still be biased by factors that are not represented in the available data.
GIGO — Garbage In = Garbage Out
This sounds comparable to simulated Structural Equation Modelling (SEM) to me. SEM can have it’s place but there are some important assumptions that must be validated. For example Science Direct indicates that “The major assumptions associated with structural equation modeling include: multivariate normality, no systematic missing data, sufficiently large sample size, and correct model specification.” It’s not clear to me how many of these transfer to this approach, but I expect they all do to some extent. When these assumptions go unchecked you might get a model that is “right”, but for the wrong reasons, and then when it’s applied outside that initial case the hidden problems can impact results but may not be immediately obvious, if at all. For example for an untested black box successfully answering the question “2+2=?” with 4, it could recognize it as a simple math problem and solving, or it could be applying a derived formula (e.g. n-1, where n is the number of digits in the incoming text string). By asking questions like “3+3=?” or “2×2=?” we can start to evaluate what confidence we might place in it.
But in this toy problem we are able to validate the results. Can we do the same for 100-year climate projections?
By analogy consider the range of effective climate sensitivity in the CMIP6 set (varies by over 3x between models) yet to some modest degree all can hindcast 20th century climate. As the IPCC indicated in the summary to working group 1 of the first assessment report “The climate system is a coupled non-linear chaotic system, and therefore the long-term prediction of future climate states is not possible.”
Why do they need “artificial intelligence” if the science is settled?
They need intelligence of some kind as it is obviously lacking.
Sounds like Joe Bastardi!
Hype chases hyperbole.
Any progress in modelling that increases the dependence on real world observations over imaginary simulations is an improvement, but using AI to simply do faster what GCM’s do so poorly in their current form is probably not a step forward if that’s all there is to this. I will await results with an open mind. Ultimately what is needed is a better understanding of the processes that still elude us in the chaotic solar/ocean/atmosphere systems that bring us sunny days and ice ages.
I think that AI will prove that even it can be just as biased as those that programmed it.
Isn’t the training on an AI system really just part of the programming. You’re just populating a bunch of unspecified variables.
Climate models do not predict climate. They predict what researchers believe climate will be. Otherwise the researchers would believe the models were in error and change them.
Neural nets are a multi dimensional curve fitting. This allows them to identify similar patterns. But unless you have something predictable to match against, the neural net will have no predictive skill.
Thus we can predict the tides because orbital mechanics are predictable. But there is no such predictable item to pattern match climate.
How do we predict ocean tides? We know from past observation that when the sun, moon, and earth are in a specific location, the tides will be a specific height. An AI neural net can solve this.
However we have no such predictors for future climate. Unlike orbital mechanics we don’t know future GHG, solar and land use. We have to guess these. As such, any future climate predictions are simply a guess based on guesses.
If AI can predict future climate, why not use AI to predict the stock market and use the winnings to finance carbon taxes and the green new deal. Why are taxes always asked to pay for these sure-fire winning solutions.
Answer: mathematically, AI cannot predict the future beyond 2 dimensions