Statistics research could build consensus around climate predictions
Philadelphia, PA—Vast amounts of data related to climate change are being compiled by research groups all over the world. Data from these many and varied sources results in different climate projections; hence, the need arises to combine information across data sets to arrive at a consensus regarding future climate estimates.
In a paper published last December in the SIAM Journal on Uncertainty Quantification, authors Matthew Heaton, Tamara Greasby, and Stephan Sain propose a statistical hierarchical Bayesian model that consolidates climate change information from observation-based data sets and climate models.
“The vast array of climate data—from reconstructions of historic temperatures and modern observational temperature measurements to climate model projections of future climate—seems to agree that global temperatures are changing,” says author Matthew Heaton. “Where these data sources disagree, however, is by how much temperatures have changed and are expected to change in the future. Our research seeks to combine many different sources of climate data, in a statistically rigorous way, to determine a consensus on how much temperatures are changing.”
Using a hierarchical model, the authors combine information from these various sources to obtain an ensemble estimate of current and future climate along with an associated measure of uncertainty. “Each climate data source provides us with an estimate of how much temperatures are changing. But, each data source also has a degree of uncertainty in its climate projection,” says Heaton. “Statistical modeling is a tool to not only get a consensus estimate of temperature change but also an estimate of our uncertainty about this temperature change.”
The approach proposed in the paper combines information from observation-based data, general circulation models (GCMs) and regional climate models (RCMs).
Regional analysis for climate change assessment. Image credit: Melissa Bukovsky, National Center for Atmospheric Research (NCAR/IMAGe)
Observation-based data sets, which focus mainly on local and regional climate, are obtained by taking raw climate measurements from weather stations and applying it to a grid defined over the globe. This allows the final data product to provide an aggregate measure of climate rather than be restricted to individual weather data sets. Such data sets are restricted to current and historical time periods. Another source of information related to observation-based data sets are reanalysis data sets in which numerical model forecasts and weather station observations are combined into a single gridded reconstruction of climate over the globe.
GCMs are computer models which capture physical processes governing the atmosphere and oceans to simulate the response of temperature, precipitation, and other meteorological variables in different scenarios. While a GCM portrayal of temperature would not be accurate to a given day, these models give fairly good estimates for long-term average temperatures, such as 30-year periods, which closely match observed data. A big advantage of GCMs over observed and reanalyzed data is that GCMs are able to simulate climate systems in the future.
RCMs are used to simulate climate over a specific region, as opposed to global simulations created by GCMs. Since climate in a specific region is affected by the rest of the earth, atmospheric conditions such as temperature and moisture at the region’s boundary are estimated by using other sources such as GCMs or reanalysis data.
By combining information from multiple observation-based data sets, GCMs and RCMs, the model obtains an estimate and measure of uncertainty for the average temperature, temporal trend, as well as the variability of seasonal average temperatures. The model was used to analyze average summer and winter temperatures for the Pacific Southwest, Prairie and North Atlantic regions (seen in the image above)—regions that represent three distinct climates. The assumption would be that climate models would behave differently for each of these regions. Data from each region was considered individually so that the model could be fit to each region separately.
“Our understanding of how much temperatures are changing is reflected in all the data available to us,” says Heaton. “For example, one data source might suggest that temperatures are increasing by 2 degrees Celsius while another source suggests temperatures are increasing by 4 degrees. So, do we believe a 2-degree increase or a 4-degree increase? The answer is probably ‘neither’ because combining data sources together suggests that increases would likely be somewhere between 2 and 4 degrees. The point is that that no single data source has all the answers. And, only by combining many different sources of climate data are we really able to quantify how much we think temperatures are changing.”
While most previous such work focuses on mean or average values, the authors in this paper acknowledge that climate in the broader sense encompasses variations between years, trends, averages and extreme events. Hence the hierarchical Bayesian model used here simultaneously considers the average, linear trend and interannual variability (variation between years). Many previous models also assume independence between climate models, whereas this paper accounts for commonalities shared by various models—such as physical equations or fluid dynamics—and correlates between data sets.
“While our work is a good first step in combining many different sources of climate information, we still fall short in that we still leave out many viable sources of climate information,” says Heaton. “Furthermore, our work focuses on increases/decreases in temperatures, but similar analyses are needed to estimate consensus changes in other meteorological variables such as precipitation. Finally, we hope to expand our analysis from regional temperatures (say, over just a portion of the U.S.) to global temperatures.”
To read other SIAM Nuggets, explaining current high level research involving applications of mathematics in popular science terms, go to http://connect.siam.org/category/siam-nuggets/.
Source article:
Matthew J. Heaton, Tamara A. Greasby, and Stephan R. Sain
SIAM/ASA Journal on Uncertainty Quantification, 1(1), 535–559 (Online publish date: December 17, 2013). The source article is available for free access at the link above through December 31, 2014.
About the authors:
Matthew Heaton is currently an assistant professor in the department of statistics at Brigham Young University, but the majority of this work was done while the author was a postgraduate scientist at the National Center for Atmospheric Research. Tamara Greasby is a postdoctoral researcher and Stephan Sain is a scientist at the Institute for Mathematics Applied to the Geosciences (IMAGe) at the National Center for Atmospheric Research (NCAR).
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.

Oh God have we not already had to put up with industrial scale stupidity with cartoonist Cook’s consensus without this. There is so much wrong with this approach I can’t even be bothered to post a critique. Don’t these people realise just how stupid they look to statisticians.
Since when does consensus mean anything in science?
Advanced GIGO for geeks.
Like all current climate modelling, it starts with the premise thar CAGW theory is correct and global temperatures will rise by 2-4 degrees C over the next century.
Exactly, Peter. Decide the answer ahead of time and find a way to make the models produce it.
Richard Feynman is either screaming or laughing his a– off right now.
This sounds like a movie speech crafted be Mel Brooks.
So much effort for a 3% gain.
My guess is that they are trying to get everyone to sing from the same page. Currently there are massive contradictions flying about, which rather shoots the notion of consensus in the foot.
While a GCM portrayal of temperature would not be accurate to a given day, these models give fairly good estimates for long-term average temperatures, such as 30-year periods, which closely match observed data.
Since when have the GCM’s closely matched observed data for 30 years?
They can’t even predict the past.
Let me help them out. It’s cooling now and will continue to cool for at least another couple of decades.
Wait a minute… I thought there already was a consensus.
It’s getting to the point where if you wanted me to explain where the environmental establishment went wrong, I wouldn’t even know where to start. It would be like trying to explain the concept of love; it baffles even the most loquacious of us. It has lost any semblance of trying to “save the planet” and now simply exists to continue its existence.
That’s “sustainability” for ya!
Statistics is a tool which one should attempt to use minimally. It is like making sausage. The output may be in more useful form, but it has been altered and adulterated with artificial ingredients. Every step by which data is processed destroys information and adds things to the data that were not originally there.
By the time you get to grand statistical amalgamations of the statistical output of multiple studies and models based on the input of still other studies you have gone way beyond the realm of processed sausage and are eating pure plastic.
“They can’t even predict the past.”
Funniest comment in a while. I’ll be stealing that one! 🙂
And I thought they ran out of ways to fudge data.
“The approach proposed in the paper combines information from observation-based data, general circulation models (GCMs) and regional climate models (RCMs).”
I would like them to only focus on observation-based data. Un-fooled around with data would be best. Data that does not fill in points that for whatever reason was not obtained. I suspect they will get different projections based on different chosen start dates.
Most likely they would discover that their projections would somewhat continue the warming/cooling cycles that we see since the end of the LIA.
How can one determine when that cyclic warming will end without knowing all the causes of the cyclic warming?
Also, I have a question:
Aren’t models supposed to be based on data/observations? Will their new, as proposed by the paper, model(s) be based partly/mostly on current models?
“Furthermore, our work focuses on increases/decreases in temperatures, but similar analyses are needed to estimate consensus changes in other meteorological variables such as precipitation.”
This sounds like they don’t trust their own prediction of a wet Saudi Arabia.
The everlasting delusion of statisticians that they can mine gold from worthless ore
simply by applying statistics to a giant boatload of crappy data and model estimates. Bayesian stats are included, no less, which aren’t even statistics, simply people giving their own guesses and weighting their belief in same. Maybe someday it wil dawn on these innocents that model output and projections are not data.
NCAR has a fever and it’s contagious?
Symptoms include delusions?
One of these models is the most accurate. It might be accurate by chance, or because it is fairly well estimating some of the important variables.
Averaging the models is a way to ensure your model is poor by straying away from the best model.
Data from these many and varied sources results in different climate projections; hence, the need arises to combine information across data sets to arrive at a consensus regarding future climate estimates…..
You’re going to take bad data….average it….and get what??
“Last fall the Climate Prediction Center of the National Oceanic and Atmospheric Administration predicted that temperatures would be above normal from November through January across much of the Lower 48 states………………”
http://www.businessweek.com/articles/2014-02-18/the-official-forecast-of-the-u-dot-s-dot-government-never-saw-this-winter-coming
I needed something to read whilst my graphics driver updates, I wish now that I had gone to the kitchen and found a very large knife and slit my throat. I am having an email and phone debate with Myles Allen and Kevin Anderson about RCP’s which are projections made by three difference computer models about Co2 emissions and currently we are above the highest which is RCP 8.5 the worst case scenario where we all get to fry next Tuesday and no matter how specific I get about the fact that it has not warmed for 17 years and 5 months its flat, they tell me I am drawing the line in the wrong place. Kevin Anderson advises Ed Davey DECC Sec of State on mitigation policies which is what Connie Hedegaarde is spending Euros 30 trillion to save 2 hundredths of 1C. Myles Allen conducts jam jar laboratory experiments on Co2 and makes extrapolations which are then picked up by computer modellers whose projections are based on those simplistic experiments without any exposure to the real world I have just read the ipcc pdf and the time it must have taken to write 17 pages of hyperbole to justify the insignificance of what they are about and which guys like Kevin Anderson believe is beyond belief. They just read what other people write and if its peer reviewed they just accept it as fact they don’t want to look close or question it because it could mean they wont have a job in six months, if it is not a conspiracy then it is certainly collusion. I have watched an horizon program on placebos today and when a consultant had a meeting about the placebo effect saying he was going to do some trials all of the other doctors in the room called him a bastard sceptic and a heretic for daring to challenge the orthodoxy he was proven right there is such a thing as the placebo effect but I don’t think I will live long enough to see these green bastards out of a job.
JohnWho:
Exactly.
Why must we arrive at a consensus? Why wouldn’t we be better served to let people do the best job they think is possible at analysis and projection, and then plan accordingly? Some may plan poorly; while others plan better, but there wouldn’t be the risk of everyone planning poorly and identically, end up like the lemmings.
Alan Robertson says:
February 19, 2014 at 1:28 pm
Wait a minute… I thought there already was a consensus.
*
Yes, but people are getting confused what the consensus is because it’s all getting out of control. Someone must have decided that they needed a consensus of the consensus. And then everyone will know what it is. 🙂