New peer reviewed paper shows just how bad the climate models really are

One of the biggest, if not the biggest issues of climate science skepticism is the criticism of over-reliance on computer model projections to suggest future outcomes. In this paper, climate models were hindcast tested against actual surface observations, and found to be seriously lacking. Just have a look at Figure 12 (mean temperature -vs- models for the USA)  from the paper, shown below:

Fig. 12. Various temperature time series spatially integrated over the USA (mean annual), at annual and 30-year scales. Click image for the complete graph

The graph above shows temperature in the blue lines, and model runs in other colors. Not only are there no curve shape matches, temperature offsets are significant as well. In the study, they also looked at precipitation, which fared even worse in correlation. The bottom line: if the models do a poor job of hindcasting, why would they do any better in forecasting? This from the conclusion sums it up pretty well:

…we think that the most important question is not whether GCMs can produce credible estimates of future climate, but whether climate is at all predictable in deterministic terms.

Selected sections of the entire paper, from the Hydrological Sciences Journal is available online here as HTML, and  as PDF ~1.3MB are given below:

A comparison of local and aggregated climate model outputs with observed data

Anagnostopoulos, G. G. , Koutsoyiannis, D. , Christofides, A. , Efstratiadis, A. and Mamassis, N. ‘A comparison of local and aggregated climate model outputs with observed data’, Hydrological Sciences Journal, 55:7, 1094 – 1110

Abstract

We compare the output of various climate models to temperature and precipitation observations at 55 points around the globe. We also spatially aggregate model output and observations over the contiguous USA using data from 70 stations, and we perform comparison at several temporal scales, including a climatic (30-year) scale. Besides confirming the findings of a previous assessment study that model projections at point scale are poor, results show that the spatially integrated projections are also poor.

Citation Anagnostopoulos, G. G., Koutsoyiannis, D., Christofides, A., Efstratiadis, A. & Mamassis, N. (2010) A comparison of local and aggregated climate model outputs with observed data. Hydrol. Sci. J. 55(7), 1094-1110.

INTRODUCTION

According to the Intergovernmental Panel on Climate Change (IPCC), global circulation models (GCM) are able to “reproduce features of the past climates and climate changes” (Randall et al., 2007, p. 601). Here we test whether this is indeed the case. We examine how well several model outputs fit measured temperature and rainfall in many stations around the globe. We also integrate measurements and model outputs over a large part of a continent, the contiguous USA (the USA excluding islands and Alaska), and examine the extent to which models can reproduce the past climate there. We will be referring to this as “comparison at a large scale”.

This paper is a continuation and expansion of Koutsoyiannis et al. (2008). The differences are that (a) Koutsoyiannis et al. (2008) had tested only eight points, whereas here we test 55 points for each variable; (b) we examine more variables in addition to mean temperature and precipitation; and (c) we compare at a large scale in addition to point scale. The comparison methodology is presented in the next section.

While the study of Koutsoyiannis et al. (2008) was not challenged by any formal discussion papers, or any other peer-reviewed papers, criticism appeared in science blogs (e.g. Schmidt, 2008). Similar criticism has been received by two reviewers of the first draft of this paper, hereinafter referred to as critics. In both cases, it was only our methodology that was challenged and not our results. Therefore, after presenting the methodology below, we include a section “Justification of the methodology”, in which we discuss all the critical comments, and explain why we disagree and why we think that our methodology is appropriate. Following that, we present the results and offer some concluding remarks.

Here’s the models they tested:

Comparison at a large scale

We collected long time series of temperature and precipitation for 70 stations in the USA (five were also used in the comparison at the point basis). Again the data were downloaded from the web site of the Royal Netherlands Meteorological Institute (http://climexp.knmi.nl). The stations were selected so that they are geographically distributed throughout the contiguous USA. We selected this region because of the good coverage of data series satisfying the criteria discussed above. The stations selected are shown in Fig. 2 and are listed by Anagnostopoulos (2009, pp. 12-13). THSJ_A_513518_O_XML_IMAGES\THSJ_A_513518_O_F0002g.jpg

Fig. 2. Stations selected for areal integration and their contribution areas (Thiessen polygons).

In order to produce an areal time series we used the method of Thiessen polygons (also known as Voronoi cells), which assigns weights to each point measurement that are proportional to the area of influence; the weights are the “Thiessen coefficients”. The Thiessen polygons for the selected stations of the USA are shown in Fig. 2.

The annual average temperature of the contiguous USA was initially computed as the weighted average of the mean annual temperature at each station, using the station’s Thiessen coefficient as weight. The weighted average elevation of the stations (computed by multiplying the elevation of each station with the Thiessen coefficient) is Hm = 668.7 m and the average elevation of the contiguous USA (computed as the weighted average of the elevation of each state, using the area of each state as weight) is H = 746.8 m. By plotting the average temperature of each station against elevation and fitting a straight line, we determined a temperature gradient θ = -0.0038°C/m, which implies a correction of the annual average areal temperature θ(H – Hm) = -0.3°C.

The annual average precipitation of the contiguous USA was calculated simply as the weighted sum of the total annual precipitation at each station, using the station’s Thiessen coefficient as weight, without any other correction, since no significant correlation could be determined between elevation and precipitation for the specific time series examined.

We verified the resulting areal time series using data from other organizations. Two organizations provide areal data for the USA: the National Oceanic and Atmospheric Administration (NOAA) and the National Aeronautics and Space Administration (NASA). Both organizations have modified the original data by making several adjustments and using homogenization methods. The time series of the two organizations have noticeable differences, probably because they used different processing methods. The reason for calculating our own areal time series is that we wanted to avoid any comparisons with modified data. As shown in Fig. 3, the temperature time series we calculated with the method described above are almost identical to the time series of NOAA, whereas in precipitation there is an almost constant difference of 40 mm per year. THSJ_A_513518_O_XML_IMAGES\THSJ_A_513518_O_F0003g.jpg

Fig. 3. Comparison between areal (over the USA) time series of NOAA (downloaded from http://www.ncdc.noaa.gov/oa/climate/research/cag3/cag3.html) and areal time series derived through the Thiessen method; for (a) mean annual temperature (adjusted for elevation), and (b) annual precipitation.

Determining the areal time series from the climate model outputs is straightforward: we simply computed a weighted average of the time series of the grid points situated within the geographical boundaries of the contiguous USA. The influence area of each grid point is a rectangle whose “vertical” (perpendicular to the equator) side is (ϕ2 – ϕ1)/2 and its “horizontal” side is proportional to cosϕ, where ϕ is the latitude of each grid point, and ϕ2 and ϕ1 are the latitudes of the adjacent “horizontal” grid lines. The weights used were thus cosϕ(ϕ2 – ϕ1); where grid latitudes are evenly spaced, the weights are simply cosϕ.

CONCLUSIONS AND DISCUSSION

It is claimed that GCMs provide credible quantitative estimates of future climate change, particularly at continental scales and above. Examining the local performance of the models at 55 points, we found that local projections do not correlate well with observed measurements. Furthermore, we found that the correlation at a large spatial scale, i.e. the contiguous USA, is worse than at the local scale.

However, we think that the most important question is not whether GCMs can produce credible estimates of future climate, but whether climate is at all predictable in deterministic terms. Several publications, a typical example being Rial et al. (2004), point out the difficulties that the climate system complexity introduces when we attempt to make predictions. “Complexity” in this context usually refers to the fact that there are many parts comprising the system and many interactions among these parts. This observation is correct, but we take it a step further. We think that it is not merely a matter of high dimensionality, and that it can be misleading to assume that the uncertainty can be reduced if we analyse its “sources” as nonlinearities, feedbacks, thresholds, etc., and attempt to establish causality relationships. Koutsoyiannis (2010) created a toy model with simple, fully-known, deterministic dynamics, and with only two degrees of freedom (i.e. internal state variables or dimensions); but it exhibits extremely uncertain behaviour at all scales, including trends, fluctuations, and other features similar to those displayed by the climate. It does so with a constant external forcing, which means that there is no causality relationship between its state and the forcing. The fact that climate has many orders of magnitude more degrees of freedom certainly perplexes the situation further, but in the end it may be irrelevant; for, in the end, we do not have a predictable system hidden behind many layers of uncertainty which could be removed to some extent, but, rather, we have a system that is uncertain at its heart.

Do we have something better than GCMs when it comes to establishing policies for the future? Our answer is yes: we have stochastic approaches, and what is needed is a paradigm shift. We need to recognize the fact that the uncertainty is intrinsic, and shift our attention from reducing the uncertainty towards quantifying the uncertainty (see also Koutsoyiannis et al., 2009a). Obviously, in such a paradigm shift, stochastic descriptions of hydroclimatic processes should incorporate what is known about the driving physical mechanisms of the processes. Despite a common misconception of stochastics as black-box approaches whose blind use of data disregard the system dynamics, several celebrated examples, including statistical thermophysics and the modelling of turbulence, emphasize the opposite, i.e. the fact that stochastics is an indispensable, advanced and powerful part of physics. Other simpler examples (e.g. Koutsoyiannis, 2010) indicate how known deterministic dynamics can be fully incorporated in a stochastic framework and reconciled with the unavoidable emergence of uncertainty in predictions.

h/t to WUWT reader Don from Paradise

0 0 votes
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

139 Comments
Inline Feedbacks
View all comments
John Q Public
December 5, 2010 9:07 pm

Haven’t we learned anything from models that try to forecast weather or the economy … or, even that sub-prime asset backed investments were safe?
Statistics, mathematics, and program code are just the new smoke and mirrors for the illiterate AGW believer.
Dumb, dumber, dumbest.

Earl Wood
December 5, 2010 9:22 pm

It is about time we’ve gotten around to comparing these models to actual data. This should have been done for each model for each publication using a model to give the reader an idea of its accuracy. This is the standard for other fields in science.

Joshua J
December 5, 2010 9:25 pm

However, the GCM’s are apparently very good at generating hockey stick graphs out of random data.

anna v
December 5, 2010 9:50 pm

Great. If I were ten years younger I would have tried to weasel my way into working with their group. They obviously know their physics and statistics.

Lucia has shown the bad fit
of model temperature outputs with reality, but I do not know whether she is aiming at publishing in peer review.

Lew Skannen
December 5, 2010 9:52 pm

This is exactly where the attack needs to concentrate. I have never understood why we keep battling AGW alarmists on territory that is no use to anyone – ie Raw Data. (was this year hotter? how much ice melted? what caused that flood? etc)
The second weakest link in the whole AGW chain is the modelling. It is so clear that the models are just feeble guesses which crumble on first contact with evidence. It is quite likely that modelling is impossible given the chaotic nature of the problem.
(The first weakest link, by the way is the steering mechanism that AGW alarmists think that they have control over. It is the giant thermostat in the sky that they are currently trying to appease by sacrificing dollars in Cancun…)

rob m
December 5, 2010 9:56 pm

The secong paragraph in the conclusion pretty much sums up why I don’t believe in AGW.

Robert
December 5, 2010 10:00 pm

I can’t think of any field where this kind of inaccuracy in modeling would be OK. No place where real money is at stake, certainly. Have these people no standards? Does nobody think they ought to check anything? Do they honestly think that trying hard is all you need?
The mind boggles.
Thank you for publishing these results. It may be painful, but not as painful as the results of taking models more seriously then they deserve.

Brian H
December 5, 2010 10:15 pm

But … hydrologists aren’t UEA – approved climatologists! What can they possibly know?

stumpy
December 5, 2010 10:20 pm

I have also looked at model rainfall and temp data for NZ and none matches the observed – its not restricted to just the USA, its a global problem!

anna v
December 5, 2010 10:32 pm

It is worth noting that this paper exposes the real reason why “anomalies” have been foisted on us. Anomalies, which are a delta(T) study the shapes of terrains ignoring the height. Under a gravitational field, for example, one expects the shapes of mountains to be similar, but can one really average the Himalayas anomaly( average height of mountains taken as base of anomaly) with a hilly country anomaly and come out with an anomaly in height that has any meaning for anything real?

December 5, 2010 10:39 pm

Demetris’ paper showing deterministic chaos in a toy model was published in 2006 and is here, rather than at the “2010” linked above. It’s important to note that the hindcast reconstruction method Demetris and his co-authors used maximized the likelihood that the GCMs would get it right.
In a conversation I had with Demetris by email awhile ago, he agreed that hindcast tests such as he’s done really could have, and should have, been done 20 years ago. That would have spared us all a lot of trouble and expense.
It’s a fair question to ask professionals such as Gavin Schmidt or Kevin Trenberth why they didn’t think to test what they have so long asserted.
Thanks for posting the work, Anthony, and Demetris may you and your excellent colleagues walk in the light of our admiration, and may some well-endowed and altruistic foundation take benevolent notice of your fine work. 🙂

Robert M
December 5, 2010 10:39 pm

It’s worse then we thought!
and…
There are way too many Roberts, Rob Ms and Me Hanging around this board…

Roger Carr
December 5, 2010 10:43 pm

Robert says: (December 5, 2010 at 10:00 pm) I can’t think of any field where this kind of inaccuracy in modeling would be OK. No place where real money is at stake, certainly.
There is real money at stake, Robert. That would seem to be the reason why “this kind of inaccuracy in modeling” is acceptable, or perhaps that should be welcomed, or perhaps even “necessary”…

english2016
December 5, 2010 10:44 pm

At what point does Al Gore get charged with perpetrating a fraud, and the nobel prize withdrawn?

ZZZ
December 5, 2010 10:46 pm

I don’t think that following the suggestion presented in this latest paper about how to model climate — namely, assuming that the change in climate is a stochastic, random process modulated by the physical processes involved — will affect the basics of the climate argument much. Instead of asserting that there will be, say, a 4C temperature rise in the next century, now alarmists will say something like “there is a 90% chance of a 4C or higher temperature rise in the next century” and the skeptics will retort that, to the contrary, the chance of that happening is much smaller — that it is less than, say, 10%. The essence of the argument will not go away, and the new stochastic climate forecasts will come to resemble current short-range weather forecasts with their predictions that tomorrow there is a certain percentage chance of rain, snow, etc. Another point worth remembering is that we do observe a great deal of non-randomness in how the climate changes over very long times — for example the beginning and end of the ice ages have followed a fairly regular 100,000 year to 130,000 year cycle over the last several million years.

Andrew30
December 5, 2010 10:47 pm

Climate models are an illustration of what the ‘climate scientist’ wants you to thinks that they know.
Comparing a climate model to reality is a simple way of illustrating the ‘climate scientists’ ignorance, amorality and hubris.

Doug in Seattle
December 5, 2010 10:57 pm

The team will argue that Koutsoyiannis and his little band of deniers in Greece are playing in the wrong sand box again. They are hydrologists – not climatologists and their article is not published in a team controlled approved journal.

Leon Brozyna
December 5, 2010 10:57 pm

Computer models? Something like the models the National Weather Service used to forecast our recent snow event here in Buffalo?
Let’s see how well that one turned out.
Wednesday morning they were still calling for us to get 2 to 4 inches of snow before the band of lake effect snow drifted south to ski country. They were doing pretty good, till that evening, when the band changed direction and drifted back to its start point over us and stayed put and kept on dumping snow so that, instead of 2-4 inches, we got 2-3 feet of heavy wet snow.
The only thing the models got right was there was going to be lake effect snow. They didn’t even get the amount right. Originally they called for 1-2 feet in ski country. My neighboring town got 42″, I only got about 30″. Ski country only got a few inches.
One good thing … I was able to burn hundreds of calories an hour shoveling all that global warming.
An added word about lake effect events (snow and even rain) … they are a bear to nail just right. Wind speed, humidity, direction, temperature all have to be just right. I can sympathize with what local forecasters have to come up with. But I don’t give that kind of understanding to the climate scientists and their models which pretend to encompass the globe and cover all possible variations.

anna v
December 5, 2010 11:00 pm

Pat Frank says:
December 5, 2010 at 10:39 pm
In a conversation I had with Demetris by email awhile ago, he agreed that hindcast tests such as he’s done really could have, and should have, been done 20 years ago. That would have spared us all a lot of trouble and expense.
It’s a fair question to ask professionals such as Gavin Schmidt or Kevin Trenberth why they didn’t think to test what they have so long asserted.

But I am sure they did have these studies. That is why they came up with the brilliant idea of anomalies, as I discuss above.
It is true that matter when subjected to forces behaves in a similar manner just because there is a limited way the gravitational and electromagnetic forces can impact it and a limited way that matter can react: elastic, inelastic, etc. That is why when we look at a terrain we need an absolute scale to be able to know whether we are looking at low level rock formations or the Alps. The scale is very important to life and limb. Similarly for waves, we need an absolute scale to know whether it is a storm in a teacup or out in the wild wild world.

anna v
December 5, 2010 11:06 pm

ZZZ says:
December 5, 2010 at 10:46 pm
I don’t think that following the suggestion presented in this latest paper about how to model climate — namely, assuming that the change in climate is a stochastic, random process modulated by the physical processes involved — will affect the basics of the climate argument much. Instead of asserting that there will be, say, a 4C temperature rise in the next century, now alarmists will say something like “there is a 90% chance of a 4C or higher temperature rise in the next century”
The models as they are now do not propagate errors and thus cannot give a consistent probability of output expected. The spaghetti plots are sleight of hand, where is the pea method of hiding this.
Once one gets models that have error propagation, trust will go up because they will be truly predicting and not handwaving.
I still believe that an analogue computer specifically designed for climate would be a solution. Or chaotic models on the lines of Tsonis et al .

December 5, 2010 11:20 pm

Thanks very much, Anthony, for this post, and Pat and all for the comments.
You may find interesting to see the accompanying Editorial by Zbigniew W. Kundzewicz and Eugene Z. Stakhiv in the same issue (go to the official journal page linked above and hit “issue 7” around the top of the page to get to the issue contents). There is also a counter-opinion paper by R. L. Wilby just below the Editorial.

Dean McAskil
December 5, 2010 11:32 pm

With my pathetic undergraduate mathematics I thought the comments in para 2 of the conclusions were self evident.
I am sure I have read studies and mathematical proofs before (I will have to hunt them down), relating to climate, financial markets and purely mathematical models, that proved conclusively that stochastic systems with many degrees of freedom simply cannot be modelled to predict extreme events. In particular for financial markets extreme events such as bubbles or busts cannot be predicted at all. This would seem analogous to climate warming runaway, or cooling for that matter.
And this doesn’t appear to me to be a difficult concept to grasp. I have never quite understood this faith that AGW proponents place in the GCM models. So I assumed it was my ignorance.

December 5, 2010 11:34 pm

very kool Dr. K
Voronoi cells. Hydrology types seem to be the only guys who use this. Nice

Phillip Bratby
December 5, 2010 11:50 pm

In a nutshell, climate models are unvalidated.
In a word, climate models are invalid.

davidmhoffer
December 5, 2010 11:50 pm

Lew Skannen;
The second weakest link in the whole AGW chain is the modelling….
(The first weakest link, by the way is the steering mechanism that AGW alarmists think that they have control over>>
The weakest link is in fact the physics. The models, the temperature records, the glaciers receding, the tree ring proxies, sea ice extent…these are all distractions from the actual physics. The AGW proponents keep changing the subject to one misleading data set to the next until their arguments is so warped that polar bear populations quadrupling becomes proof that they are going extinct due to global warming.
The fact is that they won’t discuss the physics because they can’t win the argument on the fundamentals, so they ignore them. But it won’t change the facts:
CO2 is logarithmic. The most warming that CO2 can cause is long behind us and it would take centuries of fossil fuel consumption at ten to a hundred times what we are using now to get another degree out of it over what we are already getting.
Almost no warming happens at the equator, the most happens at the poles. Most of what happens at the poles happens during winter. Most of what happens during the winter happens at the night time low.
So a really hot day at the equator goes from a day time high of +36 to +36.1 and a really cold night, in winter, at the pole, goes from -44 to -36. The lions and tigers aren’t likely to notice, and neither will the polar bears. Well, unless a climatologist shows up to study the polar bears and it warms up enough that they come out of hibernation. On the other hand, WE might notice less in that case due to a sparcity of climatologists.

1 2 3 6