One of the biggest, if not the biggest issues of climate science skepticism is the criticism of over-reliance on computer model projections to suggest future outcomes. In this paper, climate models were hindcast tested against actual surface observations, and found to be seriously lacking. Just have a look at Figure 12 (mean temperature -vs- models for the USA) from the paper, shown below:

The graph above shows temperature in the blue lines, and model runs in other colors. Not only are there no curve shape matches, temperature offsets are significant as well. In the study, they also looked at precipitation, which fared even worse in correlation. The bottom line: if the models do a poor job of hindcasting, why would they do any better in forecasting? This from the conclusion sums it up pretty well:
…we think that the most important question is not whether GCMs can produce credible estimates of future climate, but whether climate is at all predictable in deterministic terms.
Selected sections of the entire paper, from the Hydrological Sciences Journal is available online here as HTML, and as PDF ~1.3MB are given below:
A comparison of local and aggregated climate model outputs with observed data
Anagnostopoulos, G. G. , Koutsoyiannis, D. , Christofides, A. , Efstratiadis, A. and Mamassis, N. ‘A comparison of local and aggregated climate model outputs with observed data’, Hydrological Sciences Journal, 55:7, 1094 – 1110
Abstract
We compare the output of various climate models to temperature and precipitation observations at 55 points around the globe. We also spatially aggregate model output and observations over the contiguous USA using data from 70 stations, and we perform comparison at several temporal scales, including a climatic (30-year) scale. Besides confirming the findings of a previous assessment study that model projections at point scale are poor, results show that the spatially integrated projections are also poor.
Citation Anagnostopoulos, G. G., Koutsoyiannis, D., Christofides, A., Efstratiadis, A. & Mamassis, N. (2010) A comparison of local and aggregated climate model outputs with observed data. Hydrol. Sci. J. 55(7), 1094-1110.
According to the Intergovernmental Panel on Climate Change (IPCC), global circulation models (GCM) are able to “reproduce features of the past climates and climate changes” (Randall et al., 2007, p. 601). Here we test whether this is indeed the case. We examine how well several model outputs fit measured temperature and rainfall in many stations around the globe. We also integrate measurements and model outputs over a large part of a continent, the contiguous USA (the USA excluding islands and Alaska), and examine the extent to which models can reproduce the past climate there. We will be referring to this as “comparison at a large scale”.
This paper is a continuation and expansion of Koutsoyiannis et al. (2008). The differences are that (a) Koutsoyiannis et al. (2008) had tested only eight points, whereas here we test 55 points for each variable; (b) we examine more variables in addition to mean temperature and precipitation; and (c) we compare at a large scale in addition to point scale. The comparison methodology is presented in the next section.
While the study of Koutsoyiannis et al. (2008) was not challenged by any formal discussion papers, or any other peer-reviewed papers, criticism appeared in science blogs (e.g. Schmidt, 2008). Similar criticism has been received by two reviewers of the first draft of this paper, hereinafter referred to as critics. In both cases, it was only our methodology that was challenged and not our results. Therefore, after presenting the methodology below, we include a section “Justification of the methodology”, in which we discuss all the critical comments, and explain why we disagree and why we think that our methodology is appropriate. Following that, we present the results and offer some concluding remarks.
Here’s the models they tested:
Comparison at a large scale
We collected long time series of temperature and precipitation for 70 stations in the USA (five were also used in the comparison at the point basis). Again the data were downloaded from the web site of the Royal Netherlands Meteorological Institute (http://climexp.knmi.nl). The stations were selected so that they are geographically distributed throughout the contiguous USA. We selected this region because of the good coverage of data series satisfying the criteria discussed above. The stations selected are shown in Fig. 2 and are listed by Anagnostopoulos (2009, pp. 12-13). 
Fig. 2. Stations selected for areal integration and their contribution areas (Thiessen polygons).
In order to produce an areal time series we used the method of Thiessen polygons (also known as Voronoi cells), which assigns weights to each point measurement that are proportional to the area of influence; the weights are the “Thiessen coefficients”. The Thiessen polygons for the selected stations of the USA are shown in Fig. 2.
The annual average temperature of the contiguous USA was initially computed as the weighted average of the mean annual temperature at each station, using the station’s Thiessen coefficient as weight. The weighted average elevation of the stations (computed by multiplying the elevation of each station with the Thiessen coefficient) is Hm = 668.7 m and the average elevation of the contiguous USA (computed as the weighted average of the elevation of each state, using the area of each state as weight) is H = 746.8 m. By plotting the average temperature of each station against elevation and fitting a straight line, we determined a temperature gradient θ = -0.0038°C/m, which implies a correction of the annual average areal temperature θ(H – Hm) = -0.3°C.
The annual average precipitation of the contiguous USA was calculated simply as the weighted sum of the total annual precipitation at each station, using the station’s Thiessen coefficient as weight, without any other correction, since no significant correlation could be determined between elevation and precipitation for the specific time series examined.
We verified the resulting areal time series using data from other organizations. Two organizations provide areal data for the USA: the National Oceanic and Atmospheric Administration (NOAA) and the National Aeronautics and Space Administration (NASA). Both organizations have modified the original data by making several adjustments and using homogenization methods. The time series of the two organizations have noticeable differences, probably because they used different processing methods. The reason for calculating our own areal time series is that we wanted to avoid any comparisons with modified data. As shown in Fig. 3, the temperature time series we calculated with the method described above are almost identical to the time series of NOAA, whereas in precipitation there is an almost constant difference of 40 mm per year. 
Fig. 3. Comparison between areal (over the USA) time series of NOAA (downloaded from http://www.ncdc.noaa.gov/oa/climate/research/cag3/cag3.html) and areal time series derived through the Thiessen method; for (a) mean annual temperature (adjusted for elevation), and (b) annual precipitation.
Determining the areal time series from the climate model outputs is straightforward: we simply computed a weighted average of the time series of the grid points situated within the geographical boundaries of the contiguous USA. The influence area of each grid point is a rectangle whose “vertical” (perpendicular to the equator) side is (ϕ2 – ϕ1)/2 and its “horizontal” side is proportional to cosϕ, where ϕ is the latitude of each grid point, and ϕ2 and ϕ1 are the latitudes of the adjacent “horizontal” grid lines. The weights used were thus cosϕ(ϕ2 – ϕ1); where grid latitudes are evenly spaced, the weights are simply cosϕ.
It is claimed that GCMs provide credible quantitative estimates of future climate change, particularly at continental scales and above. Examining the local performance of the models at 55 points, we found that local projections do not correlate well with observed measurements. Furthermore, we found that the correlation at a large spatial scale, i.e. the contiguous USA, is worse than at the local scale.
However, we think that the most important question is not whether GCMs can produce credible estimates of future climate, but whether climate is at all predictable in deterministic terms. Several publications, a typical example being Rial et al. (2004), point out the difficulties that the climate system complexity introduces when we attempt to make predictions. “Complexity” in this context usually refers to the fact that there are many parts comprising the system and many interactions among these parts. This observation is correct, but we take it a step further. We think that it is not merely a matter of high dimensionality, and that it can be misleading to assume that the uncertainty can be reduced if we analyse its “sources” as nonlinearities, feedbacks, thresholds, etc., and attempt to establish causality relationships. Koutsoyiannis (2010) created a toy model with simple, fully-known, deterministic dynamics, and with only two degrees of freedom (i.e. internal state variables or dimensions); but it exhibits extremely uncertain behaviour at all scales, including trends, fluctuations, and other features similar to those displayed by the climate. It does so with a constant external forcing, which means that there is no causality relationship between its state and the forcing. The fact that climate has many orders of magnitude more degrees of freedom certainly perplexes the situation further, but in the end it may be irrelevant; for, in the end, we do not have a predictable system hidden behind many layers of uncertainty which could be removed to some extent, but, rather, we have a system that is uncertain at its heart.
Do we have something better than GCMs when it comes to establishing policies for the future? Our answer is yes: we have stochastic approaches, and what is needed is a paradigm shift. We need to recognize the fact that the uncertainty is intrinsic, and shift our attention from reducing the uncertainty towards quantifying the uncertainty (see also Koutsoyiannis et al., 2009a). Obviously, in such a paradigm shift, stochastic descriptions of hydroclimatic processes should incorporate what is known about the driving physical mechanisms of the processes. Despite a common misconception of stochastics as black-box approaches whose blind use of data disregard the system dynamics, several celebrated examples, including statistical thermophysics and the modelling of turbulence, emphasize the opposite, i.e. the fact that stochastics is an indispensable, advanced and powerful part of physics. Other simpler examples (e.g. Koutsoyiannis, 2010) indicate how known deterministic dynamics can be fully incorporated in a stochastic framework and reconciled with the unavoidable emergence of uncertainty in predictions.
h/t to WUWT reader Don from Paradise
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.

AGW is being demolishd from all sides. The above peer-reviewed report demolishes AGW’s GCM’s, the following demolishes the CO2 is causing global warming theory:
http://notrickszone.com/2010/12/05/conference-recap-and-my-big-disappointment/
>Nir Shaviv then explained solar activity and climate, showing that the link between solar activity, cosmic rays and cloud formation is pretty much in the bag. CERN will be soon releasing results on this, too. <
Oh, I am so underwhelmed I could just fart.
Global. GLOBAL. G.L.O.B.A.L.
If you wanted to make the GCM’s look really bad, why not get their predictions for Koolyanobbing in Western Australia? I’m sure that one would look bad too.
The idea is that it is a global phenomenon. The whole world gets warmer on average. Any predictions of local climate by a GCM is bound to be far less accurate.
Back to the drawing board guys. Better luck next time.
The short-term models that the Met Office uses regularly predict barbeque summers and mild winters – while the reality is washout summers and a Siberian Xmas. Why should we believe anything they say?
Here is Christopher Booker’s take on this, in yesterday’s Telegraph. A humorous piece:
http://www.telegraph.co.uk/comment/columnists/christopherbooker/8181558/Cancun-climate-conference-the-warmists-last-Mexican-wave.html
.
And anyway, the models DO fit the hindcast data.
It is not the models that are at fault, here, it is the faulty data – but with a little homogeonisation and rebasing, the data WILL conform to the models.
Simple really.
Can anyone run the models backwards? I want to see, how it matches
a) CET record
http://climate4you.com/CentralEnglandTemperatureSince1659.htm
b) Greenland ice core record
http://www.ncdc.noaa.gov/paleo/pubs/alley2000/alley2000.gif
As has been noted many times. The ONLY way to validate ANY model is to have it predict the future correctly a sufficient number of times that the match could not have happened by chance. Since climate models “tell” us what is supposed to happen 20, 30, 50 or 100 years in the future, they can NEVER be validated.
“we think that the most important question is not whether GCMs can produce credible estimates of future climate, but whether climate is at all predictable in deterministic terms.”
Internal Randomness generated within the various climate systems precludes the possibility of predictability as Stephen Wolfram proves in A New Kind of Science. Systems with internally generated randomness can’t be predicted, to know their future state you must observe then in real time.
Read the editorial that prof. Koutsogiannis suggested above.
The editorial is fair, impartial and stressing the need of useful modeling for applications.
There is a cognitive dissonance in people who can think that billions can be spent on the predictions of models for climate control, when the said models fail miserably when questioned on large scale details.
Moderator, this is the correct link for 10 above in the post. The one there points to the paper discussed once more.
A very interesting, and very frank work! Congratulations on your paper, Demetris and team!
I’ve come to regard GCMs as invaluable in one particular respect: They are a demonstration of how LITTLE we understand our climate system.
I’ve become increasingly frustrated with the climate science establishment’s over-reliance on climate modelling, and their wilful abuse of the Scientific Method by treating the product of models as if it were experimental data.
A climate model run is at best, manifestly, an exploration of a hypothesis, or an extrapolation of a hypothesis. It is not a TEST of that hypothesis. The only test of a climate model’s veracity with any merit whatsoever is its juxtaposition with good old, real-life observational data.
And, most importantly, that test is itself subject to acute scrutiny, requiring the model’s output to match observation for the right reasons. A model that matches observational data for the wrong reasons is not valid.
The climate science community’s collective failure, to recognise that model results are science invention rather than scientific discovery, is a fundamental flaw in the discipline and may yet prove its undoing.
Frankly if the models fail to correctly support past known factors , there is no reason to believe it can accurately predict future unknown factors when these are the same factors.
“David says:
December 6, 2010 at 1:26 am”
No, it is consistent with the political concensus that C02 is, somehow, carbon pollution. I think a better solution would be, as users of x-IBM hardware did in the UK with “mainframes” apparently in the late 70’s and early 80’s being used as anchor points in harbours for boats (Apparently there are some x-machine in the solent, around Portsmouth and Gosford).
ZZZ Says:
“Another point worth remembering is that we do observe a great deal of non-randomness in how the climate changes over very long times — for example the beginning and end of the ice ages have followed a fairly regular 100,000 year to 130,000 year cycle over the last several million years.”
No, they haven’t. They followed a approximately 40,000 cycle until about 1,000,000 years ago, then gradually shifted to an 100,000 year cycle over 200,000 years, and have followed an approximately 100,000 year cycle since then (8 cycles up till now).
This shift is obviously fundamental to understanding the causes of ice ages, but unfortunately the reason for it is not well understood (which is the accepted way to say “nobody has a clue” in a scientific context).
There are multiple studies revealing the whys & wheres of the poor performance of models. C3 site has a compilation of stories: http://www.c3headlines.com/climate-models/
Antonis Christofides: Well done. Do you foresee a bigger response from the Gavin et al than you received for your earlier paper?
@ur momisugly John Brookes
‘The idea is that it is a global phenomenon. The whole world gets warmer on average. Any predictions of local climate by a GCM is bound to be far less accurate’
But years ago when climate models were first found to be crap, the reasons given were that they were done on too big a scale. They were not precise enough. And that by investing zillions of resources they could do ever more detailed (=more localised = smaller grid cells) predictions that would therefore be much more accurate.
The resources were forthcoming…the improved models were run. And are still crap.
How can we reconcile this bit of history with your remarks that the problem now is ‘too much precision’ please?
While I love good old Anglo-Saxon terms, the accepted term for what you are calling “hindcasting” is “retrocasting.” I am not suggesting that you change terms.
anna v says:
December 5, 2010 at 11:06 pm
Anna the problem with models of chaotic systems (weather models) is that they randomly break down from the moment you start the model rolling. The Met offices have what they periods of predictive and no-predictive outcomes. That is to say that sometimes they can run the model many times with slightly different starting parameters and the outcome remains APPROXIMATELY the same for each run. Other times when they perform the same experiment the outcomes are completely chaotic.
In weather terms you will here them say that there is “some uncertanty in the forecast. Therefore, the significant breakdown can be relatively quick (within 12 hours) or relatively slow (within 3 days). Ilike the idea of an analogue computer but the same problems apply. The most important player in our ‘climate’ system ‘, as opposed to weather, is the ENSO and that cannot be predicted either in time or magnitude. Therefore, a viable climate model will almost certainly remain a mythical beast.
Simon Hopkinson says:
December 6, 2010 at 3:59 am
“A climate model run is at best, manifestly, an exploration of a hypothesis, or an extrapolation of a hypothesis. It is not a TEST of that hypothesis. The only test of a climate model’s veracity with any merit whatsoever is its juxtaposition with good old, real-life observational data.”
The situation is even worse than described. Even if one had “good old, real-life observational data,” the hypotheses that would explain and predict that data are missing. The climate models contain no hypotheses. The proof of that fact is easy. Everyone knows that a run of a climate model cannot falsify some component of that model. All that a run can show is the results that are to be expected given various assumptions. None of the assumptions are even partially confirmed in their own right apart from the model. In other words, none of the assumptions have any empirical meaning whatsoever.
by the by I also like Lucia’s analysis. I can’t always agree with her prognostications but some of her work is very good.
We should not really be surprised by these results because, contrary to what was stated in the paper, the IPCC did test the models. The problem is they also found them wanting. Buried deep in the several hundred pages of the technical section is a series of tests to which the subjected the most important models. It is a long time since I read the report but, from memory, about 11 models were tested against 6 scenarios like the 20th centrury cyclical pattern, the annual ice melt and Roman warm period. None of the models performed well on all tests and all of them had to employ fudge factors in order to make the data fit any of the hindcasts.
To the scientists’ credit none of this was hidden. The problem came in the technical summary and worse still in the summary for policy makers where all these errors miraculously turned into scientific certainty. This was when I stopped being sceptical about the science and became sceptical about the motives of those who were “managing” the science.
I believe “hindcast” is used here in the sense of “backfit”. Indeed, the climatologists’ models do not backfit the available data time series well enough to mean anything as far as I can see. So obviously the models require no further examination. Long since time to throw them in the trash. The absolute last straw, to me, is the “correcting” of the data time series to produce the desired models. This is fraud. A hoax.
This is very sad. A whole “scientific” specialty appears to have gone mad. Crying “Wolf!” so loudly and so long means no relatively sane person will believe the cry of “”Wolf!” even if the wolf is real.
We could easily have predicted this outcome simply by looking at the old Wilson Spaghetti graph. These models do a fair job of predicting each other — when zoomed out to the 1000 year scale, anyway — in the time period to which they are all “tuned”. But none of them agree on paleo climate with any reliability.
One can only assume given the paleo-divergence of these models that a future divergence is also quite likely.
Having read through many of their papers, it is my view that the Itia group at NTUA, headed by Prof. Koutsoyiannis, are putting forward the most complete, thoughtful and scientific critique of catastrophic AGW that I am aware of today.
Many congratulations to Anagnostopoulos et al for taking on the many-headed hydra of climate peer review and for publishing another good paper chiselling away at the current stagnant view of climate science.
I would think precipitation would be an even bigger headache to model. It’s dependent on clouds, prevailing winds at all relevant altitudes, temperature at all relevant altitudes AND geography just to name a few. Anyone selling you on funding them for research on predicting precipitation patterns is selling snake oil.
I mean, weathermen today have the advantage of imagery telling them where cells of precipitation are, and what direction they’re headed. Imagine taking all of that away and asking someone to tell you when it’s going to rain. Can’t do it.
Dean McAskil says:
December 5, 2010 at 11:32 pm
“I am sure I have read studies and mathematical proofs before (I will have to hunt them down), relating to climate, financial markets and purely mathematical models, that proved conclusively that stochastic systems with many degrees of freedom simply cannot be modelled to predict extreme events. In particular for financial markets extreme events such as bubbles or busts cannot be predicted at all. ”
I would modify the statement to, “cannot be predicted from the models at all.”
The existence of a bubble may not be obvious to the models, but I think we can all agree that humans who are not directly involved in the bubble, may easily discern their existence. To wit, the Dot Com Bubble, the Housing Bubble, South Sea Bubble, Mississippi Bubble, Tulip Mania, were all clearly visible to external observers.
To this extent, I’d like to rename the entire AGW mess to Climate Bubble. It sure seems to fit, with massive amounts of funds trading hands on little or no reality.