One of the biggest, if not the biggest issues of climate science skepticism is the criticism of over-reliance on computer model projections to suggest future outcomes. In this paper, climate models were hindcast tested against actual surface observations, and found to be seriously lacking. Just have a look at Figure 12 (mean temperature -vs- models for the USA) from the paper, shown below:

The graph above shows temperature in the blue lines, and model runs in other colors. Not only are there no curve shape matches, temperature offsets are significant as well. In the study, they also looked at precipitation, which fared even worse in correlation. The bottom line: if the models do a poor job of hindcasting, why would they do any better in forecasting? This from the conclusion sums it up pretty well:
…we think that the most important question is not whether GCMs can produce credible estimates of future climate, but whether climate is at all predictable in deterministic terms.
Selected sections of the entire paper, from the Hydrological Sciences Journal is available online here as HTML, and as PDF ~1.3MB are given below:
A comparison of local and aggregated climate model outputs with observed data
Anagnostopoulos, G. G. , Koutsoyiannis, D. , Christofides, A. , Efstratiadis, A. and Mamassis, N. ‘A comparison of local and aggregated climate model outputs with observed data’, Hydrological Sciences Journal, 55:7, 1094 – 1110
Abstract
We compare the output of various climate models to temperature and precipitation observations at 55 points around the globe. We also spatially aggregate model output and observations over the contiguous USA using data from 70 stations, and we perform comparison at several temporal scales, including a climatic (30-year) scale. Besides confirming the findings of a previous assessment study that model projections at point scale are poor, results show that the spatially integrated projections are also poor.
Citation Anagnostopoulos, G. G., Koutsoyiannis, D., Christofides, A., Efstratiadis, A. & Mamassis, N. (2010) A comparison of local and aggregated climate model outputs with observed data. Hydrol. Sci. J. 55(7), 1094-1110.
According to the Intergovernmental Panel on Climate Change (IPCC), global circulation models (GCM) are able to “reproduce features of the past climates and climate changes” (Randall et al., 2007, p. 601). Here we test whether this is indeed the case. We examine how well several model outputs fit measured temperature and rainfall in many stations around the globe. We also integrate measurements and model outputs over a large part of a continent, the contiguous USA (the USA excluding islands and Alaska), and examine the extent to which models can reproduce the past climate there. We will be referring to this as “comparison at a large scale”.
This paper is a continuation and expansion of Koutsoyiannis et al. (2008). The differences are that (a) Koutsoyiannis et al. (2008) had tested only eight points, whereas here we test 55 points for each variable; (b) we examine more variables in addition to mean temperature and precipitation; and (c) we compare at a large scale in addition to point scale. The comparison methodology is presented in the next section.
While the study of Koutsoyiannis et al. (2008) was not challenged by any formal discussion papers, or any other peer-reviewed papers, criticism appeared in science blogs (e.g. Schmidt, 2008). Similar criticism has been received by two reviewers of the first draft of this paper, hereinafter referred to as critics. In both cases, it was only our methodology that was challenged and not our results. Therefore, after presenting the methodology below, we include a section “Justification of the methodology”, in which we discuss all the critical comments, and explain why we disagree and why we think that our methodology is appropriate. Following that, we present the results and offer some concluding remarks.
Here’s the models they tested:
Comparison at a large scale
We collected long time series of temperature and precipitation for 70 stations in the USA (five were also used in the comparison at the point basis). Again the data were downloaded from the web site of the Royal Netherlands Meteorological Institute (http://climexp.knmi.nl). The stations were selected so that they are geographically distributed throughout the contiguous USA. We selected this region because of the good coverage of data series satisfying the criteria discussed above. The stations selected are shown in Fig. 2 and are listed by Anagnostopoulos (2009, pp. 12-13). 
Fig. 2. Stations selected for areal integration and their contribution areas (Thiessen polygons).
In order to produce an areal time series we used the method of Thiessen polygons (also known as Voronoi cells), which assigns weights to each point measurement that are proportional to the area of influence; the weights are the “Thiessen coefficients”. The Thiessen polygons for the selected stations of the USA are shown in Fig. 2.
The annual average temperature of the contiguous USA was initially computed as the weighted average of the mean annual temperature at each station, using the station’s Thiessen coefficient as weight. The weighted average elevation of the stations (computed by multiplying the elevation of each station with the Thiessen coefficient) is Hm = 668.7 m and the average elevation of the contiguous USA (computed as the weighted average of the elevation of each state, using the area of each state as weight) is H = 746.8 m. By plotting the average temperature of each station against elevation and fitting a straight line, we determined a temperature gradient θ = -0.0038°C/m, which implies a correction of the annual average areal temperature θ(H – Hm) = -0.3°C.
The annual average precipitation of the contiguous USA was calculated simply as the weighted sum of the total annual precipitation at each station, using the station’s Thiessen coefficient as weight, without any other correction, since no significant correlation could be determined between elevation and precipitation for the specific time series examined.
We verified the resulting areal time series using data from other organizations. Two organizations provide areal data for the USA: the National Oceanic and Atmospheric Administration (NOAA) and the National Aeronautics and Space Administration (NASA). Both organizations have modified the original data by making several adjustments and using homogenization methods. The time series of the two organizations have noticeable differences, probably because they used different processing methods. The reason for calculating our own areal time series is that we wanted to avoid any comparisons with modified data. As shown in Fig. 3, the temperature time series we calculated with the method described above are almost identical to the time series of NOAA, whereas in precipitation there is an almost constant difference of 40 mm per year. 
Fig. 3. Comparison between areal (over the USA) time series of NOAA (downloaded from http://www.ncdc.noaa.gov/oa/climate/research/cag3/cag3.html) and areal time series derived through the Thiessen method; for (a) mean annual temperature (adjusted for elevation), and (b) annual precipitation.
Determining the areal time series from the climate model outputs is straightforward: we simply computed a weighted average of the time series of the grid points situated within the geographical boundaries of the contiguous USA. The influence area of each grid point is a rectangle whose “vertical” (perpendicular to the equator) side is (ϕ2 – ϕ1)/2 and its “horizontal” side is proportional to cosϕ, where ϕ is the latitude of each grid point, and ϕ2 and ϕ1 are the latitudes of the adjacent “horizontal” grid lines. The weights used were thus cosϕ(ϕ2 – ϕ1); where grid latitudes are evenly spaced, the weights are simply cosϕ.
It is claimed that GCMs provide credible quantitative estimates of future climate change, particularly at continental scales and above. Examining the local performance of the models at 55 points, we found that local projections do not correlate well with observed measurements. Furthermore, we found that the correlation at a large spatial scale, i.e. the contiguous USA, is worse than at the local scale.
However, we think that the most important question is not whether GCMs can produce credible estimates of future climate, but whether climate is at all predictable in deterministic terms. Several publications, a typical example being Rial et al. (2004), point out the difficulties that the climate system complexity introduces when we attempt to make predictions. “Complexity” in this context usually refers to the fact that there are many parts comprising the system and many interactions among these parts. This observation is correct, but we take it a step further. We think that it is not merely a matter of high dimensionality, and that it can be misleading to assume that the uncertainty can be reduced if we analyse its “sources” as nonlinearities, feedbacks, thresholds, etc., and attempt to establish causality relationships. Koutsoyiannis (2010) created a toy model with simple, fully-known, deterministic dynamics, and with only two degrees of freedom (i.e. internal state variables or dimensions); but it exhibits extremely uncertain behaviour at all scales, including trends, fluctuations, and other features similar to those displayed by the climate. It does so with a constant external forcing, which means that there is no causality relationship between its state and the forcing. The fact that climate has many orders of magnitude more degrees of freedom certainly perplexes the situation further, but in the end it may be irrelevant; for, in the end, we do not have a predictable system hidden behind many layers of uncertainty which could be removed to some extent, but, rather, we have a system that is uncertain at its heart.
Do we have something better than GCMs when it comes to establishing policies for the future? Our answer is yes: we have stochastic approaches, and what is needed is a paradigm shift. We need to recognize the fact that the uncertainty is intrinsic, and shift our attention from reducing the uncertainty towards quantifying the uncertainty (see also Koutsoyiannis et al., 2009a). Obviously, in such a paradigm shift, stochastic descriptions of hydroclimatic processes should incorporate what is known about the driving physical mechanisms of the processes. Despite a common misconception of stochastics as black-box approaches whose blind use of data disregard the system dynamics, several celebrated examples, including statistical thermophysics and the modelling of turbulence, emphasize the opposite, i.e. the fact that stochastics is an indispensable, advanced and powerful part of physics. Other simpler examples (e.g. Koutsoyiannis, 2010) indicate how known deterministic dynamics can be fully incorporated in a stochastic framework and reconciled with the unavoidable emergence of uncertainty in predictions.
h/t to WUWT reader Don from Paradise
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.

Now tell me, how do we know what level of CO2 will give us 2degC of warming. An accurate figure would be appreciated.
I’m sure the guys and gals at UKCIP (http://www.ukcip.org.uk/) can tell us very accurately.
At http://ukclimateprojections.defra.gov.uk/content/view/857/500/
they provide projections of climate change, and absolute future climate for:
* Annual, seasonal and monthly climate averages.
* Individual 25 km grid squares, and for pre-defined aggregated areas.
* Seven 30 year time periods.
* Three emissions scenarios.
* Projections are based on change relative to a 1961–1990 baseline.
Now that is what I call good science.
So the models get more complicated and the computers more powerful, and all they are able to produce is just some kind of variation of the old Keeling curve.
Notice that GCMs are not able to model the PDO/AMO cycle at all, since all they are working with is the fictional “radiative forcing” concept. They are tuned to catch the 1975-2005 warming trend, but they wildly divorce with reality before and after.
Polar regions, allegedly the most sensitive areas to “increased greenhouse forcing” do not show any sign of it: Arctic shows just AMO variation and Antarctic shows even slight cooling.
http://i43.tinypic.com/14ihncp.jpg
This alone totally and unequivocally disqualifies the AGW theory. No other scientific hypotesis would survive such discrepancy between theory and observation. Shame on all scientists, who keep their mouth shut.
The editorial referred to above is here: http://www.informaworld.com/smpp/section?content=a928044622&fulltext=713240928
The Wilby paper I can’t find, unless it’s this one: http://www.informaworld.com/smpp/section?content=a928044062&fulltext=713240928
Pat Frank, thank you for your comments. You say:
Actually the 2010 is correct. It is a more recent toy model by Koutsoyiannis, in what we think is his best paper to date.
ZZZ:
The fact that you have cycles that resemble periodicity does not necessarily mean that that they are non-random. (See the 2010 paper by Koutsoyiannis linked above for a definition of randomness.) The toy model of Koutsoyiannis (in the same paper) has unpredictable cycles and trends without any difference in forcings. Having unpredictable cycles and trends can be more random than not having them, because if you have a constant long-term average “signal” plus “noise”, then you know more than if you don’t have even that. I also explain that in the epilogue of my HK Climate web site.
Dear Greeks, Well done! Very well done!
U.S. climate scientists: With more than a billion a year allocated for research … who wants answers?
Anna … You’re smart. Very smart.
Davidmhoffer … You too. [snip]
Our climate way well be too chaotic to model in the fine grain. But even the proverbially “chaotic” r^2 equation is bounded. In fact it’s closely constrained between a max and a min. Make an r^2 plot and has lots of chaotic ups and downs. But stand far enough back and all you see if a flat line. Our climate is probably like that. Certainly on multi-millennial scales. It’s almost never -20°C in LA, and never plus 35°C at the polls.
dT
anna v says: (December 5, 2010 at 10:32 pm) It is worth noting that this paper exposes the real reason why “anomalies” have been foisted on us. Anomalies, which are a delta(T) study the shapes of terrains ignoring the height. (and at December 5, 2010 at 11:00 pm) Similarly for waves, we need an absolute scale to know whether it is a storm in a teacup or out in the wild wild world.
Thank you, Anna. I always read your posts, and always either learn or find myself puzzling — my hope is that our present scientists-in-the-making do likewise.
Better still, perhaps our present crop will take note and learn from your wisdom (and gentle humour).
Is it ironic that AGW: global warming caused by human derived CO2 – is a figment of the imagination of an SiO2 based lifeform? The next element in the period is Germanium. Then Tin and Lead.
Very interesting (and very topical discussions with some real live climate modellers (all of whom are terrified of our esteemed host here and so won’t turn up) at Judith Curry’s blog.
Simply put they do not see verification and validation as a priority. Their work is ipso facto so brilliant that no checks (by observation or by external scrutiny) are needed.
Here’s the link.
Having failed to make accurate short-term forecasters (the Met office have an atrocious forecasting record …. and e.g. singularly failed to forecast the snow in Glasgow (I know because I started some outdoor work and removed all the gutters because the five day forecast was “light snow” on a single day followed by clear weather … I now have 2 foot icicles hanging from the entire roof)
So … clearly the modellers tried to justify the huge amount of money the public spent on their toy computers by claiming that “whilst we can’t predict the weather … we can predict the climate”.
HA HA HA HA HA HA HA!
What a joke they are!
‘Duh!’ as Mr Simpson would so cogently expresses his frustration with and scorn for the dumb ideas that supposedly-intelligent blokes with proper Phds and everything have persuaded the alarmist and alarmed world to take seriously for far too long.
When I was a boy, I built model airoplanes badly, being blessed with more than my share of metaphorical thumbs plus a very limited set of construction skills, but the quality of my models never stopped me dreaming they were real and capable of marvellous aeronautical feats. Sometimes they even flew.
When I grew up I realised that childhood dreams should remain in childhood, but, sadly, many clever kids never mature to full and responsible adulthood; a few of them become scientists and go on to scare the world with their clever but childish fantasies. And sadly, many ordinary people seem to like the thrill of being scared, but never willingly put themselves in any kind of actual danger, so the timid fasten on to the scaremongers for our ration of thrills and the Marxists fasten on to the models to satisfy their anti-humanity rage and control freakery. It is very interesting that many (but not all) of the alarmist tribe never go motor-racing, blue-water yachting, mountain-climbing or play vigorous contact sports and so never put themselves in a situation which might supply a good ration of adrenaline or even plain old-fashioned fun, which, apart from the joy of control or exterminating something, is not permitted by the Marxist doctrine.
And finally, we find the alarmist scientists’ juvenile climate models are c**p – wow, who woulda thunk it!!
Demetris,
I note that you did not cite the Wentz publication in the journal Science that found that the models reproduced less than one-third to one-half the increase in precipitation observed during the recent warming. Unfortunately, model based studies projecting increased risk of drought in the future fail to discuss this result and its implications for model credibility in representation of future warming associated increases in precipitation. I have wondered if they’ve been able to ignore the Wentz paper because it doesn’t report the specific model results. Your work may be more difficult to ignore, but it did not see a specific comparison of the increase in precipitation during the warm decades in the models relative to the increase seen in the observations. It would help to know if your results are consistent with those of Wentz. I’d have to read the Wentz paper again, but I assume we’d expect to see this in a comparison of the 80s or 90s with the 70s or 60s. Thanx.
Roger Carr says: “There is real money at stake, Robert.“.
True. But is there a difference between the way in which money is at stake here, compared with how it is normally at stake? When organisations build models, it is usually with the aim of making money or saving money. Their own money. So the models tend to get tested carefully, and are ditched if they aren’t up to standard.
Is it possible that the climate models’ true aim is to lose vast amounts of other people’s money? If so, then accuracy in the models would not help achieve this, would therefore not be an objective, and need not be tested for. A model could then be coded to produce certain required results, and could remain in operation for as long as it can produce those results.
Nah. Absurd. Forget it.
Another classic example of the difference between real science and ‘climate science’ – no different from Mannian Maths versus real statistics.
Any real branch of science would have a pre-requisite of testing a model by hindcasting, but with ‘climate science’, this is is not deemed to desirable or necessary, unless the raw data has been sufficiently manipulated to fit the models.
I think the UK Met Office £33m computer, currently called ‘Deep Black’, should be renamed ‘Deep Sh*t’ and taken to PC World to get their/our money back…
Apart from the fact that this should have been done, thoroughly, before a single character of any IPCC report was struck I don’t mind taxes spent out on exposing the bleeding obvious if it saves us all money.
How can there be offsets. The absolute temperature is of critical importance to the radiation balance and it must be an initial condition of the simulation. If they don’t get that right, especially with an error of 2.5-3 degress C, the error exceeds the so called CO2 forcing during the 20th century several times.
Also, I’ve spent many years fighting divergences in numerical simulations. With linear problems, it can be almost done, at least in many cases. I have always assumed that people like Gavin S. with degrees in applied math have these things under control. Now, I’m not all that confident they do.
This confirms what we knew in essence, that its darn hard to model over a long period of time any system which is under a constant state of flux as to what exactly constitutes the system and hence where the ‘model boundary’ should exist – on so many dimensions. Models work best with well defined and understood materials within a finite and often small ‘domain’ – as soon as you go beyond this the usefulness of the model quickly tails off and you just end up observing what is in essence noise.
Although very nice to have it be written up in the literature in such a precise way I must say – well done!
Fixed.
Wonderful work. Drive the stake deeper into the heart of the GCM’s, boys. This waste of money has to be stopped.
AGW is based on physics.
CAGW is based on computer models.
AGW may or may not be happening.
CAGW is plainly rubbish.
In this paper climate models were hindcast tested against actual surface observations, and found to be seriously lacking.
Let’s be clear about this. GCM’s do not match the past, which is known. The modelers disdain the past. They don’t care that their models don’t fit known data, i.e. the reality of the known past. They excuse that lack of correlation as too mundane to matter. Their theoretical framework, a gross oversimplification of a complex system, is not based on empirical data. Their model outputs fail when compared to known reality.
It’s laughable to consider whether GCM “projections” into the future are “science”. Science is the study of reality (we could argue about what “reality” is, and get bogged down in epistemology, but there is a common ground we mostly all share). The GCM’s do not comport with reality as is generally agreed upon, and as indeed the modelers themselves agree upon.
If the nut does not fit the bolt today, it is not going to suddenly fit that bolt tomorrow.
Robert says: December 5, 2010 at 10:00 pm
All the Roberts, I can think of where modelling is useful.
Self interest.
What’s the bet, that is the the parameters (ever shape shifting) that frame the financial, gambling, voting, acquiescence, alcohol (or any other addictive substance) devised enterprises FOR and OF human nature which one cares to mention. Neuro-cognitive research is in it’s infancy.
It was Friedman or von Mises that stated ‘own interest’ not ‘self interest’ I thought. But I will check.
Also what is hindcast? I may have missed the original discussion, but this is a new word. A priori and posteriori is understandable.
I would appreciate an explanation on hindcast thanks.
Not CO2. Tea For 2…
TF2.
That global climate models are not very good at local and regional levels and on shorter times scales is well known. This Science News article gives an overview of work being done to improve models. Most of the article is about modeling aerosols but the last section “Getting Local” deal with efforts to improve local projections. I did not see anything shockingly new in the Greek research paper.
http://www.sciencenews.org/view/feature/id/65734/title/The_final_climate_frontiers
REPLY: It wasn’t meant to be “shockingly new”, that’s your panic descriptor. It is an update to a previous paper. – Anthony
Mike D. says: (December 6, 2010 at 2:03 am) If the nut does not fit…
Please be more specific, Mike. Which nut? There are so many…