The Rest of the Cherries: 140 decades of Climate Models vs. Observations
by Roy W. Spencer, Ph. D.
Since one of the criticisms of our recent Remote Sensing paper was that we cherry-picked the climate models we chose to compare the satellite observations of climate variations to, here are all 140 10-year periods from all 14 climate models’ 20th Century runs we analyzed (click to see the full res. version):
As you can see, the observations of the Earth (in blue, CERES radiative energy budget versus HadCRUT3 surface temperature variations) are outside the range of climate model behavior, at least over the span of time lags we believe are most related to feedbacks, which in turn determine the sensitivity of the climate system to increasing greenhouse gas concentrations. (See Lindzen & Choi, 2011 for more about time lags).
Now, at ZERO time lag, there are a few decades from a few models (less than 10% of them) which exceed the satellite measurements. So, would you then say that the satellite measurements are “not inconsistent” with the models? I wouldn’t.
Especially since the IPCC’s best estimate of future warming (about 3 deg C.) from a doubling of atmospheric CO2 is almost exactly the AVERAGE response of ALL of the climate models. Note that the average of all 140 model decades (dashed black line in the above graph) is pretty darn far from the satellite data.
So, even with all of 140 cherries picked, we still see evidence there is something wrong with the IPCC models in general. And I believe the problem is they are too sensitive, and thus are predicting too much future global warming.

I love it. Somewhere there is a model that sometimes agrees with observation at some lag or lead, like the proverbial broken watch that is right twice a day.
Ever look at the graph of model computed cloud cover for the earth as a function of latitude? None of them follow the observed behavior, though some of them agree at some latitudes. Now these models are supposed to give trustworthy results for how clouds will change when we add CO2, even though their starting point is known to be wrong.
Dang, I read a silly comment by sharparoo, I set out to articulate why it is, go back and scan SB11, line out my thoughts, start writing a bit, decide to refresh the page only to learn I’m too slow. 🙁
So, Sharp!!! What they said!!!
But, honestly, Sharp, I don’t think Dr. Spencer considered the possibility of the screams of cherry picking to begin with. It wasn’t the purpose of the paper to show all of the possibilities and find one that may have accidentally coincided. As a net admin, I run into a similar problem. I’ve colleagues and friends that are incredibly good hackers. I’ve as much knowledge as they, but my mind doesn’t work in the manner theirs do. So, when I set up a network or a node that needs to be secure, I ask one of my friends to break into it. They let me know how hard it was and whether or not I forgot a certain approach. Dr. Spencer needs to hire an alarmist cynic to read his papers first before submitting. But even after that, it would be impossible to cover every possible potential criticism regardless of how little rationale was behind the criticism. Maybe that’s what he’s using A.D. for.
“When a model does not produce an accurate result it is wrong. When it says 2+2 = 5 it is wrong.”
Not so fast. The other 1/5th may well be hiding in the depths of the oceans…
Greg (September 22, 2011 at 4:01 pm), good one!
“Not inconsistent with”. Really nice, scientific term and concept. My ex-wife repeatedly entering a hotel lobby with my best friend is “not inconsistent with” them having a monthly two-hour lunch, but I don’t think that would play to the home team.
sharper00 says: “[yatta-yatta]…Do you have any suggestions for what’s missing from the models?”
Humility, for starters.
Another piece of excellent work Dr. Spencer. Send the Alarmists and their hockey sticks back to where they cane from!
James Sexton: rather than wasting money on a warminista, we need to develop a AGW version of the old 8-ball. You know the one that would float an answer to a question in a window? Just take all the objections we’ve run across and put them in there – the AGW crowd is so predictable that we’d probably anticipate the objection at least 60% of the time.
that’s a compelling graphic, Dr. Spencer. a regular ralph steadman autograph. i’m sure jackson pollock would praise it as the product of highly ordered minds.
if mel brooks were a statistician, that’s the sort of chart he’d make as self referential satire.
the absurd extreme would be a hockey team eating pie made from those ‘cherries’ by mickey rooney.
a more californiaesque version would be the hockey team gloating over their conquest of virginity … n/m
so many tangled threads on that chart – it deserves to be a centerfold in Sew Age!
i know this comment doesn’t contribute substantially to the scientific discussion – but beyond the obvious point that models don’t resemble the observations – climate terrorism has never been about science. it’s pure schaden-food. makes my satire-iasis act up every time…
Wouldn’t the title be more apt if it read “Spencer b-slaps pedantic critics!”?
Dr. Spencer.
You write a paper and there is outrage.
You address the outrage and there is criticism.
You reply to the criticism and there is grumbling.
From outrage to grumbling. I call that progress.
Thanks!
temp says:
September 22, 2011 at 2:34 pm
…. Hate to break it to you but models are nothing more then the ramblings of the drunk guys at the bar. Sure when your drunk they make sense… sure to the drunks rambling it makes sense to him…Sure even drunks at the bar can make profound leaps of logic and thought…”
==========================
Interesting observation as it aligns perfectly with my (unpublished) hypothesis: “The Wilderbeest Effect”. Apologies to all who may share this hypothesis.
Abstract: We are all familiar with the annual mass migrations of wildlife on the plains of Africa. The carnivores follow the wild herds and pick off the weaklings at random. Essentially this act over time improves the gene pool of those herds. In principle alcohol has the same effect on drunks, it destroys the weak brain cells (easy picking) leaving on the best cells. The effects of this action is rather a lot more noticeable in the short-term and explains why a drunk always comes up with a smart comment at just the right time to get his head knocked off.
And, I’m sure that this is equally relevant to the warmistas.
The only real point in this exercise is that 10 years of CERES data is insufficient to validate any model. Maybe if we had 200 years of CERES data and there was a model that pretty much nailed all 200 years it might be credible in predicting the next 100 so long as nothing unique happens that didn’t happen in the past 200 years. Even then, that’s a huge assumption that the past 200 years contains a complete set of all climate-altering events. There are cyclical phenomena that we know for a fact run on longer cycles than that like for instance ice ages which are currently 120,000 year cycles. And we’re near a “tipping point” back to the cold side of those cycles which is what really makes concern over global warming preposterous. Any warming we get from anthropogenic sources are properly viewed as a bit more of a safety margin for whatever set of events conspires to end interglacial periods fercrisakes.
Got to this a bit late
sharper00 at September 22, 2011 at 1:00 pm
Based on Trenberth and Dessler it seems like ENSO simulation ability is a more important factor. What do you think?
Unproven – the models that do well on ENSO are controversial and some recent literature suggest that all the models T11 and D11 cite as good modeller of ENSO are not. In fact one of the models cited by D11 as being a better fit to to the S&B relationship (MRI CGCM 2.3.2A) is described in his own reference Lin (2007) as not fitting ENSO well.
Is it possible to produce a model which reproduces both ENSO and the CERES measurements but has a low climate sensitivity?
If the three best fits to the S&B relationships don’t clearly fit ENSO well I suspect that the issue of relationship to climate sensitivity is “a bit previous” as we say (i.e. premature).
AJStrata says:
September 22, 2011 at 2:43 pm
I just realized that my comment could be understood as critical of your comment. I did not intend that. I am with your comment all the way.
sharper00 says:
September 22, 2011 at 1:00 pm
Wrong response. You were supposed to complain that groups of dozens of the models likely have a nearly identical code base and differ only initialization and maybe a few coefficients (or the monthly changes to the historical temperature record in GISS).
Then you could suggest that taking a few of the warmest, a few of the coolest, and the overall mean might work out the best.
Dave Springer;
The only real point in this exercise is that 10 years of CERES data is insufficient to validate any model.>>>
But more than sufficient to INVALIDATE the models. If they cannot get the last ten years correct, then it is impossible for them to get AN period of time correct except as pure coincidence.
Further, and model that continues to be used that doesn’t take the CERES data into account, is by default being programmed to IGNORE REALITY.
Are you listending…Dr Kevin Trenberth….
Wanna take that resignation back Dr Wolfgang Wagner…
If we consider Dessler’s purported rebuttal of Spencer, and Spencer’s rebuttal of Dessler in this post, we necessarily conclude that 2 or 3 of the 13 models used by the IPCC in their climate forecasts do a better job of reflecting recent temperatures; these are the ones which better reproduce ENSO. None of the models faithfully reproduces world temperature trends, but at least these 2 or 3 reproduce the timing (though not the magnitude) of the heat loss following a warm maximum.
This relative superiority of 2 or 3 models does not and in principle cannot prove the “truth” of these models, but does demonstrate that they are superior to the 10 or so less competent models which are currently embedded in IPCC predictions. At the least, one would want to know what future climate trends are predicted by these better models, and whether and how they differ from current IPCC predictions. Whatever they predict may not be any more accurate than a ouija board, but at least have slightly improved credibility compared to those from demonstrably worse models.
Theo Goodwin (at 2:44 above) expresses scepticism with assertions that using these 2 or 3 better models would be a “better model of Earth’s climate.” If his point is that even the best of these models has not been demonstrated to be reliable at predicting climate, I completely agree. I wouldn’t bet a dime on the predictions of the best current models, much less the economic future of the developed world. But I do think that eliminating demonstrably worse models from consideration in favor of slightly better models represents a modest scientific advance. Science can never prove the truth of a proposition, but it can prove the falsity of one; and it can discriminate between a better explanation and a worse one.
If nothing else, the Spencer paper and ensuing controversy should lead to the dumping of the worst models currently used by the IPCC.
@Theo Goodwin:
September 22, 2011 at 3:27 pm
Models, in a general sense, are simply a representation of theories and hypotheses. Newton’s Law’s could therefore be described as a model, a really good model, that experience has shown consistently ‘works’ any most every-day situations. The Law’s, and therefore the models, do a poor job when you start getting to very high speeds, really small scales and high gravational fields, in which case something else needs to be added to the ‘models’ to make them work. Therefore, AJStrata is correct to describe the tools they use to predict where satelittes should be, and are going to be in the future, as models.
Climate models have to be based on some physical theories and hypotheses, otherwise you couldn’t put together the code for them to run. The fact that the climate models aren’t doing a good job suggests there’s something missing/wrong with the underlying theories and hypotheses, just like when people started to realise that Newton’s Laws weren’t holding anymore once people tried to apply them to situations other than the everyday.
I wonder if any of the following limitations of the IPCC models provided in Chapter 8 of the 2007 Report Working Group I “The Physical Science Basis” have anything to do with the problems identified by Spencer?
“Due to the computational cost associated with the requirement of a well-resolved stratosphere, the models employed for the current assessment do not generally include the QBO.”
“In short, most AOGCMs do not simulate the spatial or intra-seasonal variation of monsoon precipitation accurately.”
“The spatial resolution of the coupled ocean-atmosphere models used in the IPCC assessment is generally not high enough to resolve tropical cyclones, and especially to simulate their intensity.”
“Blocking events are an important class of sectoral weather regimes (see Chapter 3), associated with local reversals of the mid-latitude westerlies. There is also evidence of connections between North and South Pacific blocking and ENSO variability…but these connections have not been systematically explored in AOGCMs.”
“Unfortunately, the total surface heat and water fluxes (see Supplementary Material, Figure S8.14) are not well observed.”
“These errors in oceanic heat uptake will also have a large impact on the reliability of the sea level rise projections.”
“Due to the limited resolutions of the models, many of these processes are not resolved adequately by the model grid and must therefore be parametrized. The differences between parameterizations are an important reason why climate model results differ.”
I’ve noticed a sparsity of trolls or even well meaning warmists. Poor sharpar00, everyone’s piling on.
There just aren’t enough trolls to go around!
DM says:
September 22, 2011 at 6:40 pm
“Climate models have to be based on some physical theories and hypotheses, otherwise you couldn’t put together the code for them to run.”
OK, then, do you have some formulation of the physical theory, the set of hypotheses, that was used to create the code so that the model would run? I have been asking every Warmista for that set of hypotheses. Where is it? Is it in the deep oceans with Trenberth’s missing heat?
Computer code has no cognitive content of its own. All it can do is spew out simulations which are just series of numbers. Someone then has to take those numbers and say “Oh, look, this segment of numbers tracks the observational data for La Nina.” The computer code, having no cognitive content of its own, cannot in any way be about La Nina, represent La Nina, or even sing La Nina.
I began managing models for large corporations about forty years ago. I can assure you that no computer model, however complex, amounts to a theory. After all, there are beautiful computer models of the stock market. And they are incredibly useful analytic tools. But surely you don’t think for one minute that someone actually has a theory of the stock market that consists of reasonably well-confirmed hypotheses that enable accurate prediction of stock market behavior.
What the Warmista are doing with their models is essentially the same thing that someone does when he takes a set of existing graphs for past years of something like Arctic sea ice extent and from them projects the extent for this coming Halloween. They gussy it up with all kinds of interesting differential equations and similar items and convince themselves that it amounts to a theory. It doesn’t. If you will read what I have written above, in earlier posts, you will see that a model is a pale shadow of a theory.
I am Dr. Wormarilalturooni – your deductions are quite veracious if I do say so myself. I’ve personally reviewed 42 million TPS reports on the matter and both my distinguished colleagues and I agree this Global Warming thing is an imposture to the nth degree. Cloudular stratagem to control our cerebral cortex, to us intelligentsia. I can NOT stress the fact that we might all die of chart-ular hemorrhaging long before the angry Gods of Nebulosity get to us. I’ve been working long hours in the lab designing a weather changing machine. It is almost complete.
Barry Elledge says:
September 22, 2011 at 6:18 pm
“This relative superiority of 2 or 3 models does not and in principle cannot prove the “truth” of these models, but does demonstrate that they are superior to the 10 or so less competent models which are currently embedded in IPCC predictions. At the least, one would want to know what future climate trends are predicted by these better models, and whether and how they differ from current IPCC predictions.”
Nope, sorry, but all the models show major disagreement with some set of observational data or another. That means that all of them are toast. The fact that some of them agree well with ENSO is of interest to modelers as they revise their models. But it should be of no interest to scientists. Unlike sets of physical hypotheses, aka physical theories, the individual components of a computer model have no cognitive content of their own. You cannot isolate the computer code that generates the segment of numbers that are treated as tracking ENSO and do further investigation on that code alone. Nor can you port that code to another model and treat it as the same code. Each model has to stand or fall as a whole because it is only as a whole that a model generates a simulation and has empirical import.
Please do not use the scientific word ‘predicted’ in reference to models. Instead, say that the model user “extrapolated” a result from his simulation, as one extrapolates about future sea ice from old graphs of sea ice. To do prediction, as physical hypotheses do, your tool must specify all the observable facts. Physical theory does this by implying the facts. Computer code has no way of doing this. The numbers that make up a simulation can be treated as tracking observable facts but they have no logical or methodological connection to the observable facts. The connection is totally in the mind of the model user, unencumbered by scientific method.
davidmhoffer says:
September 22, 2011 at 7:41 pm
“I’ve noticed a sparsity of trolls or even well meaning warmists. Poor sharpar00, everyone’s piling on. There just aren’t enough trolls to go around!”
I hear that Trenberth has bought a yellow submarine and that all his friends live there too. I guess they don’t have internet yet. /sarc