CMIP5 Model Temperature Results in Excel

Guest Post by Willis Eschenbach

I’ve been looking at the surface temperature results from the 42 CMIP5 models used in the IPCC reports. It’s a bit of a game to download them from the outstanding KNMI site. To get around that, I’ve collated them into an Excel workbook so that everyone can investigate them. Here’s the kind of thing that you can do with them …

42 CMIP5 climate models and HadCRUT4

You can see why folks are saying that the models have been going off the rails …

So for your greater scientific pleasure, the model results are in an Excel workbook called “Willis’s Collation CMIP5 Models” (5.8 Mb file) The results are from models running the RCP45 scenario. There are five sheets in the workbook, all of which show the surface air temperature. They are Global, Northern Hemisphere, Southern Hemisphere, Land, and Ocean temperatures. They cover the period from 1861 to 2100, showing monthly results. Enjoy.

Best to all,

w.

[UPDATE] The data in the spreadsheets is 108 individual runs from 42 models. Some models have only one run, while others are the average of two or more runs. I just downloaded the 42 individual runs data. The one-run-per-model data is here in a 1.2 Mb file called “CMIP5 Models Air Temp One Member.xlsx”. -w.

[UPDATE 2] I realized I hadn’t put up the absolute values of the HadCRUT4 data. It’s here, also as an Excel spreadsheet, for the globe, and the northern and southern hemispheres as well.

[UPDATE 3]

For your further amusement, I’ve put the RCP 4.5 forcing results into an Excel workbook here. The data is from IIASA, but they only give it for every 5-10 year span, so I’ve splined it to give annual forcing values.

Best wishes,

w.

Get notified when a new post is published.
Subscribe today!
0 0 votes
Article Rating
240 Comments
Inline Feedbacks
View all comments
December 22, 2014 9:22 pm

Willis, you’re a champion 🙂

David Norman
Reply to  The Pompous Git
December 23, 2014 3:39 am

I agree… !”it’s a Festivus miracle”!… and now for the “airing of grievances”

ferdberple
Reply to  The Pompous Git
December 23, 2014 5:19 am

They cover the period from 1861 to 2010, showing monthly results. Enjoy.
===========
typo:
They cover the period from 1861 to 2100, showing monthly results

December 22, 2014 9:38 pm

Great work Willis.
Now find the best model. enjoy

Frank
Reply to  Steven Mosher
December 22, 2014 10:30 pm

The red line?

richardscourtney
Reply to  Steven Mosher
December 22, 2014 10:34 pm

Steven Mosher
You instruct

Now find the best model.

but you fail to say what you mean by “best”.
Please say what you want: is it the model with the most misleading output?
Richard

Chris Schoneveld
Reply to  richardscourtney
December 22, 2014 11:52 pm

I suppose he means the model that resembles observations closest, by coincidence or otherwise. After all there are a number of models (with the lowest sensitivity) that almost track actual observations.

Reply to  richardscourtney
December 23, 2014 12:23 am

get off your lazy butt and decide. Then do the work.

richardscourtney
Reply to  richardscourtney
December 23, 2014 4:24 am

Steven Mosher
I asked what you meant by “best” when you wrote

Now find the best model.

and you have answered by saying in full

get off your lazy butt and decide. Then do the work.

OK. I have got off my “lazy butt” and I have decided what you meant on the basis of your past comments.
Obviously, you meant that you don’t have a clue what you are talking about and, therefore, according to you the “best” model is whatever anybody wants it to be.
The “work” was easy: answering your nonsense usually is.
Richard

Chris Schoneveld
Reply to  richardscourtney
December 23, 2014 4:29 am

Steven, due to the limiting nesting I am not sure who you are talking to, `richard or me. You are the one who raised the issue why don’t you give us the answer? I have no Excel on my iPad.

Stephen Richards
Reply to  richardscourtney
December 23, 2014 7:12 am

Mosher
You truly are an inimitable clown. Even if it mattered which model was best, and it doesn’t, all you were asked for was your definition of best. You see that as an occasion to pass intellectual insults from an English teacher to an engineer. When will you grow up?

Reply to  richardscourtney
December 23, 2014 8:02 am

Mosher, why don’t you tell us some more of your “words of wisdom” on modeling, like how you tried to bullshit that the FA-18 fuel tanks were only tested in a computer model,
http://wattsupwiththat.com/2014/12/08/climate-alarmism-secures-a-set-of-warning-signals/#comment-1809809
What a clown. This is why English majors should not comment on STEM subjects.

Reply to  richardscourtney
December 23, 2014 11:26 am

OMG, it’s always priceless when the warmistas tip their hat and show their true colors. This quote from Steven Mosher perfectly encapsulates the climastrologists. He states “get off your lazy butt and decide. Then do the work.”
That, my friends, is how they operate. Decide the outcome first, then lie, contort, obfuscate, lie some more, turture the data, adjust and voila, the work is done and ready for press release!
Thank you Stevie for the chuckle.

cd
Reply to  richardscourtney
December 23, 2014 2:32 pm

Poptech
I don’t think somebody’s qualifications, even if in an unrelated field, should have anything to do with the merit of someone’s arguments.
Although I really don’t think Mosher has made any arguments, just spouted some nebulous nonsense straight out of karate kid: “…to be strong you must first be weak…”. .

markx
Reply to  richardscourtney
December 23, 2014 6:31 pm

The best model, by Mosher’s definition, is the one whic says what you want it to say.
He clearly states elsewhere that the primary purpose of GC models is to justify policy.
He seems to think that is somehow legitimate.

Reply to  richardscourtney
December 24, 2014 12:50 pm

cd, qualifications definitely matter when you want to take someone seriously, as those who lack an education in STEM subjects like Mosher have no business making their uneducated comments about computer modeling. Especially when he has been shown to be repeatedly wrong.

Reply to  Steven Mosher
December 22, 2014 11:18 pm

Quite simply, none can be described as “best” … unless you’re asking for the best of a bad bunch, which still implies that they’re all bad.

Baa Humbug
Reply to  Steven Mosher
December 22, 2014 11:24 pm

Now find the best model

errrr Mosher don’t you mean the least worst?

Reply to  Baa Humbug
December 23, 2014 12:25 am

bonus points.
All models are wrong.
some are useful.
your FIRST task is to define the USE
your second task is to define the allowable error and defend this choice.
your third task is to calculate it.

Reply to  Baa Humbug
December 23, 2014 2:51 am

1 All models are wrong.
2 Some are useful.
Point 1 is true by definition.
Point 2 is not necessarily true at all.

Reply to  Baa Humbug
December 23, 2014 7:57 am

Incorrect Mosher, if the model is wrong it is useless for science, prediction or policy.

catweazle666
Reply to  Baa Humbug
December 23, 2014 8:00 am

Steven Mosher
bonus points.
All models are wrong.
some are useful.

Indeed.
Unfortunately, climate models are not among them.

HAS
Reply to  Baa Humbug
December 24, 2014 11:37 pm

Steven Mosher December 23, 2014 at 12:25 am
“your FIRST task is to define the USE
“your second task is to define the allowable error and defend this choice.”
To help I’ve posed the problem that needs to be addressed at Tamsin Edwards’ blog http://blogs.plos.org/models/love-uncertainty-climate-science (that is sitting in moderation – might take some time given it is Xmas). Did it there (just found the blog from Judith Curry making a reference to it) because well posing the problem from a user’s perspective derives from views of uncertainty, the subject of the current thread over there.
Do take the view that CGCMs used to forecast future weathers aren’t likely to be the best decision support tool.

Reply to  Steven Mosher
December 23, 2014 12:04 am

So BEST is s model!

Reply to  Scott Wilmot Bennett
December 23, 2014 12:13 am

So BEST is a CMIP5 Model!

Reply to  Scott Wilmot Bennett
December 23, 2014 12:32 am

err no. we compare our model of temperature against the model outputs.
there aren’t any observations of temperature, strictly philosophically speaking. real skeptics understand this.

Reply to  Scott Wilmot Bennett
December 23, 2014 7:53 am

Since you are not a real skeptic then you obviously do not understand this at all.

Editor
Reply to  Steven Mosher
December 23, 2014 12:47 am

Steven Mosher : “Great work Willis“. Agreed. “Now find the best model“. Nonsense. None of the models have any understanding of any of the major climate factors over this or any other time-scale – Earth’s orbit, the sun, cloud formation, ocean oscillations, hydrological cycle, etc, etc. None of the models actually models climate at all. You can’t have a “best” of nothing.

Reply to  Mike Jonas
December 23, 2014 1:11 am

In my understanding of Steve McIntyre’s blog article, “Unprecedented” Model Discrepancy, the models have very little in the way of fundamental physics and are largely based upon paramaterization of limited perceived physical interactions.

richard verney
Reply to  Mike Jonas
December 23, 2014 1:55 am

Willis
If the models cannot properly model; the oceans, nor properly model the clouds, it is highly unlikely that they get the hydroligical cycle right.
In fact i would go as far as saying that if a model does not properly model the oceans and/or does not properly model clouds, it cannot possibly get the hydrological cycle right.
One of the major reasons why models do not do regional well is because of the above problems/failings.
It would not surprise me if most of everything in the model is wrong, and they are even tuned to a corrupted temperature data set, causing yet further problems.

cd
Reply to  Mike Jonas
December 23, 2014 4:44 am

It’s worse than that they cannot even hope to model energy transfer through a dynamic atmosphere given the very low resolution they work at:

Even the folks at NASA admit that this is one area where they are stumped due to limitation in computing power.

Editor
Reply to  Mike Jonas
December 23, 2014 12:50 pm

Willis – My words were “None of the models have any understanding of any of the major climate factors“. So yes they have coded stuff for the sun, the hydrological cycle, etc, but because they don’t understand it they get it wrong. For the sun, for example, they include TSI and that’s it. Using their logic over past centuries, the models cannot reproduce anything like past climate. wrt the hydrological cycle, they have remarkably little connection between humidity and precipitation in spite of empirical evidence (Wifffels et al, eg.). Clouds they admit they do not understand at all. ie, they have coding for the sun, clouds, hydrological cycle, etc but no understanding.

davideisenstadt
Reply to  Steven Mosher
December 23, 2014 1:32 am

Steve yours is truly the post of an ass.
why nottry to contribute something… instead of snark.

Alx
Reply to  Steven Mosher
December 23, 2014 4:49 am

Which is the lazy one, the one who makes a meaningless reference to “best” or the one that points out the vacuous comment.
BTW there is no best model, models that are widely inconsistent with each other, and show an obvious bias in one direction in their inconsistency should all be thrown out. You don’t get credit for throwing the dice 100 times and predicting the roll correctly a few times.
There is a thing called a drawing board, climate modelers need to go back to it and forget the super computers for awhile.

DD More
Reply to  Alx
December 23, 2014 12:09 pm

There is a thing called a drawing board, climate modelers need to go back to it and forget the super computers for awhile.
And since they are run, for months on end, on MW powered computers, just think of all the CO2 we could be saving.

davideisenstadt
Reply to  Alx
December 25, 2014 10:28 am

Since you posed the question,I think the lazy one is the person is the one to whom I’m responding just now, thanks.

Reply to  Steven Mosher
December 23, 2014 5:16 am

Meh. All it does is produce results according to the standard CO2 log equation, with arbitrary parameters (mostly aerosols) to somewhat line-up with the real temp data.

rgbatduke
Reply to  Steven Mosher
December 23, 2014 8:17 am

Jeeze guys (addressing the humans replying below, not you, Steve): Give him a break!
He’s not being sarcastic! Can’t we just once not play the “let’s bait Mosher” game and take his words at face value?
I personally plan to do just that. Or well, not exactly just that. Sort-of-that. I plan to play the find the worst models game, the one that the IPCC failed to play in AR5 and steadfastly refuses to even address in the public venue.
The first step is to construct 42 distinct graphs, because sphaghetti graphs are misleading and useless. The second is to use R to assess the models one at a time. That will actually be moderately difficult because one isn’t really comparing distributions (so that the Kolmogorov-Smirnov test e.g. won’t be useful, although a variation of it might work). I may have to crack a stats book to figure out the best way to make a quantitative comparison leading to a useful p-value.
However, certain conclusions can be made instantly, just from looking at the spaghetti graph but then backed by quantitative reasoning. For example, if one computes the cumulants of the data (or the statistical moments, if you prefer) and almost any model in the set, they manifestly are very different. The variance in particular is very different. The autocorrelation appears to be quite different. Most of the models clearly represent incorrect dynamics, as the dynamics is characterized by things like autocorrelation times and variance as much as any “mean” behavior.
One piece of data I’m hoping Willis can provide is: How many model runs go into each curve? Are they Perturbed Parameter Ensemble averages, or are these single tracks from each model? If the latter, how were they selected by the owners of the model for inclusion on the site, since most of those models have been used to generate hundreds of runs? If the former, have they monkeyed at all with the scaling of the variance?
rgb

RACookPE1978
Editor
Reply to  rgbatduke
December 23, 2014 8:36 am

rgbatduke
One piece of data I’m hoping Willis can provide is: How many model runs go into each curve? Are they Perturbed Parameter Ensemble averages, or are these single tracks from each model? If the latter, how were they selected by the owners of the model for inclusion on the site, since most of those models have been used to generate hundreds of runs? If the former, have they monkeyed at all with the scaling of the variance?

Should not Mosher not only be able to answer those questions, but be enthusiastic about answering those questions?

rgbatduke
Reply to  rgbatduke
December 23, 2014 11:22 am

And why do you think he is not? Look, Mosher believes that Carbon Dioxide concentration drives the mean temperature in a monotonic way outside of all other sources of variation. So do I. So does Monckton. So does Anthony, AFAICT since he only rarely personally injects his own perceptions of things into the discussion (which is more a blessing than a curse, given the plethora of sites dominated by the views of the blog owner/manager). So does Nick Stokes. So, do many of the science-educated site participants because there are some really excellent reasons to think that it is so. Reasons that include direct measurements and observations, a fairly straightforward argument (that can and does become a lot more complex as one considers the system as a whole, so it is not certain, merely probable), good agreement with the simplest physically founded computation of its average effect and global observations. It isn’t a matter of “I want to believe” or “I have a dog in the race” it is a comparatively simple matter of physics and observation that makes the “global average temperature all things being equal should be a saturable monotonic function (most likely a natural log) of carbon dioxide concentration in the atmosphere” a probably true statement, better to believe than disbelieve given our sound knowledge of physics and the evidence. This isn’t a religious belief, however often it is argued on WUWT from little more than a religious basis — both ways.
That is a completely distinct issue from whether or not any particular General Circulation Model is an accurate predictor of future temperatures. Here, look, I’ll ask him! Since I personally think his a reasonable human being and not a troll, and since I ask politely, maybe he’ll give me a reasonable answer!
Steve, do you think any or all of the GCMs are accurate predictors of future temperature, beyond all discussion or need to compare to observation?
Who knows, maybe he will surprise you and say that the answer is: of course not! Because that is, in fact, the correct answer for any reasonable scientist or statistician. Models are useful to the precise degree that they a) correspond to past events being modelled and b) predict future events. Steve might disagree with you, or with me, as to whether or not any given model or all of the models collectively have or have not been falsified yet, but until we establish a meaningful statistical basis for a claim for falsification and agree that it is a reasonable if not correct one, we are all just hand-waving. That’s why Willis’ kindness at fighting the KNMI demons for us is so greatly appreciated. Steven McKintyre also has similar directories, but to be honest all of these sites are absolutely miserably designed and make it a royal pain to find and download the data in a usable, well-documented form. I have a directory of my own filled with files named things like “FIO-ESM_rcp45_ave.tab” which turns out to be compressed tabular data that one can, with some effort, unpack and read into R. But a single CSV files is human readable and vastly easier to parse and understand.
There is still the problem of connecting results with time. Global temperature doesn’t vary with time in a greenhouse model, it varies with greenhouse gas concentration (plus some comparatively short relaxation times) and hence one has to have a model for CO_2 as a function of time in order to model future climate or fit past climate. RCP4.5 is already systematically underestimating Mauna Loa (IMO the only reliable data we have on CO_2 concentration) by 2013 (about 1 ppm too low), 395.6 vs 396.5 but that is still within the noise, so the models should have been using CO_2 levels that closely corresponded to measured values across the modern era. I’m about to look at what it claims for 1850 to the present compared to e.g. Siple or Laws Dome data and my own interpolating model. Just eyeballing the data itself it looks to be in pretty good agreement, but then, given a concentration ballpark 285 ppm in 1850 and Mauna Loa starting in 1959 and a requirement of believable smoothness in between, it is difficult to be otherwise.
I should note well that RCP4.5 claims that the total greenhouse CO-equivalent forcing using the standard 5.35*ln(cCO_2/cCO2) formula is already around 403 ppm, and the Kyoto-equivalent forcings neglecting presumed aerosol cancellations is more like 450 ppm. This is very worrisome IMO, because one has to wonder just what is being input into the GCMs — the raw extrapolated concentrations or the now trebly modelled GHG forcing equivalents? Are GCMs the model of a model of a model, or worse? The problem here is straightforward — obviously the authors of RCP4.5 already built a model — really a whole stack of models — that made assumptions about not just CO_2 but about methane, aerosols, nitrous oxide — and extrapolated not just the concentrations indefinitely into the future but the CO_2 concentration equivalent using a presumed model for the CO_2-specific forcing. This is odd beyond compare, presuming knowledge that a) nobody actually has; and b) begs the question, when one uses this presumption in putatively quantitative computations as the basis for future forcing.
And it is all really pretty silly. Anybody can draw a line from 2014 to 2100 and say “behold”, I assume that CO_2 will do this between now and then” and then make a guess as to global average temperature (anomaly) in 2100. There is really only one reliable way to make such an estimate, and it is not building a GCM, or at least not building a GCM until the simple, reliable way to make the estimate fails. That reliable way suggests that the TCS of 3.7 is just about exactly a factor of 2 too high — it does not explain the past data without monkeying far, far too much with other stuff about which we are cosmically ignorant.
But I digress. The point is this. Steven is making a very reasonable statement — Willis has made it very easy to compare CMIP5 models to observed surface temperature. RCP4.5 is adequate as far as CO_2 in particular is concerned between 1850 and the present, hence it should permit models to do a reasonable job of modelling HadCRUT4 if one assumes that HadCRUT4 itself is a reasonable model for the global average surface temperature in between. The game is then fair, with those stipulations. If you want to play, as Willis generally asks — be specific. Which model are you addressing? Why does it succeed or fail? How are you determining success or failure (and what is its quantitative basis and statistical support)?
These are questions I’m all getting ready to ask myself because they are absolutely key to doing science instead of voicing opinions. Sure, we can look at the spaghetti and conclude that something is seriously collectively amiss in that things are not in good agreemen, but is it really amiss or just within the acceptable noise and uncertainty? We cannot answer this “collectively” for exactly the same reason that the MME mean and variance are meaningless. It has to be answered one model at a time, and one has to answer it quantitatively and using an open criterion that is subject to criticism and debate. In the end, all reasonable souls should be able to agree that the proper application of statistics to the models and data either does or does not support the assertion “this model is working to predict the data”. In the end, it all comes down to a p-value of a hypothesis test — what is the probability that model X is correct and and that the real world observation occurred? If p is low, then the null hypothesis “Model X is correct” can correspondingly be rejected with some confidence. If p is anything but low, one cannot assert that Model X is probably incorrect, but (the way hypothesis testing works) neither can one assert this as positive evidence that it is correct, because there can be a near-infinity of models that fit the data over some interval but are utterly false. All we can say is that the data does not falsify it yet, and incrementally increase our degree of belief in it compared to the large number of models that fail the test with low p-values.
So next: How do we compute the p-value of the null hypothesis for just one model curve given the data? I know a fair bit of statistics pretty darn well, but I’m going to have to think about that one. I can think of several ways to do a computation that would lead to a p-value, but they all make certain assumptions, and those assumptions are Bayesian priors for the computation and one has to have some way of defending those assumptions that isn’t just asserting that they “must” be true. Most of them will rely on using the variance of the data itself to determine when it is resolvably separated, but then one has to ask — the variance over what time interval?
This is the really, really difficult problem. There is no good reason to think that the climate is stationary neglecting CO_2. Indeed, we are pretty certain that it isn’t. That implies many time scales and many associated ranges of variation. Again, the usual thing is going to be to assume our ignorance of this dynamic outside of maybe (if we are wise) factoring it into a humble lack of certainty in our final conclusions. This is exactly what the defenders of the GCMs do when they assert that deviation hasn’t lasted long enough to reject the null hypothesis for (fill in the blank — usually for the collective MME mean but also by assumption for each model contributing to the MME mean). There is a clear 67 year harmonic signal with amplitude around 0.1 C, for example, around the general smooth rise in HadCRUT4 — this suggests that we might well be misled about the climate sensitivity if we look at the wrong part of the data and try to fit it. Hence the moving goalposts of 12 years, 15 years, 17 years, whatever, of deviation before we reject the GCMs at least collectively. Sadly, the people that argue in this way fail to recognize that that same clearly observable oscillation caused them to initialize and normalize the models themselves in the worst possible reference period, a stretch in the 1980s where the harmonic contribution produced a transient maximum slope so that now they are strongly deviating now that we are in a transient minimum slope around some sort of mean warming behavior — if one assumes that the observed oscillation is indeed a deviation from a mean warming behavior and not the result of transient phenomena that are depressing the climate from a much warmer trajectory that it “should” be on and will eventually return to!
How can one assign a probability to either one? Clearly there are many reasons to prefer the former, but one can hardly exclude the possibility of the latter. And so they hold on by their fingernails and refuse to let go, hoping that the climate will actually “suddenly” warm up and return to the predicted curve. They know this is increasingly unlikely, but it is — maybe — not yet impossibly unlikely.
Or is it? Again, the only way to tell is one model at a time. But the IPCC seems unwilling to take that step, as it would inevitably lead to throwing out models and further to strongly reducing estimates of climate sensitivity just because, well, the observed temperature is well below the models from 2000 on and is apparently diverging from them. And what about the past? We’ve been told that the models hindcast well. But is this true? Only direct comparison, one model at a time, can tell us.
rgb

maccassar
Reply to  rgbatduke
December 23, 2014 11:51 am

rgb
As is the case every single time, a well reasoned and thoughtful reply. In a lot of cases I am left wondering what the post is really all about, given the absence of some rigorous analysis. An example is the post on acidity. I am not sold on the premise by the author but have no scientific basis to challenge it.
When you weigh in, it all makes some sense.

Reply to  rgbatduke
December 23, 2014 2:19 pm

Thank you rgb.
I am not amused at the vitriol that some folks throw about. I have used some very useful models in engineering and finance that did nothing like “model” or “forecast” what was really going on but simply used empirical, testable, verifiable results to predict an outcome. Fluid dynamics can not model every little vortex, nor every bit of turbulent or laminar flow, but they give adequate information to design a pipeline. Financial models do not have to take every little nuance into account but they do a useful job at predicting income and profit simply using previous years results, backlog, fixed and variable costs.
Mosher, Eisenbach, Tisdale and a host of others provide lots of great input. It is up to the reader to apply the appropriate weighting. Vitriolic comments add little. But sometimes the humour isn’t bad.
Well, just put my skis in the car so have a very MERRY CHRISTMAS everyone and may 2015 be good for everyone – or at least as good as it can be given the climate (pun intended). 😉
Thanks for an entertaining 2014.

Reply to  Steven Mosher
December 23, 2014 3:16 pm

I said this on Finland temperature thread but maybe it belongs here?

Gunga Din
December 23, 2014 at 2:36 pm
I enjoyed the various replies.
But it seems to me that, perhaps, we need to define just what is a “model”.
The first “model” I ever built was a P-61[corrected typo] Black Widow. While it did have twin nacelles, it didn’t require a bra and there was nothing humbug about it’s combat record.
Some engineers build scale models of they are testing, say a building, scaling the strength and stresses the real thing might experience.
Computer programs are used similar design testing. Then a prototype is actually built to see if it performs as expected.
Is putting data points on a graph a globe a model? My understanding is only when the data is extrapolated to predict or project the future.
In the context of climate science, a computer generated climate model is one where something is entered into the extrapolation that will influence the future to the extent the programmer thinks it will.
The programmer may be right or he may be wrong or there may other influences under or over represented or not represented at all.
As I’ve said before, I’m just a layman, one of may that visit this site.
Those of you who aren’t “layman”, am I in the ball park?

Reply to  Gunga Din
December 23, 2014 3:19 pm

I know. Lots of other typos I didn’t correct. Consider it a “model”. 😎

December 22, 2014 10:33 pm

Lots of thanks!

Richard Keen
December 22, 2014 11:04 pm

The best model would be the observations. Nature is the best calculator of physical laws.

Reply to  Richard Keen
December 23, 2014 12:27 am

wrong.
all models are a form of data compression.

Reply to  Steven Mosher
December 23, 2014 11:31 am

Stevie, you are a troll. A trollolololol. There’s got to be a troll song for him somewhere. Maybe his Mom didn’t love him enough when he was a child. So sad.

cd
Reply to  Steven Mosher
December 23, 2014 2:52 pm

Willis hold-on:
he is a sincere, honest scientist
Is he? Even if you use the scientific method and tools of science it doesn’t make you a scientist. If someone picks up a gun does that make them a soldier? Does writing scripts make you a software engineer? Does knowledge of building regulations make you a lawyer?
There is a vast body of knowledge and expertise that is implied when someone presents themselves as a scientist hence the need for professional bodies and accredited qualifications. That doesn’t mean that unqualified persons can’t carryout sound experimental work, but then it doesn’t follow that anyone with a chemistry kit could set themselves up as a pharmacist.

cd
Reply to  Steven Mosher
December 23, 2014 3:35 pm

Willis
If you use a hammer and a saw and the methods and tools of carpentry you build something, yes, it does make you a carpenter.
I’d disagree. If you took on paid work as a carpenter on such a basis, you’d be acting in fraud – why because calling yourself a carpenter imbues a degree of competence and skill; which by any reasonable judgement requires more than being able to use the tools of the trade.
Doesn’t matter if he’s a janitor, a jerk, or a PhD physicist, the only thing that matters is the veracity of his ideas.
I would certainly agree with that. But I’ve still to see or hear anything from, at least on this thread that would suggest veracity.

Robert B
Reply to  Steven Mosher
December 23, 2014 7:07 pm

The reply by Steven Mosher is silly. You would think that there was no such thing as a model before computers. The only thing that Richard said that was wrong was that models are by definition a simpler description than reality. A simple equation (derived from approximations) can be a model.

cd
Reply to  Steven Mosher
December 24, 2014 1:50 am

Willis
At the other end of the scale, I’m a damn good carpenter
Then you’re already – significantly – more useful than about 90% of academics. Yet they seem to think you owe them a living – go figure that.

davideisenstadt
Reply to  Steven Mosher
December 25, 2014 8:16 am

our entire existence is data compression…from sight to hearing to the sensation of touch, smell, taste… so?

richard verney
Reply to  Richard Keen
December 23, 2014 1:57 am

Particularly ones that are unknown and/or not properly and/or fully understood by man.

rgbatduke
Reply to  Richard Keen
December 23, 2014 11:29 am

Hindsight is indeed 20/20, but it is also pretty useless for predicting the future. To predict the future — even so humble a future as “If I jump off of this tall place (an experiment I’ve never performed before) I wonder whether or not I’ll fall to the ground and die?” There the best model is Newton’s Law of Gravitation. Personally, I am a pretty strong believer in its general predictions and have little interesting in climbing to the highest point of my roof over the concrete driveway and testing it.
We just don’t have quite as good a model of the behavior of the future climate as a function of the unknown behavior of the future inputs to the climate and the integral of the future climate over all times between now and then as we do of gravitation. The data we have so far alone cannot tell us what will happen over the next decade or next ten decades without a model to use to extrapolate it. The only real question is whether we have any reliable model with which to perform the extrapolation, or we are back there in time trying to explain the hyperbolic orbit of a comet using Ptolemy’s epicycles or Descartes “vortices” because no Newton has yet had the critical insight required to build a functioning predictive model.
rgb

J.H.
December 22, 2014 11:04 pm

Looking at that, it is quite obvious that the models are more correct than the data.
😉

Reply to  J.H.
December 23, 2014 12:28 am

Logically this is a possibility that can’t be eliminated. every real skeptic understands this

cd
Reply to  Steven Mosher
December 23, 2014 4:47 am

You miss the irony Steven. You cannot disprove observations using models that require validation. Or are you now suggesting a hypothesis (a model) can validate reality.
Furthermore most of these models have been optimised using the same type of data series.

R2Dtoo
Reply to  Steven Mosher
December 23, 2014 10:48 am

Wow- that would be one heck of a model, since we can measure earth’s temperature to 0.01C!

rgbatduke
Reply to  Steven Mosher
December 23, 2014 11:49 am

Also, this is not true. Logically, this is a possibility that can be eliminated. The only question is whether or not it has been eliminated yet. Otherwise, science is a waste of time.
The best way to put it is that if you plotted “probability that TCS to increasing CO_2 is 3.7 C, given the data” as a function of time, there is little doubt that the probability is descending. Because probability over all hypotheses must be conserved (Cox and Jaynes, consistency) as this probability descends and the probability of still higher TCS descends more rapidly still, the probability of lower TCS has to increase.
This reasoning applies to each model, one at a time. If we compute a probability of getting the observational data given a perfect model as being, say, 0.01 for some model, say model X, in CMIP5, every real statistician or scientist recognizes that while we haven’t proven that the model is more correct than the data, we have direct evidence that if the model is correct that the data are remarkably unlikely, that instead of the world following the most probable (bundle of) trajectories, it is out there in a limiting fluctuation in phase space that is very unlikely. We would have to have an enormously good reason (Bayesian prior) to think that the model is a good model in order to continue taking it seriously, as a posterior computation would rather be inclined to decrease the prior probabilities on which the conclusion is founded rather than stubbornly cling to them in the teeth of contrary evidence.
Given an “ensemble” of models to mess with, things are actually rather worse — for the models. Now one cannot rely only a straight-up p-value per model as the basis for rejection of the null-hypothesis, as there is data dredging to consider. One has to reject much more aggressively according to Bonferroni and the number of models. Given models that aren’t independent, one has to be still more aggressive, because the existence of multiple de facto copies of a single approach replicates the error if that approach is, in fact, erroneous and hence leads one to false conclusions regarding variance and reliability. The same thing happens when one model contributes a curve that is averaged over only 3 runs from closely spaced initial conditions (PPE runs), but another model contributes a curve that is averaged over 100 PPE runs. Or if either curve is selected by anything other than random means out of a stack of PPE runs to display or consider.
That’s why I asked Willis about this — all of this stuff is explicitly ignored in AR5 (read chapter 9 of AR5) but it matters. If the curves above are all averages over 10, or 100, PPE runs, then one cannot properly consider whether or not the model contains the correct dynamics because the variance and autocorrelation of the averaged data is completely misleading compared to the variance and autocorrelation of the actual model computation, per run.
The interesting question is then: What p value would you, personally require to reject any particular model in CMIP5 as being sufficiently improbably correct as to be ignorable, at least until such a time as Nature relents and returns to a behavior that doesn’t lead to an appalling low p? The usual 0.05? 0.01? 0.001? Surely you wouldn’t continue to seriously assert that model X could be correct if the probability of observing the data given the model was 0.000001 — a one in a million shot. Yet we both would be disinclined to completely reject a model at p = 0.1, although at least I personally wouldn’t consider this to be strong evidence for the model either.
rgb

Baa Humbug
December 22, 2014 11:20 pm

Good work yet again W
Question: We know the UN IPCC uses all those models purely because of politics, but why do sceptics use all of them (and the silly black average line)?

Baa Humbug
Reply to  Willis Eschenbach
December 23, 2014 1:52 am

Thanks, Baa, but I’m not sure what your objection is.

I’ll try to explain my comment this way….In the real world – say the private sector – when a bunch of modellers present their model findings, the ones that are way off the mark would be discarded. If any are ‘kept’, they would be the ones closest to replicating reality.
The IPCC – being a UN construct – MUST keep all the models purely because of politics, and they do.
My query was why do sceptics keep all the models, why not determine which one(s) replicate reality as near as possible and use those? Averaging makes it seem like the models are closer to reality than they really are.
When I look at that chart, the ‘silly’ black line is barely 0.1 Deg off of reality as of the end of 2014. I’d doubt too many of the models come that close.
Am I being pedantic?

richard verney
Reply to  Willis Eschenbach
December 23, 2014 2:01 am

Willis
I thought that Dr Brown had completely debunked the concept of averaging the models/model runs.
The average is conceptual nonsense.

David A
Reply to  Willis Eschenbach
December 23, 2014 3:41 am

Yes, using the “modeled mean” of a group of models that consistently run wrong in the SAME direction, too warm, is of course scientific nonsense. However it is politically useful.
What happens is “scientists” who know little about the causes of AGW, can now get grant money for predicting future disaster scenarios (increased droughts tornados SL rise , hurricanes, etc) based on a T rise of the “modeled mean”.

Chris Schoneveld
Reply to  Willis Eschenbach
December 23, 2014 4:47 am

Even though models are hopelessly inadequate, it would be interesting to compare the input parameters of the one with the lowest trend (which is almost identical to the actual observations) with the highest trend. Then one can also see in what sense the lowest trend is (or better: appears) right for the wrong reasons.

richardscourtney
Reply to  Willis Eschenbach
December 23, 2014 5:02 am

Baa Humbug
You ask

My query was why do sceptics keep all the models, why not determine which one(s) replicate reality as near as possible and use those? Averaging makes it seem like the models are closer to reality than they really are.

I answer as follows.
Any model of anything is a tool.
The climate models are constructed from existing understandings of climate behaviours and climate mechanisms. Any difference between the behaviour of the climate system and a model’s emulation of climate behaviours demonstrates imperfection(s) in the understandings of climate behaviours and climate mechanisms.
Determination of an imperfection would improve understandings of climate behaviours and climate mechanisms. And any climate model is a potentially useful tool for indication of such imperfection(s). Importantly, there is no reason to suppose that the models which most nearly emulate past climate behaviour(s) are most likely to indicate the imperfect understandings.
As a sceptic I want each and every model to be assessed for the information its behaviour can provide concerning the imperfection(s) in understandings of climate behaviours and climate mechanisms.
This goes to the crux of the stupid demand from Steven Mosher at December 22, 2014 at 9:38 pm. It is imperative to define the intention of a set of models if one is to decide which is the “best” model. Is the “best” climate model that which most closely emulates past climate behaviour, or that which indicates faults in our understandings of a climate behaviour, or that which… etc.?
What can be said is that there is not – and there cannot be – any statistical validity to averaging the outputs of the climate models.
I hope this answer is adequate.
Richard

Alx
Reply to  Willis Eschenbach
December 23, 2014 5:29 am

when a bunch of modelers present their model findings, the ones that are way off the mark would be discarded. If any are ‘kept’, they would be the ones closest to replicating reality.

The issue is these models are used to forecast, you can’t pick the model that happens to work after the fact and then say the the models are good at forecasting. Private business is not inclined to bank their future on widely varying outcomes. They would either throw out the models or wait for proven model results before banking anything or use the average mean of the models.
The IPCC and alarmists have used worst case models and the average mean which is biased warm. I may have missed the IPCC using only the model showing the least warming or closest to observations in their conclusions and recommendations, if so please let me know.
In any case the average mean is way off so any company using the average mean would have had some bad years if not been out of business. I am not sure if the company basing using a model with a decent track record would have done better than the company who threw out the models and adjusted their 5 and 10 year plans annually based on observation and business sense.

Reply to  Willis Eschenbach
December 23, 2014 2:41 pm

Good grief. When I was in business we ran lots of “what if” models for different groups and then aggregated them and analyzed them and made some pretty important decisions based on those models as data came in telling us which track we are on. There are a pile of business models out there with answers to “what ifs”. They are pretty important in business. Climate is much more complex, but like Edison, if you keep trying you might find something useful.
And back to Mosh’s comment: he is absolutely correct that there might be a model out there that is actually better than the “observed” information given the machinations that the raw data has been put through to produce the “observations”.
I actually really liked that comment given I have seen the same thing in business. One of my “business” models was called a “lie detector” by the project managers in my company.

markx
Reply to  Willis Eschenbach
December 23, 2014 6:56 pm

Baa says: “Am I being pedantic?”
You are, Baa.
Willis has kindly presented all that data in a user friendly format so some smart and enthusiastic souls can now embark upon the very analysis you suggest should be done: Detailing which models may be useful, and which ones probably should be thrown out.
He has presented it as it now stands. It will be up to our intrepid analysts as to how the present their results.

richardscourtney
Reply to  Willis Eschenbach
December 26, 2014 10:44 pm

Wayne Delbeke
You say

And back to Mosh’s comment: he is absolutely correct that there might be a model out there that is actually better than the “observed” information given the machinations that the raw data has been put through to produce the “observations”.

You might be correct if Steven Mosher had said “there might be a model out there that is actually better than the “observed” information”, but HE DID NOT.
He said in total

Great work Willis.
Now find the best model. enjoy

And when asked what he meant by “best” he could not.
At no time did he mention
“observations”
or
“machinations that the raw data has been put through”
or
“a model out there that is actually better than the “observed” information” “.
He said Now find the best model.
Your imagination is being used in an attempt to defend Mosher’s meaningless comment.
Richard

Santa Baby
December 22, 2014 11:24 pm

42 chimps?

LeeHarvey
Reply to  Santa Baby
December 23, 2014 5:16 am

It was the best of times. It was the blurst of times.

HAS
December 22, 2014 11:46 pm

To help answer Mosher’s question, do they do absolute temps rather than anomalies?

Reply to  HAS
December 23, 2014 12:29 am

good question. as willis points out absolute. and they suck at it.

HAS
Reply to  Steven Mosher
December 23, 2014 11:05 am

Over the 1994-9 base period the min monthly range of model air temps is 2.6K and the max 3.7K. Must be a hard job being a gas deciding when to condense on that range of planets. I guess the physics must be different.

jolly farmer
Reply to  Steven Mosher
December 24, 2014 8:34 pm

“Sounds very sophisticated, Mr Mosher!”
“Pass me the bucket!”
Have you told the politicians that “they suck at it”?
Thought not.

thingadonta
December 23, 2014 12:10 am

According to the UNSW, if you take account of the things the models got wrong, they got it right. It’s called Model Infallibility.

jimmi_the_dalek
December 23, 2014 12:17 am

Two questions:
1) Why start in 1993?
2) Is there a estimate of error bars on the observed temperature?

rgbatduke
Reply to  Willis Eschenbach
December 23, 2014 12:23 pm

I’m not going to replot just HadCRUT4 plus error, but I have a figure that contains it here:
http://www.phy.duke.edu/~rgb/Toft-CO2-PDO.jpg
As Willis says, understated and in the case of the 19th century data, almost certainly absurd. To put it in simple terms, the error bars in the 1800s are only about twice as large as the error bars in the 2000’s. If we assume anything like central limit theorem normality, that means that there is, on average, only 4 times as much “independent” data contributing to modern error bars as there is contributing to measurements made in 1850.
In 1850 Stanley had not yet met Livingstone in the heart of Africa, the Brazilian rainforest was terra incognita, the bulk of the world’s oceans were sailed outside of well-defined sea lanes only by whalers unarmed with thermometers, Antarctica was a big, dangerous whole on the map, Tibet was unexplored, China was mostly closed to westerners, Siberia was a wilderness and the North American continent was populated only around the periphery in the East and West with huge empty lands (and few thermometers) in the middle. Now those areas are positively saturated with official and unofficial weather stations (which are still far too sparse to be actually useful and which still require extensive kriging/interpolation/infilling) but the error is only twice as small? I don’t think so.
It’s actually a shame that W4T doesn’t include error, that Wikipedia replots rarely include error, etc. And it would also be lovely if the “error” that isn’t included were somehow defined in a collectively useful way, since HadCRUT4 (for example) completely neglects the UHI effect, NASA GISS supposedly includes it (but manages to squeeze still more warming out of it), and I’m not sure what BEST does about it. That is, there is statistical/model error and systematic or neglected bias, and the former says nothing at all about the latter.
By the way, this figure is the model to beat. It is a two parameter model, one of which is common and is needed to fit and compare to the CMIP5 absolute temperature models because we Do Not Know the absolute temperature of the Earth within a degree, so HadCRUT4, my CO_2-only plus linear feedbacks and ignore everything else model, and CMIP5 all three have to agree on the zero of the vertical scale. By itself it has a residual standard error of 0.1 on 163 degrees of freedom, which means basically that there is nothing left to explain. So what you want to do (and what I plan to do) is plot this against each model, one at a time, over exactly this interval, no cherrypicking of ANY endpoints in HadCRUT4.
Note well that the blue curve is very close indeed to the CO_2-only curve of rcp8.5, and the purple curve is a bit more aggressive than rcp6.5 (and is the smoothest extrapolation of Mauna Loa that I could build within a particular form, nothing special about it). The rest of the “rcp” assumptions are a bit silly, since the data strongly suggests that all one needs to predict global temperatures is 2.62*\ln(cCO_2) + T_0 for a suitable reference temperature/concentration, accurate across all of HadCRUT4 to within around 0.1 C. In that case, forget RCP-whatever. Dial in what you expect CO_2 concentration to be in some particular year. Plug it into this formula. Congratulations! You now know the probable temperature that year if your guess as to the CO_2 concentration was correct to within about 0.1 C.
I defy anyone to build a physically defensible model of global average temperature that a) beats this model over the full range of the data; b) has one significant parameter — in my case the “2.62” that replaces the “5.35” of the standard forcing model and actually works to explain the data.
rgb

Reply to  Willis Eschenbach
December 23, 2014 3:31 pm

rgb, “If we assume anything like central limit theorem normality…” CLT normality of measurement error is assumed by virtually everyone in the surface temperature business. That’s why the published error bars are so small. Measurement error is assumed to average away.
But every single test of systematic temperature measurement error shows it to be non-normal and variable. The structure of the error violates the assumptions of the CLT. There’s no reason to think measurement error averages away. The error bars on your plot are probably a factor of 2-4 too small.

davideisenstadt
Reply to  Willis Eschenbach
December 25, 2014 8:26 am

rgbatduke:
i was tersely rebuked by richard betts (on CA) for pointing out that a one dimensional model was more accurate than the GCMs being used. His point was that GCMs model a number of climatic phenomena other than temperature. My thought was that they do a poor job of that as well.

Claude Harvey
December 23, 2014 12:18 am

Man works his tail off to give the kiddies something to play with for Christmas. How will they respond? “I wanted a green one…with a tail..and sparkles…and…!”
Merry Christmas, Uncle Willis!

Lance Wallace
Reply to  Willis Eschenbach
December 23, 2014 6:54 am

OK, at the risk of being lumped with the ungrateful kiddies, can I ask how much extra work would it be to add another of the projections, say the 6.5 or even 8.5, one of which may very well be closer to the true CO2 increase? (Maybe your initial effort can save you lots of time on a future one?)
But anyway, many thanks for a neat Christmas present!

The Ghost Of Big Jim Cooley
December 23, 2014 12:34 am

I am fascinated by human behaviour, and I really want to see just how long it will be before the divergence prompts someone, within the AGW belief community, to say something is wrong. If HadCRUt4 falls in 2015, will it be then? By 2017, the divergence could be enormous.

richard verney
Reply to  The Ghost Of Big Jim Cooley
December 23, 2014 2:06 am

If the pause (hiatus/plateau) continues through 2017, all the models will be outside the 95% confidence band so one would hope that there would be at least one scientist, within the fold, who would at that stage, stand up and be counted and acknowledge that there might be something wrong with the models and/or their projections.

The Ghost Of Big Jim Cooley
Reply to  richard verney
December 23, 2014 4:14 am

That’s my hope, yes. I would have thought that at least one person will put his or her head above the parapet by 2017. It has to be someone that is currently firmly entrenched within the idea of man-made global warming though, otherwise it doesn’t count. Even if cooling started, and went decades, people like Grant Foster and Michael Mann would never consider the idea that they might be wrong. But somewhere, there is a well-known scientist (previously voiced his/her opinions on AGW), that is uncomfortable with the divergence between models and observation. The point is, at what point do they voice it? I have had countless discussions about climate change on net forums – people HATE admitting they’re wrong, as I do. But sometimes you just have to – it’s good for the mind afterward.

David Chappell
Reply to  richard verney
December 23, 2014 4:59 am

the Ghost of BJC asks “The point is, at what point do they voice it?”
In their retirement speech when they are no longer dependent on feeding from the grant trough.

Reply to  richard verney
December 23, 2014 3:43 pm

I’ve been trying to publish a paper since April 2013, showing the results of propagating error through climate model air temperature projections. AGU 2013 meeting poster here (2.9 MB pdf).
The objections of the reviewers would be unbelievable, if I didn’t have them in black-and-white. Climate modeler reviewers have dismissed propagated error bars because they suppose the bars represent oscillations of the model between hot-house and ice-house conditions. They flat do not understand propagated physical error. To me, that lack of understanding explains a lot about the certainty with which modelers hold their results.

December 23, 2014 12:41 am

This is interesting in iteself and shows the divergency between measurements and projections.However plotting the values as anomalies hides a lot of the difference – plotting them as degree celsius relative to zero, rather than anomalies relative to the average for given period, shows even more diffeence.

Matt
December 23, 2014 1:32 am

The genitive of Willis is Willis’ – not Willis’s…
Reply: Wrong. Wrong. Wrong. Willis’s usage is correct. You form the genitive with s’ only with PLURAL nouns. Grrrr…if you’re going to be a grammar Nazi, at least be correct when you do. And I should know, if you remember for what name the “c” is an initial. ~ ctm

rogerknights
Reply to  Matt
December 23, 2014 4:29 am

CTM’s usage is the one favored by the “bible”–the Chicago Manual of Style. Also by Fowler’s Modern English Usage.

Will Nelson
Reply to  Matt
December 23, 2014 4:21 pm

Does adding an “s” onto a surname to include a whole family need to be treated as if, or in fact, plural? Like: “We didn’t invite the Nelsons for Christmas dinner, but they came anyway”. Then do we have: “Having uninvited guests for Christmas is bad enough but the Nelsons’ dog is biting the children”?

Steve Jones
December 23, 2014 1:35 am

Mr Eschenbach,
Thank you for this. There is no substitute for real data; shame the climate science community don’t treat it with the respect it deserves.
Your patience and perseverance will pay off as you are merely telling the truth.
Merry Christmas and please keep up the good work.

tonyM
December 23, 2014 1:38 am

Thank you Willis.
Are they serious in presenting results to four decimal places? Crazy.
Is there any way to find out what was forecast and what was hindcast for each run?
Merry Xmas to all.

Scarface
December 23, 2014 1:47 am

This is definetely proof of AGW: Alarmism Gone Wrong. Q.E.D.
Thanks Willis, and Merry Cristmas!

Bill
December 23, 2014 2:22 am

I wonder why the models predict s leveling off of temperatures (reducing rate of increase)at all. From the rhetoric around them by advocates I would have guessed a more linear increase predicted.

Brandon Gates
Reply to  Bill
December 23, 2014 4:53 am

Bill,
The model output Willis provides is from the RCP 4.5 scenario, which calls for emissions to level off and stabilize starting around 2070:
http://en.wikipedia.org/wiki/Representative_Concentration_Pathways#mediaviewer/File:All_forcing_agents_CO2_equivalent_concentration.png
RCP 8.5 is the “nightmare” scenario calling for a geometric increase in emissions which would produce a more linear increase in surface temperature projections.

December 23, 2014 2:33 am

Oiiii, data without interpretation!
Try these data: sea level rise without impact from man:
http://tidesandcurrents.noaa.gov/sltrends/sltrends_station.shtml?stnid=8518750
Where is the impact of man here? Nowhere!

sleepingbear dunes
Reply to  franktrades
December 23, 2014 3:41 am

I always thought of this graph when I used to hear Bloomberg freakout about runaway sea level rise.

DD More
Reply to  franktrades
December 23, 2014 1:09 pm

Frank, your ‘without impact from man:’ might just miss the reduction in width of the Hudson River. Look up a historical map of Manhattan. Now think if the Hudson River has the same flow rate with a reduced cross sectional area, what do you think the height of the water will be. You see this same effect, especially with regards to flooding at Fargo, ND on the Red River, where they keep upping the height of the levy and getting record heights for floods.

Jeef
December 23, 2014 2:41 am

What am I missing? I thought the inexorable feedbacks would lead to a runaway warming post tipping point.
The projections graph looks like temp rate of increase gets flatter.
I assume it’s some feature of the model that’s beyond my comprehension..,

Brandon Gates
December 23, 2014 3:21 am

Willis, I commend you for producing a useful tool for independent investigation. My only critique is that 1991-1994 is an inappropriate baseline period. The proper reference period for comparing CMIP5 to obs is 1986-2005 because 2006 marks the transition from historical forcings to RCPs in the model runs.

Bill
December 23, 2014 3:32 am

I wondered the same thing above. I’ve heard atmospheric carbon is increasing at least linearly, and I’ve never heard any warmists suggest that the atmospheric responce isn’t proportional.

mwh
Reply to  Bill
December 23, 2014 4:32 am

Precisely Bill, if the rate of increase of ACO2 was proportional to the Mauna Loa graph the ‘line’ would be an upward curving exponential one rather than the current nearly straight one. Surely if we are entirely responsible for the increase and mans contribution was steady then there would be a straight line, if however the rate of increase doubles at it has done several times then the amount of increase in CO2 should have similarly doubled several times producing an exponential curve after all a lot of the models have these curves in them. Something is compensating for the increasing rate, probably increasing natural sinks. I have brought this up on warmist sites to huge derision and yet noone seems to give an answer that explains the discrepancy and usually the comments bypass thinking about what I am asking.
As for models I think they are incredibly useful at representing current data and often produce fascinating insights in to our planets dynamic systems. As a predictive tool for climate they have only proved one thing, that they are useless at prediction. This at the very least should have proved a long time ago that CO2 is not the main forcing element, it never seems to be able to hold true for very long before diverging. My education was loosely science based (agriculture), but even I can recognise that the models are not being inputted with the right data and the evidence against CO2 sensitivity is stacking up very quickly (IMO).

Brandon Gates
Reply to  mwh
December 23, 2014 5:26 am

mwh,

Surely if we are entirely responsible for the increase and mans contribution was steady then there would be a straight line, if however the rate of increase doubles at it has done several times then the amount of increase in CO2 should have similarly doubled several times producing an exponential curve after all a lot of the models have these curves in them.

Recall that predicted radiative forcing due to CO2 doubling is a logarithmic relationship:
ΔF = α * ln(C/C0), α = 5.35
ΔT ≈ 0.8 * ΔF

Even with the log relationship, straight lines aren’t guaranteed since emissions aren’t constrained to a constant geometric increase, so it’s best to do some math on actual figures. Plugging in observed values from 1850-2014 we get:
ΔT = 5.35 * ln(398.43/287.40) * 0.8 = 1.4 K
Observed ΔT since 1850 is 0.9 K according to HADCRUT4. Next time someone writes, “there’s another half a degree of warming in the pipeline” or something similar, it may be based on a similar calculation to what I have just done.

Bill
Reply to  mwh
December 23, 2014 5:31 am

But my question isn’t why doesn’t the observed values don’t match co2… It’s why didn’t the model predictions? I mean the models are showing a “pause”… Just less of one than the observed…

Alx
Reply to  mwh
December 23, 2014 5:47 am

All other things being equal a mathematical relationship between CO2 and influence on temperature can be calculated. Unfortunately all other things are not equal and change in precedence and relationship over time.
This is why I find climate science off the rails, it treats CO2 and warming the same way as the polio vaccine and polio. CO2 is not a direct discreet preventative like the polio vaccine and it is childish to think so.

Brandon Gates
Reply to  mwh
December 23, 2014 6:03 am

mwh,
I should add that CO2 forcing isn’t the only game in town. For instance, solar output has increased 0.12 W/m^2 since 1880. Other well-mixed GHGs have contributed 1.85 W/m^2, ozone 0.22, black carbon soot 0.66, snow albedo reduction 0.22.
OTOH, there are offsets; -2.75 for combined aerosol effects and -0.09 for land use changes. The net from 1880, including 1.35 for CO2, is 1.63 W/m^2 * 0.8 = 1.3 K, against 0.74 K observed over the same time period, an apparent discrepancy of 0.56 K.
From observation, it’s estimated that the current energy imbalance is 0.4 W/m^2 in the down direction, times 0.8 implies 0.32 K of warming “in the pipeline”. So my calcs leave about a quarter degree unexplained, which interestingly is roughly the discrepancy between CMIP5 projections for 2014 and presently observed temps.
I saw your reply come in while I was composing this addendum to my original post. I’ll handle your specific questions separately.

Brandon Gates
Reply to  mwh
December 23, 2014 6:34 am

Bill,

But my question isn’t why doesn’t the observed values don’t match co2… It’s why didn’t the model predictions? I mean the models are showing a “pause”… Just less of one than the observed…

I’ll start by saying that present observations don’t “match” the mathematical prediction based on the IPCC’s simiplified forcing expressions. I can regress acutal temps vs. CO2 and get a tidy fit:
https://drive.google.com/file/d/0B1C2T0pQeiaSTFNEekNLWkxkMFk
Bottom graph says ΔT = 2.75 K / 2xCO2, IPCC says 1.4 to 4.5 K with the most likely value being 3 K. Again, reality lags prediction by 0.25 K. There are lots of reasons why that’s the case. I believe the most likely explanation is the thermal inertia of the oceans causing a lag in response to the external forcings.
I know that doesn’t directly answer your question. I bring it up because I think it’s important to understand that function of the oceans first before trying to understand how GCMs in the CMIP5 ensemble model them. Or don’t model them if you like.
The beginning of the answer is this: up to 2005, the GCMs used observational forcing data as input parameters. After 2005 the forward looking RCP assumptions take over as input parameters. Depending on when one starts the clock for the beginning of The Hiatus, that’s 5-7 years of Le Gran Pause the models know about from observation. After that they’re using scenario parameters for input and/or doing more of their own calculations for atmospheric/ocean coupling.
As such, the modeled trends prior to 2005 line up more with the expected long term trend, which falls between the relatively steep upward slope from 1980-2000 and the flat as a pancake trend since 2000.
I stress that I greatly oversimplify here.

Brandon Gates
Reply to  mwh
December 23, 2014 7:06 am

Bill, errata: … modeled trends prior to 2005 …
s/b subsequent to 2005. erg …

george e. smith
Reply to  mwh
December 25, 2014 11:53 pm

“””””…..
Brandon Gates
December 23, 2014 at 6:03 am
mwh,
I should add that CO2 forcing isn’t the only game in town. For instance, solar output has increased 0.12 W/m^2 since 1880. …..”””””
So just how did they measure the solar TSI to that level of precision in 1880 ??
55 years ago, the accepted best value for the value of TSI was 1353 W/m^2. It is now around 1362, and even that number has dropped from around 1366 since satellite measurements have been taken.
In 1880, there still was no satisfactory theory of Black Body radiation, So It seems quite unreasonable that they could measure TSI with that precision, back in 1880.
And now, I suspect you are going to tell us, that we can deduce what it was then from proxy’s we can evaluate today ?

Brandon Gates
Reply to  mwh
December 26, 2014 7:37 am

george e. smith,
If you know what I’m going to write before I write it, why ask the question?
It’s a rhetorical question of course. No, I wouldn’t trust TSI estimates from the 1880s over more recent estimates, but back then they could, and did, count sunspots (since 1610 according to Wikipedia, so it must be true). C14 from tree rings and Be10 from ice cores are other proxies (according to ClimateAudit, so it really must be true). KNMI has a nice plot of sunspot counts since 1750:
http://climexp.knmi.nl/data/isunspots.png
And a TSI reconstruction from 1610 through 2008 here:
http://climexp.knmi.nl/data/itsi_wls_ann.png
Regressing those two series together we get a slope of 0.0059 sunspots/Wm^-2, R^2 = 0.65. Not a … stellar … correlation but not shabby either. You can read all about what else Wang, Lean and Sheeley (2005) did here: http://sun.stanford.edu/LWS_Dynamo_2009/61797.web.pdf “Modeling the Sun’s Magnetic Field and Irradiance Since 1713”. Oh no, models, that will never do. Well, I tried.
Annnyway, reviewing the above data, the better calculation for me to have done is the linear trend from 1880-2013, which works out to 0.44 Wm^-2/century, implying a positive change of 0.58 Wm^-2 over the interval. (Ouch, five times higher than what I quoted in my previous post.) Multiply by 0.8 K/Wm^-2 and we get an implied ΔT = 0.47 K from change in solar output alone. That should make you happier, yes?

Steve from Rockwood
December 23, 2014 4:09 am

Willis, from your graph it looks the slope changes on the model graphs around 2001 in favor of less warming. Why would climate scientists hind cast their models to closely follow measured temperatures and then lower future warming forecasts (why from a science point of view)?

Manny
December 23, 2014 4:13 am

Great graph, thanks.

Mark from the Midwest
December 23, 2014 4:29 am

But wait, my first read on this data set is that it is totally inconsistent with the opinion of a number of relatives and friends that are trained in sociology, social psychology, journalism, political science, and secondary education. How could all those brilliant minds be wrong? Could my PhD, with a mere 48 semester hours of graduate level course work in statistics, be failing me? At the very least this will make for some lively holiday conversation.

1 2 3 4