New paper makes a hockey sticky wicket of Mann et al 98/99/08

NOTE: This has been running two weeks at the top of WUWT, discussion has slowed, so I’m placing it back in regular que.  – Anthony

UPDATES:

Statistician William Briggs weighs in here

Eduardo Zorita weighs in here

Anonymous blogger “Deep Climate” weighs in with what he/she calls a “deeply flawed study” here

After a week of being “preoccupied” Real Climate finally breaks radio silence here. It appears to be a prelude to a dismissal with a “wave of the hand”

Supplementary Info now available: All data and code used in this paper are available at the Annals of Applied Statistics supplementary materials website:

http://www.imstat.org/aoas/supplements/default.htm

=========================================

Sticky Wicket – phrase, meaning: “A difficult situation”.

Oh, my. There is a new and important study on temperature proxy reconstructions (McShane and Wyner 2010) submitted into the Annals of Applied Statistics and is listed to be published in the next issue. According to Steve McIntyre, this is one of the “top statistical journals”. This paper is a direct and serious rebuttal to the proxy reconstructions of Mann. It seems watertight on the surface, because instead of trying to attack the proxy data quality issues, they assumed the proxy data was accurate for their purpose, then created a bayesian backcast method. Then, using the proxy data, they demonstrate it fails to reproduce the sharp 20th century uptick.

Now, there’s a new look to the familiar “hockey stick”.

Before:

Multiproxy reconstruction of Northern Hemisphere surface temperature variations over the past millennium (blue), along with 50-year average (black), a measure of the statistical uncertainty associated with the reconstruction (gray), and instrumental surface temperature data for the last 150 years (red), based on the work by Mann et al. (1999). This figure has sometimes been referred to as the hockey stick. Source: IPCC (2001).

After:

FIG 16. Backcast from Bayesian Model of Section 5. CRU Northern Hemisphere annual mean land temperature is given by the thin black line and a smoothed version is given by the thick black line. The forecast is given by the thin red line and a smoothed version is given by the thick red line. The model is fit on 1850-1998 AD and backcasts 998-1849 AD. The cyan region indicates uncertainty due to t, the green region indicates uncertainty due to β, and the gray region indicates total uncertainty.

Not only are the results stunning, but the paper is highly readable, written in a sensible style that most laymen can absorb, even if they don’t understand some of the finer points of bayesian and loess filters, or principal components. Not only that, this paper is a confirmation of McIntyre and McKitrick’s work, with a strong nod to Wegman. I highly recommend reading this and distributing this story widely.

Here’s the submitted paper:

A Statistical Analysis of Multiple Temperature Proxies: Are Reconstructions of Surface Temperatures Over the Last 1000 Years Reliable?

(PDF, 2.5 MB. Backup download available here: McShane and Wyner 2010 )

It states in its abstract:

We find that the proxies do not predict temperature significantly better than random series generated independently of temperature. Furthermore, various model specifications that perform similarly at predicting temperature produce extremely different historical backcasts. Finally, the proxies seem unable to forecast the high levels of and sharp run-up in temperature in the 1990s either in-sample or from contiguous holdout blocks, thus casting doubt on their ability to predict such phenomena if in fact they occurred several hundred years ago.

Here are some excerpts from the paper (emphasis in paragraphs mine):

This one shows that M&M hit the mark, because it is independent validation:

In other words, our model performs better when using highly autocorrelated

noise rather than proxies to ”predict” temperature. The real proxies are less predictive than our ”fake” data. While the Lasso generated reconstructions using the proxies are highly statistically significant compared to simple null models, they do not achieve statistical significance against sophisticated null models.

We are not the first to observe this effect. It was shown, in McIntyre

and McKitrick (2005a,c), that random sequences with complex local dependence

structures can predict temperatures. Their approach has been

roundly dismissed in the climate science literature:

To generate ”random” noise series, MM05c apply the full autoregressive structure of the real world proxy series. In this way, they in fact train their stochastic engine with significant (if not dominant) low frequency climate signal rather than purely non-climatic noise and its persistence. [Emphasis in original]

Ammann and Wahl (2007)

On the power of the proxy data to actually detect climate change:

This is disturbing: if a model cannot predict the occurrence of a sharp run-up in an out-of-sample block which is contiguous with the insample training set, then it seems highly unlikely that it has power to detect such levels or run-ups in the more distant past. It is even more discouraging when one recalls Figure 15: the model cannot capture the sharp run-up even in-sample. In sum, these results suggest that the ninety-three sequences that comprise the 1,000 year old proxy record simply lack power to detect a sharp increase in temperature. See Footnote 12

Footnote 12:

On the other hand, perhaps our model is unable to detect the high level of and sharp run-up in recent temperatures because anthropogenic factors have, for example, caused a regime change in the relation between temperatures and proxies. While this is certainly a consistent line of reasoning, it is also fraught with peril for, once one admits the possibility of regime changes in the instrumental period, it raises the question of whether such changes exist elsewhere over the past 1,000 years. Furthermore, it implies that up to half of the already short instrumental record is corrupted by anthropogenic factors, thus undermining paleoclimatology as a statistical enterprise.

FIG 15. In-sample Backcast from Bayesian Model of Section 5. CRU Northern Hemisphere annual mean land temperature is given by the thin black line and a smoothed version is given by the thick black line. The forecast is given by the thin red line and a smoothed version is given by the thick red line. The model is fit on 1850-1998 AD.

We plot the in-sample portion of this backcast (1850-1998 AD) in Figure 15. Not surprisingly, the model tracks CRU reasonably well because it is in-sample. However, despite the fact that the backcast is both in-sample and initialized with the high true temperatures from 1999 AD and 2000 AD, it still cannot capture either the high level of or the sharp run-up in temperatures of the 1990s. It is substantially biased low. That the model cannot capture run-up even in-sample does not portend well for its ability

to capture similar levels and run-ups if they exist out-of-sample.

Conclusion.

Research on multi-proxy temperature reconstructions of the earth’s temperature is now entering its second decade. While the literature is large, there has been very little collaboration with universitylevel, professional statisticians (Wegman et al., 2006; Wegman, 2006). Our paper is an effort to apply some modern statistical methods to these problems. While our results agree with the climate scientists findings in some

respects, our methods of estimating model uncertainty and accuracy are in sharp disagreement.

On the one hand, we conclude unequivocally that the evidence for a ”long-handled” hockey stick (where the shaft of the hockey stick extends to the year 1000 AD) is lacking in the data. The fundamental problem is that there is a limited amount of proxy data which dates back to 1000 AD; what is available is weakly predictive of global annual temperature. Our backcasting methods, which track quite closely the methods applied most recently in Mann (2008) to the same data, are unable to catch the sharp run up in temperatures recorded in the 1990s, even in-sample.

As can be seen in Figure 15, our estimate of the run up in temperature in the 1990s has

a much smaller slope than the actual temperature series. Furthermore, the lower frame of Figure 18 clearly reveals that the proxy model is not at all able to track the high gradient segment. Consequently, the long flat handle of the hockey stick is best understood to be a feature of regression and less a reflection of our knowledge of the truth. Nevertheless, the temperatures of the last few decades have been relatively warm compared to many of the thousand year temperature curves sampled from the posterior distribution of our model.

Our main contribution is our efforts to seriously grapple with the uncertainty involved in paleoclimatological reconstructions. Regression of high dimensional time series is always a complex problem with many traps. In our case, the particular challenges include (i) a short sequence of training data, (ii) more predictors than observations, (iii) a very weak signal, and (iv) response and predictor variables which are both strongly autocorrelated.

The final point is particularly troublesome: since the data is not easily modeled by a simple autoregressive process it follows that the number of truly independent observations (i.e., the effective sample size) may be just too small for accurate reconstruction.

Climate scientists have greatly underestimated the uncertainty of proxy based reconstructions and hence have been overconfident in their models. We have shown that time dependence in the temperature series is sufficiently strong to permit complex sequences of random numbers to forecast out-of-sample reasonably well fairly frequently (see, for example, Figure 9). Furthermore, even proxy based models with approximately the same amount of reconstructive skill (Figures 11,12, and 13), produce strikingly dissimilar historical backcasts: some of these look like hockey sticks but most do not (Figure 14).

Natural climate variability is not well understood and is probably quite large. It is not clear that the proxies currently used to predict temperature are even predictive of it at the scale of several decades let alone over many centuries. Nonetheless, paleoclimatoligical reconstructions constitute only one source of evidence in the AGW debate. Our work stands entirely on the shoulders of those environmental scientists who labored untold years to assemble the vast network of natural proxies. Although we assume the reliability of their data for our purposes here, there still remains a considerable number of outstanding questions that can only be answered with a free and open inquiry and a great deal of replication.

===============================================================

Commenters on WUWT report that Tamino and Romm are deleting comments even mentioning this paper on their blog comment forum. Their refusal to even acknowledge it tells you it has squarely hit the target, and the fat lady has sung – loudly.

(h/t to WUWT reader “thechuckr”)

Share

0 0 votes
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

1.2K Comments
Inline Feedbacks
View all comments
August 14, 2010 7:40 pm

You know, this is really no surprise. Despite the fact that I felt McIntyre & McKittrick had already de-bunked the hockey stick to my satisfaction, it is still good to see this in print.
After reading the paper, key points that stick in my mind:
• Climatology is inherently a statistical endeavour
• Although climatologist may understand atmospheric physics, they don’t necessarily understand statistics & have grossly under-collaborated with statisticians in their work, which is a fundamental flaw
• Although not directly stated, it is implied that the models that climatologists hang their AGW models on are inherently flawed because they lack the proper statistical framework. It certainly explains the continued divergence between “the models” & reality.
• “response and predictor variables which are both strongly autocorrelated.” For those not versed in signal analysis, the stronger the autocorrelation function (peak at t=0), the more random the signal is. The predictor variables are the proxies – this is saying that the proxies are not much different than random noise. The response is the time signal – this is saying that temperature is close to a random response relative to the proxies. … which of course is entirely consistent with McIntyre & McKittrick , where they used a random number generator to replicate the Mann curve.
• “Commenters on WUWT report that Tamino and Romm are deleting comments even mentioning this paper on their blog comment forum. ” Tamino & Romm, as they say, “Sucks to to be you”
• Unfortunately, as damning as this is, just as
http://wattsupwiththat.com/2010/08/13/is-jim-hansens-global-temperature-skillful/#more-23402
was damning, this is a clearly a matter of faith for the believers. Don’t expect AGW to go quietly into the night. All that can be done is continue to circulate studies like McShane and Wyner 2010 & try to educate as many people as possible to buy time. With time (I am guessing by 2020, given a cold PDO & AMO going into it’s cool phase by then) it will be come clear to all that catastrophic AGW was a bogus theory & it will die. …. of course, I am sure the leftist scaremongers will have a new boogieman by then to try to scare the populus into submitting to the government.

Pofarmer
August 14, 2010 7:43 pm

“The authors of the 20- odd studies that confirmed Mann’s data are not really interested in what professional statisticians and mathematicians are saying about it.”
Then they are pretty much doomed to continue repeating the same mistakes.

August 14, 2010 7:43 pm

M. Roddy, I’m afraid you brought a knife to a Howitzer duel.

August 14, 2010 7:44 pm

Sigh, only half-way through it. Admittedly, some of the stat techniques are a bit tricky for me, (I’ll work through them.) but the paper in itself is very clear.
This paper doesn’t simply break a hockey stick, it breaks an entire sub-specialty of climatology, specifically paleoclimatology. They will either have to reprint all text books or throw the psuedo-science out the window to the trash heap to lay alongside phrenology, numerology, and astrology. Oh, the humanity!!!!

Ed Caryl
August 14, 2010 7:47 pm

Mike, explain the Antarctic.

August 14, 2010 7:47 pm

WOW! Silver bullet to the beast!

duckster
August 14, 2010 7:48 pm

Looking at the paper above…
No medieval warming period, I see. And no temperature decline post-1998?? I thought you were arguing that the world was getting cooler, and arctic ice was recovering? [Cough, cough].
I guess we can put those ones to rest then, can’t we? After the way you’ve embraced this paper!
The way it looks from here is that you guys will pretty much accept ANYTHING that throws doubt on CAGW, without worrying whether it is logically consistent with all the other things you have accepted/argued before. This does not translate into a coherent science-based system of knowledge building.
You need a theory to explain what is happening now. It needs to be falsifiable. And you have to either accept that new scientific papers fit your theory, or explain why they don’t. You would also need to follow up on Mann et al.’s commentary on this paper. Otherwise it’s just another fishing expedition.

Henry chance
August 14, 2010 7:55 pm

Looks like the tree ring circus clowns have had their final act.

Andrew30
August 14, 2010 8:00 pm

John Blake says: August 14, 2010 at 7:13 pm
GMTA = .0059 x (Year – 1880) – .52 + 2pi x Cos((Year – 1880)/60)
John, you missed the climatology bit.
valadj=[0.,0.,0.,0.,0.,-0.1,-0.25,-0.3,0.,-0.1,0.3,0.8,1.2,1.7,2.5,2.6,2.6,2.6,2.6,2.6]*0.75 ; fudge factor
yearlyadj=interpol(valadj, yrloc, x)
GMTA = GMTA +yearlyadj
Now plot GMTA, it has become a hockey stick through the science of climatology.
See also: FOI2009/FOIA/documents/harris-tree/briffa_sep98_e.pro

Andrew30
August 14, 2010 8:01 pm

Mike Roddy says:
August 14, 2010 at 7:13 pm
Here’s the definitive article on questions about the Mann Hockey Stick:
http://www.thespoof.com/news/spoof.cfm?headline=s5i64103

Mike
August 14, 2010 8:04 pm

“…our model offers support
to the conclusion that the 1990s were the warmest decade of the last millennium,…”
It is an interesting paper and may influence how proxy reconstructions are done in the future. Mann’s papers already had large error bars – maybe the should be larger; it will be interesting to see how he responds. It does not change the fact that CO2 warms the earth and we need to be thinking about what to do about our CO2 emissions.

Jean Parisot
August 14, 2010 8:06 pm

The authors of the 20- odd studies that confirmed Mann’s data are not really interested in what professional statisticians and mathematicians are saying about it. Not exactly how to win friends and influence people – what I think your going to find that politicians to whom AGW must be sold are absolutely feral statisticians regardless of their professional or academic backgrounds. If the confidence in the science cannot match the pain of solution, then it falls off of the public agenda.

RockyRoad
August 14, 2010 8:08 pm

duckster says:
August 14, 2010 at 7:48 pm
(…)
The way it looks from here is that you guys will pretty much accept ANYTHING that throws doubt on CAGW, without worrying whether it is logically consistent with all the other things you have accepted/argued before. This does not translate into a coherent science-based system of knowledge building.
———Reply:
I clicked on a link provided in one of the comments of a recent story here on WUWT and it directed me to an interview w/ Phil Jones, who acknowledged that in that past 150 or so years there had been 4 warming periods, all of about equal magnitude. It was only the last one, which we are currently enjoying, that he attributed to increases in CO2. But the big question is: Why not the other three? Did he have irrefutable proof that those were not caused by CO2? Did he have any idea what else might have caused those warming periods without benefit of anthropogenic CO2? Good questions all, but the main point is that Phil Jones has, without any verifiable reason, pinned this current warming trend on CO2.
Now, that isn’t just ANYTHING–it is Phil Jones making a mockery of science; Phil Jones is the incoherent one. HE is the one doing a great job of NOT doubting CAGW against all logic. But applying logic to Phil Jones’ statements leaves me laughing at Phil Jones. I say, man, can Phil Jones be that absolutley daft?

Bernie
August 14, 2010 8:09 pm

We should not get ahead of ourselves. I think this is a very interesting and well constructed paper. The authors certainly believe that current multi-proxy studies are seriously flawed. The assumption about the quality of the proxy data and their inclusion of Tiljander oriented presumably as Mann left it oriented are areas for further exploration. However, I do not understand enough of the statistics to start jumping up and down
That said footnote 12 is a doozie and will take some explaining.
More generally, I will be interested in how Ammann responds since he has co-authored papers with Li (2007, 2010) where more sophisticated approaches than those of Mann were used.
Mike Roddy does not know what he is talking about and I doubt that he has actually comprehended anything more than the abstract and conclusion. I will wait until Ammann, Tamino and statistically knowledgeable folks respond.
Note:
Prof Wyner dropped by CA and said a few nice things.

August 14, 2010 8:10 pm

While breaking from the reading, mainly because Adobe isn’t responding at the moment,
“MBH…a cardinal rule of statistical inference is that the method of
analysis must be decided before looking at the data. The rules and strategy of
analysis cannot be changed in order to obtain the desired result. Such a strategy
carries no statistical integrity and cannot be used as a basis for drawing sound
inferential conclusions.”
—heh, I always suspected as much.
“The degree of controversy associated with this endeavor can perhaps
be better understood by recalling Wegman’s assertion that there are very
few mainstream statisticians working on climate reconstructions (Wegman
et al., 2006). This is particularly surprising not only because the task is
highly statistical but also because it is extremely difficult.”
——Didn’t we hear those thoughts echoed by one of the climate gate white wash committees? How many times do they have to be told this is statistical work!?! Junior, leave it to the professionals!

August 14, 2010 8:12 pm

duckster says:
“Looking at the paper above… No medieval warming period, I see. ”
Duckster, are you friggin’ blind??

ZT
August 14, 2010 8:16 pm

The paper has some witty one liners too, such as:
“We assume that the data selection, collection, and processing performed by climate scientists meets the standards of their discipline.”

RockyRoad
August 14, 2010 8:17 pm

It amazes me that “climate scientists” have invented their own little band/kind/brew of math/statistics to handle their own data. I understand there are several tested and true statistics programs that would save them the pain of having to invent their own, but no, they stick their noses in the air and defy disciplines that are magnitudes older than their brief science. And to what end? To look like fools, apparently. The science isn’t settled, but really, the mathematics and the statistics doesn’t need to be re-invented–it just needs to be applied properly.

duckster
August 14, 2010 8:18 pm

@Smokey
Duckster, are you friggin’ blind??
So where exactly would you place a medieval warming period here? Asking me to accept a medieval warming period (which is what I have been asked to do here) means showing how and where it got warmer, and then how and when it got cooler. A steady downward temperature trend is not a warming period.

Frederick Michael
August 14, 2010 8:18 pm

Mike Roddy,
Have you thought about what a 40% reduction in the ocean fish biomass would mean to, say, Japan? Do you really think this could be happening without it being big news?
Sometimes a little checking is worthwhile.

August 14, 2010 8:24 pm

duckster says:
August 14, 2010 at 7:48 pm
“Looking at the paper above…
No medieval warming period, I see. And no temperature decline post-1998?? I thought you were arguing that the world was getting cooler, and arctic ice was recovering? [Cough, cough].
I guess we can put those ones to rest then, can’t we? After the way you’ve embraced this paper!”
Sis, have you actually read the paper? Are you not understanding what they are asserting? Read the little notation underneath figure 16. I’d cut and paste from the paper, but my Adobe is going belly up for some reason. Still, the paper clearly states that they are not addressing the validity of the data. (That’s probably in the next paper if it is necessary.)
What they are stating is, even if the data are correct, Mann et al. did it wrong(along with a long list of other statistician wannbees), and further, proxies have no predictive properties. Now, work backwards from that. If you require further explanations, just ask, I’d be happy to provide them to you.

Stephen Pruett
August 14, 2010 8:28 pm

This is quite important, because it comes not from “deniers”, but from objective scientists with no particular axe to grind, but who also have apparently not been influenced by climate science groupthink. It is interesting that even in the various recent “exoneration” reports and older reports as well, a recurring criticism of climate science has been the minimal collaboration with statisticians and resulting less than ideal statistical analyses. If this paper is correct, “less than ideal” may actually be “essentially useless”.

Robinson
August 14, 2010 8:29 pm

Well, all I can say is it’s about time. I mean SM and others have been poking and prodding around the statistics for many years now. It’s shocking in a way that a paper like this has been published in a statistics journal after so much time has passed and so much water has flowed under the bridge. I would have thought attempted replication would have been performed sooner.
Still, better late than never. I don’t expect this will get much traction in the mainstream, but with blogs like this, who cares?

Aaron Wells
August 14, 2010 8:31 pm

Duckster: “No medieval warming period, I see.”
Are you looking at a different graph than the one above? Looks pretty warm at the beginning of the graph.

Mike Jowsey
August 14, 2010 8:32 pm

duckster says:
August 14, 2010 at 7:48 pm
Looking at the paper above…
The main point of this paper is to debunk the maths Mann used. You can get similar hockey sticks by using random numbers. Speak to that subject please.
By shifting focus to whether or not the graph shows a MWP is a strawman and is completely irrelevant to the point of the paper. Besides, the graph (fig.16) uses the same proxy data Mann used, with correct maths. Mann’s proxy data (and maths) explicitly set out to remove the MWP so it is no surprise that his biased proxy selections camouflage the MWP. Nevertheless, fig.16 does show temperatures 1000 years ago were on a par with today’s (according to Mann’s proxies).

Verified by MonsterInsights