New paper makes a hockey sticky wicket of Mann et al 98/99/08

NOTE: This has been running two weeks at the top of WUWT, discussion has slowed, so I’m placing it back in regular que.  – Anthony

UPDATES:

Statistician William Briggs weighs in here

Eduardo Zorita weighs in here

Anonymous blogger “Deep Climate” weighs in with what he/she calls a “deeply flawed study” here

After a week of being “preoccupied” Real Climate finally breaks radio silence here. It appears to be a prelude to a dismissal with a “wave of the hand”

Supplementary Info now available: All data and code used in this paper are available at the Annals of Applied Statistics supplementary materials website:

http://www.imstat.org/aoas/supplements/default.htm

=========================================

Sticky Wicket – phrase, meaning: “A difficult situation”.

Oh, my. There is a new and important study on temperature proxy reconstructions (McShane and Wyner 2010) submitted into the Annals of Applied Statistics and is listed to be published in the next issue. According to Steve McIntyre, this is one of the “top statistical journals”. This paper is a direct and serious rebuttal to the proxy reconstructions of Mann. It seems watertight on the surface, because instead of trying to attack the proxy data quality issues, they assumed the proxy data was accurate for their purpose, then created a bayesian backcast method. Then, using the proxy data, they demonstrate it fails to reproduce the sharp 20th century uptick.

Now, there’s a new look to the familiar “hockey stick”.

Before:

Multiproxy reconstruction of Northern Hemisphere surface temperature variations over the past millennium (blue), along with 50-year average (black), a measure of the statistical uncertainty associated with the reconstruction (gray), and instrumental surface temperature data for the last 150 years (red), based on the work by Mann et al. (1999). This figure has sometimes been referred to as the hockey stick. Source: IPCC (2001).

After:

FIG 16. Backcast from Bayesian Model of Section 5. CRU Northern Hemisphere annual mean land temperature is given by the thin black line and a smoothed version is given by the thick black line. The forecast is given by the thin red line and a smoothed version is given by the thick red line. The model is fit on 1850-1998 AD and backcasts 998-1849 AD. The cyan region indicates uncertainty due to t, the green region indicates uncertainty due to β, and the gray region indicates total uncertainty.

Not only are the results stunning, but the paper is highly readable, written in a sensible style that most laymen can absorb, even if they don’t understand some of the finer points of bayesian and loess filters, or principal components. Not only that, this paper is a confirmation of McIntyre and McKitrick’s work, with a strong nod to Wegman. I highly recommend reading this and distributing this story widely.

Here’s the submitted paper:

A Statistical Analysis of Multiple Temperature Proxies: Are Reconstructions of Surface Temperatures Over the Last 1000 Years Reliable?

(PDF, 2.5 MB. Backup download available here: McShane and Wyner 2010 )

It states in its abstract:

We find that the proxies do not predict temperature significantly better than random series generated independently of temperature. Furthermore, various model specifications that perform similarly at predicting temperature produce extremely different historical backcasts. Finally, the proxies seem unable to forecast the high levels of and sharp run-up in temperature in the 1990s either in-sample or from contiguous holdout blocks, thus casting doubt on their ability to predict such phenomena if in fact they occurred several hundred years ago.

Here are some excerpts from the paper (emphasis in paragraphs mine):

This one shows that M&M hit the mark, because it is independent validation:

In other words, our model performs better when using highly autocorrelated

noise rather than proxies to ”predict” temperature. The real proxies are less predictive than our ”fake” data. While the Lasso generated reconstructions using the proxies are highly statistically significant compared to simple null models, they do not achieve statistical significance against sophisticated null models.

We are not the first to observe this effect. It was shown, in McIntyre

and McKitrick (2005a,c), that random sequences with complex local dependence

structures can predict temperatures. Their approach has been

roundly dismissed in the climate science literature:

To generate ”random” noise series, MM05c apply the full autoregressive structure of the real world proxy series. In this way, they in fact train their stochastic engine with significant (if not dominant) low frequency climate signal rather than purely non-climatic noise and its persistence. [Emphasis in original]

Ammann and Wahl (2007)

On the power of the proxy data to actually detect climate change:

This is disturbing: if a model cannot predict the occurrence of a sharp run-up in an out-of-sample block which is contiguous with the insample training set, then it seems highly unlikely that it has power to detect such levels or run-ups in the more distant past. It is even more discouraging when one recalls Figure 15: the model cannot capture the sharp run-up even in-sample. In sum, these results suggest that the ninety-three sequences that comprise the 1,000 year old proxy record simply lack power to detect a sharp increase in temperature. See Footnote 12

Footnote 12:

On the other hand, perhaps our model is unable to detect the high level of and sharp run-up in recent temperatures because anthropogenic factors have, for example, caused a regime change in the relation between temperatures and proxies. While this is certainly a consistent line of reasoning, it is also fraught with peril for, once one admits the possibility of regime changes in the instrumental period, it raises the question of whether such changes exist elsewhere over the past 1,000 years. Furthermore, it implies that up to half of the already short instrumental record is corrupted by anthropogenic factors, thus undermining paleoclimatology as a statistical enterprise.

FIG 15. In-sample Backcast from Bayesian Model of Section 5. CRU Northern Hemisphere annual mean land temperature is given by the thin black line and a smoothed version is given by the thick black line. The forecast is given by the thin red line and a smoothed version is given by the thick red line. The model is fit on 1850-1998 AD.

We plot the in-sample portion of this backcast (1850-1998 AD) in Figure 15. Not surprisingly, the model tracks CRU reasonably well because it is in-sample. However, despite the fact that the backcast is both in-sample and initialized with the high true temperatures from 1999 AD and 2000 AD, it still cannot capture either the high level of or the sharp run-up in temperatures of the 1990s. It is substantially biased low. That the model cannot capture run-up even in-sample does not portend well for its ability

to capture similar levels and run-ups if they exist out-of-sample.

Conclusion.

Research on multi-proxy temperature reconstructions of the earth’s temperature is now entering its second decade. While the literature is large, there has been very little collaboration with universitylevel, professional statisticians (Wegman et al., 2006; Wegman, 2006). Our paper is an effort to apply some modern statistical methods to these problems. While our results agree with the climate scientists findings in some

respects, our methods of estimating model uncertainty and accuracy are in sharp disagreement.

On the one hand, we conclude unequivocally that the evidence for a ”long-handled” hockey stick (where the shaft of the hockey stick extends to the year 1000 AD) is lacking in the data. The fundamental problem is that there is a limited amount of proxy data which dates back to 1000 AD; what is available is weakly predictive of global annual temperature. Our backcasting methods, which track quite closely the methods applied most recently in Mann (2008) to the same data, are unable to catch the sharp run up in temperatures recorded in the 1990s, even in-sample.

As can be seen in Figure 15, our estimate of the run up in temperature in the 1990s has

a much smaller slope than the actual temperature series. Furthermore, the lower frame of Figure 18 clearly reveals that the proxy model is not at all able to track the high gradient segment. Consequently, the long flat handle of the hockey stick is best understood to be a feature of regression and less a reflection of our knowledge of the truth. Nevertheless, the temperatures of the last few decades have been relatively warm compared to many of the thousand year temperature curves sampled from the posterior distribution of our model.

Our main contribution is our efforts to seriously grapple with the uncertainty involved in paleoclimatological reconstructions. Regression of high dimensional time series is always a complex problem with many traps. In our case, the particular challenges include (i) a short sequence of training data, (ii) more predictors than observations, (iii) a very weak signal, and (iv) response and predictor variables which are both strongly autocorrelated.

The final point is particularly troublesome: since the data is not easily modeled by a simple autoregressive process it follows that the number of truly independent observations (i.e., the effective sample size) may be just too small for accurate reconstruction.

Climate scientists have greatly underestimated the uncertainty of proxy based reconstructions and hence have been overconfident in their models. We have shown that time dependence in the temperature series is sufficiently strong to permit complex sequences of random numbers to forecast out-of-sample reasonably well fairly frequently (see, for example, Figure 9). Furthermore, even proxy based models with approximately the same amount of reconstructive skill (Figures 11,12, and 13), produce strikingly dissimilar historical backcasts: some of these look like hockey sticks but most do not (Figure 14).

Natural climate variability is not well understood and is probably quite large. It is not clear that the proxies currently used to predict temperature are even predictive of it at the scale of several decades let alone over many centuries. Nonetheless, paleoclimatoligical reconstructions constitute only one source of evidence in the AGW debate. Our work stands entirely on the shoulders of those environmental scientists who labored untold years to assemble the vast network of natural proxies. Although we assume the reliability of their data for our purposes here, there still remains a considerable number of outstanding questions that can only be answered with a free and open inquiry and a great deal of replication.

===============================================================

Commenters on WUWT report that Tamino and Romm are deleting comments even mentioning this paper on their blog comment forum. Their refusal to even acknowledge it tells you it has squarely hit the target, and the fat lady has sung – loudly.

(h/t to WUWT reader “thechuckr”)

Share

The climate data they don't want you to find — free, to your inbox.
Join readers who get 5–8 new articles daily — no algorithms, no shadow bans.
0 0 votes
Article Rating
1.2K Comments
Inline Feedbacks
View all comments
August 27, 2010 7:26 am

Henry@Francisco
I agree with everything you say. We need a balance sheet of the cooling and warming properties of each component in the air. In W/m3 @GHG relevant % range/m2/24hr.
I am sure if I had been in charge I would have come up with some general testing procedure. But for that testing you also need money. That is why I said: why don’t we get the oil companies to fund this research? CO2 is their product, after all.

RR Kampen
August 27, 2010 7:28 am

Bill Tuttle says:
August 27, 2010 at 6:43 am
..
If CO2 wasn’t the most important one in the past, why is it suddenly the most important one today?

Because it is the greenhouse gas whose concentration is rising extremely rapidly. It is doing ‘suddenly’ so. At the same time, other important drivers voor recent warming cannot be identified.
That’s my point: that CO2 increases follow a rising temperature, and do not cause it.
Both happens.
If rising CO2 levels *caused* an increase in temperatures, you would not see temperatures falling as CO2 levels continued to rise for an additional 800 years.
We wouldn’t, and: we don’t.

Steve Keohane
August 27, 2010 8:03 am

Wiglaf says: August 26, 2010 at 1:40 pm
I was told by my botanist father cherry pits have cyanide as well. I have found it interesting that cyanide supposedly smells like almonds, and that the taste of cherries and almonds is actually very similar in my experience.

August 27, 2010 10:00 am

RR Kampen: August 27, 2010 at 7:28 am
Because it is the greenhouse gas whose concentration is rising extremely rapidly. It is doing ‘suddenly’ so. At the same time, other important drivers voor recent warming cannot be identified.
So, your position is that rising CO2 was not the primary cause of warming in the past, but because it is rising *now*, it miraculously *is* the primary cause.
Both happens.
In that case, explain why temperatures have been cooling since 1999. It can’t possibly be happening if CO2 is the main driver for temperature, as is your position.
We wouldn’t, and: we don’t.
We have seen it many times in the past and we’re seeing it now.

August 27, 2010 10:39 am

Henry at Bryan
Unfortunately, what many of us here suspect, is that the global warming party is over.
e.g..(featuring now)
http://wattsupwiththat.com/2010/08/27/pre-empting-on-the-solar-curve-fit/
What is happening now in Pakistan could be due to the change in the 2nd variable of global warming/cooling given below
http://letterdash.com/HenryP/more-carbon-dioxide-is-ok-ok
don’t forget:
http://letterdash.com/HenryP/the-term-climate-change-is-hiding-the-fact-that-global-warming-has-stalled

August 27, 2010 10:40 am

Steve Keohane: August 27, 2010 at 8:03 am
I was told by my botanist father cherry pits have cyanide as well. I have found it interesting that cyanide supposedly smells like almonds, and that the taste of cherries and almonds is actually very similar in my experience.
There are a surprising number of cyanogenetic foods, too — soy, cassava (tapioca, anyone?), lima beans, spinach, bamboo shoots, and (big surprise) almonds. People who eat a *lot* of cassava say they prefer the types that concentrate more cyanide in them — they’re more flavorful.

DCA engineer
August 27, 2010 11:39 am

Henry Pool
Thank you again for your earlier replies. Here is some information on Dr. Miskolczi.
http://miskolczi.webs.com/
This Dr. Spencer’s take on it.
http://www.drroyspencer.com/2010/08/comments-on-miskolczi%e2%80%99s-2010-controversial-greenhouse-theory/
I want to thank you in advance and look forward to your reply.

DCA engineer
August 27, 2010 11:42 am
August 27, 2010 12:50 pm

Henry
@DCA engineer/Francisco/etc
I had a quick look through at what M. and RS say, and I would strongly recommend that you go through all of my postings here to understand where we differ.
I agree with RS – I doubt the work of M (too much mathematics, very few real measurements e.g. no physics&chemistry related to an indiviual components’ characteristics).
I also say that water vapor is the greatest greenhouse gas (average abt 1% in air) but even water vapor causes cooling when the sun shines…..
I am not even sure if CO2 is a greenhouse gas – i.e that its net effect is warming rather than cooling; if it is one it must be very weak one, much weaker than water vapor.
So the increase of 0,01% in CO2 concentration of the past 50 years compares to almost nothing with that of water.
On the other side of the spectrum, Ozone is a big anti greenhouse gas, i.e. it probably cools more than it warms.
If we want to know what is good and what is bad for us, we need to see a balance sheet of the cooling and warming properties of each GHG & other air components.
I favor a simple testing system as proposed by Francisco. http://wattsupwiththat.com/2010/08/17/breaking-new-paper-makes-a-hockey-sticky-wicket-of-mann-et-al-99/#comment-467920
You must be able to fluctuate your concentration of GHG (balanced with nitrogen) where the other air components stay completely constant. Then you must be able to apply the energy and measure the cooling and warming rates in
W/ M3 %GHG (relevant range) / m2/ 24 hours.
That is probably easier said than done. I have not yet thought much about how to apply the whole spectrum of wavelength energy, but knowing the spectra of the components might help us a lot to chose and apply the correct radiation.
good luck
Henry Pool

Invariant
August 27, 2010 2:41 pm

Possibly it’s time for another top story now Anthony:
McShane and Wyner has shown that current global warming may possibly not be unprecedented. Spencer and Braswell points out that “It is clear that the accurate diagnosis of short‐term feedbacks (let alone long‐term climate sensitivity) from observations of natural fluctuations in the climate system is far from a solved problem.”
http://www.drroyspencer.com/wp-content/uploads/Spencer-Braswell-JGR-2010.pdf
Is it possible to predict our climate without knowledge?
Those who have knowledge, don’t predict. Those who predict, don’t have knowledge.
– Lao Tzu, Chinese poet, 500 BC.

kim
August 27, 2010 4:21 pm

Invariant 2:41 PM.
Nice. I predict global cooling for twenty years based on the concatenation of the cooling phases of the oceanic oscillations and for a century based on the coming dearth of sunspots, the Eddy Minimum. But I confess that I have no certain knowledge of the effect of the oscillations, of CO2, and of the sun.
=================

pax
August 27, 2010 4:58 pm

Henry Pool
The CO2 concentration was about 315ppm 50 year ago and is about 385ppm today. How do you work that out to be a 0.01% increase?

Howling Winds
August 27, 2010 5:02 pm

I don’t know where the rest of you guys and gals live, but I have spent 43 summers in South Carolina, 22 of those as a farm hand. It is always “hotter than hell” here, and a few degrees in either direction is not going to make any difference to me.

jwurster
August 27, 2010 5:16 pm

I don’t think any scientific argument is going to convince warmers of future climate trends. Anyone who disagrees with them should start thinking of ways to advance the food and water supply. The Western and IPCC governments are not going to plan for cooler temperatures. It is up to people outside of the powers that be to do this planning. Advocates and ideas are needed.

Scott
August 27, 2010 6:32 pm

pax says:
August 27, 2010 at 4:58 pm

Henry Pool
The CO2 concentration was about 315ppm 50 year ago and is about 385ppm today. How do you work that out to be a 0.01% increase?

I would assume he calculated it as:
~300 ppm = 0.03%
~400 ppm = 0.04%
Difference = 0.01%
Not how I would have calculated it…..
-Scott

August 27, 2010 10:08 pm

Brian (August 27, 2010 at 7:26 am) senses a move among climatologists to attribute the cause of tragic events, such as the recent floods in Pakistan, to man-made emissions of CO2 unless this attribution can be proved false. The existence of such a movement is implied by the existence, in the IPCC’s 2007 report, of extensive coverage ( http://www.ipcc.ch/publications_and_data/ar4/wg1/en/ch9.html ) of the topic of “attribution” and non-existent coverage ( http://icecap.us/images/uploads/SPINNING_THE_CLIMATE08.pdf ) of the topic of “statistical validation.” By the evidence referenced above, attribution has replaced statistical validation in the methodology of the IPCC’s investigation into the allegation of man-made global warming.
Attribution makes the argument which, in logic, is called “Argumentum ad Ignorantiam.”
Let P designate a hypothesis. This argument is of one of the two forms:
P is unproved; therefore not P is true and
Not P is unproved; therefore, P is true.
Though P is unproved, P may nonetheless be true. Conversely, though not P is unproved, not P may nonetheless be true, Thus, Argumentum ad Ignorantiam poses a false dichotomy.
In the methodology of science, empirical testing of P plays a role that is similar to but different from the one that is played by Argumentum ad Ignorantiam for IPCC climatology. In the methodology of science if, in repeated testing, P is not found false then P gains a degree of “statistical significance.” Under no circumstance is P found true.
A methodology that replaces empirical testing by Argumentum ad Ignorantium is not a scientific methodology. Nonetheless, in its 2007 report the IPCC implies ( http://www.ipcc.ch/publications_and_data/ar4/wg1/en/ch1s1-2.html ) that its methodology is scientific in nature. By implying that its methodology is scientific when it isn’t, the IPCC errs.

August 27, 2010 11:13 pm

Scott says:
I would assume he calculated it as:
~300 ppm = 0.03%
~400 ppm = 0.04%
Difference = 0.01%
Not how I would have calculated it…..
-Scott
Correct calculation. Why not do it like that?Nobody says that the water vapor in the atmosphere is about 12000 ppm. They say it is 1%.
Comparing apples with apples?

Ammonite
August 28, 2010 4:40 am

Anthony’s reply to duckster: You really can’t argue on the basis of noise, or annual values. The mean line is the message. – Anthony
Thankyou Anthony, a cogent reply. Applying the same logic to GISS, HadCrut or satellite data clearly shows the trend in global temperature remains upward (apologies for being OT).

pax
August 28, 2010 4:44 am

Henry Pool, you obviously don’t understand percentages. You said that the *increase* in *concentration* was 0.01%, it is not, the increase in concentration is more than 20%.
According to your logic I didn’t receive a pay-raise of 15% a couple of months ago, no, I got a pay-raise of 0.000000000000000000000000000001% because I have to take the amount relative to all the monies in the world. I think I’ll take this up with my employer and see if he agrees with my reasoning 🙂
You ask: Why not do it like that? Well, maybe because nobody does it like that.

Francisco
August 28, 2010 7:14 am

pax says:
August 28, 2010 at 4:44 am
“Henry Pool, you obviously don’t understand percentages. You said that the *increase* in *concentration* was 0.01%, it is not, the increase in concentration is more than 20%.”
================
That’s true, pax, but it is no less true that the change involved only 0.01 percent of the atmosphere by volume, or, if you prefer, one ten thousandth of it, or 0.0001 of it, which went from being something else to being CO2, while 99.99 percent of it has remained stubbornly unchanged. The marvels attributed to this minuscule change are understandably met with a lot of incredulity and head scratching, especially because they seem to be 100% based on a high pile of theories and hypothesis precariously stacked upon one another (and subject to wild disagreements), and 0% based on any empirical evidence. If a similarly minuscule part of the money spent on feeding highly expensive computer models with theoretical assumptions, and promoting apocalypse at all costs, were spent in conducting some physical experiments in controlled conditions in order to get a more palpable idea of the actual effects of such a tiny change, then I would expect that the results of such experiments, if corroborating the theories, would do a lot more to convince the doubting Thomases than all the billions spent on promoting fantasy an vitriol.

August 28, 2010 7:37 am

There are different units of concentration you can use. In this case I used the unit of concentration that also contains % as unit, i.e. % m/m (but the m/m or w/w is usually not mentioned)
The wording I used is therefore correct, but I agree that if you have not studied a bit of chemistry you might misunderstand.

Vince Causey
August 28, 2010 7:45 am

RR Kampen,
“Beware. I could not prove to most people that the ratio of circumference and diameter of a circle is not a rational number. They don’t know the math.”
I get it. Most people are too dumb to understand your “proof” of CAGW. Nice argument!

August 28, 2010 8:50 am

Henry@RR Kampen
Summary of AGW &CAGW test results and theory
Let us have a planet. Let us add some CO2. Let us see if the temperature went up.
It did!!\ So that must be it!!
Are you willing to place your bets?

August 28, 2010 10:20 am

RR Kampen: August 27, 2010 at 12:22 am
Beware. I could not prove to most people that the ratio of circumference and diameter of a circle is not a rational number. They don’t know the math.
So, how would that stop you from explaining it in simpler terms? I’ve explained how a helicopter manages to land safely after its engine quits to schoolchildren who don’t know aerodynamics.

August 28, 2010 10:28 am

RR Kampen says:
August 27, 2010 at 12:22 am
Henry Pool says:
August 26, 2010 at 7:13 am
RR kampen
“….I hope the problem is not proving it to you. Anyway, you can do John Tyndall’s experiments (around 1860) in your own yard.”
Do you have a link to a description of that experiment?
“Here is another experiment: http://www.youtube.com/watch?v=SeYfl45X1wo ….”
RR Kampen, compared to what you know about it, I am just a babe in the woods with respect to the science of CO2-caused global warming or the lack of it. Please be so good and shed some light on the concerns the youtube experiment you pointed to raised in my mind.
Although the youtube experiment proves that CO2 absorbs visible light (notwithstanding that the narrator asserts that the visible radiation is IR), it is not clear to me what it proves with respect to the extent of the presence or absence of global warming.
I don’t think that anyone here objects to the fact that CO2 absorbs IR and re-radiates it in random directions. I believe that proof was provided in related discussion threads that CO2 absorbs visible light and re-radiates it into random directions as IR. The discussion revolves around the question of how much CO2 in the atmosphere it takes to cause a specific amount of warming of the global surface.
The youtube experiment shows that a large (unspecified) amount of atmospheric CO2 prevents a large (unspecified) amount of visible light from reaching the global surface.
The experiment does not illustrate what happens to the visible light that has been absorbed (although the narrator asserts that it warms the “atmosphere”). Specifically, it does not illustrate whether the absorbed visible light is re-radiated in the form of IR and how much of the IR reaches the global surface, how much of the IR is being radiated to space, and whether there is a global net-gain of energy.
Moreover, it is not clear that the experiment adequately demonstrates CO2’s opacity to IR and the absorption and re-radiation of IR.
Therefore the youtube experiment you linked to provides no insight as to the subject of the discussion, but it is a nice illustration of the opacity of CO2 to visible light.