Note: Steve McIntyre is also quite baffled by the Marcott et al paper, finding it currently unreproducible given the current information available. I’ve added some comments from him at the bottom of this post – Anthony
Guest Post by Willis Eschenbach
I don’t know what it is about proxies that makes normal scientists lose their senses. The recent paper in Science (paywalled of course) entitled A Reconstruction of Regional and Global Temperature for the Past 11,300 Years” (hereinafter M2012) is a good example. It has been touted as the latest hockeystick paper. It is similar to the previous ones … but as far as I can see it’s only similar in how bizarre the proxies are.
Nowhere in the paper do they show you the raw data, although it’s available in their Supplement. I hate it when people don’t show me their starting point. So let me start by remedying that oversight:

Figure 1. All of the proxies from M2012. The colors are only to distinguish individual records, they have no meaning otherwise.
I do love the fact that from that collection of temperature records they draw the conclusion that:
Current global temperatures of the past decade have not yet exceeded peak interglacial values but are warmer than during ~75% of the Holocene temperature history.
Really? Current global temperature is about 14°C … and from those proxies they can say what the past and present global average temperatures are? Well, let’s let that claim go for a moment and take a look at the individual records.
Here’s the first 25 of them:
Figure 2. M2012 proxies 1 to 25. Colors as in Figure 1. Note that each panel has its own vertical axis. Numbers to the left of each title are row/column.
Well … I’d start by saying that it seems doubtful that all of those are measuring the same thing. Panel 3/1 (row 3, column 1) shows the temperature decreasing for the last ten thousand years. Panels 4/4 and 4/5 show the opposite, warming for the last ten thousand years. Panel 4/3 shows four thousand years of warming and the remainder cooling.
Let’s move on to the next 25 contestants:
Figure 3. M2012 proxies 26 to 50. Colors as in Figure 1. Note that each panel has its own vertical axis. Numbers to the left of each title are row/column.
Here we see the same thing. Panels 1/1 and 4/1 show five thousand years of warming followed by five thousand years of cooling. Panel 1/5 shows the exact opposite, five thousand cooling years followed by five thousand of warming. Panel 4/5 show steady warming, panel 5/2 shows steady cooling, and panel 2/2 has something badly wrong near the start. Panel 2/4 also contains visible bad data.
Onwards, we near the finish line …
Figure 4. M2012 proxies 51 to 73. Colors as in Figure 1. Note that each panel has its own vertical axis. Numbers to the left of each title are row/column.
Panel 2/1 shows steadily rising temperatures for ten thousand years, as does panel 3/4. Panels 4/1 and 5/1, on the other hand, show steadily decreasing temperatures. Panel 4/2 has a hump in the middle. but panel 1/2 shows a valley in the middle.
Finally, here’s all the proxies, with each one shown as anomalies about the average of its last 2,000 years of data:

Figure 5. All Marcott proxies, expressed as anomalies about their most recent 2,000 years of record. Black line shows 401-point Gaussian average. N=9,288.
A fine example of their choice of proxies can be seen in the fact that they’ve included a proxy which claims a cooling about nine degrees in the last 10,000 years … although to be fair, they’ve also included some proxies that show seven degrees of warming over the same period …
I’m sorry, guys, but I’m simply not buying the claim that we can tell anything at all about the global temperatures from these proxies. We’re deep into the GIGO range here. When one proxy shows rising temperatures for ten thousand years and another shows dropping temperatures for ten thousand years, what does any kind of average of those two tell us? That the temperature was rising seven degrees while it was falling nine degrees?
And finally, their claim of turning that dogs breakfast shown in Figure 1 into an absolute global temperature and comparing it to the current 14°C average temperature estimate?
Don’t make me laugh.
I say the reviewers of this paper didn’t use their Mark I eyeball. The first thing to do when dealing with a multi-proxy study is to establish ex-ante criteria for the selection of the proxies (“ex-ante” meaning choose your criteria before looking at the proxies). Here are their claimed criteria …
This study is based on the following data selection criteria:
• Sampling resolution is typically better than ~300 yr.
• At least four age-control points span or closely bracket the full measured interval.
• Chronological control is derived from the site itself and not primarily based on tuning to other sites. Layer counting is permitted if annual resolution is plausibly confirmed (e.g., ice-core chronologies). Core tops are assumed to be 1950 AD unless otherwise indicated in original publication.
• Each time series spans greater than 6500 years in duration and spans the entire 4500 – 5500 yr B.P. reference period.
• Established, quantitative temperature proxies
• Data are publicly available (PANGAEA, NOAA-Paleoclimate) or were provided directly by the original authors in non-proprietary form.
• All datasets included the original sampling depth and proxy measurement for complete error analysis and for consistent calibration of age models (Calib 6.0.1 using INTCAL09 (1)).
Now, that sounds all very reasonable … except that unfortunately, more than ten percent of the proxies don’t meet the very first criterion, they don’t have sampling resolution that is better than one sample per 300 years. Nice try, but eight of the proxies fail their own test.
I must say … when a study puts up its ex-ante proxy criteria and 10% of their own proxies fail the very first test … well, I must say, I don’t know what to say.
In any case, then you need to LOOK AT EACH AND EVERY PROXY. Only then can you begin to see if the choices make any sense at all. And in this case … not so much. Some of them are obviously bogus. Others, well, you’d have to check them one by one.
Final summary?
Bad proxies, bad scientists, no cookies for anyone.
Regards,
w.
==============================================================
Steve McIntyre writes in a post at CA today:
Marcott et al 2013 has received lots of publicity, mainly because of its supposed vindication of the Stick. A number of commenters have observed that they are unable to figure out how Marcott got the Stick portion of his graph from his data set. Add me to that group.
The uptick occurs in the final plot-point of his graphic (1940) and is a singleton. I wrote to Marcott asking him for further details of how he actually obtained the uptick, noting that the enormous 1920-to-1940 uptick is not characteristic of the underlying data. Marcott’s response was unhelpful: instead of explaining how he got the result, Marcott stated that they had “clearly” stated that the 1890-on portion of their reconstruction was “not robust”. I agree that the 20th century portion of their reconstruction is “not robust”, but do not feel that merely describing the recent portion as “not robust” does full justice to the issues. Nor does it provide an explanation.
Read Steve’s preliminary analysis here:
[UPDATE] In the comments, Steve McIntyre suggested dividing the proxies by latitude bands. Here are those results:
Note that there may be some interesting things buried in there … just not what Marcott says.
Also, regarding the reliability of his recent data, he describes it as “not robust”. It is also scarce. Only 0.6% of the data points are post 1900, for example. This raises the question of how he compared modern temperatures to the proxies, since there is so little overlap.
Finally, about a fifth of the proxies (14 of 73) have the most recent date as exactly 1950 … they said:
Core tops are assumed to be 1950 AD unless otherwise indicated in original publication.
Seems like an assumption that is almost assuredly wrong. I don’t know if that’s a difference that makes a difference, depends on how wrong it is. If we take the error as half the distance to the next data point for each affected proxy, it averages about ninety years … pushing 1950 back to 1860 … yeah, I’ll go with “not robust” for that.
[UPDATE 2] Yes, I am shoveling gravel, one ton down, six to go … and I do get to take breaks. Here’s the result of my break, the Marcott proxies by type:
And here’s a picture of yr. unbending author playing what we used to call the “Swedish Banjo”.
Best to all,
w.


So, to summarize the issues with Marcott, Shakun, Clark, and Mix 2013:
• Their proxy data can support any conclusion you desire.
• They include Mann et al’s. (2008) tree ring reconstructions.
• 10% of their proxies have sampling resolutions of greater than 300 years.
• The “hockey stick” shape is simply an artifact produced by appending high resolution (annual) temperature data onto low resolution (140 years to >300 years) foraminifera data from marine sediment cores.
• In correspondence with Steve McIntyre Marcott says they “clearly” stated that the 1890-on portion of their reconstruction was “not robust”.
• Yet they conclude “Our global temperature reconstruction for the past 1500 years is indistinguishable within uncertainty from the Mann et al. (2) reconstruction.”
Another shining example of Climastrology at its finest by authors who are definitely “Team” players! Who knew reading goat entrails and interpreting temperature proxies were essentially the same thing?
Sorry to be a bit facetious but Fig. 5 reminded me of nothing so much as one of my grandchildren’s scribbles with coloured pencils.
Intersting article Willis.
If Marcott himself admits there is no robust data for the last 60 years of the study (ie 1890-1950), how is it that co-author Prof Clarke and Candace Major (from the body which sponsored the research) are making claims about warming spiking over ‘the last 100 years’?
http://www.dailymail.co.uk/sciencetech/article-2290138/Earth-warmest-ice-age–temperatures-rising.html?ito=feeds-newsxml
The max that could possibly be claimed of that 100 would be the *first* 40 years, ie 1850-1890, surely?
Am I missing something?
These two statements are contradictory !
Marcott to Steve McIntyre:
Marcott to Daily Mail;
Steve McIntyre reports:
Marcott is overstating his case when he says that. Here is the one use of “robust” in either the artilcle or the supplementary information.
The way I parse that is as follows:
• They used their “Standard 5×5” averaging method, and got one answer.
• Then they used a method to “in-fill” (make up) data, averaged the mix of real data and made-up data, and got a second answer.
The DIFFERENCE between the two was judged to be “probably not robust”.
This is very different from what Steve reports. I’ll cross-post this on his blog as well.
w.
Willis, I’ve parsed their coretop dating assumption. I doubt that specialists realize the effect. I’ve got most of a post done on this.
@Steve McIntyre – the modern portion of the proxies seems to be the least understandable, and the most messy. I wonder if we have a YAD061 type problem once again, where a single proxy with a modern jump becomes the most influential of the entire set?
But this work of art is Grade A Climatology.
Which has zero interest in atmospheric science or any other discipline that incorporates the scientific method.
Imagining information, from noise appears to be the forte of these wizards of Climatology.
OT, but relevant:
[snip – No, it is not. I won’t have this thread hijacked over a gun discussion – Anthony]
Steve, a point of clarification please. When they do these types of reconstructions, do they calibrate the average of proxies to global temperature (assuming tele-connection etc…), while skipping the step of evaluating/calibrating individual proxies to local temperatiure at the proxy site? If that is the case, consider my mind blown. I’ll post this question on your blog as well…
How did Marcott’s paper get published? This looks like rubbish.
Steve McIntyre says:
March 14, 2013 at 11:15 am (Edit)
I look forward to it, Steve. In the meantime, I’ve posted another update with this graphic:

All the best,
w.
Willis –
With regard to resolution criterion, while the published article states, “Sampling resolution is typically better than ~300 yr,” an earlier version has “Sampling resolutions better than ~400 yrs.”
This suggests that a 300-year resolution wasn’t intended to be a hard-and-fast criterion for selection. More interesting stuff as you look deeper.
Anthony … the reply Steve got from Marcott seemed to admit this to be true – Marcott:
First, it looks like there are at least 3 groupings with different patterns. So regional changes are THE pattern, not global.
Second, it looks like putting them all together you get a mathematically correct result but a result that has no meaning. Like saying the average height of a group of people is 6′ 7″ because you have a 3 NBA players in a crowd of 12.
clivebest says: March 14, 2013 at 12:44 pm
“These two statements are contradictory !”
Not so. Present temperatures are known from thermometers, not proxies.
Nick Stokes says:
“That doesn’t mean they aren’t measuring temps way back, and we do know what temperatures are now. Of course they can be compared – why not?”
They don’t compare temperatures they compare anomalies to “1961 – 1990” period. How did they do that – since they have no proxy data 1961-1990 ? To properly calculate anomalies you should first independently fine the average between 1961-190 at each site, then subtract this from each of the time series. Finally you make a global area weighted average on a 5×5 grid. Did they do that ?
As far as I can see they simply subtracted off a fixed temperature offset. This then conveniently allows you to overlay the instrument data, which I am predict will appear in AR5.
Therefore, IMHO their result is NOT independent of the instrument data
One of the novelties of this temperatures reconstruction from previous major ones is the use of alkenones as temperature proxies. 31 of the 73 are alkenones, gathered from both the sea and lakes. In the comments at Climate Audit, Keith DeHavelle points to a three minute video explaining in alkenone temperature proxies in layman’s terms.
Marcott et al. is a global surface (air) temperature reconstruction. Alkenones are used as a proxy for changes in water temperatures, with water temperatures then used as a proxy for surface temperatures. Alkenones are thus a double proxy. Possible issues are:-
1) The large changes in surface temperatures will lead to much smaller temperatures in sea temperatures.
2) An important factor in water temperature particular area of ocean over time is changes in ocean currents. For instance, the Gulf Stream is hugely important in the North Atlantic, where five of the proxies are located.
Willis, love your style. What you have failed to realise, unless the proxies are run with a dollar filter no output will make any sense.
Willis, in fact there was no need for you to plot out all the proxies, because Marcott has done it already in his thesis, see comments from Ian and Jean S at CA.
The Marcott thesis is at
http://ir.library.oregonstate.edu/xmlui/handle/1957/21129
and chapter 4 is a first version of the paper currently being discussed.
Look at figs 4.2 and 4.3 – no 20th century spike!!!
My impression is that these proxies are essentially random sets of data, of no particular value when it comes to past temperatures, and that naturally, if one averages random data, one gets a fairly flat line. Then, tack on recent non-proxies trends to complete the hockey stick. Seems like a lot of work for so little result.
Paul Matthews says:
March 14, 2013 at 4:04 pm
Look at figs 4.2 and 4.3 – no 20th century spike!!!
It comes even worse. Hank did a reconstruction here and it has an inverse hockey stick in the data, there are several posts by James worth to be read, enjoy:
http://suyts.wordpress.com/2013/03/14/hockey-stick-found-in-marcott-data/
http://suyts.wordpress.com/2013/03/14/more-fishing-for-hockey-sticks-in-marcott-et-al-2013/
http://suyts.wordpress.com/2013/03/10/the-hockey-stick-resurrected-by-marcott-et-al-2012/
Willis – when you plot the “Global” sets from the Temperature Stack section of the data worksheet you do get “blades” from most of the sets – primarily starting just after 1900.
And the Agassiz-Renland data shows 1.38 deg C warming from 1900-1940, or 0.83 deg C from 1900 – 1960. Which fits well, considering the lack of data points within the 1900-1950 window, with Marcott admission it is the primary driver of the hockey stick –
Marcott Agassiz-Renland:
http://tinyurl.com/Marcott-Agassiz
Marcott Temp Stack “Global”:
http://tinyurl.com/Marcott-Global
There are only a handful of remaining data sets with usable data during the 1900 to 1950 period, and with fair amount of those showing cooling during that time. The question seems to be exactly what process did Marcott use to create the “Global” data sets in the Temp Stack section.
And second, and more important it would seem, since the entirety of the data and the reconstruction behind the hockey stick portion is by the authors admission largely based on one set of data and they themselves do not consider it “robust” – why is it even included in the paper … and why was this not picked up and addressed by peer review?
I’ve started reading Marcott’s 2011 PhD dissertation, and I want to make one general point to anyone who may have the statistical/scientific background to write to him or to engage with him in public scientific discussion. He strikes me (whatever the weaknesses of the 2013 article) as a real human who should be amenable to sincere scientific discussion, unless the Mannians manage to ruin him. He now has a large opportunity for public engagement and science education for many — let’s hope he uses it well, and let’s encourage him to use it well.
i.e., let’s not assume bad faith or a Mann-style political engagement unless he should prove it later. He is a young post-doc who may have made some mis-steps as he and co-authors got drawn along in this process with “Science” mag. and preparing for AR5, etc. but he strikes me as potentially far more salvageable than someone like Mann (hard core activist from the start).
Which Ice core records were they using? Because none of the ones in the seperated out graphs look like any of the Ice core reconstructions I’ve seen. As for the rest, their assertions can only be explained if you tiljanderise the data.
About your other “endeavour”, I bet you feel cleaner after shovelling twelve tonnes of gravel than having shovelled through this muck. Having been there and done that, I feel for you :-).
Nick Stokjes writes “Not so. Present temperatures are known from thermometers, not proxies.”
To make the comparison valid between present and past, the present temperatures MUST be produced by the proxies, not the thermometers. This is not optional Nick.
For anyone who needs to see/review what Willis said last year about the Shakun et al. (2012) in Nature, which included Marcott as co-author, I put links in this comment:
http://wattsupwiththat.com/2013/03/14/another-hockey-stick/#comment-1248510
I’m getting less willing to cut Marcott, Shakun, and Clark any slack when it is apparent that these shenanigans have been going on for awhile in a variety of places. Do they ever get around to addressing critical challenges?
TTTM,
“the present temperatures MUST be produced by the proxies, not the thermometers.”
You may be thinking of treerings, where the proxies have to overlap to be calibrated. But Marcott et al are using mainly marine proxies where the temperatures are calibrated independently of air temp data.
In fact, their proxies can’t produce present temperatures reliably. Their time resolution is too low to capture fast change, and they frequently don’t go beyond a point many years past. That doesn’t mean they aren’t measuring temps way back, and we do know what temperatures are now. Of course they can be compared – why not?