Steig et al 'Antarctica Warming Paper' process is finally replicated, and dealt a blow to "robustness".

Jeff Id emailed me today, to ask if I wanted to post this with the caveat “it’s very technical, but I think you’ll like it”. Indeed I do, because it represents a significant step forward in the puzzle that is the Steig et all paper published in Nature this year ( Nature, Jan 22, 2009) that claims to have reversed the previously accepted idea that Antarctica is cooling. From the “consensus” point of view, it is very important for “the Team” to make Antarctica start warming. But then there’s that pesky problem of all that above normal ice in Antarctica. Plus, there’s other problems such as buried weather stations which will tend to read warmer when covered with snow.  And, the majority of the weather stations (and thus data points) are in the Antarctic peninsula, which weights the results. The Antarctic peninsula could even be classified under a different climate zone given it’s separation from the mainlaind and strong maritime influence.

A central prerequisite point to this is that Steig flatly refused to provide all of the code needed to fully replicate his work in MatLab and RegEM, and has so far refused requests for it. So without the code, replication would be difficult, and without replication, there could be no significant challenge to the validity of the Steig et al paper.

Steig’s claim that there has been “published code” is only partially true, and what has been published by him is only akin to a set of spark plugs and a manual on using a spark plug wrench when given the task of rebuilding an entire V-8 engine.

In a previous Air Vent post, Jeff C points out the percentage of code provided by Steig:

“Here is an excellent flow chart done by JeffC on the methods used in the satellite reconstruction. If you see the little rectangle which says RegEM at the bottom right of the screen, that’s the part of the code which was released, the thousands of lines I and others have written for the rest of the little blocks had to be guessed at, some of it still isn’t figured out yet.”

http://noconsensus.files.wordpress.com/2009/04/steigflowrev4-6-09.jpg?w=598&h=364
RegEM Satellite data flow chart. Courtesy Jeff C - click for larger image

With that, I give you Jeff and Ryan’s post below. – Anthony

Antarctic Coup de Grace

Posted by Jeff Id on May 20, 2009

I was going to hold off on this post because Dr. Weinstein’s post is getting a lot of attention right now it has been picked up on several blogs and even translated into different languages but this is too good not to post.

Ryan has done something amazing here, no joking. He’s recalibrated the satellite data used in Steig’s Antarctic paper correcting offsets and trends, determined a reasonable number of PC’s for the reconstruction and actually calculated a reasonable trend for the Antarctic with proper cooling and warming distributions – He basically fixed Steig et al. by addressing the very concern I had that AVHRR vs surface station temperature(SST) trends and AVHRR station vs SST correlation were not well related in the Steig paper.

Not only that he demonstrated with a substantial blow the ‘robustness’ of the Steig/Mann method at the same time.

If you’ve followed this discussion whatsoever you’ve got to read this post.

RegEM for this post was originally transported to R by Steve McIntyre, certain versions used are truncated PC by Steve M as well as modified code by Ryan.

Ryan O – Guest post on the Air Vent

I’m certain that all of the discussion about the Steig paper will eventually become stale unless we begin drawing some concrete conclusions. Does the Steig reconstruction accurately (or even semi-accurately) reflect the 50-year temperature history of Antarctica?

Probably not – and this time, I would like to present proof.

I: SATELLITE CALIBRATION

As some of you may recall, one of the things I had been working on for awhile was attempting to properly calibrate the AVHRR data to the ground data. In doing so, I noted some major problems with NOAA-11 and NOAA-14. I also noted a minor linear decay of NOAA-7, while NOAA-9 just had a simple offset.

But before I was willing to say that there were actually real problems with how Comiso strung the satellites together, I wanted to verify that there was published literature that confirmed the issues I had noted. Some references:

(NOAA-11)

Click to access i1520-0469-59-3-262.pdf

(Drift)

Click to access orbit.pdf

(Ground/Satellite Temperature Comparisons)

Click to access p26_cihlar_rse60.pdf

The references generally confirmed what I had noted by comparing the satellite data to the ground station data: NOAA-7 had a temperature decrease with time, NOAA-9 was fairly linear, and NOAA-11 had a major unexplained offset in 1993.

Fig_1
Fig. 1: AVHRR trend (points common with ground data).

Let us see what this means in terms of differences in trends.

Fig_2
Fig. 2: Difference in trend between AVHRR data and ground data.

The satellite trend (using only common points between the AVHRR data and the ground data) is double that of the ground trend. While zero is still within the 95% confidence intervals, remember that there are 6 different satellites. So even though the confidence intervals overlap zero, the individual offsets may not.

In order to check the individual offsets, I performed running Wilcoxon and t-tests on the difference between the satellites and ground data using a +/-12 month range. Each point is normalized to the 95% confidence interval. If any point exceeds +/- 1.0, then there is a statistically significant difference between the two data sets.

Fig_3
Fig. 3: Results of running Wilcoxon and t-tests between satellite and ground data.

Note that there are two distinct peaks well beyond the confidence intervals and that both lines spend much greater than 5% of the time outside the limits. There is, without a doubt, a statistically significant difference between the satellite data and the ground data.

As a sidebar, the Wilcoxon test is a non-parametric test. It does not require correction for autocorrelation of the residuals when calculating confidence intervals. The fact that it differs from the t-test results indicates that the residuals are not normally distributed and/or the residuals are not free from correlation. This is why it is important to correct for autocorrelation when using tests that rely on assumptions of normality and uncorrelated residuals. Alternatively, you could simply use non-parametric tests, and though they often have less statistical power, I’ve found the Wilcoxon test to be pretty good for most temperature analyses.

Here’s what the difference plot looks like with the satellite periods shown:

Fig_4
Fig. 4: Difference plot, satellite periods shown.

The downward trend during NOAA-7 is apparent, as is the strange drop in NOAA-11. NOAA-14 is visibly too high, and NOAA-16 and -17 display some strange upward spikes. Overall, though, NOAA-16 and -17 do not show a statistically significant difference from the ground data, so no correction was applied to them.

After having confirmed that other researchers had noted similar issues, I felt comfortable in performing a calibration of the AVHRR data to the ground data. The calculated offsets and the resulting Wilcoxon and t-test plot are next:

Fig_5
Fig. 5: Calculated offsets.
Fig_6
Fig. 6: Post-calibration Wilcoxon and t-tests

To make sure that I did not “over-modify” the data, I ran a Steig (3 PC, regpar=3, 42 ground stations) reconstruction. The resulting trend was 0.1079 deg C/decade and the trend maps looked nearly identical to the Steig reconstructions. Therefore, the satellite offsets – while they do produce a greater trend when not corrected – do not seem to have a major impact on the Steig result. This should not be surprising, as most of the temperature rise in Antarctica occurs between 1957 and 1970.

II: PCA

One of the items that we’ve spent a lot of time doing sensitivity analysis is the PCA of the AVHRR data. Between Jeff Id, Jeff C, and myself, we’ve performed somewhere north of 200 reconstructions using different methods and different numbers of retained PCs. Based on that, I believe that we have a pretty good feel for the ranges of values that the reconstructions produce, and we all feel that the 3 PC, regpar=3 solution does not accurately reproduce Antarctic temperatures. Unfortunately, our opinions count for very little. We must have a solid basis for concluding that Steig’s choices were less than optimal – not just opinions.

How many PCs to retain for an analysis has been the subject of much debate in many fields. I will quickly summarize some of the major stopping rules:

1. Kaiser-Guttman: Include all PCs with eigenvalues greater than the average eigenvalue. In this case, this would require retention of 73 PCs.

2. Scree Analysis: Plot the eigenvalues from largest to smallest and take all PCs where the slope of the line visibly ticks up. This is subjective, and in this case it would require the retention of 25 – 50 PCs.

3. Minimum explained variance: Retain PCs until some preset amount of variance has been explained. This preset amount is arbitrary, and different people have selected anywhere from 80-95%. This would justify including as few as 14 PCs and as many as 100.

4. Broken stick analysis: Retain PCs that exceed the theoretical scree plot of random, uncorrelated noise. This yields precisely 11 PCs.

5. Bootstrapped eigenvalue and eigenvalue/eigenvector: Through iterative random sampling of either the PCA matrix or the original data matrix, retain PCs that are statistically different from PCs containing only noise. I have not yet done this for the AVHRR data, though the bootstrap analysis typically yields about the same number (or a slightly greater number) of significant PCs as broken stick.

The first 3 rules are widely criticized for being either subjective or retaining too many PCs. In the Jackson article below, a comparison is made showing that 1, 2, and 3 will select “significant” PCs out of matrices populated entirely with uncorrelated noise. There is no reason to retain noise, and the more PCs you retain, the more difficult and cumbersome the analysis becomes.

The last 2 rules have statistical justification. And, not surprisingly, they are much more effective at distinguishing truly significant PCs from noise. The broken stick analysis typically yields the fewest number of significant PCs, but is normally very comparable to the more robust bootstrap method.

Note that all of these rules would indicate retaining far more than simply 3 PCs. I have included some references:

Click to access pca.pdf

Click to access North_et_al_1982_EOF_error_MWR.pdf

I have not yet had time to modify a bootstrapping algorithm I found (it was written for a much older version of R), but when I finish that, I will show the bootstrap results. For now, I will simply present the broken stick analysis results.

Fig_7

Fig. 7: Broken Stick Analysis on AVHRR data.

The broken stick analysis finds 11 significant PCs. PCs 12 and 13 are also very close, and I suspect the bootstrap test will find that they are significant. I chose to retain 13 PCs for the reconstruction to follow.

Without presenting plots for the moment, retaining more than 11 PCs does not end up affecting the results much at all. The trend does drop slightly, but this is due to better resolution on the Peninsula warming. The rest of the continent does not change if additional PCs are added. The only thing that changes is the time it takes to do the reconstruction.

Remember that the purpose of the PCA on the AVHRR data is not to perform factor analysis. The purpose is simply to reduce the size of the data to something that can be computed. The penalty for retaining “too many” – in this case – is simply computational time or the inability for RegEM to converge. The penalty for retaining too few, on the other hand, is a faulty analysis.

I do not see how the choice of 3 PCs can be justified on either practical or theoretical grounds. On the practical side, RegEM works just fine with as many as 25 PCs. On the theoretical side, none of the stopping criteria yield anything close to 3. Not only that, but these are empirical functions. They have no direct physical meaning. Despite claims in Steig et al. to the contrary, they do not relate to physical processes in Antarctica – at least not directly. Therefore, there is no justification for excluding PCs that show significance simply because the other ones “look” like physical processes. This latter bit is a whole other discussion that’s probably post worthy at some point, but I’ll leave it there for now.

III: RegEM

We’ve also spent a great deal of time on RegEM. Steig & Co. used a regpar setting of 3. Was that the “right” setting? They do not present any justification, but that does not necessarily mean the choice is wrong. Fortunately, there is a way to decide.

RegEM works by approximating the actual data with a certain number of principal components and estimating a covariance from which missing data is predicted. Each iteration improves the prediction. In this case (unlike the AVHRR data), selecting too many can be detrimental to the analysis as it can result in over-fitting, spurious correlations between stations and PCs that only represent noise, and retention of the initial infill of zeros. On the other hand, just like the AVHRR data, too few will result in throwing away important information about station and PC covariance.

Figuring out how many PCs (i.e., what regpar setting to use) is a bit trickier because most of the data is missing. Like RegEM itself, this problem needs to be approached iteratively.

The first step was to substitute AVHRR data for station data, calculate the PCs, and perform the broken stick analysis. This yielded 4 or 5 significant PCs. After that, I performed reconstructions with steadily increasing numbers of PCs and performed a broken stick analysis on each one. Once the regpar setting is high enough to begin including insignificant PCs, the broken stick analysis yields the same result every time. The extra PCs show up in the analysis as noise. I first did this using all the AWS and manned stations (minus the open ocean stations).

Fig_8

Fig. 8: Broken stick analysis on manned and AWS stations, regpar = 8.

Fig_9

Fig. 9: Broken stick analysis on manned and AWS stations, regpar=12.

I ran this all the way up to regpar=20 and the broken stick analysis indicates that 9 PCs are required to properly describe the station covariance. Hence the appropriate regpar setting is 9 if all the manned and AWS stations are used. It is certainly not 3, which is what Steig used for the AWS recon.

I also performed this for the 42 manned stations Steig selected for the main reconstruction. That analysis yielded a regpar setting of 6 – again, not 3.

The conclusion, then, is similar to the AVHRR PC analysis. The selection of regpar=3 does not appear to be justifiable. Additional PCs are necessary to properly describe the covariance.

IV: THE RECONSTRUCTION

So what happens if the satellite offsets are properly accounted for, the correct number of PCs are retained, and the right regpar settings are used? I present the following panel:

Fig_10

Fig. 10: (Left side) Reconstruction trends with the post-1982 PCs spliced back in (Steig’s method).

(Right side) Reconstruction trends using just the model frame.

RegEM PTTLS does not return the entire best-fit solution (the model frame, or surface). It only returns what the best-fit solution says the missing points are. It retains the original points. When imputing small amounts of data, this is fine. When imputing large amounts of data, it can be argued that the surface is what is important.

RegEM IPCA returns the surface (along with the spliced solution). This allows you to see the entire solution. In my opinion, in this particular case, the reconstruction should be based on the solution, not a partial solution with data tacked on the end. That is akin to doing a linear regression, throwing away the last half of the regression, adding the data back in, and then doing another linear regression on the result to get the trend. The discontinuity between the model and the data causes errors in the computed trend.

Regardless, the verification statistics are computed vs. the model – not the spliced data – and though Steig did not do this for his paper, we can do it ourselves. (I will do this in a later post.) Besides, the trends between the model and the spliced reconstructions are not that different.

Overall trends are 0.071 deg C/decade for the spliced reconstruction and 0.060 deg C/decade for the model frame. This is comparable to Jeff’s reconstructions using just the ground data, and as you can see, the temperature distribution of the model frame is closer to that of the ground stations. This is another indication that the satellites and the ground stations are not measuring exactly the same thing. It is close, but not exact, and splicing PCs derived solely from satellite data on a reconstruction where the only actual temperatures come from ground data is conceptually suspect.

When I ran the same settings in RegEM PTTLS – which only returns a spliced version – I got 0.077 deg C/decade, which checks nicely with RegEM IPCA.

I also did 11 PC, 15 PC, and 20 PC reconstructions. Trends were 0.081, 0.071, and 0.069 for the spliced and 0.072, 0.059, and 0.055 for the model. The reason for the reduction in trend was simply better resolution (less smearing) of the Peninsula warming.

Additionally, I ran reconstructions using just Steig’s station selection. With 13 PCs, this yielded a spliced trend of 0.080 and a model trend of 0.065. I then did one after removing the open-ocean stations, which yielded 0.080 and 0.064.

Note how when the PCs and regpar are properly selected, the inclusion and exclusion of individual stations does not significantly affect the result. The answers are nearly identical whether 98 AWS/manned stations are used, or only 37 manned stations are used. One might be tempted to call this “robust”.

V: THE COUP DE GRACE

Let us assume for a moment that the reconstruction presented above represents the real 50-year temperature history of Antarctica. Whether this is true is immaterial. We will assume it to be true for the moment. If Steig’s method has validity, then, if we substitute the above reconstruction for the raw ground and AVHRR data, his method should return a result that looks similar to the above reconstruction.

Let’s see if that happens.

For the substitution, I took the ground station model frame (which does not have any actual ground data spliced back in) and removed the same exact points that are missing from the real data.

I then took the post-1982 model frame (so the one with the lowest trend) and substituted that for the AVHRR data.

I set the number of PCs equal to 3.

I set regpar equal to 3 in PTTLS.

I let it rip.

Fig_11

Fig. 11: Steig-style reconstruction using data from the 13 PC, regpar=9 reconstruction.

Look familiar?

Overall trend: 0.102 deg C/decade.

Remember that the input data had a trend of 0.060 deg C/decade, showed cooling on the Ross and Weddel ice shelves, showed cooling near the pole, and showed a maximum trend in the Peninsula.

If “robust” means the same answer pops out of a fancy computer algorithm regardless of what the input data is, then I guess Antarctic warming is, indeed, “robust”.

———————————————

Code for the above post is HERE.

Get notified when a new post is published.
Subscribe today!
0 0 votes
Article Rating
168 Comments
Inline Feedbacks
View all comments
Pamela Gray
May 21, 2009 7:08 am

So what you are saying is that if you put in the temperature of bananas as they rot in your kitchen as a proxy for antarctic temperature, you get antarctic warming? And just to be sure I get this, if you put in the temperature of your son’s socks as they warm through the day, you would also get antarctic warming? Does this mean that the part of the code that Stulag, or whatever his name is, hasn’t published could contain an algorithm that takes whatever you put in, and adjusts it to his predicted antarctic warming? There are math tricks that do this. You plug in a number (any number, and even a banana if you want to) and the equation always equals the same thing.

Pamela Gray
May 21, 2009 7:09 am

Steig et al. My bad.

Basil
Editor
May 21, 2009 7:10 am

Claude Harvey (05:00:25) :
For the vast majority of the voting public, any attempt to judge the validity of any of this would be like serving on a jury in a trial where all the attorneys and their witnesses spoke only Greek. The jury would be left searching for clues on the expressions of the witnesses’ faces. In such a case, the jury would likely end up simply believing whatever they chose to believe regardless of the testimony (”We liked his tie!”).
Unless this paper can gain an official seal of authenticity in some quarter recognized by the public (The Goracle has his Nobel Prize), it is just so much spit in the ocean from a public perspective.
That said, I applaud this mountain of work for its own sake in the pursuit of scientific truth.

—————————-
I think the Internet is changing the rules of the game here. Efforts like this, as Adam Soereg (04:57:45) argued, show that peer review –in the traditional sense — is not working, at least not in the field of climate science. The Internet is giving rise to a different kind of “peer review,” one that is more robust, and less dominated by established cliques of authority. And that makes it truly more scientific, because science at its best is inherently anti-authoritarian.
The analogy to serving on a jury here is quite apt. In the law, juries are “lay people.” They are not experts at anything, and are not expected to be experts at anything. The experts that present testimony are “qualified” through “voir dire” before they are allowed to present their testimony. Yet, in almost every case where expert testimony is presented (I’m sure exceptions occur), there are experts on both sides of the issue. And if the testimony and evidence of the experts is such that a lay jury cannot tell the difference, or say one is likely more right than the other, then the outcome of the case should depend on something other than the expert testimony.
The difference between the courtroom analogy and modern science, or at least modern climate science, is that the cliques of authority are acting like judges who allow only one side of a dispute to present their evidence. But the issues are no longer being tried in their courtroom. It is their own fault, really, but the issues — here I’m thinking specifically of issues relating to climate change and AGW — are being tried in the court of public opinion. And just as the Internet is showing that traditional bastions of public opinion — MSM like the NYT, Times magazine, etc. — are no longer in control of the courtroom of public opinion, the same is happening with respect to politicized “science.”
The antics of traditional bastions of authority in science, like Science and Nature, are now increasingly obvious through the power of the Internet. Just as the NYT is losing influence, so are they. The “it was peer reviewed” or “it was not peer reviewed” is becoming recognized as pure and simple appeal to authority, and not indicative of any intrinsic scientific value. The public is waking up to the fact that “peer review” is now a lot like a court room where the judge (the editors of a paper) only lets one side of a dispute qualify their expert.
As long as this continues, the influence of traditional centers of authority in science will continue to erode. There will be further development of “open source” publishing in science — cf. http://www.doaj.org/ — precisely because the traditional venues are now less about promoting free and open inquiry than they are about shaping public opinion.

May 21, 2009 7:12 am

John Silver
A little while ago I researched the credentials of those that awarded Al Gore the Nobel prize. Earnest and clever people that they undoubtedly are, it was very hard to see this as being anything other than a political gesture bearing in mind the green credentials of most of them
Tonyb

Ryan O
May 21, 2009 7:14 am

Thanks to all the comments about this; I appreciate them. There are a few points that I would like to make.
1. In the course of doing this, I had several occasions where I needed to contact Steig for clarification on his methods. I personally did not ever ask for code, as I felt that (once we had a working version of RegEM) there would be much more learned by me in attempting to replicate his work from the methods description. Regardless, in every case, he was courteous and professional. On more than one occasion he freely offered to send me intermediate data not published on his website. So the personal comments about Dr. Steig I feel are unjustified – all of my interactions with him were positive.
2. Steig’s choice of 3 PCs is probably based largely on previous work that he had done with Schneider (such as attempting to find circulation patterns using PCA and doing analyses with the 37GHz satellite information). For those specific studies, 2-3 PCs were all that were needed to describe their results. This led them to believe that the temperature field could also be described by 3 PCs. I would strongly discourage the use of “fraud” or any other related terms, as I can see how their past work would have led them to this particular conclusion. I would say, rather, that the Antarctic warming paper suffered from confirmation bias, not fraud or anything close to fraud. I feel that they should have done more sensitivity studies to see if 3 PCs was a correct choice, but I would stop far short of calling this fraud.
3. RegEM and PCA, when used properly, can greatly enhance our ability to analyze sparse data accurately. They are not magic hockey-stick making algorithms. They are legitimate tools for scientific investigation. It is incumbent upon the researcher to use them properly and understand their limitations. In the Antarctic case, I feel that this was not done appropriately, but I personally would chalk that up to confirmation bias, not fraud.

Frank K.
May 21, 2009 7:27 am

John Galt (06:15:34) :
OT: Global warming may be twice as bad as previously expected
By Doyle Rice, USA TODAY
Global warming will be twice as severe as previous estimates indicate, according to a new study published this month in the Journal of Climate, a publication of the American Meteorological Society.
http://www.usatoday.com/tech/science/environment/2009-05-20-global-warming_N.htm
Because the computer models tell them so!

Oh brother!!! This is why I will never purchase a USA Today paper! Check out the same news release from this site:
http://www.sciencedaily.com/releases/2009/05/090519134843.htm
Here, buried near the end of the article, is all you need to know about this junk science:
Prinn stresses that the computer models are built to match the known conditions, processes and past history of the relevant human and natural systems, and the researchers are therefore dependent on the accuracy of this current knowledge. Beyond this, “we do the research, and let the results fall where they may,” he says. Since there are so many uncertainties, especially with regard to what human beings will choose to do and how large the climate response will be, “we don’t pretend we can do it accurately. Instead, we do these 400 runs and look at the spread of the odds.”

The money quote from above:
“…we don’t pretend we can do it accurately.”
Now, where was this in the USA Today article??
This, by the way, is a textbook example of how Global Warming stories in the media go from “…we don’t pretend we can do it accurately…” to highly accurate projections with near 100% certainty.
Unbelievable!!

hunter
May 21, 2009 7:30 am

Flanagan,
The Antarctic is not warming. It has not warmed in the past decades.
The interesting thing for AGW believers is this:
When it became obvious Antarctica was cooling, AGW promoters claimed this was *proof* of the looming apocalypse.
Now, by way of undisclosed techniques, AGW promoters claim that the Antarctic (ignoring that it has been cooling for over 20 years) has warmed (in only one small area) and that this is now *proof* that the apocalypse is looming.
So the problem for AGW believers like you is this: how do you maintain your credulity in something that is so patently contrived?

May 21, 2009 7:35 am

John Galt (06:05:34) :
Can somebody tell me what exactly is the definition of academic fraud and why Steig’s work does not qualify?
It is only fraud if there was fraudulent intent, and that has not been shown to be the case.

Richard Sharpe
May 21, 2009 7:38 am

Flanagan says:

All I see is tha any type of reconstruction, with all corrections you want, still give a warming Antarctica. When I think how many time I’ve read here and there that this conclusion was “obviously false” or “proved that data were manipulated”, I seriously wonder how the final result could be turned into a “victory” of some sort… A “coup de grace”? For skeptics?

Tell me, how useful is this function?
Flanagan(X) = { for all x, Flanagan(X) = [snip] }

Peter
May 21, 2009 7:40 am

I sent a quick summary and a link to this post to Nature.

Vg
May 21, 2009 7:45 am

Chris H : @vg Should this work/analysis not then be submitted as a letter to Nature?

John Galt
May 21, 2009 7:58 am

@Flanagan
The Steig paper has been heavily criticized by scientists, even from those known to support the AGW hypothesis.
Essentially, Steig created data where none existed and won’t release how the results were obtained. Is the computer code some sort of trade secret? Maybe Steig is planning to file for a software patent of his intellectual property. But until he release his raw data, the adjusted data, the fully documented source code, this falls under the heading of junk science.
When I was undergrad, a paper like this would not get a passing grade. But then again, I’m not a climate scientist.

John Boy
May 21, 2009 8:04 am

He’s back!
Eric Steig
Former SIL Post Doc
Faculty, Quaternary Research Center, Univ. of Washington; Associate Professor, Earth & Space Sciences, University of Washington
Ph.D., Geological Sciences, Univ. of Washington 1996
steig “AT” ess.washington.edu <== email
(206) 685-3715 (206) 543-6327
The link:
http://instaar.colorado.edu/sil/people/person_detail.php?person_ID=17
The Boy of John

John Boy
May 21, 2009 8:06 am

Another email for Steig
steig@ess.washington.edu
He’s taking questions.
The Boy of John

John Trigge
May 21, 2009 8:21 am

If Steig is to be pilloried for poor science, so to should the peer reviewers of his article.
Perhaps the work of all persons approving or involved in falsified studies should also be double-checked in order to reveal if they have created a group that merely perpetuates their own pre-conceived ideas.
There was an analysis performed on the papers for the IPCC 2007 report (I think) that demonstrated the close relationship of many persons who were ‘peer reviewers’ of each other’s work, a sure sign that they could be self-serving.

May 21, 2009 8:26 am

I see Ryan’s comment above and felt like adding my two cents.
I also have no intention of calling this fraud but am somewhat more suspicious. As you know I am completely unwelcome over at RC due to my appropriately toned discussions about the intent behind the M08 hockey stick. When I requested code politely from Dr. Steig, it did not receive a warm welcome or a professional response, probably due to my assigned rather than deserved reputation. Imagine the welcome some of the other fact checking blog hosts receive at RC.
Since that time when I started to replicate some of the work in the paper, Dr. Steig has been polite to my email correspondence although he has so far refused to release his own code which I’m sure has details we would be interested in.
In the meantime, Ryan is correct in that this is possibly just confirmation bias. However, from my experience with the hockey stick and the fact that a different number entered in the algorithm (any regpar number or PC number different than 3) will produce reduced trends, I am reasonably suspicious.
I did a summary of the reconstructions including different techniques recently which lays out the trends from hundreds of different methods. It was nearly impossible to produce a higher trend than Steig et al.
http://noconsensus.wordpress.com/2009/05/14/antarcti-summary-part-1-a-trend-of-trends/
The paper didn’t discuss trying any numbers other than regpar=3 but I don’t believe they didn’t try them and then make a conscious decision not to use them for minimally discussed and apparently incorrect reasons.

hunter
May 21, 2009 8:32 am

Frank K.,
It would be interesting to study how many articles have been publishedinthe past 5- 10 years that claim to *prove* that “AGW is twice as bad as we as we thought it was going to be”.
My brief google of the search term showed 248,000 hits. A survey of the first few pages showed articles dated from 2001 with the phrase “Global Warming Worse Than Predicted”.
I think the AGW promotion industry has been running this one up the flag pole uncahllenged far too long.
But heck, AGW promoters are still getting away with claiming that tropical cyclones have gotten worse and that the world is much hotter.

hunter
May 21, 2009 8:34 am

John Trigge,
Pielke, Sr. has pointed out this very thing in his most recent posting. I urge you to read it and to follow the .ppt link and read that, as well.
It is very revealing.
AGW depends on an echo chamber of promoters and beleivers constantly repeating, and never defending, certain basic ideas.
The conflicts of interest between reviewers, writers, profiteers, and media is strking.

John W.
May 21, 2009 8:41 am

Ryan O (07:14:49) :
Thanks to all the comments about this; I appreciate them. There are a few points that I would like to make.

You make three points attesting to Dr. Steig’s professionalism. Thank you, and thanks and kudos to him. Being “in error” isn’t the same as being fraudulent” or “being evil” I agree with your point that, in my own words, he made an understandable mistake in his methodology. We should all refrain from the ad homs.
I think it would be a good thing if he were to join in the discussion here.

Barry Foster
May 21, 2009 9:09 am

BBC Radio 4 to run a programme on “The first victims of climate change” on May 25th at 21.00. ‘The Carteret Islands – Sharks in the garden’ will undoubtedly blame man-made climate change, even though islanders have been using dynamite to blow up the surrounding reefs to kill fish – thereby removing natural barriers.

Ryan O
May 21, 2009 9:13 am

John W. (08:41:16) :
Thank you.
I would also like to point out that I would take my reconstruction with a grain of salt. I personally would not make the claim that my reconstruction represents the Truth as far as Antarctic temperatures go. The only thing I would be willing to say is that it might be true. I have not properly quantified the uncertainty in the analysis, and there are many. The calculated satellite offsets have uncertainties. The imputation of the ground station temperatures has uncertainty. The PCA on the AVHRR data has uncertainty. The method Comiso used for cloudmasking has uncertainty.
This, honestly, is my biggest criticism of Steig’s work. These uncertainties are never quantified and, for the ones that are actually mentioned in the papers, they receive only passing attention. The uncertainty in the linear trends reported are simply the +/- 95% confidence intervals (and they are not appropriately corrected for autocorrelation) of the linear regression. They assume that the satellite data is perfect, that the PC decomposition of the satellite data is perfect, that the RegEM imputation is perfect, and that the covariance between the stations is perfect and is constant in time.
Personally, I would like to see much better quantification of uncertainty in papers of this type. It helps put the conclusions in perspective.

Mark T
May 21, 2009 9:23 am

Flanagan (06:15:54) :
All I see is tha any type of reconstruction, with all corrections you want, still give a warming Antarctica.

That’s because you do not understand what these results mean. That’s OK, of course. Nobody here believes you would answer any other way, even if you did understand.
Mark

May 21, 2009 9:28 am

I have copied a post I made a few days ago about fraud in science. I do not think the IPCC set out to be fraudulent at all, but the points made in the article as to how everyone goes along with the flow are well worth reading in order to understand how we have got to this state.
This article below says it all. Frauds (or misunderstanding) peer jealousy, the need to gather more money and prestige for your dept are all major drivers in science these days-much bigger climate drivers than co2 is.
http://www.telegraph.co.uk/scienceandtechnology/5345963/The-scientific-fraudster-who-dazzled-the-world-of-physics.html
“Schön’s fraud was the largest ever exposed in physics; he ended up without a job, and was forced to leave America in disgrace. But the ease with which his fraudulent findings and grotesque errors were accepted by his peers raises troubling questions about the way in which scientists assess each other’s work, and whether there might be other such cases out there.”
We are being dazzled by reputations, unproven theories and computer models (which even the IPCC admit are flawed) whilst we set aside history and observational evidence. Yet still some believe everything they are told. I’m going to have a Mencken moment here…

Ron de Haan
May 21, 2009 9:29 am

The Climate Industrial Complex which would be non existing without the AGW HOAX:
http://online.wsj.com/article/SB124286145192740987.html

mondo
May 21, 2009 9:34 am

The test of Eric Steig’s bona fides will be how he reacts to these developments.
He could:
1. Not respond at all: This would in effect cede the game to Ryan O, Jeff ID et al
2. Respond, and defend his paper, explaining why Ryan O, Jeff ID et al are incorrect.
3. Acknowledge the work done by Ryan O, Jeff ID et al, and accept that their work enhances and refines his own work, and engage with them in producing an updated version of his paper that sets the record straight.
If Eric Steig chooses not to respond at all, it seems likely that his professional reputation will be challenged, whereas he can preserve his professional reputation by taking actions 2 or 3. I wonder which course he will choose?