Steig's Antarctic Heartburn

flaming-hot-antarctic-penguin

Art courtesy Dave Stephens

Foreword by Anthony Watts: This article, written by the two Jeffs (Jeff C and Jeff Id) is one of the more technically complex essays ever presented on WUWT. It has been several days in the making. One of the goals I have with WUWT is to make sometimes difficult to understand science understandable to a wider audience. In this case the statistical analysis is rather difficult for the layman to comprehend, but I asked for (and got) an essay that was explained in terms I think many can grasp and understand. That being said, it is a long article, and you may have to read it more than once to fully grasp what has been presented here. Steve McIntyre of Climate Audit laid much of the ground work for this essay, and from his work as well as this essay, it is becoming clearer that Steig et al (see “Warming of the Antarctic ice-sheet surface since the 1957 International Geophysical Year”, Nature, Jan 22, 2009) isn’t holding up well to rigorous tests as demonstrated by McIntyre as well as in the essay below. Unfortunately, Steig’s office has so far deferred (several requests) to provide the complete data sets needed to replicate and test his paper, and has left on a trip to Antarctica and the remaining data is not “expected” to be available until his return.

To help layman readers understand the terminology used, here is a mini-glossary in advance:

RegEM – Regularized Expectation Maximization

PCA – Principal Components Analysis

PC – Principal Components

AWS – Automatic Weather Stations

One of the more difficult concepts is RegEM, an algorithm developed by Tapio Schneider in 2001.   It’s a form of expectation maximization (EM) which is a common and well understood method for infilling missing data. As we’ve previously noted on WUWT, many of the weather stations used in the Steig et al study had issues with being buried by snow, causing significant data gaps in the Antarctic record and in some burial cases stations have been accidentally lost or confused with others at different lat/lons. Then of course there is the problem of coming up with trends for the entire Antarctic continent when most of the weather station data is from the periphery and the penisula, with very little data from the interior.

Expectation Maximization is a method which uses a normal distribution to compute the best probability of fit to a missing piece of data.  Regularization is required when so much data is missing that the EM method won’t solve.  That makes it a statistically dangerous technique to use and as Kevin Trenberth, climate analysis chief at the National Center for Atmospheric Research, said in an e-mail: “It is hard to make data where none exist.” (Source: MSNBC article) It is also valuable to note that one of the co-authors of Steig et al, Dr. Michael Mann, dabbles quite a bit in RegEm in this preparatory paper to Mann et al 2008 “Return of the Hockey Stick”.

For those that prefer to print and read, I’ve made a PDF file of this article available here.

Introduction

This article is an attempt to describe some of the early results from the Antarctic reconstruction recently published on the cover of Nature which demonstrated a warming trend in the Antarctic since 1956.   Actual surface temperatures in the Antarctic are hard to come by with only about 30 stations prior to 1980 recorded through tedious and difficult efforts by scientists in the region.  In the 80’s more stations were added including some automatic weather stations (AWS) which sit in remote areas and report the temperature information automatically.  Unfortunately due to the harsh conditions in the region many of these stations have gaps in their records or very short reporting times (a few years in some cases).  Very few stations are located in the interior of the Antarctic, leaving the trend for the central portion of the continent relatively unknown.  The location of the stations is shown on the map below.

2jeffs-steig-image1

In addition to the stations there are satellite data from an infrared surface temperature measurement which records the temperature of the actual emission from the surface of the ice/ground in the Antarctic.  This is different from the microwave absorption measurements as made from UAH/RSS data which measure temperatures in a thickness of the atmosphere.  This dataset didn’t start until 1982.

Steig 09 is an attempt to reconstruct the continent-wide temperatures using a combination of measurements from the surface stations shown above and the post-1982 satellite data.  The complex math behind the paper is an attempt to ‘paste’ the 30ish pre-1982 real surface station measurements onto 5509 individual gridcells from the satellite data.  An engineer or vision system designer could use several straightforward methods which would insure reasonable distribution of the trends across the grid based on a huge variety of area weighting algorithms, the accuracy of any of the methods would depend on the amount of data available.  These well understood methods were ignored in Steig09 in favor of RegEM.

The use of Principal Component Analysis in the reconstruction

Steig 09 presents the satellite reconstructions as the trend and also provides an AWS reconstruction as verification of the satellite data rather than a separate stand alone result presumably due to the sparseness of the actual data.  An algorithm called RegEM was used for infilling the missing data. Missing data includes pre 1982 for satellites and all years for the very sparse AWS data.  While Dr. Steig has provided the reconstructions to the public, he has declined to provide any of the satellite, station or AWS temperature measurements used as inputs to the RegEM algorithm.  Since the station and AWS measurements were available through other sources, this paper focuses on the AWS reconstruction.

Without getting into the detail of PCA analysis, the algorithm uses covariance to assign weighting of a pattern in the data and does not have any input whatsoever for actual station location.  In other words, the algorithm has no knowledge of the distance between stations and must infill missing data based solely on the correlation with other data sets.  This means there is a possibility that with improper or incomplete checks, a trend from the peninsula on the west coast could be applied all the way to the east.  The only control is the correlation of one temperature measurement to another.

If you were an engineer concerned with the quality of your result, you would recognize the possibility of accidental mismatch and do a reasonable amount of checking to insure that the stations were properly assigned after infilling.  Steig et. al. described no attempts to check this basic potential problem with RegEM analysis.  This paper will describe a simple method we used to determine that the AWS reconstruction is rife with spurious (i.e. appear real but really aren’t) correlations attributed to the methods used by Dr. Steig.  These spurious correlations can take a localized climactic pattern and “smear” it over a large region that lacks adequate data of its own.

Now is where it becomes a little tricky.  RegEM uses a reduced information dataset to infill the missing values.  The dataset is reduced by Principal Component Analysis (PCA) replacing each trend with a similar looking one which is used for covariance analysis.  Think of it like a data compression algorithm for a picture which uses less computer memory than the actual but results in a fuzzier image for higher compression levels.

2jeffs-steig-image2

While the second image is still visible, the actual data used to represent the image is reduced considerably.  This will work fine for pictures with reasonable compression, but the data from some pixels has blended into others.  Steig 09 uses 3 trends to represent all of the data in the Antarctic.  In it’s full complexity using 3 PC’s is analogous to representing not just a picture but actually a movie of the Antarctic with three color ‘trends’ where the color of each pixel changes according to different weights of the same red, green and blue color trends (PC’s).  With enough PC’s the movie could be replicated perfectly with no loss.  Here’s an important quote from the paper.

“We therefore used the RegEM algorithm with a cut-off parameter K=3. A disadvantage of excluding higher-order terms (k>3) is that this fails to fully capture the variance in the Antarctic Peninsula region.  We accept this tradeoff because the Peninsula is already the best-observed region of the Antarctic.”

http://www.climateaudit.org/wp-content/uploads/2009/02/regpar9.gif

Above: a graph from Steve McIntyre of ClimateAudit where he demonstrates how “K=3 was in fact a fortuitous choice, as this proved to yield the maximum AWS trend, something that will, I’m sure, astonish most CA readers.

K=3 means only 3 trends were used, the ‘lack of captured variance’ is an acknowledgement and acceptance of the fuzziness of the image.  It’s easy to imagine that it would be difficult to represent a complex movie image of Antarctic with any sharpness from 1957 to 2006 temperature with the same 3 color trends reweighted for every pixel.  In the satellite version of the Antarctic movie the three trends look like this.

2jeffs-steig-image3

Note that the sudden step in the 3rd trend would cause a jump in the ‘temperature’ of the entire movie.  This represents the temperature change between the pre 1982 recreated data and the after 1982 real data in the satellite reconstruction.  This is a strong yet overlooked hint that something may not be right with the result.

In the case of the AWS reconstruction we have only 63 AWS stations to make the movie screen, by which the trends of 42 surface station points are used to infill the remaining data.  If the data from one surface station is copied to the wrong AWS stations the average will overweight and underweight some trends. So the question becomes, is the compression level too high?

The problems that arise when using too few principal components

Fortunately, we’re here to help in this matter.  Steve McIntyre again provided the answer with a simple plot of the actual surface station data correlation with distance.  This correlation plot compares the similarities ‘correlation’ of each temperature station with all of the 41 other manual surface stations against the distance between them.  A correlation of 1 means the data from one station is exactly equal to the other.  Because A -> B correlation isn’t a perfect match for B->A there are 42*42 separate points in the graph.  This first scatter plot is from measured temperature data prior to any infilling of missing measurements.  Station to station distance is shown on the X axis.  The correlation coefficient is shown on the Y axis.

2jeffs-steig-image4

Since this plot above represents the only real data we have existing back to 1957, it demonstrates the expected ‘natural’ spatial relationship from any properly controlled RegEM analysis.  The correlation drops with distance which we would expect because temps from stations thousands of miles away should be less related than those next to each other.  (Note that there are a few stations that show a positive correlation beyond 6000 km.  These are entirely from non-continental northern islands inexplicably used by Steig in the reconstruction.  No continental stations exhibit positive correlations at these distances.)  If RegEM works, the reconstructed RegEM imputed (infilled) data correlation vs. distance should have a very similar pattern to the real data.  Here’s a graph of the AWS reconstruction with infilled temperature values.2jeffs-steig-image5

Compare this plot with the previous plot from actual measured temperatures.  Now contrast that with the AWS plot above.  The infilled AWS reconstruction has no clearly evident pattern of decay over distance.  In fact, many of the stations show a correlation of close to 1 for stations at 3000 km distant!  The measured station data is our best indicator of true Antarctic trends and it shows no sign that these long distance correlations occur.  Of course, common sense should also make one suspicious of these long distance correlations as they would be comparable to data that indicated Los Angeles and Chicago had closely correlated climate.

It was earlier mentioned that the use of 3 PCs was analogous to the loss of detail that occurs in data compressions.   Since the AWS input data is available, it is possible to regenerate the AWS reconstruction using a higher number of PCs.  It stood to reason that spurious correlations could be reduced by retaining the spatial detail lost in the 3 PC reconstruction.  Using RegEM, we generated a new AWS reconstruction using the same input data but with 7 PCs.  The distance correlations are shown in the plot below.

2jeffs-steig-image6

Note the dramatic improvement over that shown in the previous plot.  The correlation decay with distance so clearly seen in the measured station temperature data has returned.  While the cone of the RegEM data is slightly wider than the ‘real’ surface station data, the counterintuitive long distance correlations seen in the Steig reconstruction have completely disappeared.  It seems clear that limiting the reconstruction to 3 PCs resulted in numerous spurious correlations when infilling missing station data.

Using only 3 principal components distorts temperature trends

If Antarctica had uniform temperature trends across the continent, the spurious correlations might not have a large impact in the overall reconstruction.  Individual sites may have some errors, but the overall trend would be reasonably close.  However, Antarctica is anything but uniform.  The spurious correlations can allow unique climactic trends from a localized region to be spread over a larger area, particularly if an area lacks detailed climate records of its own.  It is our conclusion is that is exactly what is happening with the Steig AWS reconstruction.

Consider the case of the Antarctic Peninsula:

  • The peninsula is geographically isolated from the rest of the continent
  • The peninsula is less than 5% of the total continental land mass
  • The peninsula is known to be warming at a rate much higher than anywhere else in Antarctica
  • The peninsula is bordered by a vast area known as West Antarctica that has extremely limited temperature records of its own
  • 15 of the 42 temperature surface stations (35%) used in the reconstruction are located on the peninsula

If the Steig AWS reconstruction was properly correlating the peninsula stations temperature measurements to the AWS sites, you would expect to see the highest rates of warming at the peninsula extremes.  This is the pattern seen in the measured station data.  The plot below shows the temperature trends for the reconstructed AWS sites for the period of 1980 to 2006.  This time frame has been selected as this is the period when AWS data exists.  Prior to 1980, 100% of AWS reconstructed data is artificial (i.e. infilled by RegEM).

2jeffs-steig-image7

Note how warming extends beyond the peninsula extremes down toward West Antarctica and the South Pole.  Also note the relatively moderate cooling in the vicinity of the Ross Ice Shelf (bottom of the plot).  The warming once thought to be limited to the peninsula appears to have spread.  This “smearing” of the peninsula warming has also moderated the cooling of the Ross Ice Shelf AWS measurements.  These are both artifacts of limiting the reconstruction to 3 PCs.

Now compare the above plot to the new AWS reconstruction using 7 PCs.

2jeffs-steig-image8

The difference is striking.  The peninsula has become warmer and warming is largely limited to its confines.  West Antarctica and the Ross Ice Shelf area have become noticeably cooler.  This agrees with the commonly-held belief prior to Steig’s paper that the peninsula is warming, the rest of Antarctica is not.

Temperature trends using more traditional methods

In providing a continental trend for Antarctica warming, Steig used a simple average of the 63 AWS reconstructed time series.  As can be seen in the plots above, the AWS stations are heavily weighted toward the peninsula and the Ross Ice Shelf area.  Steig’s simple average is shown below.  The linear trend for 1957 through 2006 is +0.14 deg C/decade.  It is worth noting that if the time frame is limited to 1980 to 2006 (the period of actual AWS measurements), the trend changes to cooling, -0.06 deg C/decade.

2jeffs-steig-image9

We used a gridding methodology to weight the AWS reconstructions in proportion to the area they represent.  Using the Steig’s method, 3 stations on the peninsula over 5% of the continent’s area would have the same weighting as three interior stations spread over 30% of the continent area.  The gridding method we used is comparable to that utilized in other temperature constructions such as James Hansen’s GISStemp.  The gridcell map used for the weighted 7 PC reconstruction is shown here.

2jeffs-steig-image10

Cells with a single letter contain one or more AWS temperature stations.  If more than one AWS falls within a gridcell, the results were averaged and assigned to that cell.  Cells with multiple letters had no AWS within them, but had three or more contiguous cells containing AWS stations.  Imputed temperature time series were assigned to these cells based on the average of the neighboring cells.  Temperature trends were calculated both with and without the imputed cells.  The reconstruction trend using 7 PCs and a weighted station average follow.

2jeffs-steig-image11

The trend has decreased to 0.08 deg C/decade.  Although it is not readily apparent in this plot, from 1980 to 2006 the temperature profile has a pronounced negative trend.

Temporal smearing problems caused by too few PCs?

The temperature trends using the various reconstruction methods are shown in the table below.  We have broken the trends down into three time periods; 1957 to 2006, 1957 to 1979, and 1980 to 2006.  The time frames are not arbitrarily chosen, but mark an important distinction in the AWS reconstructions.  There is no AWS data prior to 1980.  In the 1957 to 1980 time frame, every single temperature point is a product of the RegEM algorithm.   In the 1980 to 2006 time frame, AWS data exists (albeit quite spotty at times) and RegEM leaves the existing data intact while infilling the missing data.

We highlight this distinction as limiting the reconstruction to 3 PCs has an additional pernicious effect beyond spatial smearing of the peninsula warming.   In the table below, note the balance between the trends of the 1957 to 1979 era vs. that of the 1980 to 2006 era. In Steig’s 3 PC reconstruction, moderate warming that happened prior to 1980 is more balanced with slight cooling that happened post 1980.  In the new 7 PC reconstruction, the early era had dramatic warming, the later era had strong cooling.  It is believed that the 7 PC reconstruction more accurately reflects the true trends for the reasons stated earlier in this paper.  However, the mechanism for this temporal smearing of trends is not fully understood and is under investigation.  It does appear to be clear that limiting the selection to three principal components causes warming that is largely constrained to a pre-1980 time frame to appear more continuous and evenly distributed over the entire temperature record.

Reconstruction

1957 to 2006 trend

1957 to 1979 trend (pre-AWS)

1980 to 2006 trend (AWS era)

Steig 3 PC

+0.14 deg C./decade

+0.17 deg C./decade

-0.06 deg C./decade

New 7 PC

+0.11 deg C./decade

+0.25 deg C./decade

-0.20 deg C./decade

New 7 PC weighted

+0.09 deg C./decade

+0.22 deg C./decade

-0.20 deg C./decade

New 7 PC wgtd imputed cells

+0.08 deg C./decade

+0.22 deg C./decade

-0.21 deg C./decade

Conclusion

The AWS trends which this incredibly long post was created from were used only as verification of the satellite data.  The statistics used for verification are another subject entirely.  Where Steig09 falls short in the verification is that RegEM was inappropriately applying area weighting to individual temperature stations.  The trends from the AWS reconstruction clearly have blended into distant stations creating an artificially high warming result.  The RegEM methodology also appears to have blended warming that occurred decades ago into more recent years to present a misleading picture of continuous warming.  It should also be noted that every attempt made to restore detail to the reconstruction or weight station data resulted in reduced warming and increased cooling in recent years.  None of these methods resulted in more warming than that shown by Steig.

We don’t yet have the satellite data (Steig has not provided it) so the argument will be:

“Silly Jeff’s you haven’t shown anything, the AWS wasn’t the conclusion it was the confirmation.”

To that we reply with an interesting distance correlation graph of the satellite reconstruction (also from only 3 PCs).  The conclusion has the exact same problem as the confirmation.  Stay tuned.

2jeffs-steig-image13

(Graph originally calculated by Steve McIntyre)

Get notified when a new post is published.
Subscribe today!
0 0 votes
Article Rating
135 Comments
Inline Feedbacks
View all comments
Jeff C.
March 1, 2009 7:58 am

chris y (05:38:11) :
“These give similar positive trends for 1957-1979, and similar negative trends for 1980-2006. The positive and negative trends have almost identical absolute values of 0.21 C/decade. Yet the total trend for 1957-2006 comes out at +0.08 C/decade. Is this a consequence of the error bars of the resulting trends? End effects in the data? Or something else?”
This is a good question and was something I investigated when writing this up. Applying trend lines is tricky and can be subject to manipulation and cherry picking of start/end dates. We chose the two periods (1957-1979 and 1980-2006) because their is a clear distinction between these two time frames. Prior to 1980, every single data point is imputed from the surface stations (no true measured values and AWS did not exist). After 1980, the data includes real measured temperatures from the AWS and RegEM infills missing dates.
If you view the plot below, you can see their is a step in the values around the 79/80 breakpoint. I beleive this is caused to some extent by RegEM, but I don’t fully understand it. However, it is clear that there is a pronounced downward trend from 1980 on. This coincides with the inclusion of actual measured temperatures in the AWS reconstruction.
http://i404.photobucket.com/albums/pp127/jeffc1728/7PCwghtdreconwithtrends.jpg

John F. Hultquist
March 1, 2009 8:15 am

[snip, noted thank you, but this is off topic and I’m trying to keep this thread centered]

March 1, 2009 8:17 am

Dr. Bob,
We didn’t attempt to test the significance of the trend. Any test of significance is usually based on the variance of the output. If we had some kind of significance test for the quality of the method I think it would be more meaningful. From memory, Steig reported a +/- 0.07 C/decade trend with a 95% significance repeatedly stated to be highly robust on RC. We tweaked the method a bit and got a 0.06C/Decade drop in trend. It isn’t like we looked for ways to reduce the trend either, we just made an effort to not overweight concentrations of station data. Perhaps 8 PC’s is the way to go and it might make the trend even flatter.
Allen63,
The plot by SteveM used his own method for getting the expectation maximization to converge. I’m not sure if Jeff C has tried it but I couldn’t get convergence from the algorithm much beyond 8 PC’s – this is also a very interesting point to me but I haven’t proven out why. I can almost do post on it now. If you notice SteveM’s method has a flatter slope than RegEM at 7.

Wyatt A
March 1, 2009 8:21 am

Anthony or Jeff or Jeff,
Can anyone point me to a good URL to get a better understanding of Principle Component Analysis?
Thanks,
REPLY: Here are a few – Anthony
http://www.cs.otago.ac.nz/cosc453/student_tutorials/principal_components.pdf
http://www.snl.salk.edu/~shlens/pub/notes/pca.pdf
http://en.wikipedia.org/wiki/Principal_components_analysis

Jack Simmons
March 1, 2009 8:21 am

It would be great fun to use the same methods to project the temperature trends of North America by using only the data collected in a peninsula known for its warmth and a few data points, around 30, from the interior.
Using Baja California, it would be interesting to see what the correlations would be for weather stations in Calgary, Chicago, and so forth. Then, show what these calculations would come up with for the missing data in say, Denver.
Great article.

Pamela Gray
March 1, 2009 8:23 am

Love the idea about using a control. Would Australia work? It is also surrounded by water. How about choosing stations near the Arctic using the same distance and spacial pattern? Granted, that area is surrounded by land so that is a variable that will change what you get. But still, it seems reasonable to me to produce a known control to show just how far off the back-filling method can be compared to the actual data log from all the stations.

March 1, 2009 8:44 am

It may be instructive to study the Wiki entry on PCA here:
http://en.wikipedia.org/wiki/Principal_components_analysis
In essence, PCA transforms simple data sets into linear combinations of data sets. X, Y, and Z become D = aX + bY +cZ. Thus new imaginary data is created. The coefficients of the linear equations are selected mathematically:

PCA is theoretically the optimal linear scheme, in terms of least mean square error, for compressing a set of high dimensional vectors into a set of lower dimensional vectors and then reconstructing the original set.

The imaginary data was used instead of the real data in Steig’s original paper and in the paper above by “the two Jeffs” (I apologize for not knowing who Jeff C is).
For many the use of imaginary data is sufficient and satisfying, but not for me. I prefer real data, not the imaginary kind. I like my data raw, uncooked, and not jiggered into fanciful linear combinations.
There are always “missing data” in any study. Nobody (nor any machine) can collect all the possible readings from every point in space and time. There is always measurement error, to some degree. Such are the realities of empirical science.
Those limitations do not justify the use of imaginary data. I am sorry to see so much hoorah about imaginary trends from imaginary data. I am sure all the authors and all the readers of these studies have much better things to do with their time.
Such frivolity and foolishness would be more tolerable if the imaginary science were not being used to impose Draconian economic and social oppression in the real world.

robert brucker
March 1, 2009 8:52 am

[take this issue to another thread please]

Pamela Gray
March 1, 2009 8:54 am

See the following:
http://oceancurrents.rsmas.miami.edu/southern/antarctic-cp.html
Near the bottom of the article, it describes a warm pool/cold pool oceanic wave connected to the Antarctic Current that stretches anywhere from 8 to 10 years in oscillation, but could be shorter or longer because it hasn’t been studied for very long. This wave has a major affect on SST and land temps, as well as precipitation, in the higher SH latitudes, including the Antarctic.
hmmmmm
I would want to graph any Antarctic peninsula temp changes to this wave current. Of all the places that this wave would most likely cause a temp fluctuation, it would be in and around that peninsula.

March 1, 2009 9:04 am

A small but important note: next time please include page numbers in long papers! It makes discussion about the document much easier!

dearieme
March 1, 2009 9:07 am

“It would be great fun to use the same methods to project the temperature trends of North America by using only the data collected in a peninsula known for its warmth and a few data points, around 30, from the interior.” I agree, Jack, but we must be sure to include data from a few neighbouring islands – e.g. Bermuda, Hawaii and Cuba.

Jeff C.
March 1, 2009 9:18 am

Allen63 (06:03:48) :
“By the way, why exactly 7 PC?”
Let me follow up on Jeff Id’s reply. RegEM runs in Matlab which is pricey. Steve McIntyre wrote a limited version for R (statistical analysis software in the public domain) to allow people who don’t have access to Matlab to experiment with the program. As Jeff mentioned, the Matlab version doesn’t work much beyond 8 PCs. I also this reconran it with 8 and got virtually the same results as with 7 PCs.
We beleive 7 PCs is reasonable number to use based on the distance correlation plots shown above. When we first started looking at this, we thought RegEM was not sophisticated enough to properly correlate the temperature series without explicit site to site distance information. The 3 PC distance correlation plot seemed to confirm this belief. Once we started testing higher numbers of PCs in RegEM, we made a startling discovery. RegEM could perform proper distance correlation, but when limited to 3 PCs, it was throwing that information away.

March 1, 2009 9:34 am

Anthony, thank you for hosting such a fascinating forum, for making available more lyrical expositions of pithy data, and also for your Resources and Glossary sections. I appreciate being able to send my less mathematical friends to WUWT. If only we could get our legislators to read WUWT…
To the Jeffs. More, please! Wonderful stuff.

Bernie
March 1, 2009 9:36 am

Pamela and Jack:
Perhaps one could use Tasmania as the equivalent of the Antarctica peninsula. The additional benefit is that it has the same rough shape as Antarctica and distribution of weather stations (see Steve McIntyre’s and CA discussion on shapes and PCA methods)

RickA
March 1, 2009 9:37 am

Wow. Very impressive work Jeff C and Jeff Id. I have been following along and it has been more fun than a good mystery novel. Keep up the good work.

Paul Maynard
March 1, 2009 9:43 am

This may seem trite but I agree with Mike D.
When the hockey stick first appeared in TAR, surely the first reaction of the IPCC, all scientists and indeed historians should have been “how strange this result is when set against the plethora of recorded history since the Egyptians”. Instead, they swallowed it hook line and sinker and it required M&M to trash the stats rather than for everyone to observe the obvious inconsistency.
When Steig and the gang produced their results, they could just have looked at Gistemp for the South Pole
http://data.giss.nasa.gov/cgi-bin/gistemp/gistemp_station.py?id=700890090008&data_set=1&num_neighbors=1
Then they could have debated the theory that allowed a warming trend except at the South Pole. Of course that would have been really inconvenient.
Regards
Paul

Pamela Gray
March 1, 2009 9:45 am

I don’t have much quibble with that peninsula area of Antarctic warming up. Several factors are at play. Temperature station growth, urban heat, seabed volcanic activity, and oceanic oscillation patterns. If my AIMS memory serves me, Antarctica is not exactly a hotbed of atmospheric CO2. If it is being touted as an indicator of global pollution and warming due to CO2 in the atmosphere, tell me how that is so in Antarctica?

Jeff C.
March 1, 2009 9:53 am

“I apologize for not knowing who Jeff C is”
No apology needed. My background is systems engineering in a technical field unrelated to climate science. One of my primary roles is the development of verification algorithms. These are techniques for post-processing measured test data in order to extrapolate information for parameters beyond those we can easily measure. Sort of like climate science where you get to create data where it doesn’t exist. The difference being if we get it wrong the product fails and people get fired.

Ken Hall
March 1, 2009 10:05 am

Phil: “Steig et al must be burning the midnight oil to come up with an answer to this.”
——————————
If only that were necessary. Unfortunately for truthful and honest reportage, the mainstream media will report Steig et al as unquestionable truth and not even touch the counter reports such as this with a barge-pole.
So as far as the lay-man in the street is concerned, Stieg et al is accurate, correct and beyond question.
Stieg will not lose sleep over his theory being shown to be based upon much faulty and unreliable science, because the mainstream media and peer-reviewers do not care if the science is shoddy and wrong. So long as they get the headlines, they are happy.

REPLY:
Barring their acceptance of publishing a rebuttal, perhaps we should consider a full page ad in Nature. I think we could garner enough financial support from readers to make that happen. – Anthony

Scottie
March 1, 2009 10:06 am

My knowledge of Higher Math would fit on the back of the proverbial postage stamp but I think I’ve managed to follow this one, so thanks, Jeffs, for making it understandable to a thicko!
I don’t want to accuse anyone of dishonest science (no, seriously) but it worries me that Steig reports only a rising temperature trend when there seems to be ample evidence from all sorts of other sources that we can expect 30 years of rising temps followed by 30 years of falling temps which seems to be pretty much what his research and this paper shows.
I’m not naive enough to think all scientists are perfect but what is the gain in the long-term for any of us — ‘Warm-mists’ included — if we are caught “looking the wrong way” when the temps turn down?
BTW, I fancy the idea of using the US as a control. At least you could guarantee to be able to cross-check theory against reality.

Robert Bateman
March 1, 2009 10:14 am

This is a case example of attempting to reconstruct data to a high resolution in which no consistent high resolution data was taken. While the end product is an image, it is still a DIGITAL image composed of data. Filling in gives you more of the same blur of missing data. What would be far more appropriate here is to use a median filter to put as much data points in from actual data sets, rather than strict interpolation.
The bottom line is that an image is simply a graphical representation of the data set, operated on or not. If the operation is a kludge, the image shows it.
When somebody comes along with some advanced image restoration techniques such as drizzle, SWARP, MaxEnt, LR or others, then we’ll have a better handle on what’s going on. In the meantime, blurred images don’t have much to say.

DAV
March 1, 2009 10:29 am

Interesting article. Well, done.
The underlying assumption in the EM algorithm is that data dropouts are normally distributed, i.e., non-systemic. It is effectively stereotyping. Filling in a station that never was is like determining how a person would answer — if only he had been born — by using the demographics of presumably surrounding hypothetical neighbors. We all should know the dangers of stereotyping.
The process works well for photographs — even for systemic errors such as dead or hot pixels — because the eye will average them anyway. The value of each pixel is small compared to the overall image. It only works because 1) you just need the average and 2) there are a large number of surrounding pixels.
The process fails miserably for gaping holes.
RegEM is essentially the equivalent of normalizing then rotating the data for maximum variable separation by minimizing cross-correlations. Using RegEM might mean your guesses are more refined when applicable but it can’t perform magic, either . The end result is still a guess.
I’m reminded of a short-lived program a few years ago about the CIA (“The Agency,” IIRC) where, in one episode, technicians completely reconstructed a face from what couldn’t have been more than 16 pixels obtained from a reflection from the back of a rearview mirror imaged by a liquor store security camera. Pure Fantasy.

John Peter
March 1, 2009 10:33 am

I have been following this and other articles revealing faults in the calculations etc. in papers particularly submitted by AGW supporters with increasing degrees of incredulity. When I worked for a Scottish company owned by a US corporation we were accused of manufacturing missile carrieres for Iraq. A US government auditor was sent over and found no such evidence. Is it not possible for a relevant US government auditing department to send in neutral observers to investigate potential fraud or consistent reported manipulations of evidence in organisations either owned by or obtaining funding from US government funds? Surely NASA would be a case in point.

Julius StSwithin
March 1, 2009 10:33 am

Excellent work. Thanks to all of you. A few brief comments:
1. You would have thought that after the Mann/Wegman sequence Nature would have been more circumspect.
2. I am surprised that the inter-station correlation drops to zero with large distances. Working with similar data in other parts of the world
I have found that R2 drops to a low (0.1 to 0.3) value and stays there for all distances. This residual value reflects the fact that all stations have similar seasonal patterns.
3. There is one general point the two Jeffs might like to comment on which I have not seen discussed elsewhere. In any regression of Y = a1.X1 + a2.X2 + … +b calculated values of Y will have a lower variance than observed Y. This is clearly seen in PC3 above but applies to all proxy data sets.

Bob Buchanan
March 1, 2009 10:42 am

Paul M
I looked at the GISS plot you linked.
Something very weird is going on since there is an obvious (to me anyway) step increase in year-to-year variability about the same time we started getting satellite data.
What’s going on?
Regards
Bob B