Art courtesy Dave Stephens
Foreword by Anthony Watts: This article, written by the two Jeffs (Jeff C and Jeff Id) is one of the more technically complex essays ever presented on WUWT. It has been several days in the making. One of the goals I have with WUWT is to make sometimes difficult to understand science understandable to a wider audience. In this case the statistical analysis is rather difficult for the layman to comprehend, but I asked for (and got) an essay that was explained in terms I think many can grasp and understand. That being said, it is a long article, and you may have to read it more than once to fully grasp what has been presented here. Steve McIntyre of Climate Audit laid much of the ground work for this essay, and from his work as well as this essay, it is becoming clearer that Steig et al (see “Warming of the Antarctic ice-sheet surface since the 1957 International Geophysical Year”, Nature, Jan 22, 2009) isn’t holding up well to rigorous tests as demonstrated by McIntyre as well as in the essay below. Unfortunately, Steig’s office has so far deferred (several requests) to provide the complete data sets needed to replicate and test his paper, and has left on a trip to Antarctica and the remaining data is not “expected” to be available until his return.
To help layman readers understand the terminology used, here is a mini-glossary in advance:
RegEM – Regularized Expectation Maximization
PCA – Principal Components Analysis
PC – Principal Components
AWS – Automatic Weather Stations
One of the more difficult concepts is RegEM, an algorithm developed by Tapio Schneider in 2001. It’s a form of expectation maximization (EM) which is a common and well understood method for infilling missing data. As we’ve previously noted on WUWT, many of the weather stations used in the Steig et al study had issues with being buried by snow, causing significant data gaps in the Antarctic record and in some burial cases stations have been accidentally lost or confused with others at different lat/lons. Then of course there is the problem of coming up with trends for the entire Antarctic continent when most of the weather station data is from the periphery and the penisula, with very little data from the interior.
Expectation Maximization is a method which uses a normal distribution to compute the best probability of fit to a missing piece of data. Regularization is required when so much data is missing that the EM method won’t solve. That makes it a statistically dangerous technique to use and as Kevin Trenberth, climate analysis chief at the National Center for Atmospheric Research, said in an e-mail: “It is hard to make data where none exist.” (Source: MSNBC article) It is also valuable to note that one of the co-authors of Steig et al, Dr. Michael Mann, dabbles quite a bit in RegEm in this preparatory paper to Mann et al 2008 “Return of the Hockey Stick”.
For those that prefer to print and read, I’ve made a PDF file of this article available here.
Introduction
This article is an attempt to describe some of the early results from the Antarctic reconstruction recently published on the cover of Nature which demonstrated a warming trend in the Antarctic since 1956. Actual surface temperatures in the Antarctic are hard to come by with only about 30 stations prior to 1980 recorded through tedious and difficult efforts by scientists in the region. In the 80’s more stations were added including some automatic weather stations (AWS) which sit in remote areas and report the temperature information automatically. Unfortunately due to the harsh conditions in the region many of these stations have gaps in their records or very short reporting times (a few years in some cases). Very few stations are located in the interior of the Antarctic, leaving the trend for the central portion of the continent relatively unknown. The location of the stations is shown on the map below.
In addition to the stations there are satellite data from an infrared surface temperature measurement which records the temperature of the actual emission from the surface of the ice/ground in the Antarctic. This is different from the microwave absorption measurements as made from UAH/RSS data which measure temperatures in a thickness of the atmosphere. This dataset didn’t start until 1982.
Steig 09 is an attempt to reconstruct the continent-wide temperatures using a combination of measurements from the surface stations shown above and the post-1982 satellite data. The complex math behind the paper is an attempt to ‘paste’ the 30ish pre-1982 real surface station measurements onto 5509 individual gridcells from the satellite data. An engineer or vision system designer could use several straightforward methods which would insure reasonable distribution of the trends across the grid based on a huge variety of area weighting algorithms, the accuracy of any of the methods would depend on the amount of data available. These well understood methods were ignored in Steig09 in favor of RegEM.
The use of Principal Component Analysis in the reconstruction
Steig 09 presents the satellite reconstructions as the trend and also provides an AWS reconstruction as verification of the satellite data rather than a separate stand alone result presumably due to the sparseness of the actual data. An algorithm called RegEM was used for infilling the missing data. Missing data includes pre 1982 for satellites and all years for the very sparse AWS data. While Dr. Steig has provided the reconstructions to the public, he has declined to provide any of the satellite, station or AWS temperature measurements used as inputs to the RegEM algorithm. Since the station and AWS measurements were available through other sources, this paper focuses on the AWS reconstruction.
Without getting into the detail of PCA analysis, the algorithm uses covariance to assign weighting of a pattern in the data and does not have any input whatsoever for actual station location. In other words, the algorithm has no knowledge of the distance between stations and must infill missing data based solely on the correlation with other data sets. This means there is a possibility that with improper or incomplete checks, a trend from the peninsula on the west coast could be applied all the way to the east. The only control is the correlation of one temperature measurement to another.
If you were an engineer concerned with the quality of your result, you would recognize the possibility of accidental mismatch and do a reasonable amount of checking to insure that the stations were properly assigned after infilling. Steig et. al. described no attempts to check this basic potential problem with RegEM analysis. This paper will describe a simple method we used to determine that the AWS reconstruction is rife with spurious (i.e. appear real but really aren’t) correlations attributed to the methods used by Dr. Steig. These spurious correlations can take a localized climactic pattern and “smear” it over a large region that lacks adequate data of its own.
Now is where it becomes a little tricky. RegEM uses a reduced information dataset to infill the missing values. The dataset is reduced by Principal Component Analysis (PCA) replacing each trend with a similar looking one which is used for covariance analysis. Think of it like a data compression algorithm for a picture which uses less computer memory than the actual but results in a fuzzier image for higher compression levels.
While the second image is still visible, the actual data used to represent the image is reduced considerably. This will work fine for pictures with reasonable compression, but the data from some pixels has blended into others. Steig 09 uses 3 trends to represent all of the data in the Antarctic. In it’s full complexity using 3 PC’s is analogous to representing not just a picture but actually a movie of the Antarctic with three color ‘trends’ where the color of each pixel changes according to different weights of the same red, green and blue color trends (PC’s). With enough PC’s the movie could be replicated perfectly with no loss. Here’s an important quote from the paper.
“We therefore used the RegEM algorithm with a cut-off parameter K=3. A disadvantage of excluding higher-order terms (k>3) is that this fails to fully capture the variance in the Antarctic Peninsula region. We accept this tradeoff because the Peninsula is already the best-observed region of the Antarctic.”

Above: a graph from Steve McIntyre of ClimateAudit where he demonstrates how “K=3 was in fact a fortuitous choice, as this proved to yield the maximum AWS trend, something that will, I’m sure, astonish most CA readers.”
K=3 means only 3 trends were used, the ‘lack of captured variance’ is an acknowledgement and acceptance of the fuzziness of the image. It’s easy to imagine that it would be difficult to represent a complex movie image of Antarctic with any sharpness from 1957 to 2006 temperature with the same 3 color trends reweighted for every pixel. In the satellite version of the Antarctic movie the three trends look like this.
Note that the sudden step in the 3rd trend would cause a jump in the ‘temperature’ of the entire movie. This represents the temperature change between the pre 1982 recreated data and the after 1982 real data in the satellite reconstruction. This is a strong yet overlooked hint that something may not be right with the result.
In the case of the AWS reconstruction we have only 63 AWS stations to make the movie screen, by which the trends of 42 surface station points are used to infill the remaining data. If the data from one surface station is copied to the wrong AWS stations the average will overweight and underweight some trends. So the question becomes, is the compression level too high?
The problems that arise when using too few principal components
Fortunately, we’re here to help in this matter. Steve McIntyre again provided the answer with a simple plot of the actual surface station data correlation with distance. This correlation plot compares the similarities ‘correlation’ of each temperature station with all of the 41 other manual surface stations against the distance between them. A correlation of 1 means the data from one station is exactly equal to the other. Because A -> B correlation isn’t a perfect match for B->A there are 42*42 separate points in the graph. This first scatter plot is from measured temperature data prior to any infilling of missing measurements. Station to station distance is shown on the X axis. The correlation coefficient is shown on the Y axis.
Since this plot above represents the only real data we have existing back to 1957, it demonstrates the expected ‘natural’ spatial relationship from any properly controlled RegEM analysis. The correlation drops with distance which we would expect because temps from stations thousands of miles away should be less related than those next to each other. (Note that there are a few stations that show a positive correlation beyond 6000 km. These are entirely from non-continental northern islands inexplicably used by Steig in the reconstruction. No continental stations exhibit positive correlations at these distances.) If RegEM works, the reconstructed RegEM imputed (infilled) data correlation vs. distance should have a very similar pattern to the real data. Here’s a graph of the AWS reconstruction with infilled temperature values.
Compare this plot with the previous plot from actual measured temperatures. Now contrast that with the AWS plot above. The infilled AWS reconstruction has no clearly evident pattern of decay over distance. In fact, many of the stations show a correlation of close to 1 for stations at 3000 km distant! The measured station data is our best indicator of true Antarctic trends and it shows no sign that these long distance correlations occur. Of course, common sense should also make one suspicious of these long distance correlations as they would be comparable to data that indicated Los Angeles and Chicago had closely correlated climate.
It was earlier mentioned that the use of 3 PCs was analogous to the loss of detail that occurs in data compressions. Since the AWS input data is available, it is possible to regenerate the AWS reconstruction using a higher number of PCs. It stood to reason that spurious correlations could be reduced by retaining the spatial detail lost in the 3 PC reconstruction. Using RegEM, we generated a new AWS reconstruction using the same input data but with 7 PCs. The distance correlations are shown in the plot below.
Note the dramatic improvement over that shown in the previous plot. The correlation decay with distance so clearly seen in the measured station temperature data has returned. While the cone of the RegEM data is slightly wider than the ‘real’ surface station data, the counterintuitive long distance correlations seen in the Steig reconstruction have completely disappeared. It seems clear that limiting the reconstruction to 3 PCs resulted in numerous spurious correlations when infilling missing station data.
Using only 3 principal components distorts temperature trends
If Antarctica had uniform temperature trends across the continent, the spurious correlations might not have a large impact in the overall reconstruction. Individual sites may have some errors, but the overall trend would be reasonably close. However, Antarctica is anything but uniform. The spurious correlations can allow unique climactic trends from a localized region to be spread over a larger area, particularly if an area lacks detailed climate records of its own. It is our conclusion is that is exactly what is happening with the Steig AWS reconstruction.
Consider the case of the Antarctic Peninsula:
- The peninsula is geographically isolated from the rest of the continent
- The peninsula is less than 5% of the total continental land mass
- The peninsula is known to be warming at a rate much higher than anywhere else in Antarctica
- The peninsula is bordered by a vast area known as West Antarctica that has extremely limited temperature records of its own
- 15 of the 42 temperature surface stations (35%) used in the reconstruction are located on the peninsula
If the Steig AWS reconstruction was properly correlating the peninsula stations temperature measurements to the AWS sites, you would expect to see the highest rates of warming at the peninsula extremes. This is the pattern seen in the measured station data. The plot below shows the temperature trends for the reconstructed AWS sites for the period of 1980 to 2006. This time frame has been selected as this is the period when AWS data exists. Prior to 1980, 100% of AWS reconstructed data is artificial (i.e. infilled by RegEM).
Note how warming extends beyond the peninsula extremes down toward West Antarctica and the South Pole. Also note the relatively moderate cooling in the vicinity of the Ross Ice Shelf (bottom of the plot). The warming once thought to be limited to the peninsula appears to have spread. This “smearing” of the peninsula warming has also moderated the cooling of the Ross Ice Shelf AWS measurements. These are both artifacts of limiting the reconstruction to 3 PCs.
Now compare the above plot to the new AWS reconstruction using 7 PCs.
The difference is striking. The peninsula has become warmer and warming is largely limited to its confines. West Antarctica and the Ross Ice Shelf area have become noticeably cooler. This agrees with the commonly-held belief prior to Steig’s paper that the peninsula is warming, the rest of Antarctica is not.
Temperature trends using more traditional methods
In providing a continental trend for Antarctica warming, Steig used a simple average of the 63 AWS reconstructed time series. As can be seen in the plots above, the AWS stations are heavily weighted toward the peninsula and the Ross Ice Shelf area. Steig’s simple average is shown below. The linear trend for 1957 through 2006 is +0.14 deg C/decade. It is worth noting that if the time frame is limited to 1980 to 2006 (the period of actual AWS measurements), the trend changes to cooling, -0.06 deg C/decade.
We used a gridding methodology to weight the AWS reconstructions in proportion to the area they represent. Using the Steig’s method, 3 stations on the peninsula over 5% of the continent’s area would have the same weighting as three interior stations spread over 30% of the continent area. The gridding method we used is comparable to that utilized in other temperature constructions such as James Hansen’s GISStemp. The gridcell map used for the weighted 7 PC reconstruction is shown here.
Cells with a single letter contain one or more AWS temperature stations. If more than one AWS falls within a gridcell, the results were averaged and assigned to that cell. Cells with multiple letters had no AWS within them, but had three or more contiguous cells containing AWS stations. Imputed temperature time series were assigned to these cells based on the average of the neighboring cells. Temperature trends were calculated both with and without the imputed cells. The reconstruction trend using 7 PCs and a weighted station average follow.
The trend has decreased to 0.08 deg C/decade. Although it is not readily apparent in this plot, from 1980 to 2006 the temperature profile has a pronounced negative trend.
Temporal smearing problems caused by too few PCs?
The temperature trends using the various reconstruction methods are shown in the table below. We have broken the trends down into three time periods; 1957 to 2006, 1957 to 1979, and 1980 to 2006. The time frames are not arbitrarily chosen, but mark an important distinction in the AWS reconstructions. There is no AWS data prior to 1980. In the 1957 to 1980 time frame, every single temperature point is a product of the RegEM algorithm. In the 1980 to 2006 time frame, AWS data exists (albeit quite spotty at times) and RegEM leaves the existing data intact while infilling the missing data.
We highlight this distinction as limiting the reconstruction to 3 PCs has an additional pernicious effect beyond spatial smearing of the peninsula warming. In the table below, note the balance between the trends of the 1957 to 1979 era vs. that of the 1980 to 2006 era. In Steig’s 3 PC reconstruction, moderate warming that happened prior to 1980 is more balanced with slight cooling that happened post 1980. In the new 7 PC reconstruction, the early era had dramatic warming, the later era had strong cooling. It is believed that the 7 PC reconstruction more accurately reflects the true trends for the reasons stated earlier in this paper. However, the mechanism for this temporal smearing of trends is not fully understood and is under investigation. It does appear to be clear that limiting the selection to three principal components causes warming that is largely constrained to a pre-1980 time frame to appear more continuous and evenly distributed over the entire temperature record.
| Reconstruction |
1957 to 2006 trend |
1957 to 1979 trend (pre-AWS) |
1980 to 2006 trend (AWS era) |
| Steig 3 PC |
+0.14 deg C./decade |
+0.17 deg C./decade |
-0.06 deg C./decade |
| New 7 PC |
+0.11 deg C./decade |
+0.25 deg C./decade |
-0.20 deg C./decade |
| New 7 PC weighted |
+0.09 deg C./decade |
+0.22 deg C./decade |
-0.20 deg C./decade |
| New 7 PC wgtd imputed cells |
+0.08 deg C./decade |
+0.22 deg C./decade |
-0.21 deg C./decade |
Conclusion
The AWS trends which this incredibly long post was created from were used only as verification of the satellite data. The statistics used for verification are another subject entirely. Where Steig09 falls short in the verification is that RegEM was inappropriately applying area weighting to individual temperature stations. The trends from the AWS reconstruction clearly have blended into distant stations creating an artificially high warming result. The RegEM methodology also appears to have blended warming that occurred decades ago into more recent years to present a misleading picture of continuous warming. It should also be noted that every attempt made to restore detail to the reconstruction or weight station data resulted in reduced warming and increased cooling in recent years. None of these methods resulted in more warming than that shown by Steig.
We don’t yet have the satellite data (Steig has not provided it) so the argument will be:
“Silly Jeff’s you haven’t shown anything, the AWS wasn’t the conclusion it was the confirmation.”
To that we reply with an interesting distance correlation graph of the satellite reconstruction (also from only 3 PCs). The conclusion has the exact same problem as the confirmation. Stay tuned.
(Graph originally calculated by Steve McIntyre)












When this was discussed on CA, I asked about the possible criteria that could be used to determine the optimum number of PC’s to use, and what they said.
It seems clear from the data that 3 were not enough to represent the station data.
There are 2 criteria mentioned on this web site which explains the principal components representation method.
http://www.statsoft.com/textbook/stfacan.html
I am referring to the Kaiser Criterion
“The Kaiser criterion. First, we can retain only factors with eigenvalues greater than 1.”
and the Scree Criterion.
“Catell suggests to find the place where the smooth decrease of eigenvalues appears to level off to the right of the plot. “
I posted a similar post at Climate Audit and decided to post this here too as Climate Audit appears to be down,
This is an excellent analysis and rework of the methods. However, as the Jeffs note, it analyzes the AWS data which Steig et al. indicate were only used for corroborative purposes. As documented by Steve McIntyre (from statements on Realclimate). In his post
http://www.climateaudit.org/?p=5312#comment-329096
the actual AVHRR data that the Steig et al. paper were based upon have not been provided.
This raises the question of why Nature published this paper at all (not to mention gave it a cover) when the paper violates Nature’s own editorial policies for “Availability of data and materials”
(editorial policies available at: http://www.nature.com/authors/editorial_policies/availability.html ), which clearly state:
“An inherent principle of publication is that others should be able to replicate and build upon the authors’ published claims. Therefore, a condition of publication in a Nature journal is that authors are required to make materials, data and associated protocols promptly available to readers without preconditions. Any restrictions on the availability of materials or information must be disclosed to the editors at the time of submission. Any restrictions must also be disclosed in the submitted manuscript, including details of how readers can obtain materials and information. If materials are to be distributed by a for-profit company, this should be stated in the paper.”
“Supporting data must be made available to editors and peer-reviewers at the time of submission for the purposes of evaluating the manuscript. Peer-reviewers may be asked to comment on the terms of access to materials, methods and/or data sets; Nature journals reserve the right to refuse publication in cases where authors do not provide adequate assurances that they can comply with the journal’s requirements for sharing materials.”
“After publication, readers who encounter refusal by the authors to comply with these policies should contact the chief editor of the journal (or the chief biology/chief physical sciences editors in the case of Nature). In cases where editors are unable to resolve a complaint, the journal may refer the matter to the authors’ funding institution and/or publish a formal statement of correction, attached online to the publication, stating that readers have been unable to obtain necessary materials to replicate the findings.”
It was bad enough that Steig refused to provide the actual code for replication. As the AVHRR data are not available and apparently were not available to the reviewers, one has to ask how did this paper get published when it violates Nature’s policies?
“Rocket Man (11:21:30) :
I would think we would still need some ground based stations for calibration purposes, just to keep the satellites honest.”
OK, then what do we do to keep the ground based stations honest?
Having spent nearly 40 years in the measurement business, long term absolute accuracy is very hard to achieve. How often are these sensors calibrated? By what means? In the cold of Antarctica, I suspect a lot of things drift with time. And not all in the same direction.
GIGO big time.
J&J’s work does an excellent job of showing Steig may have to go back to the drawing board. We need more of this.
Allen63 (06:03:48) :
By the way, why exactly 7 PC?
Why not use as many PCs as possible to get the best fit possible?
I might ask “Why use more than one?” What we are seeing here, I think, is a misuse of PCA. Adding PC’s is a bit like adding variables to a regression equation: the goodness of fit, R-squared, always increases, with each variable added to the equation. But eventually, the equation gets overspecified, or overdetermined, and the variables are not adding anything meaningful to the analysis.
Where is the table that shows the proportion of the covariance matrix explained by each PC? Especially for those who are not familiar with PCA, I think this is crucial info. Almost everybody who has ever heard of “regression analysis” has a basic understanding of “R-squared.” (That is not always good, I know, but what I am getting at here is roughly equivalent to whatever good R-squared tells us.) In the same way, it is pretty easy to understand that PC’s are ranked in the order of their contribution to the covariance matrix.
Now it is easy to understand, from the third figure, why Steig et al didn’t stop at just one, and did stop at three. The first, and most significant PC, is negative. Since that isn’t what Steig et al set out to show, they couldn’t stop there. So they went on to PC3, the most positive PC, and stopped there. What Jeff & Jeff have done, so nicely, with the fourth figure is demonstrate that PC3 is just an artifact, and is not indicative of a underlying natural or physical process.
Which brings me full circle here. PCA, when used correctly, is intended to extract signals from multivariate data that are thought to be significant, in this case natural process. It is not the combination of PC’s that we are after, it is the individual PC’s that are supposed to be representative of something. Without articulating what those individual PC’s might represent — which is either something meaningful, or otherwise is just “noise” — then PCA, as we’re seeing it with Steig et al, is nothing but a technique of massaging the data to get it to say what we want it to say.
Or, as I prefer to put it: If you torture the data, it will confess, even to crimes it did not commit.
Until I hear a convincing explanation why we shouldn’t stop at PC1, then all I know is that the strongest PC is negative.
To Jeff and Jeff,
A very understandable and effective article. In many of these studies the influence of an engineer would keep the results grounded in reality. Having to design something for the real world and being held accountable for the results makes it very difficult to perform this type of analysis without thinking of all of the ramifications and variance.
It is unfortunate that it takes one to apply a rigorous analysis of the system under study.
Also, thanks Anthony for sponsoring this great article.
Thanks to Anthony and everyone for the supportive comments. BTW: if someone can find that elusive big oil check, I can be bought. Just kidding 🙂
There were several good questions and examples here that I’m not sure I can answer cleanly, if I left some unattended which really need addressing I apologize. There were several pertinent comments about the number of PC’s from people who have clearly got some experience.
In my opinion PC1 is all that’s required if the stations are weighted according to the area they cover. However, if the stations are weighted that way, I believe PC1 breaks down to a least squares fit of the average. Nice simple math that anyone can appreciate. If the stations are simply tossed in the number masher without concern for area weighting (as was done in Steig 09) higher PC’s are clearly required to infill the data correctly.
Anyway, I’ll keep stopping by because this thread is still very active but if I miss something from the comments you can find me pretty easily and ask again on my much smaller but still fast growing blog.
Retired Engineer (15:01:44) :
The statement:
“I would think we would still need some ground based stations for calibration purposes, just to keep the satellites honest.”
was made by HasItBeen4YearsYet? (12:54:42) . I was just quoting him.
I completely agree with you that the accuracy of the ground based stations is open to questions due to calibration issues. Working in Aerospace, we regularly calibrate every piece of instrumentation on a fixed schedule. Yearly is the most common, but some are longer and some are shorter. We regularly find errors in the instruments brought in for calibration, which sometimes requires us to go back and repeat a test.
In addition to the siting issues the Surfacestations.org project has brought to light, I would love to see the calibration data on the temperature sensors as well. How often are they calibrated? How often are they out of calibration when they are calibrated and how much are they off? What is the true accuracy of the temperature sensors?
Perhaps the sensors in Antarctica would be a good place to look at these issues as there are so few stations to evaluate. My guess is that the errors in these sensors swamp whatever temperature signal they are trying to tease out of the data.
Basil and others,
There is plenty more to come on this. Jeff and I focused on one particular aspect, the flaws in the AWS reconstruction (primarily the false long distance correlations) and what happens to the trends if you include the omitted information that caused the false correlations (the higher order PCs).
Steve McIntyre has done some very insightful posts on the spatial patterns of the PC coefficients. In the paper, Dr. Steig claims:
Steve has shown that the spatial patterns are driven by the shape of Antarctica, and any relationship to the “Southern Hemisphere atmospheric circulation” is most-likely incidental. Even worse, citations from climate journals over the past 30 years warned against drawing premature conclusions from the spatial patterns of the PC coefficients. One author likened it to “the observations of children who see castles in the clouds”. Thus Dr. Steig’s justification for stopping at three PCs appears weak.
Another fascinating aspect is that every single data point in the satellite reconstruction can be described with an accuracy of 10E-8 by three principal components. This is over 3 million data points (600 months x 5509 locations). The takeaway is that the contents of the entire satellite reconstruction are fitted values. Unlike the AWS reconstruction which contains real measured values and infilled values, the entire satellite reconstruction is infilled! The measured satellite data wasn’t supplemented, it was replaced. We would like to compare the raw satellite measurements to the fitted values in the satellite reconstruction. However, this is the data Dr. Steig has not seen fit to release.
We also have the very last plot in this article which shows the satellite reconstruction distance correlation is just as bad as the AWS reconstruction, if not worse.
For those interested, I would urge you to follow the almost daily deconstruction of this paper at Climate Audit or Jeff Id’s site http://noconsensus.wordpress.com/.
Very well presented study.
Thanks and congratulations to all involved in its production, but we should perhaps temper our enthusiasm with the thought that the results seen here and their implications will never receive the publicity splash that Steig got from his faulty piece. … and that is truly sad.
thefordprefect (13:23:13) :
“No infill. No reconstruction. Just the data:”
Exactly. So why didn’t Dr. Steig just use this to support his claim instead of turning to the statistical sausage grinder?
The answer is the stations cited are clustered on the peninsula and around the coast. There are only three interior stations, Vostok, Amundsen-Scott (south pole) and Byrd. Three stations for 90% of the area of the continent.
Dr. Comiso of NASA put out a very thorough paper in 2000 that detailed the results of the satellite temperature measurements enhanced using a technique known as cloud masking. His findings, laid out in painstaking detail in the paper, show that the vast majority of the continent interior was cooling. Yes, some areas were warming, but they were largely restricted to the coastal belts. Here is the link, check it out. http://ams.allenpress.com/archive/1520-0442/13/10/pdf/i1520-0442-13-10-1674.pdf
Dr Steig needed to show that Comiso 2000 was wrong. He couldn’t do that with the station data because it doesn’t exist, hence the foray with RegEM.
Wow, what great work on an obviously complicated subject. Kudos to all involved. It’s just fantastic to have so many committed to efforts like these to “stress” test the shaky science that is coming out those $50 billion in funds to prove global warming is really “global warming.”
Do these writers of this paper have real names? What are they please?
I am being badgered by the AGW crowd over the anonymous nature of their posting. I felt it should stand or fall on its technical merits but..
Jeff C. (16:08:51) :
Jeff,
I’m glad there’s more to come, and I’m not trying to detract from what you and Jeff Id (and Steve, and Hu, and whomever else) are doing. But somewhere along the way I’d like to see it brought out that the statement
“The first three principal components are statistically separable and can be meaningfully related to important dynamical features of high-latitude Southern Hemisphere atmospheric circulation, as defined independently by extrapolar instrumental data.”
is meaningless unless the “important dynamical features…yada yada” are related to specific PC’s in a plausible and meaningful way. Let’s get the first off the table. What “important dynamical feature” created the negative PC1? I’m guessing, rather, that PC1 is nothing more than a weighted proxy of temperature, so that it doesn’t really explain anything, or tell us what anything that we didn’t already know: that without massaging the data to make data where there is none, the weight of the evidence indicates that Antarctica has been cooling (except the Peninsula).
While I think you guys are doing good work exposing the statistical nonsense in all this, the real travesty is that this was accepted for publication in a science journal without, it appears, any kind of real science underpinning the analysis. It was, in the end, all about getting the data to confess to a crime it did not commit.
David,
Who’s badgering?
For those wondering about that peninsula, just google Antarctic Ocean Current and you will get lots of good info. That peninsula is far more affected by cyclic components of this huge current, including a warm and cold wave phase. The information is easy to digest and will be very educational in terms of taking “global warming” and “the ice is melting” stuff with a grain of salt water.
DAV said: “The underlying assumption in the EM algorithm is that data dropouts are normally distributed, i.e., non-systemic.”
I really doubt if the dropouts are totally non-systemic. One of the reason for data dropout is burial of the sensor by snow. I.e., the stations with the worst weather will tend to have more missing data, IMO. I know that there’s no established correlation between snow accumulation and temperature, but it might bear investigation.
timbrom said: “Has anyone calculated the energy required to melt all the ice in Antartica…? At a rough guess I’d hazard that it would take a couple of years, at least.”
I understand that the esteemed Dr. Steig himself, in a candid moment, stipulated that it would take a very long time for all the accumulated Antarctic ice to melt, based on the slow warming trend he is certain exists. He’s been a lot less open, lately, which I find regrettable, since I know he’s well regarded by some reasonable scientists. I’m hoping he’ll relent, but he may be under some pressure from other interested parties.
There is a considerable amount of misinformation propagated about the greenhouse effect by people from both sides of the debate. This is true, but all the graphs and data will not convince joe sixpack of anything. Joe sixpack worries about more taxes, does not want to give up his hemi powered pickup, and has no idea what a carbon foot print is. While science provides us with the needed information, Rush, and those like him constantly regail against it. Joe has disregarded the science and believes that climate change is natural, from a wobble in the Earths orbit, a change in the sun, or the end of the last ice age. He may even believe there is no climate change at all. Joe believes there is no reason to change anything, as natural, we did not cause it, we can not change it, therefore nothing can be done about it. Going green is foolishness. An electric or hydroelectric car is pointless. If joe sixpack was the republican hero, than Homer Simpson is his archtype. To evolve from the horse drawn wagon to the gas powered automobile was easy, to evolve beyond that is more than Joe can handle. Maybe he has a point, Consider this, In one mans life span we have gone from horse drawn wagon, to the automobile, to the jet airplane, then to walking on the moon. Throw in the atom bomb, Vietnam, and computers, maybe it is more than Joe can handle. With the last President, science was ignored. Perhaps now with President Obama things will change. If the U.S. continues its downward trend regarding it economic downturn, then the climate change problem will be put on the back burner. Europe and Asia could lead the way for change. Greed is greed though. What drove us drives them as well. To Joe, anything that changes his lifestyle is a loss of freedom. A seatbelt use law is like a communist plot. Even though it is for everyones good, to mandate any change for going greener, would be viewed as some sort of government restriction on Joes freedom. His motto is family, land and his rifle. Any change to deal with climate change will have to be done with out him. Alkataba
“the results seen here and their implications will never receive the publicity splash that Steig got from his faulty piece.”
Unless Nature withdraws the paper.
Works like this one are important to those few who actually seek truth for its own sake. Unfortunately, the general public has neither the will nor the capacity to seek out truth on the subject of “Manmade Global Warming”. For that, they depend on “the court of world opinion” where the foreman of the jury is the news media. The “attorneys” who present the case on the side of the affirmative are the likes of Al Gore and Dr. James Hansen, and they have been very effective at capturing the jury foreman’s imagination. The “attorneys” on the side of the negative, whoever they may be, have yet capture the jury’s attention, but works like this one will be important evidence in the event said “attorneys” ever actually do make an effective appearance.
In the meantime, the jury foreman (news media) is suffering a problem with leadership of the jury. He’s having a devil of a time keeping a straight face as he presents “weather stories” of record cold temperatures left and right while simultaneously presenting “climate stories” of “we’re all going to burn up and die (at least those of us who don’t drown beforehand)”. In the final analysis, nature may well decide the outcome while the court dithers; even the dullest bulb on the human tree can read a thermometer and comes equipped with an uncanny ability to detect an icicle working its way down his or her back.
Amazing! Over 100 posts over several days and not one counter argument from the AGW crowd! I expected them to be all over this like a rash crying foul.
Good work Jeff’s!
alkataba (21:52:51) :
What does “science was ignored” under Bush mean? I tend to understand this to mean the technocratic planners were not allowed to plan the economy and lives of the people.
When science is applied to a messianic mission I’ll be slow to take it up, because for every polio example there is a DDT example.
Does this mean I ignore science? Nope. It means when it is politicized I will slow my adoption of its strictures.
Jeff and Jeff
Very good work. I raised this question at RC (“Antarctic warming is robust” thread, comment 353) and Gavin failed to answer it as pointed out by ApolytonGP, comment 364.
Steig etal has to be wrong, because their ‘reconstruction’ has the peninsula warming at 0.11 C/decade when the observations show it’s about 0.5 C/decade (their error bars are +- 0.04, so their actual error is about ten times their error bars!) As you say, they seem to accept this. But given that they are out by a factor of 5 on the peninsula, what hope can we have of the accuracy of their ‘reconstruction’ in the interior?
But there are some crucial numbers you haven’t given us yet as far as I can see.
Steig et al say that over 1956-2007 W antarctica warmed at 0.17C/decade and the peninsula at 0.11.
What are the results for these separate regions in your 7-PC reconstruction? This may be difficult because they dont define these areas exactly but you should be able to get a rough answer. To put it another way, can you produce a picture like the Nature front cover using (a) 3PCs (b) 7 PCs?
(And, if you have time, what happens with more PCS? And what happens if you exclude data from the Islands like Grytviken etc?…)
[snip, off topic – nothing to do with Antarctica]
Thanks again for the support.
Paul S makes a good point.
I’m curious about the amazing silence on Tamino and RC. They usually take any shot they can at WUWT yet not one comment on their threads. I left a comment last night and it was clipped. Someone always takes a shot a Tamino, yet there’s nothing.
I would like to explore this a little further. I did a short post requesting simple reasonable, polite, on topic questions to RC regarding our result with a copy of the request placed in my comment thread. It will be interesting to see if anyone can get through and how they reply.
http://noconsensus.wordpress.com/2009/03/02/the-stunning-sound-of-silence-requests-for-reply/