Guest Post by Willis Eschenbach
A while back, I noticed an oddity about the Hadley Centre’s HadISST sea ice dataset for the Arctic. There’s a big change in variation from the pre- to the post-satellite era. Satellite measurements of ice areas began in 1979. Here is the full HadISST record, with the monthly variations removed.
Figure 1. Anomaly in the monthly sea ice coverage as reported by the HadISST, GISST, and Reynolds datasets. All data are from KNMI. Monthly average variations from the overlap period (1981-1994) have been subtracted from each dataset. All data are from KNMI (see Monthly Observations).
There’s a few points of note. First, the pre-1953 data is pretty useless, much of it is obviously not changing from year to year. Second, although the variation in the GISST dataset is doesn’t change in 1979, the variation in the HadISST dataset changes pretty radically at that point. Third, there is a large difference between the variability of the Reynolds and the GISST datasets during the period of their overlap.
I had filed this under unexplained curiosities and forgotten about it … until the recent publication of a paper called Observations reveal external driver for Arctic sea-ice retreat, by Notz and Marotzke, hereinafter N&M2012
Why did their paper bring this issue to the fore?
Well, the problem is that the observations they use to establish their case are the difference in variability of the HadISST during period 1953-1979, compared to the HadISST variations since that time. They look at the early variations, and they use them as “a good estimate of internal variability”. I have problems with this assumption in general due to the short length of time (25 years), which is way too little data to establish “internal variability” even if the data were good … but it’s not good, it has problems.
To their credit, the authors recognized the problems in N&M2012, saying:
Second, from 1979 onwards the HadISST data set is primarily based on satellite observations. We find across the 1978/1979 boundary an unusually large increase in sea-ice extent in March and an unusually large decrease in sea-ice extent in September (Figures 1b and 1d). This indicates a possible inconsistency in the data set across this boundary.
Ya think? I love these guys, “possible inconsistency”. The use of this kind of weasel words. like “may” and “might” and “could” and “possible”, is Cain’s mark on the post-normal scientist. Let me remove the GISST and Reynolds datasets and plot just the modern period that they use, to see if you can spot their “possible inconsistency” between the 1953-1979 and the post 1979 periods…
Figure 2. As in Figure 1, for HadISST only.
The inconsistency is clearly visible, with the variability of the pre- and post-1979 periods being very different.
As a result, what they are doing is comparing apples and oranges. They are assuming the 1953-1979 record is the “natural variability”, and then they are comparing that to the variability of the post-1979 period … I’m sorry, but you just can’t do that. You can’t compare one dataset with another when they are based on two totally different types of measurements, satellite and ground, especially when there is an obvious inconsistency between the two.
In addition, since the GISST dataset doesn’t contain the large change in variability seen in the HadISST dataset, it is at least a working assumption that there is some structural error in the HadISST dataset … but the authors just ignore that and move forwards.
Finally, we have a problematic underlying assumption that involves something called “stationarity”. The stationarity assumption says that the various statistical measures (average, standard deviation, variation) are “stationary”, meaning that they don’t change over time.
They nod their heads to the stationarity problem, saying (emphasis mine):
For the long-term memory process, we estimate the Hurst coefficient H of the pre-satellite time series using detrended fluctuation analysis (DFA) [Peng et al., 1994]. Only a rough estimate of 0.8 < H < 0.9 is possible both because of the short length of the time series and because DFA shows non-stationarity even after removal of the seasonal cycle.
Unfortunately, they don’t follow the problem of non-stationarity to its logical conclusion. Look, for example, at the variability in the satellite record in the period 1990-2000 versus the period 2000-2005. They are quite different. In their analysis, they claim that a difference in variability pre- to post-1979 establishes that human actions are the “external driver” … but they don’t deal with the differences pre- and post-2000, or with the fact that their own analysis shows that even the variability of the pre-1979 data is not stationary.
Finally, look at the large change in variability in the most recent part of the record. The authors don’t mention that … but the HadISST folks do.
03/DECEMBER/2010. The SSM/I satellite that was used to provide the data for the sea ice analysis in HadISST suffered a significant degradation in performance through January and February 2009. The problem affected HadISST fields from January 2009 and probably causes an underestimate of ice extent and concentration. It also affected sea surface temperatures in sea ice areas because the SSTs are estimated from the sea ice concentration (see Rayner et al. 2003). As of 3rd December 2010 we have reprocessed the data from January 2009 to the present using a different sea ice data source. This is an improvement on the previous situation, but users should still note that the switch of data source at the start of 2009 might introduce a discontinuity into the record. The reprocessed files are available from the main data page. The older version of the data set is archived here.
08/MARCH/2011. The switch of satellite source data at the start of 2009 introduced a discontinuity in the fields of sea ice in both the Arctic and Antarctic.
Curious … the degradation in the recent satellite data “probably causes an underestimate of ice extent and concentration,” and yet it is precisely that low recent ice concentration that they claim “reveals an external driver” …
In any case, when I put all of those problems together, the changes in variability in 1979, in 2000, and in 2009, plus the demonstrated non-stationarity pre-1979, plus the indirect evidence from the GISST and Reynolds datasets, plus the problems with the satellites affecting the critical recent period, the period they claim is statistically significant in their analysis … well, given all that I’d say that the N&M2012 method (comparing variability pre- and post-1979) is totally inappropriate given the available data. There are far too many changes and differences in variability, both internal to and between the datasets, to claim that the 1979 change in variability means anything at all … much less that it reveals an “external driver” for the changes in Arctic sea ice.
w.
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.

Obviously another example of groupthink analysis by an unsupervised graduate student. Forward!
“Ya think? I love these guys, “possible inconsistency”. The use of this kind of weasel words. like “may” and “might” and “could” and “possible”, is Cain’s mark on the post-normal scientist.
Struth mate!
Let me tell you an issue I’d like to see someone look into … I find it funny that the variability of the arctic ice between 1997 and 2007 seems awfully constant and compressed when compared to 1979-1997 and 2007 to present. Though it does make a nice tight trend downwards .. the variability just seems very uncharacteristic when compared to the rest of the record.
Kinda makes me think that the data has been cooked. Now .. that couldn’t be .. I mean the CAGW crowd would never cook the data … right??
There could be a man-made external forcing. Check NoTricksZone tomorrow.
http://notrickszone.com/
I hope they will let the snow crab fishing boats which are iced in and cannot go after them , that the ice has melted
But if they don’t torture the data, then how is that little rascal Mankind to confess it’s sins? Poor widdle earthy worthy got all broke-d cuz the bad mans got all pwoffity woffitty, uh huh…
You’ve all heard of the 350 org, I hereby announce the 500 org.
“Curious … the degradation in the recent satellite data ‘probably causes an underestimate of ice extent and concentration,’ and yet it is precisely that low recent ice concentration that they claim ‘reveals an external driver’ …”
In the vernacular and as detailed by the musical group Tag Team, “Whomp, there it is.”
Admittedly, though, N&M2012 asserts that their purported negative, linear trend is across the entire sample time period – on the presumption the pre-satellite and satellite data reported similar magnitudes attributed to internal variability. However, Mr. Eschenbach presented a strong argument that the collection methodologies associated with the two data sets render subsequent comparisons largely suspect.
Which narrative would you regard as more accurate – (a) the paragraph you read without your glasses in which you confidently reported seven out of every ten words (due to the blur factor) or (b) the paragraph you read with your glasses in which you confidently reported ten out of every ten words (due to the clarity factor)? Now, add a gouge to the glasses (i.e., a recent and significant degradation in performance) after the dog knocked you over with one last sentence to read and even this accuracy decreases, as well.
Common sense, albeit subjective, suggests that any comparison between the two data sets must default to the use of a broad brush rather than a narrow, more precise one (Monte-Carlo simulations notwithstanding due to the lies, damned lies, and statistics factor).
“The use of this kind of weasel words. like “may” and “might” and “could” and “possible”, is Cain’s mark on the post-normal scientist.”
Not really. Having written my share of scientific reports, I can say that those words are part of the normal lexicon. It is the absence of those words, not their presence, that identifies pseudoscience. The disconnect lies in skipping from a scientific paper directly to policy, conveniently omitting the qualifiers to achieve the desired result.
But I agree; the paper is severely flawed, as you state. . Notz & Marotz note a discontinuity, assume that it represents human effects, base their analytical statistics on short-term data before the discontinuity, then cite the short-term post-discontinuity data as proof of human effects. This is nothing more than a fanciful way of begging the question. Worse, they attribute the presence of a sudden discontinuity to human CO2 emissions, though the latter are gradual and the former is abrupt. This is post-doc, post hoc thinking, proof that they’ve assumed an understanding of the Arctic ice system that they don’t really possess.
They also state “Arctic sea ice is currently retreating rapidly…” though as of this date the NORSEX Arctic Ice area is within a hairbreadth of the 1979-2006 average. http://arctic-roos.org/observations/satellite-data/sea-ice/observation_images/ssmi1_ice_area.png
Too much grant money, chasing too few opportunities for real climate science.
The 2007 onwards increase in variability does seem a real phenomena resulting from increased ice ‘melt’ and increased ice formation on an annual basis.
jorgekafkazar says:
May 3, 2012 at 7:50 pm
I disagree. A scientific statement would be something like “the value of the variable is 3.6 ± 1.2 units (one standard deviation).
The corresponding pseudo-scientific statement is “the value of the variable may be greater than 6”.
Both statements could be true. The value could in fact be 6.5. But only one of the statements is science.
I see this all the time, in forms like Hansen’s infamous claim that the increase in CO2 could lead to a 20 metre increase in sea level by the end of the century. Well, yes, it could, and I could win the lottery tomorrow, but that’s not science.
Science is generally distinguished, in fact, by the use of error estimates rather than weasel words like “may occur” and “could lead to”.
However, Jorge, I do agree with you that “those words are part of the normal lexicon”.
That’s the problem …
w.
The human drivers in this case are the “scientists” playing with numbers. Crunch ’em this way, then crunch ’em that, until the only “logical” explanation is fumes from their Cadillacs.
They call me Baby Driver
And once upon a pair of wheels
Hit the road and I`m gone ah
What`s my number
I wonder how your engine feels.
Willis is quite correct here in both his analysis and in the lexicon of science speak. My take on it it is simply another run as making the data fit the theory.
Hi Willis,
the pre-1953 data is not pretty useless.
The artificial step change in the early 50s stands out so obviously, that its use is clear: cheat.
Another artificial step change may be around 1973, perhaps inserted with an instrument change, but as always in the direction to support alarm.
This is brilliant — it dismisses the statistical analysis used in a peer-reviewed paper in a prestigious journal without actually doing any statistics! Only a man without the intellectual blinders of advanced degree in science, without experience in scientific research, or any without any kind background in mathematics more advanced than algebra could so effortlessly dismiss the scientific work of a couple of egghead PhD research geeks. We should all reread D&K (2000) and pondor admiringly on Mr. Eschenbach’s achievements in science and statistics.
Willis:
Nice work again. And your conclusion that the changes in the time series are the results of changes in measurement techniques and changes in instrument performance, not CO2 forcing or any other global warming terminology appear logical and supported by the data. The authors’ rush to man-caused change is simply not evident in the data. Again, nice work and thank you for all your hard work in our behalf.
This is brilliant — it dismisses the statistical analysis used in a peer-reviewed paper in a prestigious journal without actually doing any statistics!
People don’t realize how much poor or bad science gets published. A prominent scientist I know very well, who evaluates published studies for the purpose of formulating public policy tells me that most of the published studies in his field are worthless. Specifically, the conclusions they draw aren’t warranted by the data they present.
Which, in a nutshell, is Willis’s critique above.
peter says:
May 3, 2012 at 9:25 pm
Ah, yes, the argument from authority. Please notice, folks, that it appears he cannot find any errors in my logic or my facts or my claims. As a result, he is reduced to saying but, but, but the authors have PhDs! It’s peer reviewed! The journal is prestigious! All overlaid with a thick coat of satire.
My achievements in science and statistics are a matter of public record, both in the peer reviewed journals and here on this blog. I have four peer-reviewed publications in the scientific journals, including a peer-reviewed “Communications Arising” in Nature magazine. They had no problem with my qualifications or achievements. Regarding my math-fu, it’s strong, and there is always more to learn. I got credit for a year of college calculus when I was still in high school, and never looked back. I wrote my first computer program in 1963, and have continued programming to this day. And yes, I do make stupid math errors sometimes, and when I do, it’s very public. The Argus-eyed intarwebs assure that I get called on it and admit it and continue to learn.
But that’s not the point, peter. It’s not about me … and it’s not about the authors or the journals either. It’s about the science, the ideas. Nothing else matters. Doesn’t matter who said it, doesn’t matter if they wrote it on the bathroom wall. All that matters is, is it true? Is it consistent? Does it hold up under attack?
Notice, folks, that peter cannot find any fault in the actual science and statistical concepts and arguments that I have presented here.
peter, my statistics are reasonably good, but the underlying problem with their analysis is not statistical. Or rather, the problem is statistical but it lies in the area of the logic of statistics, rather than what you call the doing of statistics. That’s why I’m not “doing any statistics”, even though I can do them. That’s not where the problem lies.
The problem is that the authors are comparing two separate, non-stationary datasets taken by different groups of people at different times using different measuring tools. Here’s the problem:
You can’t draw any conclusions from comparing the variations of two such datasets.
It’s a fools errand. There is no reason to assume that the mean, variance, or distribution of the two are related at all. They may be related, but if they are not, this does not “reveal an external driver”, that’s nonsense.
Finally, appeals to authority such as you have made above, even if they are as clever and sarcastic as yours, don’t get any traction here. At WUWT we do science. If you have objections to my science, to my logic, to my data, to my mathematics, to my statistics, then QUOTE MY WORDS THAT YOU OBJECT TO and explain my error, which I know to be manifold.
Anything else, and you’re just spinning your wheels …
w.
>>
This is brilliant — it dismisses the statistical analysis used in a peer-reviewed paper in a prestigious journal without actually doing any statistics!
>>
Brilliant is perhaps an over-statement but it is a very good article by Willis.
He does not dismiss that statistical analysis itself, he is quite clearly showing why any such analysis is invalid before it is even calculated because the data are incomparable AND are explicitly said to be so by the authors that publish the data.
Not only is this paper flawed science , it is knowingly flawed and hence dishonest science.
If GRL was as “prestigious” as you suggest, they would not publish such obviously false rubbish.
Arctic ice extent depends on Norther Atlantic SST
http://i51.tinypic.com/oqlpxi.jpg
and North Atlantic SST goes 30 years up and 30 years down
http://i51.tinypic.com/k8epd.jpg
so the pre-satellite steady ice extent is nonsense. There is a study which observes Barents sea ice following the AMO in the same wave pattern.
One hint – NW passage was open in 1942-44 and then only in 2007. Both on the top of the warm SST peaks.
peter says:
May 3, 2012 at 9:25 pm
On the contrary, Peter. What this case shows is the worthlessness of peer review – how was it that none of the peer reviewers picked up these relatively straightforward problems?
You’d expect the peer reviewers to be knowledgeable about the subject matter of the paper, right? In which case, you’d also expect them to be aware of the available datasets, right? If so, shouldn’t they have been asking the quite basic questions that Willis has asked here? And at least requiring the authors to address those questions properly?
This kind of paper simply brings the whole system into disrepute.
This makes the case for the abandonment of peer review for something more open, where anyone who is interested can make open comments on a draft paper for all to see and where the authors’ response is also done openly. The true spirit of open enquiry.
RE
Willis Eschenbach says:
May 3, 2012 at 11:00 pm
peter says:
May 3, 2012 at 9:25 pm
———-
Willis pins the issue brilliantly as usual. The big problem for the peter-warmists of the world is that they have so few real world data they can hope to conjour into a form that even remotely looks like it supports the failed CAGW hypothesis that they feel absolutely compelled to defend this type of Post Normal Science (PNS) to the last man as it were – because if they don’t they know the entire edifice of “evidence” will disappear in a flash of light and puff of smoke.
The Shakun et. al. paper delusion falls into the ‘same sh*t different shovel’ category in its shameless cherry picking of proxies, as does the entire contorted, twisted and battered land-based temp record, the entire catalogue of IPCC spin, the various attempts by Santer and friends to cover for the missing ‘hotspot’ using wind shear measurements in preference to radiosonde thermometer readings, the Mann-schtick trick flick etc.etc.
peter, are you not in any way embarrassed by your ‘play the man’ attacks? Do you really have such little sham and insight? And you cultists wonder why people generally have become so suspicious with your entire story-line.
First of all, the referenced paper is completely wrong about the “beginning” of the satellite era. There is an almost continuous record of Arctic and Antarctic ice from satellites,stretching back to the early 1960s’.
My team has personally helped the NSIDC with Nimbus I, II, and III HRIR data. Beyond that, it has recently been discovered that there is a film copy of the Nimbus AVCS (visible light) images.
We need to put this 1979 boundary crap to bed.
Makes me wonder if we have almost exhausted the limits of truly extraordanary scientific discoveries? So they are having to make stuff up to keep winning prizes and getting money off the gullable. Most branches of science seem to be at the same place, few to no new dicoveries, just the occational improvement, occational backward steps.
Part which intrigues me is the statement from the ‘HadIIST folks’ that SST are ‘SSTs are estimated from the sea ice concentration’.
By concentration do they mean thickness or extent or a combination of both? Does this take account of wind driven ice piled thicker or more closely driven together and changing the extent? If th temperature of the sea is 0.5 or 1 deg C colder or warmer than the overlying ice – that wouldn’t seem to figure in their calculations so how do they correct for that?
Is it simply more of the guesstimation that seems to be a theme of the Post Normal Science philosophy that too many climate scientists seem o be adherents of ?