Lovejoy's 99% 'confidence' vs. measurement uncertainty

By Christopher Monckton of Brenchley

It is time to be angry at the gruesome failure of peer review that allows publication of papers, such as the recent effusion of Professor Lovejoy of McGill University, which, in the gushing, widely-circulated press release that seems to accompany every mephitically ectoplasmic emanation from the Forces of Darkness these days, billed it thus:

“Statistical analysis rules out natural-warming hypothesis with more than 99 percent certainty.”

One thing anyone who studies any kind of physics knows is that claiming results to three standard deviations, or 99% confidence, requires – at minimum – that the data underlying the claim are exceptionally precise and trustworthy and, in particular, that the measurement error is minuscule.

Here is the Lovejoy paper’s proposition:

“Let us … make the hypothesis that anthropogenic forcings are indeed dominant (skeptics may be assured that this hypothesis will be tested and indeed quantified in the following analysis). If this is true, then it is plausible that they do not significantly affect the type or amplitude of the natural variability, so that a simple model may suffice:

clip_image002 (1)

ΔTglobet is the measured mean global temperature anomaly, ΔTantht is the deterministic anthropogenic contribution, ΔTnatt is the (stochastic) natural variability (including the responses to the natural forcings), and Δεt is the measurement error. The last can be estimated from the differences between the various observed global series and their means; it is nearly independent of time scale [Lovejoy et al., 2013a] and sufficiently small (≈ ±0.03 K) that we ignore it.”

Just how likely is it that we can measure global mean surface temperature over time either as an absolute value or as an anomaly to a precision of less than 1/30 Cº? It cannot be done. Yet it was essential to Lovejoy’s fiction that he should pretend it could be done, for otherwise his laughable attempt to claim 99% certainty for yet another me-too, can-I-have-another-grant-please result using speculative modeling would have visibly failed at the first fence.

Some of the tamperings that have depressed temperature anomalies in the 1920s and 1930s to make warming this century seem worse than it really was are a great deal larger than a thirtieth of a Celsius degree.

Fig. 1 shows a notorious instance from New Zealand, courtesy of Bryan Leyland:

clip_image004

Figure 1. Annual New Zealand national mean surface temperature anomalies, 1990-2008, from NIWA, showing a warming rate of 0.3 Cº/century before “adjustment” and 1 Cº/century afterward. This “adjustment” is 23 times the Lovejoy measurement error.

 

clip_image006clip_image008

Figure 2: Tampering with the U.S. temperature record. The GISS record from 1990-2008 (right panel) shows 1934 0.1 Cº lower and 1998 0.3 Cº higher than the same record in its original 1999 version (left panel). This tampering, calculated to increase the apparent warming trend over the 20th century, is more than 13 times the tiny measurement error mentioned by Lovejoy. The startling changes to the dataset between the 1999 and 2008 versions, first noticed by Steven Goddard, are clearly seen if the two slides are repeatedly shown one after the other as a blink comparator.

Fig. 2 shows the effect of tampering with the temperature record at both ends of the 20th century to sex up the warming rate. The practice is surprisingly widespread. There are similar examples from many records in several countries.

But what is quantified, because Professor Jones’ HadCRUT4 temperature series explicitly states it, is the magnitude of the combined measurement, coverage, and bias uncertainties in the data.

Measurement uncertainty arises because measurements are taken in different places under various conditions by different methods. Anthony Watts’ exposure of the poor siting of hundreds of U.S. temperature stations showed up how severe the problem is, with thermometers on airport taxiways, in car parks, by air-conditioning vents, close to sewage works, and so on.

(corrected paragraph) His campaign was so successful that the US climate community were shamed into shutting down or repositioning several poorly-sited temperature monitoring stations. Nevertheless, a network of several hundred ideally-sited stations with standardized equipment and reporting procedures, the Climate Reference Network, tends to show less warming than the older US Historical Climate Network.

That record showed – not greatly to skeptics’ surprise – a rate of warming noticeably slower than the shambolic legacy record. The new record was quietly shunted into a siding, seldom to be heard of again. It pointed to an inconvenient truth: some unknown but significant fraction of 20th-century global warming arose from old-fashioned measurement uncertainty.

Coverage uncertainty arises from the fact that temperature stations are not evenly spaced either spatially or temporally. There has been a startling decline in the number of temperature stations reporting to the global network: there were 6000 a couple of decades ago, but now there are closer to 1500.

Bias uncertainty arises from the fact that, as the improved network demonstrated all too painfully, the old network tends to be closer to human habitation than is ideal.

clip_image010

Figure 3. The monthly HadCRUT4 global temperature anomalies (dark blue) and least-squares trend (thick bright blue line), with the combined measurement, coverage, and bias uncertainties shown. Positive anomalies are green; negative are red.

Fig. 3 shows the HadCRUT4 anomalies since 1880, with the combined anomalies also shown. At present, the combined uncertainties are ±0.15 Cº, or almost a sixth of a Celsius degree up or down, over an interval of 0.3 Cº in total. This value, too, is an order of magnitude greater than the unrealistically tiny measurement error allowed for in Lovejoy’s equation (1).

The effect of the uncertainties is that for 18 years 2 months the HadCRUT4 global-temperature trend falls entirely within the zone of uncertainty (Fig. 4). Accordingly, we cannot tell even with 95% confidence whether any global warming at all has occurred since January 1996.

clip_image012

Figure 4. The HadCRUT4 monthly global mean surface temperature anomalies and trend, January 1996 to February 2014, with the zone of uncertainty (pale blue). Because the trend-line falls entirely within the zone of uncertainty, we cannot be even 95% confident that any global warming occurred over the entire 218-month period.

Now, if you and I know all this, do you suppose the peer reviewers did not know it? The measurement error was crucial to the thesis of the Lovejoy paper, yet the reviewers allowed him to get away with saying it was only 0.03 Cº when the oldest of the global datasets, and the one favored by the IPCC, actually publishes, every monthy, combined uncertainties that are ten times larger.

Let us be blunt. Not least because of those uncertainties, compounded by data tampering all over the world, it is impossible to determine climate sensitivity either to the claimed precision of 0.01 Cº or to 99% confidence from the temperature data.

For this reason alone, the headline conclusion in the fawning press release about the “99% certainty” that climate sensitivity is similar to the IPCC’s estimate is baseless. The order-of-magnitude error about the measurement uncertainties is enough on its own to doom the paper. There is a lot else wrong with it, but that is another story.

5 1 vote
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

268 Comments
Inline Feedbacks
View all comments
F.A.H.
April 12, 2014 4:46 am

the Knife 10:07 PM.
No offense intended at all in this comment. I am a huge fan of Einstein but that particular quote can not be reliably attributed to him, although it is widely reported to have been said by him. I have tried to find it and the most frequently given source is an attribution by Alice Calaprice in Quotable Einstein (1996, 2005), in which she attributes it to an issue of the Sunday Times (whose date I have unfortunately forgotten, but I think it was in the 50’s or 60’s. The issue is available online and I have looked at it (I think in JSTOR or some such) and I have been unable to find it. I am afraid it may be a widely attributed quote for which there is no reliable source. It sounds like something he would have said, if he had been asked. If you can find the source I would be grateful.

Gerald Machnee
April 12, 2014 5:04 am

Mosher says 10,000.
Yes, but most of them have no meaning or history.
How many of them have 100 years of history and ca be used to show a long term history?
Not many.

Doug Huffman
April 12, 2014 5:21 am

Dilettantism, per se, in no way diminishes the quality of ones science, indeed it removes hypothetical confounding conflicts of interest. Some suggest that science is harmed by the loss of the non-professional scientist.

Bertram Felden
April 12, 2014 5:22 am

To sum up Monckton’s argument one need only go back to statistics 1.01. Garbage in = garbage out.

Julian Flood
April 12, 2014 5:32 am

My Lord,
Mosh above mentions that he, too, has used the assumption that CO2 emissions can serve as a proxy for all anthropogenic warming influences. This, surely, cannot be sensible.
While CO2 forcing is logarithmic, what about UHI? Does it saturate? The paper admits that the understanding of aerosols is inadequate, so is lumping aerosol warming/cooling in with CO2 forcing justifiable? Take my own favourite outsiders in the AGW race, dissolved silica and oil/surfactant pollution of the ocean surface. The first will initially make a steady change in light isotope C left in the atmosphere as less heavy C is sequestered by non-DMS-producing diatoms, the latter will reduce cloud cover as salt and DMS aerosol production is suppressed and warming enhanced by surface effects, but both will saturate and the effect will level off, whereas CO2 will, according to theory, keep warming forever.
The notion of climate sensitivity, defined as the response to a doubling of atmospheric CO2, does not make sense if one simply throws all warming influences into the pot. If humans were only effecting physical processes it might do so, but not where biology is involved.
JF

Tom J
April 12, 2014 5:35 am

Hate to say it but I think everybody is overreacting. Our poor Professor Lovejoy is not really a Killjoy after all. Let us read this line carefully:
“Statistical analysis rules out natural-warming hypothesis with more than 99 percent certainty.”
See, all his analysis has done is rule out any natural warming hypothesis that is ‘more’ than 99% certain. A natural warming hypothesis that is ‘less’ than 99% certain has not been ruled out at all and since a hypothesis, by its very definition, cannot be considered a ‘law’ his analysis essentially cannot rule out any natural warming hypothesis whatsoever.
Now, certainly we cannot expect our good professor to be such an incredibly blithering idiot, such a stupendously stupid individual, and to still be employed at such a renowned, and enlightened university of higher knowledge, if our good professor were to imbecilically state that his research does the exact opposite of what it does.

DirkH
April 12, 2014 5:40 am

In his precision claim of 0.03 K Lovejoy ignores that temperatures cannot be meaningfully averaged. Cauchy type distribution, law of Large Numbers does not hold. Averaging a thousand stations or what have you does therefore not improve the precision as would be expected with normal distributions.
http://chiefio.wordpress.com/2012/12/10/do-temperatures-have-a-mean/

Janice
April 12, 2014 5:46 am

[With apologies to A. Einstein]
No amount of peer-review can ever prove me right; a single well-thought root-cause-analysis of measurement uncertainty can prove me wrong.

HankHenry
April 12, 2014 5:52 am

I wish people would be clearer. When they speak of “surface temperature” they are usually referencing surface air temperature. Considering the typical temperature of the abyss, this is not a minor point. If you wanted to know how the temperature of a car’s coolant varies, measuring temperatures at the tailpipe wouldn’t be the best approach. I like still like Willis’s article on ocean heat as measured by Argo floats http://wattsupwiththat.com/2014/03/02/argo-temperature-and-ohc/ . If accurate, I’m not sure his figure of .03 degrees of warming per decade isn’t worrisome.

commieBob
April 12, 2014 6:17 am

Here’s a simple experiment I give my students:
Create some data using y = mx + b + noise(10%)
Do a linear regression.
The students will find the following:
The r-value and p-value will be really good.
Both m’ and b’ will be significantly wrong.
The standard error of the estimate will be garbage.
standard error of estimate
Just because a statistician says something with 99.9% confidence it doesn’t mean that it isn’t junk. Wm Briggs has commented widely on the problem, here’s one example: link

Jim s
April 12, 2014 6:22 am

I find the entire concept of a Global Average Surface Temperature absurd. Assuming for one second that such an entity exists. How would one go about measuring it? Just watch the weather report on your local TV and you will see different temperatures for reporting stations with in a few miles of one another. And that is for urban/suburban areas, rural and wilderness areas have even fewer reporting stations. Now recall that the Earth’s surface is over 70% water with very few reporting stations (ships) , the problem of computing the mythical Global Surface Temperature becomes unsolvable. But it gets even worse, since every year an Annual Global Mean Temperature is reported to an accuracy of 0.1C. Really? Wow.
Now a useful entity that might acutely exists would be a total atmospheric heat content. And with enough satellites measuring temperatures/pressures at various altitudes you might even be able to compute that.
Another thing that annoys me is the criticism our learned author receives for using satellite data. Satellite data is the BEST DATA we have. It’s 24/7 global data with no agenda!
I am dumb founded that someone with a PhD in physics would claim 99% certainly on a meaningless non existent entity which is unmeasurable with our current instrumentation. Had I turned in such a paper to my first year physics teacher I would have been given a F and possibly even dismissed from the class.

DR
April 12, 2014 6:37 am

At least Mosher finally admitted BEST alters data. Everyone knew that, but really Mosher, why should anyone take you seriously?
http://www.woodfortrees.org/plot/best/from:1979/offset:-0.40/mean:60/plot/rss-land/from:1979/mean:60

Mark Bofill
April 12, 2014 6:42 am

Stated succinctly, our statistical hypothesis on the natural variability is that its extreme probabilities (Pr<3%) are bracketed by a modified Gaussian with qD between 4 and 6 and with standard deviation (and uncertainties) given by the scaling of the multiproxies in fig. 7:
σ125 = 0.20±0.03 K.

emphasis added

I’m not conversant with Haar fluctuations and fluctuation analysis, so I’m going to be puzzling over the math in this paper for awhile. Still, the section above made my hackles rise. We’re getting the standard deviation by scaling multiproxies and coming up with 0.2 plus or minus 0.03K huh.
I question whether one can make such determinations from the proxies. Instruments, sure ok. Proxies? Maybe so, but I’d like to find something to convince me of it before I swallow that.

Rich Lambert
April 12, 2014 6:44 am

I propose that the names of all reviewers of peer reviewed papers appear on the papers.

Jean Parisot
April 12, 2014 6:47 am

When did climate science reach the conclusion that land use changes were less important then CO2 as an anthropological forcing?

Coach Springer
April 12, 2014 6:59 am

When it’s been warmer before, certainty that it’s not natural is not science, it’s a statistical mind game filled with error and assumption.

Damian
April 12, 2014 7:00 am

As the saying goes When All Else Fails, Manipulate the Data.

jeanparisot
April 12, 2014 7:05 am

Since we aren’t warming, doesn’t that address the natural warming hypotheses with at least 95% certainty?

Mark Bofill
April 12, 2014 7:10 am

The other thing I’m wondering is, what conclusion would we come to applying this statistical technique to the Little Ice Age? It seems to me that the temperature variance of the LIA isn’t established (when I look at the most dreaded <a href=http://en.wikipedia.org/wiki/File:2000_Year_Temperature_Comparison.png wikipedia page for example), and the result would depend heavily on the proxies you used. But if you buy that temperatures were .5C lower than ‘normal’ in the depths of the LIA, I find it hard to swallow that we can say with 99% confidence that running .85C high can’t possibly be natural variation.

Mark Bofill
April 12, 2014 7:11 am

oops, formatting error.
The other thing I’m wondering is, what conclusion would we come to applying this statistical technique to the Little Ice Age? It seems to me that the temperature variance of the LIA isn’t established (when I look at the most dreaded wikipedia page for example), and the result would depend heavily on the proxies you used. But if you buy that temperatures were .5C lower than ‘normal’ in the depths of the LIA, I find it hard to swallow that we can say with 99% confidence that running .85C high can’t possibly be natural variation.

Carrick
April 12, 2014 7:25 am

One thing anyone who studies any kind of physics knows is that claiming results to three standard deviations, or 99% confidence, requires – at minimum – that the data underlying the claim are exceptionally precise and trustworthy and, in particular, that the measurement error is minuscule.

Three sigma is 99.7% not 99%. In high energy, typically five sigma is required though, not three.
Discussed here.

John Archer
April 12, 2014 7:32 am

Lord Monckton,
Shouldn’t the note to your Figure 2 rather read as follows?

Figure 2: Tampering with the U.S. temperature record. The GISS record from 1990-2008 (right panel) shows 1934 0.1 Cº higher lower and 1998 0.3 Cº lower higher than the same record in its original 1999 version (left panel).

But maybe you’re writing from Australia, in which case of course it makes perfect sense. 🙂
By the way, I think this sort of ‘tampering’ by the klimatophiliacs should be given a suitable name. How about calling it a ‘savile data massage‘?

Jan Lindström
April 12, 2014 7:33 am

Well done, my Lord. Don´t forget this one either: Beenstock, M., Reingewertz, Y., and Paldor, N.: Polynomial cointegration tests of anthropogenic impact on global warming, Earth Syst. Dynam., 3, 173-188, doi:10.5194/esd-3-173-2012, 2012. “This means, however, that as with all hypotheses, our rejection of AGW is not absolute; it might be a false positive, and we cannot rule out the possibility that recent global warming has an anthropogenic footprint. However, this possibility is very small, and is not statistically significant at conventional levels”. In your face Lovejoy!

Carbon500
April 12, 2014 7:34 am

‘Pokerguy’: I have to to disagree regarding what you describe as ‘arcane words’.
I thought Lord Monckton’s description of the offending Lovejoy paper as an ‘effusion’ and a ‘mephitically ectoplasmic emanation from the Forces of Darkness’ hugely entertaining.
Sadly the use of colourful English seems to be in decline. Shades of meaning and subtlety are thus being progressively removed, and so we move ever closer to ‘newspeak’.
Yes, I agree that his description isn’t what is typically seen in a scientific paper – but thank goodness that some welcome humour is brought into the discussion!

Greg Goodman
April 12, 2014 7:37 am

Magma says:
A short comparison
S. Lovejoy: Physics PhD; climate, meteorological and statistical expert; 500+ publications
C. Monckton: Classics BA; nil; 1 (not peer reviewed)
===
Thank you Magma. So we are correct to conclude, from his level of qualification and experience, that Lovejoy’s paper is deliberately and deceitfully misleading. In other words a scientific fraud.
It is long over due that this sort of behaviour is called out and those writing, publishing an peer-reviewing such scientific fraud are called to account.