UPDATE – BOMBSHELL: audit of global warming data finds it riddled with errors

I’m bringing this back to the top for discussion, mainly because Steven Mosher was being a cad in comments, wailing about “not checking”, claiming McLean’s PhD thesis was “toast”, while at the same time not bothering to check himself. See the update below. – Anthony


Just ahead of a new report from the IPCC, dubbed SR#15 about to be released today, we have this bombshell- a detailed audit shows the surface temperature data is unfit for purpose. The first ever audit of the world’s most important temperature data set (HadCRUT4) has found it to be so riddled with errors and “freakishly improbable data”  that it is effectively useless.

From the IPCC:

Global Warming of 1.5 °C, an IPCC special report on the impacts of global warming of 1.5 °C above pre-industrial levels and related global greenhouse gas emission pathways, in the context of strengthening the global response to the threat of climate change, sustainable development, and efforts to eradicate poverty.

This is what consensus science brings you – groupthink with no quality control.

HadCRUT4 is the primary global temperature dataset used by the Intergovernmental Panel on Climate Change (IPCC) to make its dramatic claims about “man-made global warming”.  It’s also the dataset at the center of “ClimateGate” from 2009, managed by the Climate Research Unit (CRU) at East Anglia University.

The audit finds more than 70 areas of concern about data quality and accuracy.

But according to an analysis by Australian researcher John McLean it’s far too sloppy to be taken seriously even by climate scientists, let alone a body as influential as the IPCC or by the governments of the world.

Main points:

  • The Hadley data is one of the most cited, most important databases for climate modeling, and thus for policies involving billions of dollars.
  • McLean found freakishly improbable data, and systematic adjustment errors , large gaps where there is no data, location errors, Fahrenheit temperatures reported as Celsius, and spelling errors.
  • Almost no quality control checks have been done: outliers that are obvious mistakes have not been corrected – one town in Columbia spent three months in 1978 at an average daily temperature of over 80 degrees C.  One town in Romania stepped out from summer in 1953 straight into a month of Spring at minus 46°C. These are supposedly “average” temperatures for a full month at a time. St Kitts, a Caribbean island, was recorded at 0°C for a whole month, and twice!
  • Temperatures for the entire Southern Hemisphere in 1850 and for the next three years are calculated from just one site in Indonesia and some random ships.
  • Sea surface temperatures represent 70% of the Earth’s surface, but some measurements come from ships which are logged at locations 100km inland. Others are in harbors which are hardly representative of the open ocean.
  • When a thermometer is relocated to a new site, the adjustment assumes that the old site was always built up and “heated” by concrete and buildings. In reality, the artificial warming probably crept in slowly. By correcting for buildings that likely didn’t exist in 1880, old records are artificially cooled. Adjustments for a few site changes can create a whole century of artificial warming trends.

Details of the worst outliers

  • For April, June and July of 1978 Apto Uto (Colombia, ID:800890)  had an average monthly temperature of  81.5°C, 83.4°C and 83.4°C respectively.
  • The monthly mean temperature in September 1953 at Paltinis, Romania is reported as -46.4 °C (in other years the September average was about 11.5°C).
  • At Golden Rock Airport, on the island of St Kitts in the Caribbean, mean monthly temperatures for December in 1981 and 1984 are reported as 0.0°C. But from 1971 to 1990 the average in all the other years was 26.0°C.

More at Jo Nova


The report:

Unfortunately, the report is paywalled. The good news is that it’s a mere $8.

The researcher, John McLean, did all the work on his own, so it is a way to get compensated for all the time and effort put into it. He writes:

This report is based on a thesis for my PhD, which was awarded in December 2017 by James Cook University, Townsville, Australia. The thesis1 was based on the HadCRUT4 dataset and associated files as they were in late January 2016. The thesis identified 27 issues of concern about the dataset.

The January 2018 versions of the files contained not just updates for the intervening 24 months, but also additional observation stations and consequent changes in the monthly global average temperature anomaly right back to the start of data in 1850.
The report uses January 2018 data and revises and extends the analysis performed in the original thesis, sometimes omitting minor issues, sometimes splitting major issues and sometimes analysing new areas and reporting on those findings.

The thesis was examined by experts external to the university, revised in accordance with their comments and then accepted by the university. This process was at least equivalent to “peer review” as conducted by scientific journals.

I’ve purchased a copy, and I’ve reproduced the executive summary below. I urge readers to buy a copy and support this work.

Get it here:

Audit of the HadCRUT4 Global Temperature Dataset


EXECUTIVE SUMMARY

As far as can be ascertained, this is the first audit of the HadCRUT4 dataset, the main temperature dataset used in climate assessment reports from the Intergovernmental Panel on Climate Change (IPCC). Governments and the United Nations Framework Convention on Climate Change (UNFCCC) rely heavily on the IPCC reports so ultimately the temperature data needs to be accurate and reliable.

This audit shows that it is neither of those things.

More than 70 issues are identified, covering the entire process from the measurement of temperatures to the dataset’s creation, to data derived from it (such as averages) and to its eventual publication. The findings (shown in consolidated form Appendix 6) even include simple issues of obviously erroneous data, glossed-over sparsity of data, significant but questionable assumptions and temperature data that has been incorrectly adjusted in a way that exaggerates warming.

It finds, for example, an observation station reporting average monthly temperatures above 80°C, two instances of a station in the Caribbean reporting December average temperatures of 0°C and a Romanian station reporting a September average temperature of -45°C when the typical average in that month is 10°C. On top of that, some ships that measured sea temperatures reported their locations as more than 80km inland.

It appears that the suppliers of the land and sea temperature data failed to check for basic errors and the people who create the HadCRUT dataset didn’t find them and raise questions either.

The processing that creates the dataset does remove some errors but it uses a threshold set from two values calculated from part of the data but errors weren’t removed from that part before the two values were calculated.

Data sparsity is a real problem. The dataset starts in 1850 but for just over two years at the start of the record the only land-based data for the entire Southern Hemisphere came from a single observation station in Indonesia. At the end of five years just three stations reported data in that hemisphere. Global averages are calculated from the averages for each of the two hemispheres, so these few stations have a large influence on what’s supposedly “global”. Related to the amount of data is the percentage of the world (or hemisphere) that the data covers. According to the method of calculating coverage for the dataset, 50% global coverage wasn’t reached until 1906 and 50% of the Southern Hemisphere wasn’t reached until about
1950.

In May 1861 global coverage was a mere 12% – that’s less than one-eighth. In much of the 1860s and 1870s most of the supposedly global coverage was from Europe and its trade sea routes and ports, covering only about 13% of the Earth’s surface. To calculate averages from this data and refer to them as “global averages” is stretching credulity.

Another important finding of this audit is that many temperatures have been incorrectly adjusted. The adjustment of data aims to create a temperature record that would have resulted if the current observation stations and equipment had always measured the local temperature. Adjustments are typically made when station is relocated or its instruments or their housing replaced.

The typical method of adjusting data is to alter all previous values by the same amount. Applying this to situations that changed gradually (such as a growing city increasingly distorting the true temperature) is very wrong and it leaves the earlier data adjusted by more than it should have been. Observation stations might be relocated multiple times and with all previous data adjusted each time the very earliest data might be far below its correct value and the complete data record show an exaggerated warming trend.

The overall conclusion (see chapter 10) is that the data is not fit for global studies. Data prior to 1950 suffers from poor coverage and very likely multiple incorrect adjustments of station data. Data since that year has better coverage but still has the problem of data adjustments and a host of other issues mentioned in the audit.

Calculating the correct temperatures would require a huge amount of detailed data, time and effort, which is beyond the scope of this audit and perhaps even impossible. The primary conclusion of the audit is however that the dataset shows exaggerated warming and that global averages are far less certain than have been claimed.

One implication of the audit is that climate models have been tuned to match incorrect data, which would render incorrect their predictions of future temperatures and estimates of the human influence of temperatures.

Another implication is that the proposal that the Paris Climate Agreement adopt 1850-1899 averages as “indicative” of pre-industrial temperatures is fatally flawed. During that period global coverage is low – it averages 30% across that time – and many land-based temperatures are very likely to be excessively adjusted and therefore incorrect.

A third implication is that even if the IPCC’s claim that mankind has caused the majority of warming since 1950 is correct then the amount of such warming over what is almost 70 years could well be negligible. The question then arises as to whether the effort and cost of addressing it make any sense.

Ultimately it is the opinion of this author that the HadCRUT4 data, and any reports or claims based on it, do not form a credible basis for government policy on climate or for international agreements about supposed causes of climate change.


Full report here


UPDATE: 10/11/18

Some commenters on Twitter, and also here, including Steven Mosher, who said McLean’s thesis/PhD was “toast” seem to doubt that he was actually allowed to submit his thesis, and/or that it was accepted, thus negating his PhD. To that end, here is the proof.

McLean’s thesis appears on the James Cook University website:  “An audit of uncertainties in the HadCRUT4 temperature anomaly dataset plus the investigation of three other contemporary climate issues“, submitted for Ph.D. in physics from James Cook University (2017).

And, he was in fact awarded a PhD by JCU for that thesis.

Larry Kummer of Fabius Maximus directly contacted the University to confirm his degree. Here is the reply.

ADDED:

JOHN MCLEAN here.
For Mr Mosher,

I don’t insult and I don’t accuse without investigation. And if I don’t know I try to ask.

(a) Data files
If you want copies of the data that I used in the audit, as they were when I downloaded them in January, go to web page https://robert-boyle-publishing.com/audit-of-the-hadcrut4-global-temperature-dataset-mclean-2018/ and just scroll down.

Or download the latest versions of the files from yourself from the CRU and Hadley Centre, namely https://crudata.uea.ac.uk/cru/data/temperature/ and https://www.metoffice.gov.uk/hadobs/hadsst3/data/download.html. (The fact that file names are always the same and it’s confusing is one of the fidnings of the audit.)

(b) Apto Uto not used? Figure 6.3 shows that it is used, the lower than expected spikes are because of other stations in the same grid cell and the vale of the cell is the average anomaly for all such stations.

(c) What stations are used and what are not?
The old minimum of 20 years of the 30 from 1961 to 1990 was dropped a few HadCRUT versions back. It then went to 15 years with no more than 5 missing in any decade. HadCRUT4 reduced it again to 14.

best wishes

John

5 1 vote
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

512 Comments
Inline Feedbacks
View all comments
BallBounces
October 11, 2018 3:57 pm

“James Cook” sounds like a white male — which discredits the whole post.

John Tillman
Reply to  BallBounces
October 11, 2018 4:16 pm

Not just any white male, but the Royal Navy captain of voyages of discovery, famous for bringing venereal diseases to Pacific islands:

comment image

Killed in Hawaii, with the flesh scraped off his bones and returned to his crew wrapped in leaves, as a sign of respect. The British sailors didn’t know about the respect bit, however.

Robert
October 11, 2018 4:19 pm

I haven’t read all of the posts so this may have already been addressed.
I mentioned this audit on another website and the response was that all of the correction have been made prior to the temperature data being added to the data set. Therefore this audit was of no use as all of the problems had been corrected.
It was also stated that other temperature data sets were used and produced the same results. Therefore the problems pointed out could not have resulted in significant problems. Now I have learned quite a bit about climate science over the past several years but I will be the first to admit that the depth of my ignorance is still huge. Are these valid criticisms? They do not sound like valid criticisms to me.

Scott Bennett
Reply to  Robert
October 11, 2018 5:44 pm

No it is not valid Criticism, that HadCRUT4 has errors and uncertainties is accepted by all sides of the debate.

The recent “criticism” is simply misdirection because it didn’t address the specific issue raised in the audit (Though it did concern the use of poor raw data.) which is, that though correction were made and stations removed, major systematic errors* remain in the current 2018 database that can not be corrected without creating even larger uncertainty!

More importantly this problem is only one of the 25 major issues discovered!

In short, the world may be warming since 1850 or not but HadCRUT4 has nothing to say about it except weak talking points** for the IPCC!

*There is a real problem that remains in the database caused by the use of uncorrected raw data that is used in the correction process itself. A systematic error that continues to this day.
**”Authors of the IPCC’s Fifth Climate Assessment Report (2013) admitted during the review process for that report that no audit of the HadCRUT4 dataset or any associated dataset had been undertaken.”

MarkW
Reply to  Robert
October 12, 2018 7:41 am

The point is that most of the problems detailed can’t be corrected.
How do you correct for wrong or missing raw data?
How do you correct for woefully inadequate surface area coverage?

The claim is that this data can be used to determine the temperature of the entire earth to within 0.1C.
Just examining the lack of coverage is enough to disprove that claim. These other problems just make the claim more ridiculous.

John McLean
October 11, 2018 6:44 pm

JOHN MCLEAN here.
For Mr Mosher,

I don’t insult and I don’t accuse without investigation. And if I don’t know I try to ask.

(a) Data files
If you want copies of the data that I used in the audit, as they were when I downloaded them in January, go to web page https://robert-boyle-publishing.com/audit-of-the-hadcrut4-global-temperature-dataset-mclean-2018/ and just scroll down.

Or download the latest versions of the files from yourself from the CRU and Hadley Centre, namely https://crudata.uea.ac.uk/cru/data/temperature/ and https://www.metoffice.gov.uk/hadobs/hadsst3/data/download.html. (The fact that file names are always the same and it’s confusing is one of the fidnings of the audit.)

(b) Apto Uto not used? Figure 6.3 shows that it is used, the lower than expected spikes are because of other stations in the same grid cell and the vale of the cell is the average anomaly for all such stations.

(c) What stations are used and what are not?
The old minimum of 20 years of the 30 from 1961 to 1990 was dropped a few HadCRUT versions back. It then went to 15 years with no more than 5 missing in any decade. HadCRUT4 reduced it again to 14.

best wishes

John

Anthony Banton
Reply to  John McLean
October 12, 2018 1:28 am

“If you want copies of the data that I used in the audit, as they were when I downloaded them in January”

And what makes you think those files were included in Hadcrut without being QC’d first?
– and are not the files as originally sent by the national Met service involved?
And (correctly) recorded by the UKMO as being corrupted on receipt.
IOW: Please show that a particular file was used for Hadcrut as sent.

Staggering confirmation bias on display (yes I know of your “stance” regarding AGW)
The UKMO is well used to QC procedures – it uses them routinely in its “day job” – that massive quantities of data that arrives continually 24/7 for inclusion in it’s NWP models.
It’s just a basic requirement.

It would be staggering incompetence on the part of the UKMO (as Nick says – full of actual Phd recipients) – and anyone with knowledge of NWP observational data knows that they are certain to have errors.
So they don’t bother?
Really?
Oh, well if you say so – and the echo-chamber agrees, so it will enter into the Naysayers bible of myths.

Whiskey
October 11, 2018 6:51 pm

[snip -fake email address -mod]

John Tillman
Reply to  Whiskey
October 11, 2018 7:07 pm

Had you read more of the comments, you’d have noted that the alleged “defect” wasn’t.

But even had it been a fault, what about the dozens of other problems identified in the PhD thesis?

Have you considered the possibility that there aren’t any defects to be found?

But, again, please, if the defects are so obvious to you, share them with us.

Thanks!

Robert
Reply to  Whiskey
October 11, 2018 7:11 pm

You must not have actually read much of this post or you would be aware that those objections have been demonstrated to be invalid. But if you are like most alarmist that I have encountered you simply ignore facts which prove to be inconvenient.

Reply to  Whiskey
October 11, 2018 8:36 pm

“you would be aware that those objections have been demonstrated to be invalid”
Where? The objection is that he is critiquing raw data, not originally CRU’s, which then goes through a QC filter, which he doesn’t investigate.

But yes, I downloaded the thesis (some days ago). What stuck out for me was this quote:
“This thesis makes little attempt to quantify the uncertainties exposed by this investigation, save for some brief mention of the impact certain issues might have on error margins, because numerous issues are discussed, and it would be an enormous task to quantify the uncertainties associated with the many instances of each. It has been left to others to quantify the impact of incomplete data, inconsistencies, questionable assumptions, very likely data errors and questionable adjustments of the recorded data.

WTF? “left to others?”. How can you get a PhD saying that I did the proofreading, but calculations were to hard. And if a PhD project can’t do it, who are those others?

It isn’t an enormous task at all. HADCRUT isn’t rocket science. You just write a program that emulates it, and then see what happens when you correct hat you think is wrong with the data. I wrote a program years ago which I have run every month, before the major results come in (details and code here). I have done that for seven years. They are in good agreement. In particular the simplest of my methods, TempLS grid, gives results which are very close to HADCRUT. If I used 5° cells and hadsst3 instead of ERSST, I could make the agreement exact. I wouldn’t expect to get a PhD from doing that, let alone saying it was too hard.

scross
Reply to  Whiskey
October 11, 2018 8:39 pm

Well, some of us may prefer to carefully read through the entire report (all 135 pages of it), and maybe at least spot check some of the data for ourselves (there’s quite a bit of that data) before we make any comments about it. So there’s that. Also, while some here have pointed out potentially valid criticisms of the report itself, others here have pointed out that those criticisms may not actually be accurate. So there’s that, too.

Editor
Reply to  Whiskey
October 11, 2018 11:06 pm

Whiskey wrote:

> Wow. We are up to 406 comments and the only critique comes from Mosh and Stokes,

Look for a comment posted 7 minutes before yours, i.e. https://wattsupwiththat.com/2018/10/11/bombshell-audit-of-global-warming-data-finds-it-riddled-with-errors/#comment-2487365

Whiskey
October 11, 2018 7:34 pm

[snip – fake email address -mod]

John Tillman
Reply to  Whiskey
October 11, 2018 7:56 pm

Whiskey,

It is true. And it still appears that you didn’t read the relevant comments.

You missed the comments showing that HadCRU doesn’t do a quality control audit on the “data” before using them.

Nick asserted that they don’t use two of the sites cited in the paper, and couldn’t find a third, but that doesn’t mean that all “data” used by HadCRU have been checked. Nick was also wise enough to say that he “personally don’t use HADCRUT back to 1850”.

Whiskey
October 11, 2018 7:51 pm

[snip -fake email address -mod]

Chris Hanley
Reply to  Whiskey
October 11, 2018 8:16 pm

You’re writing gibberish, lay off the whiskey for a while — that’s just a suggestion.

Whiskey
October 11, 2018 8:13 pm

Your arguments are so weak, which prob’ly why Stokes didnt bother with them. 3 out of 3 cases are not found in the dataset, i forget which one. That’s 100%. So you should take it on yourself, oh mighty arguer, to come here with data, you know, sort of that sciency thing.

Find some data used that is wrong. Do it. Do it!

Then you won’t look so weak and irrelevant.

And as for the older parts of the data set, before 1900, what does it really matter? Until someone with a better data set comes up, that’s all we have. You and your friends should put a grant in to do it better. Last one like that was BEST. How did that turn out?

[fake- non functional email is against our WUWT commenting policy – mod]

Caligula Jones
Reply to  Whiskey
October 12, 2018 11:19 am

Hard to jibe:

“So you should take it on yourself, oh mighty arguer, to come here with data”

with

“prob’ly” and “i forget which one.”

Mad arguing skillz you got there. Did you get your name from imbibing?

Mairon62
October 11, 2018 8:59 pm

Granted I’m only 56 years old, but it just seems to me that most all the “problems” that the political left has tasked itself to solve don’t actually exist on a global basis or don’t exist at all…oh, but wait until you see their bill for services rendered. It seems that I may have accidentally stumbled upon a truly global problem; rent-seeking parasites masquerading as experts and leaders.

Chris
Reply to  Mairon62
October 12, 2018 12:23 am

Mairon62,

More vague hand-waving on WUWT. The usual proof by assertion, such as “…that most all the “problems” that the political left has tasked itself to solve don’t actually exist on a global basis or don’t exist at all.”

MarkW
Reply to  Chris
October 12, 2018 7:53 am

It’s not so much that the problems don’t exist. Global poverty obviously does exist.
The problem is that the solutions pushed by liberals never solve these problems and almost always make them worse.

John McLean
October 11, 2018 9:29 pm

JOHN MCLEAN – Update.

CRUTEM4 documentation should apply to CRUTEM4 dataset, not necessarily to HadCRUT4. But is seems it doesn’t apply to even CRUTEM4 because an extract of the grid cell for Apto Uto shows strange values too.

The grid cell extract of
ABC
DEF
GHI
where Apto Uto is in the central cell is listed below, with Year and Month, then cells A to I as per above.

1978 1 0.20 0.00 -0.05 0.25 0.93 0.23 0.17 0.13 0.20
1978 2 0.80 0.70 0.78 0.95 1.10 1.43 1.13 0.87 0.20
1978 3 0.50 0.40 0.65 -0.30 0.17 0.23 0.07 0.27 -0.30
1978 4 0.50 -0.20 0.15 -0.35 8.65 -1.03 -0.83 -0.27 -0.50
1978 5 0.10 0.30 0.30 -0.10 0.22 0.20 -0.57 0.10 -0.80
1978 6 0.10 -0.10 -0.32 -0.25 9.02 -0.23 -0.50 -0.17 -0.60
1978 7 0.10 -0.30 -0.05 -0.40 9.22 0.63 -0.13 0.07 0.00
1978 8 0.40 -0.10 -0.37 0.40 -0.18 -1.00 -0.13 -0.27 -0.80
1978 9 0.00 0.20 0.03 -0.10 0.14 -0.07 -0.03 0.20 -0.50
1978 10 0.10 0.10 0.07 0.00 0.46 -0.30 0.27 -0.13 -0.40
1978 11 0.20 0.60 0.43 0.25 0.62 0.30 0.27 -0.27 -0.20
1978 12 -0.20 0.10 -0.15 -0.20 0.12 -0.00 -0.03 -0.33 -0.50

Note the odd values in the central grid cell in April, June and July of that year. I’ve also checked the other stations reporting data for that grid cell in that month and none vary from their averages much more than 2C.

It seems that Mr Mosher has assumed
(a) that what’s said about CRUTEM4 definitely holds true for HadCRUT4 (okay, the doc for HadCRUT4 isn’t really clear)
and (b) that documentation from the CRU correctly describes CRUTEM4.

An apology seems to be called for.

John Endicott
Reply to  John McLean
October 12, 2018 8:18 am

An apology requires honesty and integrity. I wouldn’t hold your breath waiting for it.

Caligula Jones
Reply to  John Endicott
October 12, 2018 11:17 am

Yes, there are two parts to a civil debate: expertise and good will.

Somehow, climate alarmists, even when they can exhibit the first, can rarely claim the second.

Geoff Sherrington
October 11, 2018 11:10 pm

For those seeking more depth in the early history of CRU, Temperatures, Phil Jones, etc., I delved into the detail here.
http://joannenova.com.au/2012/01/that-famous-email-explained-and-the-first-volunteer-global-warming-skeptic/
Geoff

Geoff Sherrington
October 11, 2018 11:19 pm

It is astounding to see some here questioning whether the Author, John McLean, knew the difference between raw observations and adjusted opbservations, are straining their credibility. A person invited to submit a Ph.D. thesis at a recognised modern University can probably bes assumed to know enough of the chosen topic to avoid making stupid, elementart errors. Ref Nick Stokes who brushed it aside by saying “These are errors in the raw data files as supplied by the sources named. The MO publishes these unaltered, as they should. But they perform quality control before using them.” and Mosh with “A) data suppliers can apply QC and then document how they QCed. This is done with flags typically”.

You think John McLean does not know this?
Yet, he proceeds to provide evidence of a significant problem. This is proper, because it is real.
I can’t count how many times since 1992 I have written that the data in question is unfit for purpose, with the main purpose in my criticisms being that it is used to construct a global temperature average when it cannot possibly do this accurately enough for use for most puirposes.

Reply to  Geoff Sherrington
October 11, 2018 11:41 pm

“A person invited to submit a Ph.D. thesis at a recognised modern University can probably bes assumed to know enough of the chosen topic to avoid making stupid, elementart errors.”
So we have to assume JM is right, because he has been invited to submit a PhD? And so believe him when he says that HADCRUT, which is full of established PhDs, is making stupid elementary errors.

John Bills
Reply to  Nick Stokes
October 12, 2018 12:10 am

The elephant Nick Stokes:
https://doi.org/10.1029/2018JD028355

MarkW
Reply to  Nick Stokes
October 12, 2018 7:47 am

Nick takes the goal posts and proceeds to run with them.

Geoff never said we need to assume that JM is right because he was invited to submit a PhD. What he said was that he could be assumed to know enough to not make the basic mistakes that you accused him of.

Please go away until you are mature enough to argue honestly.

Venter
Reply to  MarkW
October 13, 2018 7:07 pm

Nick is someone who’ll do anything to justify whatever [the] AGW crowd do. Honour, truth, a sense of shame etc.don’t appear in the picture for him. He’ll twist himself worse than a pretezel to move goalposts and discussions to avoid the truth. [pruned].

Patrick MJD
Reply to  Venter
October 13, 2018 10:44 pm

If Stokes did work for the CSIRO and is now retired, he can say what the h3ll he likes, he will still get his pension and perks (Govn’t agency). So he can’t be in it for the money.

John McLean
Reply to  Geoff Sherrington
October 12, 2018 2:59 am

Geoff, the situation is even simpler. It can be shown that obvious errors are included in the HadCRUT4 dataset. Figure 6.3 of the audit shows that Apto Uto is included.
Also, if the CRU had integrity it would show the data files that it was sent AND it would show a file with revised value and explain what was adjusted and why.

Geoff Sherrington
October 11, 2018 11:27 pm

There have been many man-months of work over the years by a group of us who cannot see Australia’s land data showing warming at more than 0.5 deg C versus the official 0.9 deg C roughly for the century starting 1910.
We are sticking by that.
Australian data goes into HadCRUT4. It has a large influence on estimates of Southern Hemisphere temperatures. This estimate has errors that should be corrected.
We have done several Australian land temperature studies showing data problems. One of them is here. Geoff.
http://www.geoffstuff.com/explanation_chris_gilham.pdf http://www.waclimate.net/year-book-csir.html

Anthony Banton
October 12, 2018 1:03 am

“You missed the comments showing that HadCRU doesn’t do a quality control audit on the “data” before using them.”

There may be “comments”
But that does o equal evidence.
Except on WUWT of course.

A C Osborn
Reply to  Anthony Banton
October 12, 2018 3:10 am

Mr Banton, please show us the data to back up your statements.
Disprove Mr McLean’s findings with actual data instead of snide remarks.

Anthony Banton
Reply to  A C Osborn
October 14, 2018 12:26 am

Mr Osborne:

That is the job of the accuser (obviously).
To provide evidence.
And he patently hasn’t done it.
A file stored from a Nat Met service is NOT evidence of it being included without being QC’d.

There is an easy and straightforward way to check.
To write to the UKMO and ask.
Smacks of …. “I have found a smoking gun” … and the accusation is good enough.
Lets not spoil it by actually getting to the truth.
That you do not see that is of course a given.

And – my “snide remarks” are well deserved here as yet again denizens are entirely unsceptical of sceptics while being entirely critical of the rest.

Anthony Banton
Reply to  A C Osborn
October 14, 2018 12:41 am

One other “common-sense” thought that is missing in Mclean’s “analysis”.
As I’ve stated above.
The UKMO is an organisation that for ALL it’s data – checks for errors (QCs) as a MUST requirement, else the outcome is fatally altered – weather data assimilation is their “day-job”.
They wouldn’t have in place a routine QC software? really? …..

You really need to be ideologically motivated to jump to that conclusion.
And to not do the easy thing and clarify with the UKMO that that is in indeed what they do, is the classic put up an unsupported accusation into the naysayer book of myths via gaining uncritical hugs and kisses in the Blogosphere.

I also note that Mr Mclean has not answered my objections.

Scott Bennett
Reply to  Anthony Banton
October 15, 2018 3:13 am

==>Anthony Banton

I’d also like complete clarification along with you and Dr Mclean, as he did ask the same question in his paper. Why are outliers included in the station data files and why weren’t suitable quality control measure applied to the data?

A greater problem to be addressed is why such obvious outliers are included in the station data files. It would seem that not all the national meteorological services that supply the data to the CRU, nor the CRU itself, apply suitable quality control measures to the data in question. – Mclean, John D. 2017

I can confirm and have verified that outliers such as Apt(Airport) Otu are in the published files supplied in association with CRUTEM4. These files are included in the latest 2018 version and the implausible monthly mean above 80.0C is there for all to see.

These station files list the normals and standard deviations which again backup the statement made in the Dr Mclean’s paper:

The data files published in association with the CRUTEM4 dataset do not include a set of station data files that have been corrected or data removed so it is not possible to determine the changes that have been made or to verify that the data has [been]modified as described. What is clear however is that the inclusion of some erroneous values when calculating long-term average temperature and standard deviations has negative repercussions on the CRUTEM4 dataset. – Mclean, John D. 2017

The questions now to ask, is what station data has been modified and what were those changes… as it isn’t at all obvious to anyone who examines the dataset!

Hivemind
October 12, 2018 1:37 am

“… riddled with errors”

I read this as “… fiddled with errors” at first.

Anthony Banton
Reply to  Larry
October 12, 2018 11:35 am

“Write up a clear and comprehensive response. I am sure that Watts will publish it (or JoNova, or me, whoever your choose).”

No, how about Mr Mclean write to the UKMO and put questions to them.

First question…. Do the files as sent by the Nat Met services go un-QC’d into Hadcrut?

Is it me or is that not an obvious thing to do … and indeed an obvious thing for his Phd referees to ask of him.
(being as there is a presumption of incompetence on his part)
I cant think of a better way to get the truth – unless of course….
The “C” word comes into thinking.

Venter
Reply to  Anthony Banton
October 15, 2018 9:50 am

UKMO were more gracious than you. They responded acknowldging mistakes. Whereas you twisted you were busy throwing mud at McLean without even bothering to read his thesis and acknowledge it’s merits. Who’s looking like a fool now?

Scott Bennett
October 12, 2018 5:31 am

==>John McLean

Hi John,

Firstly, thank you for commenting. It is a brave scientist who dares to show up at WUWT!

As a layman I’m struggling to work out what data was used and when.

On the CRU web site* under the heading – “Land Stations used by the Climatic Research Unit within CRUTEM4” – there is a link to the station files,** Below the link it says:

The file gives the locations and names of the stations used at some time (i.e. in the gridding that is used to produce CRUTEM4) during the period from 1850 to 2010. All these stations have sufficient data to calculate 30-year averages for 1961-90…

I downloaded all the relevant files and the station data is there for Apto Uto but it is not included in this site list – as has been pointed out here. This has become the major bone of contention with you paper and nobody seems capable of moving beyond this one issue, despite there being many others of major importance.

It is interesting if confusing that even with the data provided, the exact averages for HadCRUT4 and HadSST3 can not be replicated!!

The reason given in the FAQ is as follows:

Both these are ensemble datasets. This means that there are 100 realizations of each in order to sample the possible assumptions involved in the structure of the various components of the error… All 100 realizations are available at the above Hadley Centre site, but we have selected here the ensemble median. For the gridded data this is the ensemble median calculated separately for each grid box for each time step from the 100 members. For the hemispheric and global averages this is again the median of the 100 realizations. The median of the gridded series will not produce the median of the hemispheric and global averages, but the differences…

This seems absurd to me and incredibly opaque but what would I know!

The other disclaimer is the admission of several “variance adjustment” (Monthly updates, NMSs and the moving 30-year baseline) that change the data from year to year.

It would seem to be an impossible task to audit the provided data when even the provider has declared that their result can not be duplicated! ;-(

Scott
Reply to  Scott Bennett
October 12, 2018 5:35 am

Whoops! Links mentioned above:
* https://crudata.uea.ac.uk/cru/data/temperature/crutem4/landstations.htm
**https://crudata.uea.ac.uk/cru/data/temperature/crutem4/crutem4_asof020611_stns_used_hdr.dat

Solomon Green
October 12, 2018 6:31 am

My thanks to Mr. Watts and Dr. McLean. I am still studying the paper but even the little that I have followed makes it well worth while the $8.

October 12, 2018 1:20 pm

“Cad”.
That’s the most generous term that could be used.

Geoff Sherrington
October 12, 2018 10:47 pm

John McLean notes on page 7 of his report, ” The frequency of the upward or
downward adjustments are irrelevant on these scales; it is the size of the adjustment that
matters. For example, five adjustments downwards by 1.0°C are not cancelled out by five
adjustments upwards by 0.2°C.”

This is an important point that I do not have the computing power to deal with myself.

An adjustment applied to a selected portion of a temperature/time series can have an effect depending on 3 main factors –

1. Magnitude. How large the change is, e.g. delete 0.5 deg C, replace with 0.75 deg C.
2. Duration. The duration of the change, e.g. a change to a time period one year long has less effect than a change applied to 10 years long.
3. Leverage. How far from the pivot point the change is. Current methodology keeps the most recent observation as the fulcrum point, so a change to a block of data dated dated 100 years ago will swing the outcome more than a chnage of similar smagnitude and duration made 1 year ago (that is like a torque in foot pounds can have the same pounds but many more foots).

Of course sign is part of this, as noted.

So, each adjusted time series needs to be examined for the total effect of the adjustment considering sign, magnitude, duration and leverage. This is what I see as an analog of torque. The program to do this is not daunting to write, but easy access to each temperature-time series used in HadCRUT4 has to be there, preferably cleaned of the other errors mentioned.

Running this quality control test would put to rest a whole lot of speculation about the speculated effects of adjustments and homogenizations. It is a complete answer to Mosh’s claims that adjustments to BEST made the past warmer. They might have, but the full test does not seem to have been done yet.

In my book, it is reprehensible for BEST and HadCRUT (and probably others like GISS) to have endured this long before they do the definitive demonstration using this methodology. Maybe they can redeem some reputation by doing it immediately. Geoff.

barry
October 14, 2018 4:53 am

If HadCRU uses different data to GISS and NOAA and BEST, if the each have different methodologies, and the results are very similar….

If raw and adjusted data get similar results. If different datasets (GHCN daily – GSOD etc) get similar results…

How bad is HadCRU really?

Remember this skeptic effort?

https://noconsensus.wordpress.com/2010/03/24/thermal-hammer/

“First the obvious, a skeptic, denialist, anti-science blog published a greater trend than Phil Climategate Jones. What IS up with that?”

We read from above (Stokes) that McLean is referring to the raw data, which gets tripped of these errors in the processing. I did not get that clarification in the post here or the abstract.

We learn upthread that station data (from at least one station) that is faulty is not used.

I’ve seen a lot of back-slapping and cheering, but not much of the skepticism that should be driving interest in scientific findings every time. Why is it left to Mosh and Stokes to cast a critical eye on this? And why are skeptics waiting for anyone else to do due diligence?

scross
Reply to  barry
October 14, 2018 3:17 pm

As I’ve already stated elsewhere in this thread, some of us may be carefully taking our time to review things before making comments either way. Meanwhile Mosher, for one, was apparently so eager to stick the knife in that it looks like he got a lot of his criticism wrong, as others in this thread have already pointed out, should you care to peruse the whole thing.

barry
Reply to  barry
October 15, 2018 6:07 am

I didn’t see one person say they would digest it carefully before commenting (you mentioned this as a possibility). But there were a great many who took the conclusions as read, praising the author for getting to “the truth”, and scorning the Met Office.

I doubt we’ll get any commentary from regulars who digest the report with a critical eye. I’d be delighted to be proved wrong, but there’s just no reason to believe it.

“Meanwhile Mosher, for one, was apparently so eager to stick the knife in that it looks like he got a lot of his criticism wrong”

There were only 2 points of criticism from him in this thread, so he couldn’t have got “a lot of” his criticism wrong.

He said no data, no code. McLean THEN replied by offering links here, but not to code. Looks like it wasn’t supplied with the report.
He said Apto Uto station not used. Stokes checked and says it’s not used after 1970. Mosh may have been wrong about that, then. But the faulty data was in the 1970s.

The failure to mention that it was raw data that was so messy seems to be a pretty damning indictment of the report. I’m prepared to give some benefit of doubt, but I can’t see much changing with other temp records of different method and provenance having similar results.

Philip Schaeffer
Reply to  barry
October 15, 2018 6:54 am

Gawd I wish we could have a sensible discussion on that level. You’ve got right to the heart of the issue.

I know I’m probably dreaming, but what I would love to see here is an article detailing exactly what the complaints about this paper are, in detail, and detailed responses to each bit from McLean.

It would be a shame to let this just slip by when we actually have a chance to sort this out and get some clarity.

Venter
Reply to  Philip Schaeffer
October 15, 2018 9:47 am

While you lot are moaning and throwing mud at McLean’s work, Met Office admitted that he flagged issues and they did corrections. So eat some crow, barry and Philip.

Philip Schaeffer
Reply to  Venter
October 15, 2018 11:46 am

The argument isn’t about whether or not the data set is perfect.

scross
Reply to  barry
October 15, 2018 1:23 pm

“He said no data, no code. McLean THEN replied by offering links here, but not to code. Looks like it wasn’t supplied with the report.”

Links to the data can be found near the bottom of the following page. AFAIK these have been there since before the WUWT discussion even began, and I assume that they are also buried somewhere in the document itself, but I haven’t looked. Links to code may be in there somewhere, too, but again I haven’t looked.

https://robert-boyle-publishing.com/product/audit-of-the-hadcrut4-global-temperature-dataset-mclean-2018/

Concerning Apto Uto, Mosher said the following:

“The reason why CRU does not USE Apto Uto is because it does NOT have the required number of years
in the base period. For CRU this is 1951-1980 and a station MUST HAVE 20 of those 30 years
OR IT IS DROPPED”

McLean responded thusly:

“(c) What stations are used and what are not?
The old minimum of 20 years of the 30 from 1961 to 1990 was dropped a few HadCRUT versions back. It then went to 15 years with no more than 5 missing in any decade. HadCRUT4 reduced it again to 14.”

Now I noticed immediately that Mosher seemed to be using the wrong decade range, from what I’d seen so far of the report plus the relevant web sites, only I said nothing because I wasn’t quite sure. But McLean seems to be confirm this and also points out that Mosher seems to be using completely incorrect (or at least outdated) criteria in his criticism. So it comes as no surprise to me that others note Apto Uto is actually in there somewhere, while Mosher insists that it simply can’t be, using his criteria.

Before making any further comments here, you do know that we’ve been down this path at least once before, right, with HadCRUT3? While McLean may very well be the first person to do a deep quality analysis of HadCRUT4, others did similar work with HadCRUT3 years ago and found problems with it similar to what McLean has found. Those folks were generally mostly ignored, though (except by sites like WUWT), but McLean’s work may be harder to ignore since it seems to be drawing a lot of attention.

barry
Reply to  scross
October 15, 2018 2:43 pm

Yes, the baseline period is wrong (it’s GISS), but I suspected brain fart, and it seems McLean did too judging by the update.

Before making any further comments here, you do know that we’ve been down this path at least once before, right, with HadCRUT3?

Yes. I’ve seen criticism of the temp records for over a decade: McIntyre noticing a problem with 2000s data in the US temp record and a correction from the institute acknowledging him; Anthony Watts’ surfacestations project and photographic evidence of site bias; endless focus on adjustments that cool the record for a particular weather station (but never, it seems, showing the adjustments that war a particular station – it took Stokes to point out that there were as many cool as warm adjustments); criticism of the SST constructions, station ‘drop out’ and the rest.

But I’ve also seen skeptics knuckle down and construct their own temp records – people who did more than notice problems, they acted on it. I’ve seen BEST produce a temp record much the same as the others. I’ve seen Jeff Condon and Roman M come up with a temp record from raw data that ran warmer than HadCRU. I’ve sen Anthony Watts publish a paper highlighting min/max biases in the US records, but that corroborated the US mean temp record.

I also see that these developments are soon forgotten whenever a ‘bombshell’ appears. It seems that it is enough that someone has criticised for a wave of congratulations to appear, so that it seems that there is only ever one scandal after another.

We’ve known for years that the data are not perfect. And for years the various compilers of the official records have said the same. This is old ground. What’s new in this report that will make a substantive impact? The Met Office replied to it that the small number of errors out of millions of datum would not significantly change the results.

You have suggested that – maybe – some regulars are taking their time digesting the paper and will eventually comment. That there is a sober coterie doing due diligence.

Whether or not that’s true, none have said so or recommended patience. If it was an ‘enemy’ paper being attacked there would already be specific comments on it.

No, if any AGW skeptics are going through it, they’re not commenting, and I doubt we’ll see substantive commentary from them before this thread shuts down, leaving Mosh and Stokes as the few voices of doubt in a tide of approval.

Venter
Reply to  barry
October 15, 2018 6:51 pm

Oh Yeah, we’ve seen that. Every error won’t make a difference. Every adjustment won;t make a difference. But hey presto, sum of all will be ” It’s worse than ever “. That’s been your lot’s modus operandi.

Scott Bennett
Reply to  scross
October 15, 2018 2:51 pm

Since I had the 2018 version of the data and the APTO_OTU (Otu Airport?*) file open on my desktop, here are some details:

APTO_OTU has 41 years of data 1947 – 1988.
The 80C outlier is from 1978.
13 years are missing and 11 years are incomplete.
The normals were calculated from 1961-1988
The standard deviations are from 1947 – 1988

Here is header above the observations:

Number= 800890
Name= APTO_OTU
Country= COLOMBIA
Lat= 7.0
Long= 74.7
Height= 630
Start year= 1947
End year= 1988
First Good year= 1947
Source ID= 79
Source file= Jones
Jones data to= 1988
Normals source= Data
Normals source start year= 1961
Normals source end year= 1988
Normals= 24.1 24.4 24.6 27.8 24.6 27.9 28.0 24.6 24.4 24.1 24.1 24.0
Standard deviations source= Data
Standard deviations source start year= 1947
Standard deviations source end year= 1988
Standard deviations= 0.6 0.6 0.5 11.9 0.5 11.8 12.0 0.6 0.5 0.6 0.6 0.7
Obs:…”

*Otu airport has that Lat/Long.

Gary Palmgren
October 14, 2018 1:17 pm

I know of only one study on global warming based on temperature readings that I consider legitimate. Tony Heller at realclimatescience.com one time looked at the temperature record from continental US (only data set he considers reliable). He computed the temperature trend at each station and then averaged the trends to find an over cooling trend from the data. This directly contradicts the government sponsored rising temperature claims and is thermodynamically valid. Averaging temperatures from areas of different heat capacities is not.

Regardless of heat capacity, heat does flow from something with a higher temperature to something cooler. This is why I claim Mr Heller’s averaging of temperature trends at different stations is valid. If global warming was real this analysis would show it and there is no legitimate reason not to process the temperature data by first computing trends for each station. Step changes at one site would actually be a good indicator of site changes.

Now if we could just quite throwing half the data away and compute minimum daily temperature trends separately from maximum daily temperature trends. I have read here at WUWT that nighttime minimum temperatures are increased by buildings increasing mixing more than the temperature highs for each day. Keeping both might allow some measure of changes in the Urban Heat Island effect at each site to get a measure of site changes.

1 3 4 5