Guest Post by Willis Eschenbach
Recapping the story begun at WUWT here and continued at WUWT here, data from the temperature station Darwin Zero in northern Australia was found to be radically adjusted and showing huge warming (red line, adjusted temperature) compared to the unadjusted data (blue line). The unadjusted data showed that Darwin Zero was actually cooling over the period of the record. Here is the adjustment to Darwin Zero:
Figure 1. The GHCN adjustments to the Darwin Zero temperature record.
Many people have written in with questions about my analysis. I thank everyone for their interest. I’m answering them as fast as I can. I cannot answer them all, so I am trying to pick the relevant ones. This post is to answer a few.
• First, there has been some confusion about the data. I am using solely GHCN numbers and methods. They will not match the GISS or the CRU or the HadCRUT numbers.
• Next, some people have said that these are not separate temperature stations. However, GHCN adjusts them and uses them as separate temperature stations, so you’ll have to take that question up with GHCN.
• Next, a number of people have claimed that the reason for the Darwin adjustment was that it is simply the result of the standard homogenization done by GHCN based on comparison with other neighboring station records. This homogenization procedure is described here (PDF).
While it sounds plausible that Darwin was adjusted as the GHCN claims, if that were the case the GHCN algorithm would have adjusted all five of the Darwin records in the same way. Instead they have adjusted them differently (see below). This argues strongly that they were not done by the listed GHCN homogenization process. Any process that changed one of them would change all of them in the same way, as they are nearly identical.
• Next, there are no “neighboring records” for a number of the Darwin adjustments simply because in the early part of the century there were no suitable neighboring stations. It’s not enough to have a random reference station somewhere a thousand km away from Darwin in the middle of the desert. You can’t adjust Darwin based on that. The GHCN homogenization method requires five well correlated neighboring “reference stations” to work.
From the reference cited above:
“In creating each year’s first difference reference series, we used the five most highly correlated neighboring stations that had enough data to accurately model the candidate station.”
and “Also, not all stations could be adjusted. Remote stations for which we could not produce an adequate reference series (the correlation between first-difference station time series and its reference time series must be 0.80 or greater) were not adjusted.”
As I mentioned in my original article, the hard part is not to find five neighboring stations, particularly if you consider a station 1,500 km away as “neighboring”. The hard part is to find similar stations within that distance. We need those stations whose first difference has an 0.80 correlation with the Darwin station first difference.
(A “first difference” is a list of the changes from year to year of the data. For example, if the data is “31, 32, 33, 35, 34”, the first differences are “1, 1, 2, -1”. It is often useful to examine first differences rather than the actual data. See Peterson (PDF) for a discussion of the use of the “first-difference method” in climate science.)
Accordingly, I’ve been looking at the candidate stations. For the 1920 adjustment we need stations starting in 1915 or earlier. Here are all of the candidate stations within 1,500 km of Darwin that start in 1915 or before, along with the correlation of their first difference with the Darwin first difference:
WYNDHAM_(WYNDHAM_PORT) = -0.14
DERBY = -0.10
BURKETOWN = -0.40
CAMOOWEAL = -0.21
NORMANTON = 0.35
DONORS_HILL = 0.35
MT_ISA_AIRPORT = -0.20
ALICE_SPRINGS = 0.06
COEN_(POST_OFFICE) = -0.01
CROYDON = -0.23
CLONCURRY = -0.2
MUSGRAVE_STATION = -0.43
FAIRVIEW = -0.29
As you can see, not one of them is even remotely like Darwin. None of them are adequate for inclusion in a “first-difference reference time series” according to the GHCN. The Economist excoriated me for not including Wyndham in the “neighboring stations” (I had overlooked it in the list). However, the problem is that even if we include Wyndham, Derby, and every other station out to 1,500 km, we still don’t have a single station with a high enough correlation to use the GHCN method for the 1920 adjustment.
Now I suppose you could argue that you can adjust 1920 Darwin records based on stations 2,000 km away, but even 1,500 km seems too far away to do a reliable job. So while it is theoretically possible that the GHCN described method was used on Darwin, you’ll be a long, long ways from Darwin before you find your five candidates.
• Next, the GHCN does use a good method to detect inhomogeneities. Here’s their description of their method.
To look for such a change point, a simple linear regression was fitted to the part of the difference series before the year being tested and another after the year being tested. This test is repeated for all years of the time series (with a minimum of 5 yr in each section), and the year with the lowest residual sum of the squares was considered the year with a potential discontinuity.
This is a valid method, so I applied it to the Darwin data itself. Here’s that result:
Figure 2. Possible inhomogeneities in the Darwin Zero record, as indicated by the GHCN algorithm.
As you can see by the upper thin red line, the method indicates a possible discontinuity centered at 1939. However, once that discontinuity is removed, the rest of the record does not indicate any discontinuity (thick red line). By contrast, the GHCN adjusted data (see Fig. 1 above) do not find any discontinuity in 1941. Instead, they claim that there are discontinuities around 1920, 1930, 1950, 1960, and 1980 … doubtful.
• Finally, the main recurring question is, why do I think the adjustments were made manually rather than by the procedure described by the GHCN? There are a number of totally independent lines of evidence that all lead to my conclusion:
1. It is highly improbability that a station would suddenly start warming at 6 C per century for fifty years, no matter what legitimate adjustment method were used (see Fig. 1).
2. There are no neighboring stations that are sufficiently similar to the Darwin station to be used in the listed GHCN homogenization procedure (see above).
3. The Darwin Zero raw data does not contain visible inhomogeneities (as determined by the GHCN’s own algorithm) other than the 1936-1941 drop (see Fig. 2).
4. There are a number of adjustments to individual years. The listed GHCN method does not make individual year adjustments (see Fig. 1).
5. The “Before” and “After” pictures of the adjustment don’t make any sense at all. Here are those pictures:
Figure 3. Darwin station data before and after GHCN adjustments. Upper panel shows unadjusted Darwin data, lower panel shows the same data after adjustments.
Before the adjustments we had the station Darwin Zero (blue line line with diamonds), along with four other nearby temperature records from Darwin. They all agreed with each other quite closely. Hardly a whisper of dissent among them, only small differences.
While GHCN were making the adjustment, two stations (Unadj 3 and 4, green and purple) vanished. I don’t know why. GHCN says they don’t use records under 20 years in length, which applies to Darwin 4, but Darwin 3 is twenty years in length. In any case, after removing those two series, the remaining three temperature records were then adjusted into submission.
In the “after” picture, Darwin Zero looks like it was adjusted with Sildenafil. Darwin 2 gets bent down almost to match Darwin Zero. Strangely, Darwin 1 is mostly untouched. It loses the low 1967 temperature, which seems odd, and the central section is moved up a little.
Call me crazy, but from where I stand, that looks like an un-adjustment of the data. They take five very similar datasets, throw two away, wrench the remainder apart, and then average them to get back to the “adjusted” value? Seems to me you’d be better off picking any one of the originals, because they all agree with each other.
The reason you adjust is because records don’t agree, not to make them disagree. And in particular, if you apply an adjustment algorithm to nearly identical datasets, the results should be nearly identical as well.
So that’s why I don’t believe the Darwin records were adjusted in the way that GHCN claims. I’m happy to be proven wrong, and I hope that someone from the GHCN shows up to post whatever method that they actually used, the method that could produce such an unusual result.
Until someone can point out that mystery method, however, I maintain that the Darwin Zero record was adjusted manually, and that it is not a coincidence that it shows (highly improbable) warming.
Sponsored IT training links:
Want to pass HP0-J33 at first try? Gets certified 000-377 study material including 199-01 dumps to pass real exam on time.
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.



Gail Combs (05:44:57), you raise an interesting and often misunderstood point:
What is used to adjust the data is not the absolute value of the data. It is the change in the data. As you point out, someplace five miles away may get very different temperatures from where you are.
However, what is related is the change in temperature. If you are having a hot month, it is very likely that someplace five miles away is having a hot month as well. It is this relationship that is used for the adjustment.
However, that doesn’t mean that the adjustment is valid. As someone pointed out above, what it can do is re-institute a UHI signal on a station which has been moved.
Someone else pointed out that Darwin is an outlier. While this is true, such outliers should be a huge warning flag to whoever is using the algorithm. People often say “the exception proves the rule” without realizing that “prove” in this expression has the meaning used in the phrase “proving ground”, as in “testing ground”. The real meaning of the phrase is “the exception tests the rule”. In the current case, the rule seems to be failing the test …
wobble (06:53:46) :
Not according to the GHCN document, An Overview of the Global Historical Climatology Network Temperature Database, which says (emphasis mine):
There are many flaws in the Peterson algorithm, both theoretical and practical.
I’m kinda weird, in that I do all the scientific work that I do simply because I’m passionate about climate science. I have persevered and gone forward with my analysis not because of JJ, but in spite of him. These kinds of studies take time, as unlike Peterson and Hanson and their ilk, I have a regular job that does not involve climate science.
I don’t think JJ is a bad guy, I just get upset by constant cavilling with no corresponding productive input. Lead, follow, or get out of the way.
Ryan Stephenson (08:27:36) :
The problem is that the method you suggest assumes that there really were some kind of problems needing adjustment in those years. I find nothing, either in the record itself or in the station metadata, indicating any problem in those years. The Aussies make adjustments based (very loosely, it appears) on the metadata … but they don’t make a single adjustment in any of the years adjusted by GHCN.
temp (12:33:00) :
My apologies for the lack of clarity, temp, you raise an important point. For the 1920 step, I used the correlation for the period 1900-1940, or as much of that period as the station covered. There is nothing I could find in their documentation which describes what period they actually used.
Nor does their documentation specify a couple of other important things. One is whether the correlation is statistically significant. If we flip two coins, and then flip two other coins, they may come up all heads (correlation = 1.0 between the two sets of two). However, because of the length of the dataset (N=2), the correlation is not statistically significant. A ten year temperature dataset is generally not long enough to establish significance.
The other is the effect of autocorrelation on the significance. (Autocorrelation measures how much this year’s data resembles last year’s data.) Temperature records tend to have fairly high autocorrelation, a warm year is likely to be followed by a warm year and vice versa. As a result, we are more likely to get a spurious correlation between temperature series.
I chose a forty-year period because of those two factors, which make a shorter time period more likely to produce spurious correlations. I did not use the entire period of the overlap between two stations because the correlation on that is likely to be lower than on the 40 year period, and I wanted to give the algorithm every chance. However, I still found nothing resembling Darwin’s record in the “neighboring” stations.
Which brings me back to my point that nature is not homogeneous. It is freckled and fractured, a “ghost of rags and patches”. Trying to force it into a homogeneous strait-jacket can be done … but what we end up with, almost by definition, will be “un-natural”. Nature specializes in jumps and changes, in instant inversions and retroversions, and not in bland homogeneity. As the poet said:
Homogenize that and see how it all looks … not pretty.
Willis Eschenbach (14:52:53) :
“”according to the GHCN document,…the correlation between first-difference station time series and its reference time series must be 0.80 or greater””
Thanks! Again, great job.
@Willis: I think maybe you misunderstand me. I am not suggesting that the algorithm that detects the discontinuities was right or wrong. That doesn’t matter to me. What matters to me is if the adjustment that was made, automatically or manually, and whether that adjustment was reasonable. Up till now you have been asking “the Team” to justify their adjustments, and waiting…. an waiting….. But I have suggested an approach that will indicate if their adjustments were wrong or not without their help.
So lets assume the algorithm that detects the discontinuities is doing so “willy-nilly”. So what? The Teams algorithm tells us that between 1961 and 1980 there was no need for any adjustment – because the black line is perfectly flat here. This means that the algorithm is telling us that the site was perfectly stable during that time and that the temperature measurments were valid. So let us look at the raw data between those dates. We can pick out the specific years that are between the adjustments that were made – about 17 years altogether – then we can plot a line of regression between them. The gradient of that line of regression will give us the rate of change of temperature. I haven’t done the plot, but just looking at it suggests about 2Celsius per century during that period*. But the gradient of the adjusted data shows a gradient of 6Celsius per century. These two facts are in direct contradiction. The site is detected to have been apparently stable for 17 years, by the Teams own algorithm, so the measured rate of change during those 17 years should be valid, by the Teams own criteria. But their own adjustment method comes up with a rate of change three times higher than this. The adjustment method fails this simple sanity check. There is far more adjustment than is necessary.
*naturally the measured rate of change is not attributable to any specific source. Could be CO2, could be UHI, could be something else.
@Willis: OK, I tried to do it myself, but GISS/GHCN are both unable to deliver me the raw data 😉
I read the results from the graph you have shown of the raw data during about 1962 to 1979 and stuffed them into Excel, which was then used to add a linear line of regression. The gradient of that line came out at 3.5Celsius per century. Very high, certainly, but definitely not 6Celsius per century. The algorithm used to make the adjustment fails the sanity check – the adjustments used at Darwin are not reasonable.
Before anybody says “wow, 3.5Celsius per century- we’re all going to fry” remember this is data over only a 17 year period, it is preceeded by a long cooling period of about 1Celsius per century lasting some 50 years or more, and we aren’t able to indentify the source of the warming trend – it may not be CO2. I think if you did the same analysis of the full dataset using this approach then you would probably see that temperatures today are about the same as they were in 1880. Shame we can’t get hold of the raw data as it seems to have been, um, taken off line.
Phil. (23:26:41) :
This is shown by the data, the record low at the airport is 10.4, the record at the PO was 13.4.
Which data? I can’t find any 10.4 or 13.4 for Darwin in v2.mean.
My reply to, Willis Eschenbach (15:36:39).
Well, I’m pretty sure that they didn’t use a signficance measure or they’d certainly of mentioned that. It seems like they used a straight cut off of 0.8. You could certainly argue not using a p-value cut-off is a weakness.
I’m also pretty sure your 40 year period is likely to wrong. Remember in total, they are only looking for 25 years or so of data from the station (certainly not 40). You’ve linked the pdf to one of their new papers, but if you go back and look at the refs in that, I’m pretty sure that they talk about a 5 year window analysis, and at the average global level that has given us something that looks similar to what they’ve produced (not exactly).
No homogenization method is going to be perfect or “natural”, but the fact of the matter is that the data does need to be homogenized (though I will point out that if you simply take all of the data and average it and ignore everything you still get a nice bi-step warming this century, including a late 20th century warming trend, resulting in now being as warm as any other time this century).
For your point to be relevant:
1. You have to claim no use of the data is “good” and that we should essentially ignore all of the surface data.
or
2. You need to suggest a method that produces a more “natural” result that doesn’t show warming.
My reply to wobble (12:57:25).
If he REALLY wants the details about this particular station, I’d suggest he’d do what has already been suggested- that he contact GHCN and ask them what they did.
Ryan Stephenson (17:10:25)
“Shame we can’t get hold of the raw data as it seems to have been, um, taken off line”
If you are after the NOAA GHCN V2 data you can find it and download it from here.
NOAA GHCN V2 data
and it might be worth having a look at the links in this thread.
Reproducing Willis Eschenbach’s WUWT Darwin analysis
KevinUK
Ryan Stephenson (17:10:25)
If you (and anyone else) are interested in getting access to the NOAA GHCN data using a very friendly user interface that amongst other things enables you to easily select and then export raw and/or adjusted data for a selected WMO station to an Excel spreadsheet/CSV file, then click the link below to the TEKTemp system (as I’ve called it).
You’ll need to register first before you can login and use it, but its simple as you just need to supply a valid email address (to be used as your username) and a password on the TEKTemp registration page
The TEKTemp system
Note this system is currently running on my own test web server and so may not always be on-line. I hope to port it to a proper web hosting account shortly
KevinUK
KevinUK (13:23:59) :
In total, I have found 194 instances of WMO stations where “cooling” has been turned into “warming” by virtue of the adjustments made by NOAA to the raw data.
And how many instances of turning warming into cooling?
temp (18:47:44) :
BWA-HA-HA-HA …
Sorry, temp, that just slipped out, couldn’t help it …
But heck, why not? I’ve tried this path many times without any joy, but this time might be different. I just sent the following to the address given on the NCDC web site for questions about GHCN, ncdc.info@noaa.gov .
So … we’ll see. I’m not holding my breath. I’ll report back with the answer. My bet is on the polite brushoff, but since we’re now in the post-Climategate era, who knows? I’d be happy to be surprised.
temp (18:44:00) :
Do you have a citation for this? Is it gridded, or simply an average? How is the averaging done?
For the record, I should say that I do think that the planet warmed in the 20th century … and the 19th century … and the 18th century …
I also think that we are no warmer than the earlier warm period in the 1930s-40s.
And the ride through the fun house begins … in reply to my email of questions for GHCN, I just got the following:
I’ve resent the email to the address in the automated reply, ncdc.orders …
I’m sorry this is a bit late, may not be relevant to Willis Eschenbachs work or may have already been covered. WUWT has been so busy lately it’s hard to keep abreast of things. I do feel however that it would of interest to Australian station researchers. The late John L. Daly examined Darwin station and here is what he found.
But there is something peculiar about Darwin, a tropical site. It shows an overall cooling. But that cooling was mostly done in a period of only 3 years between 1939 and 1942.
http://www.john-daly.com/darwin.htm
@temp 18:44.
“No homogenization method is going to be perfect or “natural”, but the fact of the matter is that the data does need to be homogenized ”
Unfortunately the homogenization DOES need to be perfect. We are looking to adjust data that only persists for about 15 years between significant site changes, and the underlying trend we are looking for during those periods is only 1Celsius per century – i.e. 0.15Celsius when the difference in annual means can be anything up to 1.5Celsius. So a very small error in the homogenisation process will kill the accuracy of that data.
So is the total homogenisation process perfect? Well the approach I have used to do a sanity check of the data at Darwin uses the Teams own algorithm to detect when the site is perceived to be “stable” and not requiring any homogenisation. The problem is that these periods are only about 15 years long, and during that time the data can be going up and down more or less randomly by 1.5Celsius. Fortunately, at Darwin between 1962 and 1979 the rate of rise is very steep. Therefore we can plot a line of regression between the points and even though there is considerable uncertainty in exactly where the line should run, we can see that it is certainly closer to 3.5Celsius/century rather than 6Celsius per century. No amount of fiddling will allow you to impose a 6Celsius per century rate of rise on the raw data between 1962 and 1979, during which time the Teams own algorithm tells you the site was stable.
So Darwin fails this simple sanity check – the algorithm as applied here is not working. Period. OK, so now you say “ah, but Darwin is unusual – the algorithm works in ‘usual’ cases”. Problem is – how can you prove that? In the “usual” cases the rate of rise in the adjusted data is much smaller. If I applied the same sanity check approach to a “normal” station I couldn’t say anything with any certainty because there would be so much uncertainty about where exactly the line of regression should be fitted, so I wouldn’t even try. So we can’t sanity check the algorithm in the “usual” cases – we have to take it on faith that the algorithm is working. Most mathematicians would throw out the algorithm at this point – they would say it fails one sanity check and those it might pass are highly disputable. They would take a look at the algorithm again and make sure it was doing something sensible with the Darwin data. They would use the Darwin data to expose the flaws in their algorithm.
“the fact of the matter is that the data does need to be homogenized ”
Well, why is there a need? We can see from the Teams own algorithm that the sites are really only stable for about 15 years before some massive adjustment is necessary. We can see that doing a line of regression to find the rate of change of the temperature during those 15 year periods is problematic because there isn’t enough data in 15 years to be sure that the gradient we get is correct. But is replacing the data with data from another station a valid method of getting a long, unbiased trend?
Well remember what we have just said. The sites are changing every 15 years according to the Team own algorithm. But the algorithm looks for correlation between several sites in order to backfill dubious data in the site under inspection. But how can the algorithm reliably do that? If all the sites are likely to be changing every 15 years, how can you reliably correlate one site to another? Remember that both the site under investigation and the sites used for homogenisation are changing very often – if this wasn’t the case then you could just throw out the data from the sites that had changed and stick with the sites that show no site changes. So the algorithm is probably backfilling data between all the sites in a given region, taking data from site 1 and giving it to site 2 which gives data to site 3 and also to site 4 in some crazy difficult to follow method that might very well result in some circular changes being made (unless of course they are using the method “all sites are created equal, but some are more equal than others”!). The data may match over some specific 15 year period to an extent, but is that sufficiently reliable? Is it any more reliable than fitting a line of regression to 15 years of stable data? Well the algorithm doesn’t seem to look for perfect correlation anyway, and the correlation it is looking for is based on the annual changes in the means i.e. a weather signal not a climate signal. This of course sounds rather dubious as a process, and then we have the sanity check peformed above which shows us that the algorithm is known to fail anyway.
There is another underlying problem with the homogenisation process. It tends to be biased in favour of stations that suffer the worst UHI effect. Most sites at the beginning of the 19th century would have been edge of town locations with low-rise building, little tarmac and no heavy traffic. During the 20th century all these factors would have changed contributing to significant UHI (which is why simply averaging all the data only proves there was warming of the measurment kit, not that the wamring was specifically related to CO2). Thos sites that were moved during the 20th century were almost certainly moved from urban areas (where they got in the way of development) to rural areas. However, the Teams algorithm automatically detects these sites because of the discontinuities in the readings caused by the site change, and seeks to replace the data from elsewhere. Where is it most likely to get that data? From those sites that have not changed much. Those are sites that were on the edge of time, have seen massive development around them, but then have not been moved. The site then suffers from UHI – but the Teams algorithm can’t detect that because it can only see sudden changes in the means either side of data from a given year. So the algorithm detects these urbanised sites as “good” sites and then uses these to replace the data at the “bad” rural sites. This is exactly the opposite of what logic would tell you should be the case.
Fundamentally Willis has shown that the algorithm is definitely wrong, that the output of the algorithm cannot be relied upon in any cases and that the underlying logic of the algorithm is faulty. We can go on to say that the high frequency of site changes and the unknown precise impact of UHI at each site makes it impossible to rely on the surface temperature measurements as a means of detecting climate change of 1Celsius per century reliably and should be abandoned. The best we can do is to learn from these mistakes and set up true climate monitoring stations to give us data reliably for the next 100 years unimpacted by UHI or site changes.
Nick Stokes (20:34:00) :
“KevinUK (13:23:59) :
In total, I have found 194 instances of WMO stations where “cooling” has been turned into “warming” by virtue of the adjustments made by NOAA to the raw data.
And how many instances of turning warming into cooling”
Nick why don’t you visit my thread on ”diggingintheclay‘ and you’ll find out?
The link to the thread on ‘Physically unjustifiable NOAA GHCN adjustments’ is here
As you’ll see I found 216 instances in which NOAA’s GHCN adjustments resulted in ‘warming being tirned into cooling’ and I list the Top 30 worse cases. The worst case is Mayo Airport, Yukon, Canada. As with Darwin and many other stations there is no physically justifiable reason as to why its raw data needs to be adjusted. It is an almost complete set of raw data spanning 1925 to 1990 with (according to NOAA’s GHCN station inventory data), no station moves and no duplicate series. Its raw data trend looks relative ‘flat’ but is in fact slightly positive at +1.3 deg. C/century. For some reason NOAA feels the need to adjust this raw data so that all data prior to 1983 is increased using a ‘staircase’ type step function (see the thread for the charts) similar to Darwin (and Edson – see thread) except in this case the adjustments result in a strong cooling trend of -4.5 deg. C/century.
I’ve checked the raw data for Mayo Airport on the Environment Canada web site here and it shows that raw data is available from 1984 to 2009 and it all looks fine with occasional missing data so why did it ‘drop out’ of NOAA’s adjustment analysis after 1983?
Perhaps you’d like to have a look at some of the other stations on my two lists? It’s easy to extract the data for individual WMO stations using my TEKTemp admin system (see earlier post for link).
KevinUK
KevinUK (02:46:24) :
I loaded Mayo A data from the Climate Explorer data into Excel, plotted the adjustments and see what you mean by ’staircase’.
Jesus… there must be some kind of ‘statistical explanation’ – or whatever the hell scientists call it – to substantiate why the algorithm does not introduce spurious trends, and does, in fact, qualitatively improve on the ‘original’ data. And no – that the adjustment trends largely cancel themselves out over whole dataset is not a sufficient explanation, I think.
Actually, would it be possible and not too difficult to group and at least average the trends (or the resulting anomalies) somehow – without weighting them and not within grid but, say, by some of the countries?
“Do you have a citation for this? Is it gridded, or simply an average? How is the averaging done?”
No reference. I did it myself. Took 10 minutes. It is an easy thing to do, and I doubt it is publishable in of to itself, but we are working on doing something with it (essentially taking an opposite direction with respect to constructing global temp trends).
If you don’t believe me, do it yourself, or ask somebody that you trust to do it (ask a regular contributor to this blog to do it). Doing the simple average is really really easy.
For averaging w/ no gridding, I simply took the values for all the stations availible at any time point and averaged them.
Gridding adds layers of compexitity, but I did it as simply as I could.
With the completely raw data just averaged, I don’t get the the late 1930’s to be especially warm. The other “warm” period in this century shifts by about 30 years or so. I contribute this mostly to changes in the distribution of stations. In early years, they are mostly European and N. American. Post-WWII, more stations from other parts of the world come in, and you see warming (there might be a urban heat island and/or green house gas effect contributing). From there, the average decreases pretty substantially, but then it goes back up.
Ryan Stephenson (02:35:00)- I just want to point out that did I offer another option. That is to simply throw out all of the surface data.
Generally thought, I disagree with you. Through different data analysis methods people routinely pull good information out of data that has serious flaws.
Possibly, this homogenization method isn’t the best and has issues, which is why I suggested from the beginning that something that would generate real interest would be a “better” homogenzation method that doesn’t show global warming.
I spotted another logical flaw in these homogenisation approaches.
The temperature data is only measured to the nearest 0.1Celsius. In order to detect a correlation between two sites, you would have to allow variations of +/-0.1Celsius to allow for variations that are purely due to the random uncertainty in the measurement due to the low resolution of thermometers. This means that to allow correlation between two sites we must allow discrepancies of (at least) 0.1Celsius that occur between the two sets of data.
However, the climate signal we are normally looking for (not at Darwin, which seems to be a special case!) is of the order of 0.1Celsius per decade. So if we use a correlation algorithm which says “oh, if the sites don’t correlate right down to 0.1Celsius we don’t care” and we use this to correct for sites that have moved every ten years then we are permitting uncertainties in the correlation which are as big as the alleged climate change signal that we are looking for.
That isn’t mathematically sound.
Probably the correlation looks okay, because the charts go up and down in the right places, but the precision of the correlation can’t possibly be sufficient to reliably extract a climate trend signal from the merged data.
KevinUK (02:46:24) : Read your site. That is actually good work, though I’m not sure you conclusions are right.
You seem to see exactly about what you’d expect. “Corrections” lead to as much warming as cooling.
I don’t think for example, you can claim there is no reason for the GHCN to correct the data in the manner they have. There are legitimate issues with stations where you’d expect there to be artificial induced warming and cooling, and if those cases can be identified, then a correction should be made.
You could make the argument over long periods of times, they should cancel out (which is why I thought it would be interesting to just look at absolute averages), which is what your results seem to suggest.
Janis B (05:51:10) :]]
“Actually, would it be possible and not too difficult to group and at least average the trends (or the resulting anomalies) somehow – without weighting them and not within grid but, say, by some of the countries”
I’m working on this right now. I’ll be following up with a further serie sof threads on digginthclay over the next few days. You’ll be very surprised by the number of WMO stations that drop out of the NOAA GHCN analysis in the last two decades or so.
If you arent aware of them I’ll be reproducing the analyses of Giorgio Gillestri and RomanM very shortly. I had hoped GG would be prepared to extend his interesting analysis but it seems he comes from the Eric Steig school of the scientific method where you give up on your analysis once you’ve reached your pre-conceived conclusion. Pity, but I’m happy to pick up where he left off and do a somewhat more detailed analysis which will include diaggregating his analysis into NH/SH and high, mid and low latitudes in each hemisphere.
KevinUK