Foreword: The focus of this essay is strictly altitude placement/change of GHCN stations. While challenge and debate of the topic is encouraged, please don’t let the discussion drift into other side issues. As noted in the conclusion, there remain two significant issues that have not been fully addressed in GHCN. I believe a focus on those issues (particularly UHI) will best serve to advance the science and understanding of what GHCN in its current form is measuring and presenting, post processing. – Anthony

By Steven Mosher, Zeke Hausfather, and Nick Stokes
Recently on WUWT Dr. McKitrick raised several issues with regard to the quality of the GHCN temperature database. However, McKitrick does note that the methods of computing a global anomaly average are sound. That is essentially what Zeke Hausfather and I showed in our last WUWT post. Several independent researchers are able to calculate the Global Anomaly Average with very little differences between them.
GISS, NCDC, CRU, JeffId/RomanM, Tamino, ClearClimateCode, Zeke Hausfather, Chad Herman, Ron Broberg, Residual Analysis, and MoshTemp all generally agree. Given the GHCN data, the answer one gets about the pace of global warming is not in serious dispute. Whether one extrapolates as GISS does or not, whether one uses a least squares approach or a spatial averaging approach, whether one selects a 2 degree bin or a 5 degree bin, whether one uses an anomaly period of 1961-90 or 1953-1982, the answer is the same for virtually all practical purposes. Debates about methodology are either a distraction from the global warming issues at hand or they are specialist questions that entertain a few of us. Those specialist discussions may refine the answer or express our confidence in the result more explicitly, but the methods all work and agree to a high degree.
As we noted before, the discussion should therefore turn and remain focused on the data issues. How good is GHCN as a database and how serious are its shortcomings? As with any dataset, those of us who analyze data for a living look for several things. We look for errors, we look for bias, we look at the sampling characteristics, and we look at adjustments. Dr. McKitrick’s recent paper covers several topics relative to the make up and changes in GHCN temperature data. In particular he covers changes over time in the sampling of GHCN stations. He repeats a familiar note: over time the stations representing the temperature data set have changed. There is, as most people know, a fall off in stations reporting shortly after 1990 and then again in 2005. To be sure there are other issues that he raises as well. Those issues, such as UHI, will not be addressed here. Instead, the focus will be on one particular issue: altitude. We confine our discussion to that narrow point in order to remove misunderstandings and refocus the issue where it rightly belongs.
McKitrick writes:
Figure 1-8 shows the mean altitude above sea level in the GHCN record. The steady increase is consistent with a move inland of the network coverage, and also increased sampling in mountainous locations. The sample collapse in 1990 is clearly visible as a drop not only in numbers but also in altitude, implying the remote high-altitude sites tended to be lost in favour of sites in valley and coastal locations. This happened a second time in 2005. Since low-altitude sites tend to be more influenced by agriculture, urbanization and other land surface modification, the failure to maintain consistent altitude of the sample detracts from its statistical continuity.
There are several claims here.
- The increase in altitude is consistent with a move inland and out of valleys
- The increase in altitude is consistent with more sampling in mountainous locations.
- Low level sites tend to be influenced by agriculture, urbanization and other land use modifications
A simple study of the metadata available in the GHCN database shows that the stations that were dropped do not have the characteristics that McKitrick supposes. As Nick Stokes documents, the process of dropping stations is more related to dropping coverage in certain countries rather than a direct effort to drop high altitude stations . McKitrick also get the topography specifics wrong. He supposes that the drop in thermometers shifts the data out of mountainous inland areas into the valleys and low level coastal areas, areas dominated by urbanization and land use changes. That supposition is not entirely accurate as a cursory look at the metadata shows.
There are two significant periods when stations are dropped; Post 1990 and again in 2005. As Stokes show below.
FIGURE 1: Station drop and average altitude of stations.
The decrease in altitude is not caused by a move into valleys, lowland and coastal areas. As the following figures show, the percentage of coastal stations is stable, mountainous stations are still represented and the altitude loss more likely comes from the move out of mountainous valleys .
A simple summary of the total inventory shows this
| ALL STATIONS | Count | Total | Percent |
| Coastal | 2180 | 7280 | 29.95 |
| Lake | 443 | 7280 | 6.09 |
| Inland | 4657 | 7280 | 63.97 |
TABLE 1: Count of Coastal Stations
The greatest drop in stations occurs in the 1990-1995 period and the 2005 period, as shown above McKitrick supposes that the drop in altitude means a heavier weighting for coastal stations. The data do not support this
| Dropped Stations 90-95 | Count | Total | Percent |
| Coastal | 487 | 1609 | 30.27 |
| Lake | 86 | 1609 | 5.34 |
| Inland | 1036 | 1609 | 64.39 |
| Dropped in 2005-06 | |||
| Coastal | 104 | 1109 | 9.38 |
| Lake | 77 | 1109 | 6.94 |
| Inland | 928 | 1109 | 83.68 |
TABLE 2: Count of Coastal Stations dropped
The great march of the thermometers was not a trip to the beach. Neither was the drop in altitude the result of losing a higher percentage of “mountainous” stations.
FIGURE 2: Distribution of Altitude for the entire GHCN Inventory
| Minimum | 1st Qu | Median | Mean | 3rd Qu | Max | NA |
| -224.0 | 38.0 | 192.0 | 419.9 | 533.0 | 4670 | 142 |
TABLE 3: descriptive statistics for Altitude of the entire dataset
We can assess the claim about the march of thermometers down the mountains in two ways. First, by looking at the actual distribution of dropped stations.
FIGURE 3 Distribution of altitude for stations dropped in 1990-95
| Minimum | 1st Qu | Median | Mean | 3rd Qu | Max | NA |
| -21.0 | 40.0 | 183.0 | 441 | 589.2 | 4613.0 | 29 |
TABLE 4: Descriptive statistics for the Altitude of dropped stations
The character of stations dropped in the 2005 time frame are slightly different. That distribution is depicted below
FIGURE 4 Distribution of altitude for stations dropped in 2005-06
| Minimum | 1st Qu | Median | Mean | 3rd Qu | Max | NA |
| –59 | 143.0 | 291.0 | 509.7 | 681.0 | 2763.0 | 0 |
TABLE 5: Descriptive statistics for the Altitude of dropped stations 2005-06
The mean of those dropped is slightly higher than the average station. That hardly supports the contention of thermometers marching out of the mountains. We can put this issue to rest with the following observation from the metadata. GHCN metadata captures the topography surrounding the stations. There are four classifications FL, HI, MT and MV: flat, hilly, mountain and mountain valley. The table below hints at what was unique about the dropout.
| Type | Entire Dataset | Dropped after90-95 | Dropped 2005-06 | Total of two major movements |
| Flat | 2779 | 455 (16%) | 504 (23%) | 959 (43%) |
| Hilly | 3006 | 688 (23%) | 447 (15%) | 1135 (38%) |
| Mountain | 61 | 15 (25%) | 3 (5%) | 18 (30%) |
| Mountain Valley | 1434 | 451(31%) | 155 (11%) | 606 (42%) |
TABLE 6 Station drop out by topography type
There wasn’t shift into valleys as McKitrick supposes, but rather mountain valley sites were dropped. Thermometers left the flatlands and the mountainous valleys. That resulted in a slight decrease in the overall altitude.
That brings us to McKitrick’s third critical claim. McKitrick claims that the dropping of thermometers over weights places more likely to suffer from urbanization and differential land use. “Low level sites tend to be influenced by agriculture, urbanization and other land use modifications.” The primary concern that Dr. McKitrick voices is that the statistical integrity of the data may have been compromised. That claim needs to be turned into a testable hypothesis. What exactly has been compromised? We can think of two possible concerns. The first concern is that by dropping higher altitude mountain valley stations one is dropping stations that are colder. Since temperature decreases with altitude this would seem to be a reasonable concern. However, it is not. Some people make this claim, but McKitrick does not. He doesn’t because he is aware that the anomaly method prevents this kind of bias. When we create a global anomaly we prevent this kind of bias from entering the calculation by scaling the measurements of station by the mean of that station. Thus, a station located at 4000m may be at -5C, but if that station is always at -5C its anomaly will be zero. Likewise, a station at sea level in Death Valley that is constantly 110F will also have an anomaly of zero. Anomaly captures the departure from the mean of that station.
What this means is that as long as high altitude stations warm or cool at the same rate as low altitude stations, removing them or adding them will not bias the result.
To answer the question of whether dropping or adding higher altitude stations impacts the trend we have several analytical approaches. First, we could add back in stations. But we can’t add back in GHCN stations that were discontinued. The alternative is to add stations from other databases. Those studies indicate that adding addition stations does not change the trends:
http://moyhu.blogspot.com/2010/07/using-templs-on-alternative-land.html
http://moyhu.blogspot.com/2010/07/arctic-trends-using-gsod-temperature.html
http://moyhu.blogspot.com/2010/07/revisiting-bolivia.html
http://moyhu.blogspot.com/2010/07/global-landocean-gsod-and-ghcn-data.html
The other approach is to randomly remove more stations from GHCN and measure the effect. If we fear that GHCN has biased the sample by dropping higher altitude stations, we can drop more stations and measure the effect. There are two ways to do this. A Monte Carlo approach and an approach that divides the existing data into subsets:
Nick Stokes has conducted the Monte Carlo experiments. In his approach stations are randomly removed and global averages are recomputed. Stations were removed based on a randomization approach that preferentially removed high altitude stations. This test gives us an estimate of the Standard Error as well.
| Period | Trend of All | Re-Sampled | s.d |
| 1900-2009 | 0.0731 | 0.0723 | 0.00179 |
| 1979-2009 | 0.2512 | 0.2462 | 0.00324 |
| Mean Altitude | 392m | 331m |
Table 7 Monte Carlo test of altitude sensitivity
This particular test consists of selecting all the stations whose series end after 1990. There are 4814 such stations. The sensitivity to altitude reduction was performed by randomly removing higher altitude stations. The results indicate little to no interaction between altitude and temperature trend in the very stations end after the 1990 period.
The other approach, dividing the sample, was approached in two different ways by Zeke Hausfather and Steven Mosher. Hausfather, approached the problem using a paired approach. Grid cells are selected for processing if the have stations both above and below 300m. This eliminates cells that are represented by a single station. Series are then constructed for the stations that lie above 300m and below 300m.
| Period | Elevation > 300m | Elevation <300m |
| 1900-2009 | .04 | .05 |
| 1960-2009 | .23 | .19 |
| 1978-2009 | .34 | .28 |
Table 8. Comparison of trend versus altitude for paired station testing
FIGURE 5: Comparison of temperature Anomaly for above mean and below mean stations
This test indicates that higher elevation stations tend to see higher rates of warming rather than lower rates of warming. Thus, dropping them, does not bias the temperature record upward. The concern lies in the other direction. If anything the evidence points to this: dropping higher altitude stations post 1990 has lead to a small underestimation of the warming trend.
Finally, Mosher, extending the work of Broberg tested the sensitivity of altitude by dividing the existing sample in the following way, by raw altitude and by topography.
- A series containing all stations.
- A series of lower altitude stations Altitude < 200m
- A series of higher altitude stations Altitude >300m
- All Stations in Mountain Valleys
- A series of stations at very high altitude. Altitude >400m
The results of that test are shown below
FIGURE 6 Global anomaly. Smoothing performed for display purpose only with a 21 point binomial filter
The purple series is the highest altitude stations. The red series lower elevation series. Green is the mountain valley stations. A cursory look at the “trend” indicates that the higher elevation stations warm slightly faster than the lower elevation, confirming Hausfather. Dropping higher elevation stations, if it has any effect whatsoever works to lower the average. Stations at lower altitudes tend to warm less rapidly than stations at higher elevations. So quite the opposite of what people assume, the dropping of higher altitude stations is more likely to underestimate the warming rather than over estimate the warming.
Conclusion:
The distribution of altitude does change with time in GHCN v2.mean data. That change does not signal a march of thermometers to places with higher rates of warming. The decrease in altitude is not associated with a move toward or away from coasts. The decrease is not clearly associated with a move away mountainous regions and into valleys, but rather a movement out of mountain valley and flatland regions. Yet, mountain valleys do not warm or cool in any differential manner. Changing altitude does not bias the final trends in any appreciable way.
Regardless of the differential characteristics associated with higher elevation, changes in temperature trends is not clearly or demonstrably one of them. For now, we have no evidence whatsoever that marching thermometers up and down hills makes any contribution to a overestimation of the warming trend.
Dr. McKitrick presented a series of concerns with GHCN. We have eliminated the concern over changes in the distribution of altitude. That merits a correction to his paper. The concerns he raised about latitude, and airports and UHI will be addressed in forthcoming pieces. Given the preliminary work done on airports. (and here) and latitude to date, we can confidently say that the entire debate will come down to two basic issues: UHI and adjustments, the issues over latitude changes and sampling at airports will fold into those discussions. So, here is where the debate stands. The concerns that people have had about methodology have been addressed. As McKitrick notes, the various independent methods get the same answers. The concern about altitude bias has been addressed. As we’ve argued before, the real issue with temperature series is the metadata, its related microsite and UHI issues and adjustments made prior to entry in the GHCN database.
Special thanks to Ron Broberg for editorial support.
References:
A Critical Review of Global Surface Temperature Data Products. Ross McKitrick, Ph.D. July 26, 2010
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.





Mosher: “STATIONS WERE NOT DROPPED”
In Canada (using GHCN v2) they went from 500 stations reporting Max/Min in 1975 to 40 in the 2000’s and then even lower.
The GHCN v2 mean station count for Canada reporting hit 700 in 1965 and started dropping in 1981 and is now in the 30’s.
There were more stations in the the late 1870’s to 1899 reporting than in the 2000’s.
Question for a statistician:
How significant is the calculated global average warming trend of the measured local anomalies against the null hypothesis that the actual global average warming trend = 0? I motivate that question with the following observations: We have a very large, very fluid object – the global atmosphere – and we want to know whether its temperature trend is positive, negative or 0. So we randomly (or not so randomly) distribute thermometers through it and periodically (or irregularly) observe each thermometer over periods of time randomly (or not so randomly) distributed both in duration and in starting/ending. This gives us a large number of temperature records, from each of which we can extract a trend. Each trend should come with an error estimate. Some of the trends will be positive, some negative and some, within the bounds of error, will be 0. Now we combine all these local trends into one giant global trend, which should be stated with some estimate of error. It, too, could be positive, negative, or, within the bounds of error, 0. So, to return to my initial question, given the actual errors of the actual thermometers used to create GHCN, and the actual distribution of these thermometers in space and time; and considering their readings as ESTIMATES of the actual local temperatures at the time of the readings; what is the expected error of the calculated global mean trend as an ESTIMATOR of the actual global mean trend? And is the calculated global mean trend significantly different from 0, given the structure of the sample in relation to the population, and the errors inherent in the measurements?
Station altitude per se is not the issue, it’s how representative that station record is of climatic temperature variations in the region (not city) surrounding it. That seems to be the crucial point that is missed in the sausage method of stitching together anomalies from an ever-variable set of stations and then pointing to “the observed anomaly trend,” as if it were an inherent climatic feature.
Anomalies are not legitimately interchangable from station to station, unless there is complete spatial homogeneity of temperature variabilty. That is almost never the case. The march of thermometers that led to a sharp decrease in “rural” stations in key regions makes the constructed anomaly series scientifically invalid. And as Anna V points out, quite beside the central thermal energy issue.
There is something wholly totemic in the premise that, because a similar result is obtained by different people using the same broken data set and only somewhat different processing techniques, the result is thereby validated. Climate data crunchers need to learn a lesson from dumpster divers in recognizing the difference between rotten fruit and something perhaps bruised, but edible. I agree completely with E.M. Smith when he calls the GHCN data crappy and concludes: “Admiring the uniformity of the crappiness does not yield much comfort.”
JT says:
August 21, 2010 at 7:48 am
You pose a very good question to which I have time to provide only a brief answer.
The crux of the matter lies in the fact that bona fide climatic time-series do NOT exhibit a consistent trend. What is obtained by fitting such to available data is highly variable values that depend not only on record-length but also upon start time. In other words, the stochastic structure of climatic variations is such that the trend plus noise model simply doen not fit. Because that structure involves oscillatory components, which may be modelled as autoregressions of fairly high order, there is no analytic formula quantifying the uncertainty. You can be sure, however, that the confidence intervals are very much wider than those for an AR(1) process, which AGW alarmists often invoke. Hope this helps.
Here is a great post from a guy who has worked with Gisstemp.
He is a fellow who helped me with my google earth code.
Station dropout. If the concern that dropped stations will impact the trend ( raise them) Then there is a simple way to test that, as EM Smith noted. Look at what Gisstemp gives you if only look at long lived stations.
Say you start with 1000 stations, then more stations become available through 1990 and you get to 7000. After 1990, you are back down to 1500 or so.
Concerned about the stations gone missing? well DONT ADD THEM IN. If taking them out raises the trend, then never put them in to begin with. If there is an effect to taking them out, then there should be an effect to putting them in, adding in all those super high altitude and high latitude sites should have an impact ( they dont)
http://oneillp.wordpress.com/2010/06/25/the-effect-of-station-dropout-on-gistemp/#more-910
BillyBob says:
August 20, 2010 at 5:05 pm (Edit)
Mosher: “STATIONS WERE NOT DROPPED”
In Canada (using GHCN v2) they went from 500 stations reporting Max/Min in 1975 to 40 in the 2000′s and then even lower.
###########################################
You still dont get it. When the GHCN was compiled and published, the published
WHAT THEY HAD. Its not like they were collecting data all along and dropped stations in the early 90s. After the compilation of the HISTORICAL data (GHCN.. H=historical) they wanted to turn it into a database with monthly updates, so they focused on incorporating stations using CLIMAT. stations are not dropped. Its not a conspiracy.
http://journals.ametsoc.org/doi/abs/10.1175/1520-0477(1997)078%3C2837:AOOTGH%3E2.0.CO;2
This chart gives a better idea of what’s been happening to the temperature record.
You can see where the raw & urban readings begin to diverge.
http://homeclimateanalysis.blogspot.com/2009/12/continuous-stations.html
Kevan Hashemi said…
By special request: the trend we get from stations that reported during at least 35 out of the 40 years from 1960 to 2000. Analysis code is here, and the graph is http://www.hashemifamily.com/Kevan/Climate/Cont_80_1960_2000.gif . And using the same code, but a different reference period, we have the trend from stations that reported for at least 20 years during the period 1840 to 1880 as well, http://www.hashemifamily.com/Kevan/Climate/Cont_80_1840_1880.gif
Mr. Mosher, with all due respect, it appears you have not understood the argument made by E.M. Smith. After carefully reading all his posts on the subject, and your analysis here, I’m afraid Mr. Smith has the better argument.
The basic problem is that you, and all the others who believe in the GISS and other datasets, I don’t mean to imply that you are alone in this, take the data as valid. It indeed may be valid for that moment in time when the observation was made.
But, when the long-term data for a given site consists of multiple stations from different locations, e.g. a station move, or an instrument upgrade, or any of the other changes that occur in far too many stations, there are insurmountable problems. For example, if one has 10 years of data in one location, then 10 subsequent years after relocating the instrument, then yet another 10 years at still a different location, one cannot establish the 30-year norm or mean for purposes of computing an anomaly. Mr. Watts has posted on WUWT several examples of really bad observations that illustrate this.
All the effort expended in defending the GHCN should be better targeted to curing the defects that are created from splicing together disparate data segments.
The summation by E.M. Smith, that of admiring the uniformity of the crappiness, is completely on target. That phrase gets my vote for Climate Comment of the Year.
Smokey says:
August 22, 2010 at 1:29 pm
The USHCN yearly chart you present is very interesting and highly similar to a QCed compilation made via a much-different sampling scheme. Is the numerical data series available? I’d like to do a comprehensive cross-comparison with said compilation and share the results with you.
sky,
I got that chart here. It’s apparently from a June ’09 article. Sorry I can’t be more helpful. ☹