Calculating global temperature

I’m happy to present this essay created from both sides of the aisle, courtesy of the two gentlemen below. Be sure to see the conclusion. I present their essay below with only a few small edits for spelling, format, and readability. Plus an image, a snapshot of global temperatures.  – Anthony

http://veimages.gsfc.nasa.gov/16467/temperature_airs_200304.jpg
Image: NASA The Atmospheric Infrared Sounder (AIRS) instrument aboard NASA’s Aqua satellite senses temperature using infrared wavelengths. This image shows temperature of the Earth’s surface or clouds covering it for the month of April 2003.

By Zeke Hausfather and Steven Mosher

There are a variety of questions that people have about the calculation of a global temperature index. Questions that range from the selection of data and the adjustments made to data, to the actual calculation of the average. For some there is even a question about whether the measure makes any sense or not. It’s not possible to address all these questions in one short piece, but some of them can be addressed and reasonably settled. In particular we are in a position to answer the question about potential biases in the selection of data and biases in how that data is averaged.

To move the discussion onto the important matters of adjustments to data or, for example, UHI issues in the source data it is important to move forward on some answerable questions. Namely, do the methods for averaging data, the methods of the GISS, CRU and NCDC bias the result? There are a variety of methods for averaging spatial data, do the methods selected and implemented by the big three bias the result?

There has been a trend of late among climate bloggers on both sides of the divide to develop their own global temperature reconstructions. These have ranged from simple land reconstructions using GHCN data

(either v2.mean unadjusted data or v2.mean_adj data) to full land/ocean reconstructions and experiments with alternative datasets (GSOD , WDSSC , ISH ).

Bloggers and researchers who have developed reconstructions so far this year include:

Roy Spencer

Jeff Id

Steven Mosher

Zeke Hausfather

Tamino

Chad

Nick Stokes

Residual Analysis

And, just recently, the Muir Russell report

What is interesting is that the results from all these reconstructions are quite similar, despite differences in methodologies and source data. All are also quite comparable to the “big three” published global land temperature indices: NCDC , GISTemp , and CRUTEM .

[Fig 1]

The task of calculating global land temperatures is actually relatively simple, and the differences between reconstructions can be distilled down to a small number of choices:

1. Choose a land temperature series.

Ones analyzed so far include GHCN (raw and adjusted), WMSSC , GISS Step 0, ISH , GSOD , and USHCN (raw, time-of-observation adjusted, and F52 fully adjusted). Most reconstructions to date have chosen to focus on raw datasets, and all give similar results.

[Fig 2]

It’s worth noting that most of these datasets have some overlap. GHCN and WMSSC both include many (but not all) of the same stations. GISS Step 0 includes all GHCN stations in addition to USHCN stations and a selection of stations from Antartica. ISH and GSOD have quite a bit of overlap, and include hourly/daily data from a number of GHCN stations (though they have many, many more station records than GHCN in the last 30 years).

2. Choosing a station combination method and a normalization method.

GHCN in particular contains a number of duplicate records (dups) and multiple station records (imods) associated with a single wmo_id. Records can be combined at a single location and/or grid cell and converted into anomalies through the Reference Station Method (RSM), the Common Anomalies Method (CAM), and First Differences Method (FDM), or the Least Squares Method (LSM) developed by Tamino and Roman M . Depending on the method chosen, you may be able to use more stations with short records, or end up discarding station records that do not have coverage in a chosen baseline period. Different reconstructions have mainly made use of CAM (Zeke, Mosher, NCDC) or LSM (Chad, Jeff Id/Roman M, Nick Stokes, Tamino). The choice between the two does not appear to have a significant effect on results, though more work could be done using the same model and varying only the combination method.

[Fig 3]

3. Choosing an anomaly period.

The choice of the anomaly period is particularly important for reconstructions using CAM, as it will determine the amount of usable records. The anomaly period can also result in odd behavior of anomalies if it is too short, but in general the choice makes little difference to the results. In the figure that follows Mosher shows the difference between picking an anomaly period like CRU does, 1961-1990, and picking an anomaly period that maximizes the number monthly reports in a 30 year period.  The period that maximizes the number of monthly reports over a 30 year period turns out to be 1952-1983.  1953-82 (Mosher). No other 30 year period in GHCN has more station reports. This refinement, however, has no appreciable impact.

[Fig 4]

4. Gridding methods.

Most global reconstructions use 5×5 grid cells to ensure good spatial coverage of the globe. GISTemp uses a rather different method of equal-size grid cells. However, the choice between the two methods does not seem to make a large difference, as GISTemp’s land record can be reasonably well-replicated using 5×5 grid cells. Smaller resolution grid cells can improve regional anomalies, but will often result in spatial bias in the results, as there will be large missing areas during periods when or in locations when station coverage is limited. For the most part, the choice is not that important, unless you choose extremely large or small gridcells. In the figure that follows Mosher shows that selecting a smaller grid does not impact the global average or the trend over time. In his implementation there is no averaging or extrapolation over missing grid cells. All the stations within a grid cell are averaged and then the entire globe is averaged. Missing cells are not imputed with any values.

[Fig 5]

5. Using a land mask.

Some reconstructions (Chad, Mosh, Zeke, NCDC) use a land mask to weight each grid cell by its respective land area. The land mask determines how much of a given cell ( say 5×5) is actually land. A cell on a coast, thus, could have only a portion of land in it. The land mask corrects for this. The percent of land in a cell is constructed from a 1 km by 1 km dataset. The net effect of land masking is to increase the trend, especially in the last decade. This factor is the main reason why recent reconstructions by Jeff Id/Roman M and Nick Stokes are a bit lower than those by Chad, Mosh, and Zeke.

[Fig 6]

6. Zonal weighting.

Some reconstructions (GISTemp, CRUTEM) do not simply calculate the land anomaly as the size-weighted average of all grid cells covered. Rather, they calculate anomalies for different regions of the globe (each hemisphere for CRUTEM, 90°N to 23.6°N, 23.6°N to 23.6°S and 23.6°S to 90°S for GISTemp) and create a global land temp as the weighted average of each zone (weightings 0.3, 0.4 and 0.3, respectively for GISTemp, 0.68 × NH + 0.32 × SH for CRUTEM). In both cases, this zonal weighting results in a lower land temp record, as it gives a larger weight to the slower warming Southern Hemisphere.

[Fig 7]

These steps will get you a reasonably good global land record. For more technical details, look at any of the many http://noconsensus.wordpress.com/2010/03/25/thermal-hammer-part-deux/different  http://residualanalysis.blogspot.com/2010/03/ghcn-processor-11.html models  http://rankexploits.com/musings/2010/a-simple-model-for-spatially-weighted-temp-analysis/ that have been publicly  http://drop.io/treesfortheforest released http://moyhu.blogspot.com/2010/04/v14-with-maps-conjugate-gradients.html

].

7. Adding in ocean temperatures.

The major decisions involved in turning a land reconstruction into a land/ocean reconstruction are choosing a SST series (HadSST2, HadISST/Reynolds, and ERSST have been explored  http://rankexploits.com/musings/2010/replication/ so far), gridding and anomalizing the series chosen, and creating a combined land-ocean temp record as a weighted combination of the two. This is generally done by: global temp = 0.708 × ocean temp + 0.292 × land temp.

[Fig 8]

8. Interpolation.

Most reconstructions only cover 5×5 grid cells with one or more station for any given month. This means that any areas without station coverage for any given month are implicitly assumed to have the global mean temperature. This is arguably problematic, as high-latitude regions tend to have the poorest coverage and are generally warming faster than the global average.

GISTemp takes a somewhat different approach, assigning a temperature anomaly to all missing grid boxes located within 1200 km of one or more stations that do have defined temperature anomalies. They rationalize this based on the fact that “temperature anomaly patterns tend to be large scale, especially at middle and high latitudes.” Because GISTemp excludes SST readings from areas with sea ice cover, this leads to the extrapolation of land anomalies to ocean areas, particularly in the Arctic. The net effects of interpolation on the resulting GISTemp record is small but not insignificant, particularly in recent years. Indeed, the effect of interpolation is the main reason why GISTemp shows somewhat different trends from HadCRUT and NCDC over the past decade.

[Fig 9]

9. Conclusion

As noted above there are many questions about the calculation of a global temperature index. However, some of those questions can be fairly answered and have been fairly answered by a variety of experienced citizen researchers from all sides of the debate. The approaches used by GISS and CRU and NCDC do not bias the result in any way that would erase the warming we have seen since 1880. To be sure there are minor differences that depend upon the exact choices one makes, choices of ocean data sets, land data sets, rules for including stations, rules for gridding, area weighting approaches, but all of these differences are minor when compared to the warming we see.

That suggests a turn in the discussion to the matters which have not been as thoroughly investigated by independent citizen researchers on all sides:

A turn to the question of data adjustments and a turn to the question of metadata accuracy and finally a turn to the question about UHI. Now, however, the community on all sides of the debate has a set of tools to address these questions.

Get notified when a new post is published.
Subscribe today!
0 0 votes
Article Rating
194 Comments
Inline Feedbacks
View all comments
July 13, 2010 2:20 pm

Much ado about nothing, clearly. A change of 1 degree C over 130 years is of negligible importance — particularly when the underlying mechanisms of the change remain uncertain and largely unknown.
The Earth has seen much more radical climate changes in the past, long before humans made the scene. All the hoopla over these global climate metrics approach mental masturbatory status.

July 13, 2010 2:23 pm

And for those interested, I recently completed a detailed overview of the different Sea Surface Temperture datasets. Refer to:
http://bobtisdale.blogspot.com/2010/07/overview-of-sea-surface-temperature.html

July 13, 2010 2:24 pm

Anthony,
Could I suggest hyperlinking all the URLs to the word that immediately precedes them? It might make the text a bit more readable.
REPLY: You are quite welcome. If either of you want to revise your manuscript, I’ll be happy to repost it. I posted it as it was provided. – Anthony
Reply 2: I just did the linking for Mosh. Maybe I can get him to clean the kitchen in return. ~ ctm

Varco
July 13, 2010 2:25 pm

Excellent post, entirely in character for this fine blog.

July 13, 2010 2:26 pm

Well, Mosh, it was nailbiting reading through that – I kept wondering when you would mention UHI. Finally you did. But how easy is that to correct for? It must be difficult, otherwise HadCRUT3 would have done it 😉
Rich.

July 13, 2010 2:34 pm

Rich,
UHI is a tough one, simply because it can depend so much on micro-site effects that are difficult to quantify.
There has been a fair bit of work on trying to use macro-site characteristics (through various urbanity proxies like GRUMP and satellite nightlights), e.g.:
http://rankexploits.com/musings/2010/uhi-in-the-u-s-a/
http://rankexploits.com/musings/2010/in-search-of-the-uhi-signal/
That said, there is plenty of work still to be done. Now that everyone and their mother has created an anomaly and gridding tool, various bloggers can do their own analysis using existing metadata, CRN rating, etc.

latitude
July 13, 2010 2:45 pm

I guess Jones was right after all. At least they all show no statistical warming for the past 15 years.
But 1 degree in over 100 years?
How much of that 1 degree rise is from adjusting the earlier temps down?

jmc
July 13, 2010 2:47 pm

All graphs show an upward trend since 1970, but see figures 11 to 20 of “The Cause of Global Warming” by Vincent Gray, January 2001 here. The conclusion is:
“Global temperature measurements remote from human habitation and activity show no evidence of a warming during the last century. … The small average and highly irregular individual warming displayed by surface measurements is therefore caused by changes in the thermal environment of individual measurement stations over long periods of time, and not by changes in the background climate.”

R. Gates
July 13, 2010 2:48 pm

Superb article…exactly why WUWT is #1.
Thanks!

David A. Evans
July 13, 2010 3:11 pm

How RAW is RAW?
DaveE.

rbateman
July 13, 2010 3:12 pm

I would like a link to just one of those plotted dataset files. The yearly mean temp, or the anomaly data plus the mean temp it is based upon.
Something that Leif told me about regarding a standard in plotting data, to keep things in perspective.

carrot eater
July 13, 2010 3:13 pm

Nicely done, gentlemen. This is a good summary, and should serve as a useful reference.
Two things: it is not entirely clear from your language, but I’m pretty sure that GISS always interpolates to the center of the box, regardless of whether the box has stations or not. Of course, when the (sub)box does have stations of its own, then anything being interpolated from afar will have little influence.
Second, it’s worth emphasizing again that most (looking over it, probably all) of your graphs presented here use unadjusted land station data. Quite simply, global warming is not an artifact of adjustments. This appears to continuously be surprising to some people.

EthicallyCivil
July 13, 2010 3:16 pm

The starting point for all of these analyzes seems to be the “we’ve cooled the 40’s for your protection” post-Hansen baseline, the one that lowered past temperatures based on the need for homogeneity between urbanized and non-urbanized sights. Remember the one with the logic — “gee we know there is no UHI, thus if rural sites aren’t showing the same temperature growth as urban sites they must have been reporting too hot in the past” that way the temperature growth is homogenized (and as such a UHI signal becomes *very* hard to find, as only a residual remains).
Or am I wrong?

George E. Smith
July 13, 2010 3:19 pm

So what happened to the story in the headline:- “”” Calculating global temperature
Posted on July 13, 2010 by Steven Mosher “””
So finally I thought we were going to find out how to measure global Temperature; specially with that impressive; but rather limited range NASA AIRS picture.
Too bad their instrument only goes from -81 to + 47 deg C. Why not cover the full +/- 90 deg C range that can cover from Vostok lows; to asphalt surface highs.
But then all hell breaks loose, and apparnetly they lose sight of the headlien; and all we find is graphs of anomalies; not real Temperatures.
Yes measuring the earth’s surface Temperature is simple; you put a thermometer in the center of each global grid cell, and you read them all simultaneously; then you multiply each readingt by the cell area; add them all up and divide by the global surface area.
Well I didn’t see anything like that happening here.
Of course you have to choose the cell size properly so that you have at least one cell sample for each half cycle of the highest spatial frequency that occurs in the global temperature map.
Looking at the daily weather map for the SF Bay area nine counties; it appears that you need a cell size no bigger than about 1 km on a side or else you will be undersampling.
Well of course there probably haven’t been that many thermometers made since Merucry was discovered.
then there’s that 1961 to 1990 base period for all those anomaly graphs; what is that all about. If they didn’t measure the correct global temperature during the base period; then of course the anomalies don’t have any real reference either.
Why not admit, that ALL of these records; are simply the result of statistical manipulations on the output of a certain set of thermometers; and that any connection between the result of that computation and the actual temperature of the earth is without any scientific foundation; it’s a fiction.
And it’s a meaningless fiction since there isn’t any scientific relationship between the temperature at any earth location and the local flux of energy into or out of planet earth. Each different type of terrain results in different thermal processes, and there is no common link between energy flow, and local Temperature.
Energy flow is a consequence of Temperature differences between locations; it is not simply related to the atual Temperature at any place; or at any time.
And anyone who thinks a 1200 km cell size is sufficient just doesn’t understand the problem.

BillyBob
July 13, 2010 3:22 pm

What kind of warming “have we seen”?
Without the min/max we don’t don’t know if:
1) Both he min and max have gone up
2) or the min has gone up and the max has not
3) or the max has gone up and the min has not
If you use the top 50 US state temperature records as a proxy, and not 25 of the top 50 max records are int he 1930’s, one could conclude that the max is lower than the 1930’s and the min has gone up.
If all the warming is raised min’s caused by UHI and station dropout, we have zero to worry about.
So, Zeke and Steve, is there a reliable record for min/max anywhere?

Editor
July 13, 2010 3:23 pm

Very good article, well written and logical in its format.
I remain to be convinced of the merits of a global temperature comprised of thousands of stations that record one micro climate, then move or are subsumed by uhi and are consequently recording a completely different micro climate to the one they started off with. As for the idea that we have accurate global ocean temperature records dating back over a century-where are they supposed to have come from?
Be that as it may, how do we explain the fact that CO2 is being blamed for rising temperatures since 1880 yet our instrumental records show temperatures have been increasing for 350 years?
This is Co2 superimposed on CET showing linear regression
http://c3headlines.typepad.com/.a/6a010536b58035970c0120a7c87805970b-pi
These are other historic data sets with linear regression
http://i47.tinypic.com/2zgt4ly.jpg
http://i45.tinypic.com/125rs3m.jpg
Other historic datasets from around the world are collected on my site here, together with weather observations, diaries and articles.
http://climatereason.com/LittleIceAgeThermometers/
The giss and cru records merely plugged into the end of a steadily rising trend established two hundred years previously. They did not record the start of the rising trend.
tonyb

George E. Smith
July 13, 2010 3:27 pm

On #4 Gridding Methods; just what the heck grid are we talking about ?
I thought both Hadcrud, and GISStemp used data from some small number of thermometers spread around the world; so what the heck are these grid cells and what do 5 x 5 and 3 x 3 grids mean ?

July 13, 2010 3:28 pm

Hansen seems to think that a global temperature is not that easy …
http://eureferendum.blogspot.com/2009/12/that-elusive-temperature.html

John Trigge
July 13, 2010 3:30 pm

All of these reconstructions rely on data from dubious sources as they start well before the (probable) unbiassed satellite records.
The Surface Stations project (thanks, Anthony) confirms garbage in, garbage out.
Given the siting issues Anthony and others have PROVEN, how much credence can be placed on ANY of the temperatures graphs.
Until the World agrees on, finances and impliments a global, open, transparent, long-term and unbiassed temperature measurement system, all of these reconstructions will be able to be dismisssed by one side whilst lauded by the other.
AND NONE OF THEM PROVE THAT CO2 IS THE CAUSE

John from CA
July 13, 2010 3:30 pm

Thanks — its great to see a comparison of approaches that generally produce the same result for a change. I was beginning to worry that we might have to pick a different computer model behind curtain #3 every day.
Couple of questions:
– why aren’t rogue measurements discarded as we would discard a looney response to a market research study?
– the charts don’t indicate a degree of accuracy — do they need to?

George E. Smith
July 13, 2010 3:31 pm

What is the reason for “Correcting” for UHI.
Are not UHI regions of higher temperature than their rural surroundings ? And if they are then they surely must be contributing to higher temperature readings; so what is to correct; the temperature is going up because of UHI; and Mother Gaia takes all of that into consideration, when she decides what the earth’s temperature should be.
So why are we doing it differently; is not correction simply making up false data ?

DirkH
July 13, 2010 3:32 pm

Here’s a guy who did a very simple analysis of raw data who comes to the conclusion that there is no discernible trend:
http://crapstats.wordpress.com/2010/01/21/global-warming-%e2%80%93-who-knows-we-all-care/
He compensates the death of thermometers by using, for each year-to-year interval, only the thermometers that exist in both years – a so-called “constant set”.
(NOTE: I added the image for DirkH, since it is interesting – Anthony)
From Crapstats blog

1 2 3 8