Calculating global temperature

I’m happy to present this essay created from both sides of the aisle, courtesy of the two gentlemen below. Be sure to see the conclusion. I present their essay below with only a few small edits for spelling, format, and readability. Plus an image, a snapshot of global temperatures.  – Anthony

http://veimages.gsfc.nasa.gov/16467/temperature_airs_200304.jpg
Image: NASA The Atmospheric Infrared Sounder (AIRS) instrument aboard NASA’s Aqua satellite senses temperature using infrared wavelengths. This image shows temperature of the Earth’s surface or clouds covering it for the month of April 2003.

By Zeke Hausfather and Steven Mosher

There are a variety of questions that people have about the calculation of a global temperature index. Questions that range from the selection of data and the adjustments made to data, to the actual calculation of the average. For some there is even a question about whether the measure makes any sense or not. It’s not possible to address all these questions in one short piece, but some of them can be addressed and reasonably settled. In particular we are in a position to answer the question about potential biases in the selection of data and biases in how that data is averaged.

To move the discussion onto the important matters of adjustments to data or, for example, UHI issues in the source data it is important to move forward on some answerable questions. Namely, do the methods for averaging data, the methods of the GISS, CRU and NCDC bias the result? There are a variety of methods for averaging spatial data, do the methods selected and implemented by the big three bias the result?

There has been a trend of late among climate bloggers on both sides of the divide to develop their own global temperature reconstructions. These have ranged from simple land reconstructions using GHCN data

(either v2.mean unadjusted data or v2.mean_adj data) to full land/ocean reconstructions and experiments with alternative datasets (GSOD , WDSSC , ISH ).

Bloggers and researchers who have developed reconstructions so far this year include:

Roy Spencer

Jeff Id

Steven Mosher

Zeke Hausfather

Tamino

Chad

Nick Stokes

Residual Analysis

And, just recently, the Muir Russell report

What is interesting is that the results from all these reconstructions are quite similar, despite differences in methodologies and source data. All are also quite comparable to the “big three” published global land temperature indices: NCDC , GISTemp , and CRUTEM .

[Fig 1]

The task of calculating global land temperatures is actually relatively simple, and the differences between reconstructions can be distilled down to a small number of choices:

1. Choose a land temperature series.

Ones analyzed so far include GHCN (raw and adjusted), WMSSC , GISS Step 0, ISH , GSOD , and USHCN (raw, time-of-observation adjusted, and F52 fully adjusted). Most reconstructions to date have chosen to focus on raw datasets, and all give similar results.

[Fig 2]

It’s worth noting that most of these datasets have some overlap. GHCN and WMSSC both include many (but not all) of the same stations. GISS Step 0 includes all GHCN stations in addition to USHCN stations and a selection of stations from Antartica. ISH and GSOD have quite a bit of overlap, and include hourly/daily data from a number of GHCN stations (though they have many, many more station records than GHCN in the last 30 years).

2. Choosing a station combination method and a normalization method.

GHCN in particular contains a number of duplicate records (dups) and multiple station records (imods) associated with a single wmo_id. Records can be combined at a single location and/or grid cell and converted into anomalies through the Reference Station Method (RSM), the Common Anomalies Method (CAM), and First Differences Method (FDM), or the Least Squares Method (LSM) developed by Tamino and Roman M . Depending on the method chosen, you may be able to use more stations with short records, or end up discarding station records that do not have coverage in a chosen baseline period. Different reconstructions have mainly made use of CAM (Zeke, Mosher, NCDC) or LSM (Chad, Jeff Id/Roman M, Nick Stokes, Tamino). The choice between the two does not appear to have a significant effect on results, though more work could be done using the same model and varying only the combination method.

[Fig 3]

3. Choosing an anomaly period.

The choice of the anomaly period is particularly important for reconstructions using CAM, as it will determine the amount of usable records. The anomaly period can also result in odd behavior of anomalies if it is too short, but in general the choice makes little difference to the results. In the figure that follows Mosher shows the difference between picking an anomaly period like CRU does, 1961-1990, and picking an anomaly period that maximizes the number monthly reports in a 30 year period.  The period that maximizes the number of monthly reports over a 30 year period turns out to be 1952-1983.  1953-82 (Mosher). No other 30 year period in GHCN has more station reports. This refinement, however, has no appreciable impact.

[Fig 4]

4. Gridding methods.

Most global reconstructions use 5×5 grid cells to ensure good spatial coverage of the globe. GISTemp uses a rather different method of equal-size grid cells. However, the choice between the two methods does not seem to make a large difference, as GISTemp’s land record can be reasonably well-replicated using 5×5 grid cells. Smaller resolution grid cells can improve regional anomalies, but will often result in spatial bias in the results, as there will be large missing areas during periods when or in locations when station coverage is limited. For the most part, the choice is not that important, unless you choose extremely large or small gridcells. In the figure that follows Mosher shows that selecting a smaller grid does not impact the global average or the trend over time. In his implementation there is no averaging or extrapolation over missing grid cells. All the stations within a grid cell are averaged and then the entire globe is averaged. Missing cells are not imputed with any values.

[Fig 5]

5. Using a land mask.

Some reconstructions (Chad, Mosh, Zeke, NCDC) use a land mask to weight each grid cell by its respective land area. The land mask determines how much of a given cell ( say 5×5) is actually land. A cell on a coast, thus, could have only a portion of land in it. The land mask corrects for this. The percent of land in a cell is constructed from a 1 km by 1 km dataset. The net effect of land masking is to increase the trend, especially in the last decade. This factor is the main reason why recent reconstructions by Jeff Id/Roman M and Nick Stokes are a bit lower than those by Chad, Mosh, and Zeke.

[Fig 6]

6. Zonal weighting.

Some reconstructions (GISTemp, CRUTEM) do not simply calculate the land anomaly as the size-weighted average of all grid cells covered. Rather, they calculate anomalies for different regions of the globe (each hemisphere for CRUTEM, 90°N to 23.6°N, 23.6°N to 23.6°S and 23.6°S to 90°S for GISTemp) and create a global land temp as the weighted average of each zone (weightings 0.3, 0.4 and 0.3, respectively for GISTemp, 0.68 × NH + 0.32 × SH for CRUTEM). In both cases, this zonal weighting results in a lower land temp record, as it gives a larger weight to the slower warming Southern Hemisphere.

[Fig 7]

These steps will get you a reasonably good global land record. For more technical details, look at any of the many http://noconsensus.wordpress.com/2010/03/25/thermal-hammer-part-deux/different  http://residualanalysis.blogspot.com/2010/03/ghcn-processor-11.html models  http://rankexploits.com/musings/2010/a-simple-model-for-spatially-weighted-temp-analysis/ that have been publicly  http://drop.io/treesfortheforest released http://moyhu.blogspot.com/2010/04/v14-with-maps-conjugate-gradients.html

].

7. Adding in ocean temperatures.

The major decisions involved in turning a land reconstruction into a land/ocean reconstruction are choosing a SST series (HadSST2, HadISST/Reynolds, and ERSST have been explored  http://rankexploits.com/musings/2010/replication/ so far), gridding and anomalizing the series chosen, and creating a combined land-ocean temp record as a weighted combination of the two. This is generally done by: global temp = 0.708 × ocean temp + 0.292 × land temp.

[Fig 8]

8. Interpolation.

Most reconstructions only cover 5×5 grid cells with one or more station for any given month. This means that any areas without station coverage for any given month are implicitly assumed to have the global mean temperature. This is arguably problematic, as high-latitude regions tend to have the poorest coverage and are generally warming faster than the global average.

GISTemp takes a somewhat different approach, assigning a temperature anomaly to all missing grid boxes located within 1200 km of one or more stations that do have defined temperature anomalies. They rationalize this based on the fact that “temperature anomaly patterns tend to be large scale, especially at middle and high latitudes.” Because GISTemp excludes SST readings from areas with sea ice cover, this leads to the extrapolation of land anomalies to ocean areas, particularly in the Arctic. The net effects of interpolation on the resulting GISTemp record is small but not insignificant, particularly in recent years. Indeed, the effect of interpolation is the main reason why GISTemp shows somewhat different trends from HadCRUT and NCDC over the past decade.

[Fig 9]

9. Conclusion

As noted above there are many questions about the calculation of a global temperature index. However, some of those questions can be fairly answered and have been fairly answered by a variety of experienced citizen researchers from all sides of the debate. The approaches used by GISS and CRU and NCDC do not bias the result in any way that would erase the warming we have seen since 1880. To be sure there are minor differences that depend upon the exact choices one makes, choices of ocean data sets, land data sets, rules for including stations, rules for gridding, area weighting approaches, but all of these differences are minor when compared to the warming we see.

That suggests a turn in the discussion to the matters which have not been as thoroughly investigated by independent citizen researchers on all sides:

A turn to the question of data adjustments and a turn to the question of metadata accuracy and finally a turn to the question about UHI. Now, however, the community on all sides of the debate has a set of tools to address these questions.

Get notified when a new post is published.
Subscribe today!
0 0 votes
Article Rating
194 Comments
Inline Feedbacks
View all comments
BillyBob
July 13, 2010 11:17 pm

Mosher: “2. GHCN has daily min/max. others have as well.”
2. GHCN has daily min/max. others have as well.
GHCN v2 max/min for Canada ( for example) drops from 500-600 stations to 20-30 in the 1990s. Its a joke.
The raw data shows the max is cooling by the way.
For examples this is raw GHCN V2 max data for June/July/Aug ranked in 10 year periods for the USA. The 30s were the hottest (max) period.
Decade JJA
1930 – 1939 30.24
1929 – 1938 30.19
1931 – 1940 30.17
1928 – 1937 30.07
1932 – 1941 30.06
1933 – 1942 29.99
1934 – 1943 29.93
1927 – 1936 29.87
1935 – 1944 29.77
1925 – 1934 29.71
1926 – 1935 29.68
1936 – 1945 29.65
1924 – 1933 29.53
1917 – 1926 29.49
1893 – 1902 29.49
1952 – 1961 29.47
1916 – 1925 29.46
1892 – 1901 29.46
1913 – 1922 29.44
1951 – 1960 29.42
1937 – 1946 29.41
1923 – 1932 29.40
1994 – 2003 29.40

July 13, 2010 11:31 pm

EC: That’s what’s bothered me. How can we have AGW that exempts the US
There are a couple of answers to that question.
The one I’ll go with is that, according to the IPCC AR4 WG1,
the “A” part of “GW” is only, in just the last several decades,
beginning to stand out from other natural forcings.
http://www.ipcc.ch/publications_and_data/ar4/wg1/en/spmsspm-understanding-and.html
In other words, the earlier highs were dominated by natural causes. The current highs are to some degree greater than zero a product of increasing CO2 (as determined by modeling analysis )

DirkH
July 14, 2010 12:41 am

Steven Mosher says:
July 13, 2010 at 11:03 pm
“DirkH says:
July 13, 2010 at 3:50 pm (Edit)
“And right after 82, a steep temp rise (and declining thermometer population).
Not accusing anyone of anything, just saying.”
That’s largely incorrect.
1. the answer is not changed by station drop off and
2. The stations do not drop that quickly after 1982.
3. If I change the period you will still get a rise.
4. If I pick 1000 stations for the whole period you will still see a rise.”
Your points (1),(2),(3) and (4) might be correct, but they don’t interfere with my words:
“And right after 82, a steep temp rise (and declining thermometer population).”
Your “That’s largely incorrect.” talks about a possible conclusion that i intentionally did not write down.

KenB
July 14, 2010 12:48 am

This whole issue is so over complicated by failure to either transparently clean up poor temperature records and audit the sites for compliance and, then rate them properly for adjustment of things like UHI, along with cavalier one size fits all global extrapolation that is also on the face of it “not transparent” Its hardly surprising we are where we are.
I can also understand the frustration of Anthony when a site is proved to be so badly sited that it is withdrawn from the present system, but its rotten data is left to contaminate the historical record. (If that’s wrong feel free to correct A)
When looking back over historical hottest, coldest records its bad enough that some of these may be due to historical errors, deliberate skewing to maintain a locations claim (that happened in Australia) (Source BOM History The Weather Watchers) or poorly sited equipment that is poorly maintained.
But then applying guesswork and adjusting temperatures down in historical records can also skew modern interpretation, especially where trends are constantly used to illustrate extreme views. It takes very little to adjust a model bias, and when trust is lost within the scientific process, suspicion abounds.
Hopefully with co-operation on sites such as this eventually, some real consensus on data interpretation and weighting methods to be applied might be reached and confidence returned.
my 2 cents !

DirkH
July 14, 2010 12:53 am

Ron Broberg says:
July 13, 2010 at 4:00 pm
“Dirk H: Here’s a guy who did a very simple analysis of raw data who comes to the conclusion that there is no discernible trend:
That guy freely admits that he did no geographic weighting. GHCN has a high percentage of US stations – and a low percentage of Southern Hemisphere stations. ”
Is the SH warming faster than the NH? It didn’t seem so to me in GISS’s famous global anomaly maps. Here’s one from Dec 2008:
http://global-warming.accuweather.com/2009/01/despite_recent_trends_giss_sti_1.html
So global warming seems to affect foremost landmasses with a lack of thermometers. Hmm, what could one do to find out more?
Add thermometers? I don’t know if that is a scientific answer, though, me not being a scientist…

stephen richards
July 14, 2010 2:50 am

the mechansim by whch GHG warm the planet is a physical theory. A physical theory that engineers use in the everyday construction of devices that we all enjoy. Its a physical theory which, Monkton, Lindzen, Christy, spencer, all agree with. we bicker over the MAGNITUDE of the effect, but no serious skeptic denies the basics of radiative transfer.
Over the magnitude once the physical is inserted into a system as complex as the climate. It is possible that there is no effect from CO² in the climate but no-one has yet completed a full mathematical model (SteveMc’s engineering study) which proves it one way or the other.
On another note, Mosh et al thanks for this example and I really do appreciate all this effort. I think a clear statement of purpose, etc at the beginning would have negated many of the comments received and I am utterly convinced that the introduction would have mentioned all that George Smith wrote.
I believe you know he is correct in what he says but that was not the purpose of your essay. In the final analysis it is only energy balance / lapse rate that matters and globa temp is for the public/media.
Lastly, I have become aware that GISS have been adjusting historic data along with the latest data and that these adjustments have tended to lower the historic data relative to the later. Am I correct, if so how has this been accounted for in your examples?

Ryan
July 14, 2010 3:33 am

Amusing article that comes to the conclusion that all the simple arithmetic is done correctly. So it seems even climate scientists can add up.
Still, it puzzles me that the theory states that the additional CO2 should make most impact when the insolation is highest but they insist on averaging temperatures over a year. Seems to me that a more sensitive apporach would be to look only at temperature measurements on the Longest Day for both Northern and Southern hemispheres to see if there is any real evidence that CO2 is doing its evil work, rather than watering down the possibility of detection by adding in all the winter temperatures too. This approach would also mean you would retain the natural variation of temperatures due to “weather” which would allow mathemiticians to consider the statistical significance of any anomoly.

carrot eater
July 14, 2010 3:53 am

stephen richards
GISS adjustments do not enter into the above, at all. The source data are taken from a source upstream from GISS.
As for what GISS does, see the effect here.
http://clearclimatecode.org/gistemp-urban-adjustment/

carrot eater
July 14, 2010 3:55 am

DirkH
Can you clarify what methods and data your source used? It’s very difficult to tell.
By parsing language, I kind of think he used New Zealand data only, and may have used the First Difference Method for the calculation, but it’s not obvious to me.

Malaga View
July 14, 2010 4:46 am

George E. Smith says:
July 13, 2010 at 3:19 pm
Why not admit, that ALL of these records; are simply the result of statistical manipulations on the output of a certain set of thermometers; and that any connection between the result of that computation and the actual temperature of the earth is without any scientific foundation; it’s a fiction.

AGREED: It is a fiction.
George E. Smith says:
July 13, 2010 at 3:31 pm
is not correction simply making up false data?

AGREED: That is what scientists seem to do these days.
George E. Smith says:
July 13, 2010 at 5:31 pm
The point is that NOWHERE in this process, can the result be connected to the planet to “Calculate the Global Temperature” It simply calculates the variations of some quite arbitrary set of thermometers from themselves.

AGREED: End of story.
George E. Smith says:
July 13, 2010 at 5:45 pm
The GISStemp process, and the HADcrud process calculate GISStemp and HADcrud respectively; and nothing else. They have no connection to the mean global temperature of the planet; which in turn has no connection to the energy balance of the earth energy budget.

AGREED: Anomalies just seem to be about fear and obscuration.
Luis Dias says:
July 13, 2010 at 6:14 pm
George Smith, that’s quite an astonishing nihillist (and paranoid) vision of reality. I never thought I’d see that kind of thing. Even here.

Welcome to the real world.
rbateman says:
July 13, 2010 at 5:34 pm
If we plotted the yearly mean temp instead of the anomaly, set the bottom of the graph at ZERO, set the top at 2x the mean, then we’d see how this Global Temp scare is making a mountain out of a molehill.

AGREED: A very simple, very sensible and very correct way to look at their data

Dusty Rhodes
July 14, 2010 4:57 am

A very interesting post. I was particularly surprised at the lack of difference in the results from the various methods used thereby eliminating method as a problem area.
Would overlaying the standard deviation of at least the baseline data be helpful in interpreting the graphs?
Slightly O/T but I was looking at the 234 year long Central England Temperature record (link below) and noticed that the global average curve was almost entirely below the CET one. Now given that (just considering the northern hemisphere) there is a great deal more of the earth’s (warmer) surface south of UK than north of it I would have thought, intuitively, that the global curve would therefore have been above CET. Or am I missing something fundamental?
Thoughts anyone?
http://www.decc.gov.uk/assets/decc/statistics/climate_change/1_20100319151831_e_@_surfacetemperaturesummary.pdf

July 14, 2010 5:04 am

Global temperature from 1979 plotted on a normal y-axis.

July 14, 2010 5:18 am

Ron Broberg says at 11:31 pm:
“The current highs are to some degree greater than zero a product of increasing CO2 (as determined by modeling analysis )”
That may or may not be correct, but it should be kept in mind that it is an assumption based on a computer model. It is not real world data, and quoting the IPCC still doesn’t make it anything more than a conjecture.

July 14, 2010 5:27 am

There has to be a large component of guessing in all these reconstructions. If the old thermometer method was adequate, why was there a change to thermocouples and thermistors? As the daily sampling rate went from 1 a day to hundreds a day, did not this cause different smoothing assumptions, removal of spikes, etc? Why should the pre- and post- mercury be able to be spliced; what was the duration and magnitude of the splice? Would one not expect spike removal in recent times to drive down maxima? How can you use old records when the metadata sheets are still being studied and adjustments made as we speak? How can you define a generic term “rural” and know it was always thus through instrumented time? A station in the middle of a vast paddock will respond to the heat of a lamp used to light it for reading before sunrise – is this everywhere quantified and corrected? Even the height of the grass growing around it can cause substantial change.
The error terms, when teated in a high quality manner such as for data that really matter for health or safety, are so large that it is impossible to draw solid conclusions. I’m with George E. Smith – July 13, 2010 at 5:45 pm
I think the comment of Luis Dias – July 13, 2010 at 6:14 pm is harsh and unjustified. The more temperature series I plot, the more I find stations with no change over the last 40 years, within reasonable interpretation. It takes only one such station, in theory, to disprove global warming; but when large numbers of them exist, then the concept of global warming has to explain a negative temperature driver at each one, which global warming theorists have failed to do.

tallbloke
July 14, 2010 5:29 am

Mosh says:
“underlying mechanism. Well, the results are consistent with and confirm the theory of GHG warming, espoused BEFORE this data was collected.”

Which of these two graphs suggests the better correlation Mosh?
http://tallbloke.files.wordpress.com/2010/07/soon-arctic-tsi.jpg
Take your time…