A new must read paper: McKitrick on GHCN and the quality of climate data

This new paper by Dr. Ross McKitrick of the University of Guelph is a comprehensive review of the GHCN surface and sea temperature data set. Unlike many papers (such as the phytoplankton paper in Nature, complete code is made available right from the start, and the data is freely available.

There is a lot here that goes hand in hand with what we have been saying on WUWT and other climate science blogs for months, and this is just a preview of the entire paper.This graph below caught my eye, because it tells one part of the GHCN the story well.

Figure 1-7: GHCN mean latitude of monitoring stations. Data are grouped by latitude band and the bands are weighted by geographical area. Data source: GHCN. See Appendix for calculation details.

1.2.3. Growing bias toward lower latitudes

The decline in sample has not been spatially uniform. GHCN has progressively lost more and more high latitude sites (e.g. towards the poles) in favour of lower-latitude sites. Other things being equal, this implies less and less data are drawn from remote, cold regions and more from inhabited, warmer regions. As shown in Figure 1-7, mean laititude declined as more stations were added during the 20th century.

Here’s another interesting paragraph:

2.4. Conclusion re. dependence on GHCN

All three major gridded global temperature anomaly products rely exclusively or nearly exclusively on the GHCN archive. Several conclusions follow.

  • They are not independent as regards their input data.
  • Only if their data processing methods are fundamentally independent can the three series be considered to have any independence at all. Section 4 will show that the data processing methods do not appear to change the end results by much, given the input data.
  • Problems with GHCN, such as sampling discontinuities and contamination from urbanization and other forms of land use change, will therefore affect CRU, GISS, and NOAA. Decreasing quality of GHCN data over time implies decreasing quality of CRU, GISS and NOAA data products, and increased reliance on estimated adjustments to rectify climate observations.

From the summary: The quality of data over land, namely the raw temperature data in GHCN, depends on the validity of adjustments for known problems due to urbanization and land-use change. The adequacy of these adjustments has been tested in three different ways, with two of the three finding evidence that they do not suffice to remove warming biases.

The overall conclusion of this report is that there are serious quality problems in the surface temperature data sets that call into question whether the global temperature history, especially over land, can be considered both continuous and precise. Users should be aware of these limitations, especially in policy sensitive applications.

Read the entire preview paper here (PDF), it is well worth your time.

h/t to E.M. Smith

0 0 votes
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

164 Comments
Inline Feedbacks
View all comments
John Whitman
August 3, 2010 10:28 am

From Dr. Ross McKitrick’s paper section “1.3 Increasing magnitude of adjustments employed to try and fix the problems of sampling discontinuities”:

“The second, and more obvious feature is the chimney-brush appearance of the graph, indicating a massive increase in the volatility of the adjustments after 1990. The instability in the record dwarfs the size of the century-scale global warming signal, with fluctuations routinely going to ±0.5 degrees C from one year to the next. The southern hemisphere (bottom panel) is particularly noisy.”


——————-
Dr. McKitrick,
I think you have made blog history with the term “the chimney-brush” in describing the ‘GHCN Delta = Adjusted GHCN – Unadjusted GHCN’ graph (your Figure 1-10).
I had not heard the term before. I like it.
John

August 3, 2010 10:32 am

The loss of temperature stations to the global temperature record is well known, even though the stations continue to exist and collect the data. The loss of higher elevation/cooler area stations noted here is even more alarming. Here in Canada we have 35 stations going into the record (if I got the number right) at present, though thousands are being recorded. Has Jones/Hansen/Schmidt, NOAA/HadCrut/GISS put out reasons for this data loss?

John Finn
August 3, 2010 10:38 am

David Ball says:
August 3, 2010 at 7:48 am

John Finn says:
August 3, 2010 at 6:36 am

You haven’t been paying attention in class, have you.
Well – yes I have actually – for several years as it happens and I can’t recall anyone showing that the trend at airport stations is significantly different to the trend at rural stations. Nick Stokes, in this post, suggests that airport and rural trends were similar
Nick Stokes says:
August 3, 2010 at 7:25 am
Zeke looked at this in detail, and found that a reconstruction using airports vs non-airports gave very similar results. I looked less thoroughly, but found the same.

My own observations are almost certainly less thorough than Zeke’s and Nick’s, but one in particular may be of interest. The Armagh Observatory in Northern Ireland is cited as an example of the perfect site for temperature measurments. David Archibald refers to it in his own research. The observatory is in a rural setting which is relatively unchanged in the past 2 centuries. Aldergove airport in Belfast is just a few miles away. Temperatures have been measured at Aldergrove since 1880.
Between 1975 and 2004 the Armagh (rural) trend is actually greater than the Aldergrove (airport) trend. The difference is even more pronounced if you go back to 1880. The Armagh trend is ~1.5 times the Aldergrove trend. Aldergrove data is part of the GISS station database. I’m not sure if this data is already adjusted for UH but if it is, it looks as though UH effect has been over-estimated.

E.M.Smith
Editor
August 3, 2010 10:57 am

Peter Stroud says: An excellent paper that should be read by all IPCC sponsored scientists. One thing puzzles me though. On page 11 we see the percentage GHCN stations located at airports from 1890 – 2009. But according to WikiAnswers the first airport was built in 1909, College Park Maryland.
That is an artifact of how GHCN stores the Metadata. (Explained in the link in the paper). Basically, GHCN has a broken metadata structure. You can only assign a “present status” to a location, not a status-by-year. So if a bit of dirt becomes an airport, it will be marked as an airport for all time. Similarly, an “urban” station will be “urban” even if in 1800 it was a cow field. The airport at Templehof Germany (you know, Berlin Air Lift…) is being converted to a shopping mall. Whenever the metadata are changed, it will suddenly evaporate from ever having been an airport and for all time in GHCN. Just another stupidity in the data…
So you can make a report of AirStation FLAGS, but not of actual airports, by year. This means it’s an approximation (that will be conservative) in that you must choose to “start time” a bit after 1900 as zero airports with the knowledge that the slope of airport increase is actually much stronger than shown in the metadata.

E.M.Smith
Editor
August 3, 2010 11:04 am

Ackos says:
Are similar bias present between high and low altitude? Rural v city?

Yes.
http://chiefio.wordpress.com/2009/11/16/ghcn-south-america-andes-what-andes/
http://chiefio.wordpress.com/2009/11/17/ghcn-the-under-mountain-western-usa/
http://chiefio.wordpress.com/2009/12/01/ncdc-ghcn-africa-by-altitude/
and a whole lot more at:
http://chiefio.wordpress.com/category/ncdc-ghcn-issues/page/3/
and other pages in that category.
BTW, Sinan Unur has great stuff well worth watching. The animations are very informative.

CRS, Dr.P.H.
August 3, 2010 11:10 am

Thanks for posting, Anthony! I just opened the paper for a moment, and this popped out immediately:
——-
The GHCN website does not report which sites are included in the monthly updates, but recent tabulations using GHCN records shows that:
• coverage has fallen everywhere (including the US)
• the sample has become increasingly skewed towards airport sites;
• the sample has migrated from colder latitudes to warmer latitudes;
• the sample has migrated to lower altitudes.
—-
…as you’ve been reporting for many months!
Yeah, I can’t wait to see this pop up on CNBC, CBS, BBC etc. They are too busy reporting about plunging phytoplankton populations due to, what, maybe a 0.5 C increase in temperature? IF there is even a temperature increase!

E.M.Smith
Editor
August 3, 2010 11:20 am

JamesS says:
If it gets warmer because of increased urbanization, well then, it gets warmer. Are they trying to make the numbers look like what they might be if nothing had changed?

I found that a particularly interesting question, as the first step ought to be looking for warming in the base data ( I can’t call what they produce “raw” – it isn’t.) If there is no warming even WITH UHI, then there really is no warming.

I wonder what these graphs would look like if only the raw data were used? When I look at the raw data for site in out-of-the-way places in West Virginia, or other places that haven’t seen any major development over the years, there is no AGW signature visible in those records. What’s being done to the raw data for the sake of “accuracy” is a crime against data.

Also, if you look at long lived stations, there is no ‘warming signal’. It only shows up in the constantly changing kaleidoscope of station change.
Here’s what it looks like in the “base data”. The “unadjusted” GHCN (that has been adjusted during various “Quality Control” steps…) with full artifact inclusion using a very simple “self to self” anomaly where each thermometer is compared only to itself and only inside the same month. A very clean comparison. (graph in link). It’s pretty much dead flat in summer months, but the winters warm from the horridly cold ones of the Little Ice Age ( “year without a summer”..) to a flat recent trend.
http://chiefio.wordpress.com/2010/07/31/agdataw-begins-in-1990/
Then 1990 hits and it all goes very crazy as the instruments are changed in odd ways. The onset of the “new” methods (new “duplicate numbers” for old stations) is in 1986, then the old ones are dropped in 1990 (ish) so you get a ‘feathering’ of the change to blend the hockey blade onto the shaft…
http://chiefio.wordpress.com/2010/08/02/agw-jumping-sharks-since-1986/

E.M.Smith
Editor
August 3, 2010 11:32 am

John Finn says:
There is much criticism of the use of weather stations at airports. Is there actually any evidence that the temperature trends at airport stations are signiicantly greeater than the trend at nearby rural stations. I’m sure there must be some that are, but equally I’ve noticed some that quite definitley aren’t.

The effect varies a bit with wind speed, but one of the problems is that large airports are near large urban centers, so you often can simply have the city UHI blowing over the airport, so it’s both wind speed AND direction (and what’s nearby…) that matters most. But overall, yes, they are hot places. You can do A/B/ compares via Wunderground and see it for yourself some times.
from comments here:
http://chiefio.wordpress.com/2009/08/26/agw-gistemp-measure-jet-age-airport-growth/
“An interesting paper (from 1989) discussing heat island issues at airports: (see pg 18 of 45, re the Akron-Canton airport)
https://kb.osu.edu/dspace/bitstream/1811/23307/1/V089N2_001.pdf
It also uses the phrase “airport heat island.”

AREA FORECAST DISCUSSION
NATIONAL WEATHER SERVICE MEMPHIS TN
305 PM CDT WED JUL 21 2004
.DISCUSSION…
PERSISTENCE SEEMS TO BE THE WEATHER WORD RIGHT NIGHT. UPPER RIDGE
STILL SQUARELY PLACED OVER THE MIDSOUTH WHILE HIGH PRESSURE IS
OVER THE SOUTHEASTERN UNITED STATES. EXPECT DAYTIME CUMULUS TO
DISSIPATE AFTER SUNSET AND CIRRUS FROM THE STORMS IN THE MIDWEST
SHOULD THIN OUT AS THEY WORK INTO THE RIDGE. THUS…EXPECT A
MOSTLY CLEAR NIGHT. WINDS WILL BECOME LIGHT AND TEMPERATURES DROP
INTO THE LOWER 70S…EXCEPT FOR THE MEMPHIS AIRPORT HEAT ISLAND
WHERE IT WILL ONLY GET DOWN TO THE MID 70S.
That implies a couple of degrees of AHI (low 70s moved to mid 70s).

and

The urban heat island effect at Fairbanks, Alaska
MAGEE N. (1) ; CURTIS J. (1) ; WENDLER G. (1) ;
Geophysical Institute, University of Alaska Fairbanks, Fairbanks, Alaska,
Abstract
Using climatic data from Fairbanks and rurally situated Eielson Air Force Base in Interior Alaska, the growth of the Fairbanks heat island was studied for the time period 1949-1997. The climate records were examined to distinguish between a general warming trend and the changes due to an increasing heat island effect. Over the 49-year period, the population of Fairbanks grew by more than 500%, while the population of Eielson remained relatively constant. The mean annual heat island observed at the Fairbanks International Airport grew by 0.4°C, with the winter months experiencing a more significant increase of 1.0°C. Primary focus was directed toward long-term heat island characterization based on season, wind speed, cloud cover, and time of day. In all cases, the minima temperatures were affected more than maxima and periods of calm or low wind speeds, clear winter sky conditions, and nighttime exhibited the largest heat island effects.
Journal Title
Theoretical and applied climatology ISSN 0177-798X
Source
1999, vol. 64, no1-2, pp. 39-47 (11 ref.)

John Whitman
August 3, 2010 11:37 am

Schrodinger’s Cat was the initial condition but apparently it was a female and pregnant when put in the experimental box . . . . now it is Schrodinger’s Cats in the experimental box.
John

John Whitman
August 3, 2010 11:39 am

oops, wrong thread for Schrodinger’s Cat post,
John

August 3, 2010 11:41 am

JamesS says:
August 3, 2010 at 6:27 am
I wish someone would explain why “adjustments for known problems due to urbanization and land-use change” are made to the raw data from a station. If the point of the exercise is to measure the temperature at a given spot, then why would one change the values based on the above parameters? The temperature there is the temperature there — period. If it gets warmer because of increased urbanization, well then, it gets warmer. Are they trying to make the numbers look like what they might be if nothing had changed?

JamesS you might be interested in this series of posts by Dr. William M. Briggs on the whole Homogenization business.
http://wmbriggs.com/blog/?p=1459
There is links to the other 4 parts at the top of Part 1, but here is the kicker line for you:

Scenario 1: fixed spot, urban growth
The most difficult scenario first: our thermometer is located at our precise spot and never moves, nor does it change characteristics (always the same, say, mercury bulb), and it always works (its measurement error is trivial and ignorable). But the spot itself changes because of urban growth. Whereas once the thermometer was in an open field, later a pub opens adjacent to it, and then comes a parking lot, and then a whole city around the pub.
In this case, we would have an unbroken series of temperature measurements that would probably—probably!—show an increase starting at the time the pub construction began. Should we “correct” or “homogenize” that series to account for the possible urban heat island effect?
No.
At least, not if our goal was to determine the real average temperature at our spot. Our thermometer works fine, so the temperatures it measures are the temperatures that are experienced. Our series is the actual, genuine, God-love-you temperature at that spot. There is, therefore, nothing to correct. When you walk outside the pub to relieve yourself, you might be bathed in warmer air because you are in a city than if you were in an open field, but you aren’t in an open field, you are where you are and you must experience the actual temperature of where you live. Do I make myself clear? Good. Memorize this.
Scenario 2: fixed spot, longing for the fields
But what if our goal was to estimate what the temperature would have been if no city existed; that is, if we want to guess the temperature as if our thermometer was still in an open field? Strange goal, but one shared by many. They want to know the influence of humans on the temperature of the long-lost field—while simultaneously ignoring the influence of humans based on the new city. That is, they want to know how humans living anywhere but the spot’s city might have influenced the temperature of the long-lost field.

August 3, 2010 11:57 am

Kilty

Very useful video, except the first hundred years or so is like watching paint dry. (link added by me)

I think it is necessary to watch that paint dry when so as to be able to put in context the absurdity of statements such as:

Since around 1960, for mysterious reasons, trees have stopped responding to temperature increases in the same way they apparently did in previous centuries.

previous centuries! Give me a break. See Idiocrats at work.

August 3, 2010 12:24 pm

I appreciate the comments and feedback. A few people have emailed with a request that I draw conclusions about how all the discontinuities affect the trends. However I am not sure that it is possible to do such a thing. The main task at this point is to give people a better understanding of where these famous graphs come from. The implicit standard I have in the back of my mind is that of building official national statistics. If a sampling frame changes dramatically, the statistical agency is supposed to terminate one series and start another, and advise users that they are not necessarily continuous.
I also hope to emphasize the difference between ex ante and ex post tests of data adjustments. Thus far people have focused on getting a clear list of the adjustments. But this just gives you, in effect, the list of ingredients on the bottle. You still need to test whether the medicine cures the disease. The Muir Russell team got the distinction correct (p. 155):

Comparison of all of these plots demonstrates that adjustments which have been made are largely immaterial to the overall shape of the temperature series and are not on their own the cause of the rising temperature trend in recent years.
We have not addressed the opposite question of whether the adjustments made are adequate to allow for non climatic effects. This is entirely a matter for proper scientific study and debate and outside the scope of this review.

E.M.Smith
Editor
August 3, 2010 12:26 pm

FWIW, I do use all the GHCN stations.
Also, per the folks saying that colder vs warmer stations don’t mater because it’s the trend that matters: That would be true for a pure “self to self” thermometer anomaly process (such as the ones I use) if you did not then have to splice the series of anomalies together.
IMHO, it’s “all about the splice” (once “adjustments” are disposed of…)
Now in many of the codes, like GIStemp, they have two issues. One is “the splice” of patching all these series together (and there are splice artifacts to be found). The other is that they do their “grid/box” anomalies with one box of thermometers in the baseline and a different set in the present (and an even different set out of the baseline in the past historical settings). So you not only get “the splice” but you get the anomaly calculated by comparing your Jaguar to an old Chevy and saying the car is hotter now than it was then.
And yes, if you all do your “reconstructions” using similar techniques, you all find the same errors.
JamesS says:
I’ve been in the software and database development field for 27 years, so I know a little bit about data and analyzing same. Perhaps the problem here is a more basic one than climate scientists will admit: there isn’t enough data to derive a global average temp.

It’s worse than that. Temperature is an intensive variable. Calculating a Global Average Temperature will always be a bogus number as you need other things (like, oh, specific heat, heat of fusion, all that water cycle stuff…) to have meaning. I have more on the math of it in this link, but unless you are a math geek, it will just cause you to glaze. (Even if I think it is foundationally important):
http://chiefio.wordpress.com/2010/07/17/derivative-of-integral-chaos-is-agw/
But yes, there is simply not enough data both spacially and temporally to do the deed, so they must make stuff up. And with predictable consequences.
Vorlath says:
I’m just saying it’s amazing how often the selection, monitoring and reporting of temperature (but NOT the temperature itself) seems to match the warming trend (temperature). After a while, one has to wonder if this is just coincidence or if there is actually correlation and causation.

Another interesting graph is this one:
http://chiefio.files.wordpress.com/2009/12/jetfuel_globaltemp.png
from this posting:
http://chiefio.wordpress.com/2009/12/15/of-jet-exhaust-and-airport-thermometers-feed-the-heat/
(with reference to original source in the article).
Direct correspondence of jet fuel with ‘warming’…
Oh, and on the issue of “by altitude” or “by latitude” station dropout not mattering, this ignores two very important points:
1) It means you take another SPLICE (see above).
2) It means you are taking another Apples vs Oranges anomaly.
and there is a more subtle point:
It means you have one class of station (colder and volatile) in the record in one period of time and replace it with a different class of station (warmer and less volatile) in another. This, btw, it the issue I’d hinted at before but not posted on. If you let me swap a volatile series for a non-volatile AND let me pick when, I can tune that to things like the PDO and give you any “shape” of data warming you want via judicious splicing. If the trend is going against you, move to non-volatile, if it’s in your favor, go volatile. This, btw, is a standard technique in stock trading. That the issue of site volatility with temperature regimes is ignored is a lethal failure of GHCN. (If it is shown to not be ignored, but to be a deliberate move to non-volatile stations near water and at low altitude, then it’s worse…)
So you put volatile stations in during 1950 to 1980 while the PDO is cold and “lock in” very cold excursions due to the higher volatility. Then, when things are in the warm phase of the PDO, move to low volatility stations. It’s now impossible for them to find as much cold as the old stations did during the next cold PDO. They lack the volatility range to get there.
So the key elements of the “magic” are:
1) The Splice.
2) The “Jag vs Chevy” comparison.
3) High volatility in the past cold PDO with low volatility substitution recently.
There is also a ‘process change’ that happens with the change of “Duplicate Number” that happens between 1986 and 1990 which clips cold going excursions rather dramatically (either that, or we’ve never had a cold winter since, and I think 2009 pretty much showed we have… see the “hair graph” in the “jump the shark” link); but I’ve not yet finished finding out exactly what that process change is.
But don’t worry, I’ll get to it. I never stop.
I’ve taken to calling this approach of a bunch of distributed biases “The Distributed Salami Technique”. Similar to the Salami Technique used in financial shenanigans, but with the method spread over many steps and many individually minor “issues”. Sort of a steganographic approach to fudge. Though the question still exists if this is deliberate or if Hansen is just a “Clever Hans”en… 😉

DirkH
August 3, 2010 12:36 pm

Climate science is a game where one side is funded with tens of billion Dollars, creating data, models and publications. The other side is unfunded and it’s their job to find out where the well-funded side has pulled which trick. And we know in advance that after one trick has been exposed they’ll use another one. Maybe they have entire departments with the sole purpose of brainstorming new techniques of deception. And they’ll never be held responsible.

Bernie
August 3, 2010 12:48 pm

In Ross’s paper he makes generous statements about and includes some of the contributions of well known temp mavens – what I termed Serious Amateurs with Strong Data Analysis Skills or more neutrally Independent Researchers. But he refers to them as bloggers. I feel “bloggers” diminishes the substance of their contributions and suggested that he come up with a less pejorative label. However, this is only my opinion and I would be interested in how others feel.

Gail Combs
August 3, 2010 1:07 pm

Ross McKitrick says:
August 3, 2010 at 12:24 pm
I appreciate the comments and feedback. A few people have emailed with a request that I draw conclusions about how all the discontinuities affect the trends. However I am not sure that it is possible to do such a thing. The main task at this point is to give people a better understanding of where these famous graphs come from. The implicit standard I have in the back of my mind is that of building official national statistics. If a sampling frame changes dramatically, the statistical agency is supposed to terminate one series and start another, and advise users that they are not necessarily continuous.
I also hope to emphasize the difference between ex ante and ex post tests of data adjustments.
_________________________________________________________-
A very simple explanation of what you mean by ex ante and ex post tests, would go a long way to making this report completely understandable to the lay people this report needs to reach. Perhaps including a glossary of terms would help.
Again thanks for a very good well written report.

Rational Debate
August 3, 2010 1:25 pm

First, I admit I haven’t read the paper yet.
One question however – when I’ve mentioned the changes in locations to a preponderance of lower latitude locations to AGW believers, one rebuttal comes up at times – that since temps are measured in terms of anomalies, and supposedly the higher latitudes are warming faster than the lower latitudes, removing higher latitude sites would actually result in lower anomalies rather than higher ones… comments please?

Dave Andrews
August 3, 2010 1:50 pm

Mosher,
So you are saying that stations that have ONLY 15 years of full data during the 1961- 90 period are an adequate basis upon which to determine temperature trend? No matter what your, and Zeke’s and others, reconstructions purport to show surely you need more data than this to produce a sound result?

PhilJourdan
August 3, 2010 1:53 pm

A question on Dr. McKitrick’s paper (I have not read all the comments so it may have been addressed). On page 4, at the bottom, he writes (quoting Folland and Parker 1995):

“abrupt transition from the use of uninsulated or partially insulated buckets to the use of engine inlets” in December 1941

However on page 5, first paragraph, he writes:

It is immediately apparent that the Parker-Folland analysis was incorrect to assume that bucket-measurements gave way to engine intake measurements “abruptly” in 1945,

I was curious as to when the change actually occurred. At the start or end of WWII?

August 3, 2010 1:57 pm

Wow! I’ve spent 1/2 hour reading through this work. Amazing! I’m just surmizing that the data used for input to the processing code is readily available and clearly referenced in the work.
THE WAY REAL SCIENCE AN ANALYSIS SHOULD NOW BE DONE. No excuses. NO processed data HIDING behind computer codes NOT available…
No data sets “dissappearing”, “We’ll, we had it when we did the work. Whoops, lost it, so sorry…but you can trust our work, here’s the conclusion(s).” Dang, as the old saying goes, “Adults get PAID to do this?”

August 3, 2010 2:17 pm

Ross McKitrick:
1. In the paper you wrote, “The so-called Optimal Interpolation (OI) method used by GISS employs satellite measures to interpolate SST data for complete global coverage.”
GISS uses the Reynolds OI.v2 SST data, but the dataset originates at NCDC. So you may want to change GISS to NCDC.
2. A note regarding the “1945 Discontinuity” discussed in Thompson et al: The discontinuity also appears in numerous other datasets. Refer to:
http://bobtisdale.blogspot.com/2009/03/large-1945-sst-discontinuity-also.html
And to:
http://bobtisdale.blogspot.com/2009/03/part-2-of-large-sst-discontinuity-also.html
3. Thanks for the reference to my post on page 28.
4. You wrote, “GISS uses another NOAA product, the Reynolds et al. (2008) Optimal Interpolation version 2 (OI.v2) data base. This is based on ICOADS up to 1998. Thereafter, like Hadley, they switch to a subset that are continuously updated. The updated subset is weighted towards buoy data since many shipping records are provided in hard copy. OI.v2 also uses AVHRR satellite retrievals to improve the interpolation for unsampled regions. Unlike the ERSST data set the satellite input is still used in OI.v2.”
Should “Reynolds et al (2008)” be Reynolds et al (2002)? Als s far as I know, the satellite data has always been a part of the NCDC Optimum Interpolation dataset and I believe it is the primary data.
The combined use of satellite and in situ SST data was first discussed in Reynolds (1988):
ftp://ftp.emc.ncep.noaa.gov/cmb/sst/papers/reynolds_1988.pdf
Reynolds and Marsico (1993) mention Optimum Interpolation but don’t describe it:
ftp://ftp.emc.ncep.noaa.gov/cmb/sst/papers/blend_w_ice.pdf
Optimum Interpolation is discussed in detail in Reynolds and Smith (1994):
ftp://ftp.emc.ncep.noaa.gov/cmb/sst/papers/oiv1.pdf
The OI.v2 version is introduced in Reynolds et al (2002):
ftp://ftp.emc.ncep.noaa.gov/cmb/sst/papers/oiv2.pdf
5. During the discussion of ERSST.v3 & .v3b you wrote, “This edition was called ERSST v3. However they noted that it reduced the trend slightly and deemed this effect a cold bias, so the satellite data were removed for version v3b.”
The trend wasn’t the problem. The satellite data changed which year and month had the record high temperatures, which brings nitpicky to an extreme. Refer to their webpage here:
http://www.ncdc.noaa.gov/oa/climate/research/sst/ersst_version.php
There they write, “While this small difference did not strongly impact the long term trend, it was sufficient to change rankings of warmest month in the time series, etc. Therefore, the use of satellite SST data was discontinued.”
The changes in rankings were illustrated in Table 6 of Smith et al (2008):
http://www.ncdc.noaa.gov/oa/climate/research/sst/papers/SEA.temps08.pdf
The irony in that is, in Smith and Reynolds (2004), the note how they perceive the satellite-based data. They write, “Although the NOAA OI analysis contains some noise due to its use of different data types and bias corrections for satellite data, it is dominated by satellite data and gives a good estimate of the truth.”
So if the “truth” changes which year has the second or third highest SST anomaly, they delete the truth. Link to Smith and Reynolds (2004):
http://www.ncdc.noaa.gov/oa/climate/research/sst/papers/ersst-v2.pdf
Thanks for the preview of the paper, Ross, and again, thanks for referring to my work.
Regards

August 3, 2010 2:26 pm

Phil – good catch; my typo. The bucket transition was assumed to occur abruptly at 1941 in the Folland-Parker adjustment analysis. The abrupt SST change in 1945 was a different issue discussed in the Thompson et al. Nature paper.

Kevin Kilty
August 3, 2010 3:11 pm

Carrick says:
August 3, 2010 at 8:34 am
Kevin:
I believe most people here understand what you say, but the point is also that larger portions of the Earth, especially those considered to be most important to the detection of warming, are doing with fewer actual measurements. This does make the results more dependent on errors in measurement and adjustment, wouldn’t you say?
I agree with you, and actually this is my biggest gripe with the temperature reconstructions: Very few of them make any attempt at all to generate realistic error bars. This is experimental data, and the mean (and metrics derived from that) is meaningless without a statement of the uncertainty in the measurement. Of course the opposite is also true: Skeptics are wont to point out enumerable warts in the surface data sets, without ever sitting down and demonstrating whether they would amount to a hill of beans, numerically.

I agree about measurements needing best value plus uncertainty. I don’t know all the details in the weighting given to sparse stations. I can imagine in the worst case that sparse stations might have too much weight because they are allowed to represent too large an area. I might point out that a number of skeptics attempt to do many things with this data, including seeing if it amounts to a hill of beans, but are sometimes stymied.

Few “actual measurements” may increase the error bars somewhat (not so much, as you note I have included error bars in my figure), but it wouldn’t explain a large systematic effect with latitude.

The operative word here is “large”. I agree with you that the shift to lower latitudes ought to produce a more moderate signal, but this I say without having thought about every influence that may occur. There was a time I would have thought that homogenization would not have had a systematic effect too, but after reading how NCDC does homogenization I now have serious doubts.

If you think it does, the onus is on you as the critic to demonstrate how a sparsity of stations could explain this temperature trend.

Whoa there, friend. First, you are implying that changing station distribution is the only “fly in the ointment.” It is not. Second, how come the onus is on me as critic? Maybe I should have tried that line on my dissertation committee?
It’s not my data. I’m not paid to do this work. I’m not one trying to guide policy using it. And I am certainly not one claiming these temperature series represent some sort of gold standard. Using sparse data to extrapolate trends over large regions deserves comment.

Dr A Burns
August 3, 2010 3:21 pm

This is an interesting comment from p36:
‘(IPPC AR4) Summary for Policymakers stated:
“Urban heat island effects are real but local, and have a negligible influence (less than 0.006°C per decade over land and zero over the oceans) on these values.”
The 0.006°C is referenced back to Brohan et al. (2006), where it is merely an assumption about the standard error, not the size of the trend bias itself. ‘
Warmists often quote the UHI effect as being negligible … theeir source is yet another IPCC error !