On ‘denying’ Hockey Sticks, USHCN data, and all that – part 2

In part one of this essay which you can see here, I got quite a lot of feedback on both sides of the climate debate. Some people thought that I was spot on with criticisms while others thought I had sold my soul to the devil of climate change. It is an interesting life when I am accused of being in cahoots with both “big oil” and “big climate” at the same time. That aside, in this part of the essay I am going to focus on areas of agreement and disagreement and propose a solution.

In part one of the essay we focus on the methodology that was used that created a hockey stick style graph illustrating missing data. Due to the missing data causing a faulty spike at the end, Steve McIntyre commented, suggesting that it was more like the Marcott hockey stick than it was like Mann’s:

Steve McIntyre says:

Anthony, it looks to me like Goddard’s artifact is almost exactly equivalent in methodology to Marcott’s artifact spike – this is a much more exact comparison than Mann. Marcott’s artifact also arose from data drop-out.

However, rather than conceding the criticism, Marcott et al have failed to issue a corrigendum and their result has been widely cited.

In retrospect, I believe McIntyre is right in making that comparison. Data dropout is the central issue here and when it occurs it can create all sorts of statistical abnormalities.

Despite some spirited claims in comments in part one about how I’m “ignoring the central issue”, I don’t dispute that data is missing from many stations, I never have.

It is something that has been known about for years and is actually expected in the messy data gathering process of volunteer observers, electronic systems that don’t always report, and equipment and or sensor failures. In fact there is likely no weather network in existence that has perfect data without some being missing. Even the new U.S. Climate Reference Network, designed to be state-of-the-art and as perfect as possible has a small amount of missing data due to failures of uplinks or other electronic issues, seen in red:

CRN_missing_data

Source: http://www.ncdc.noaa.gov/crn/newdaychecklist?yyyymmdd=20140101&tref=LST&format=web&sort_by=slv

What is in dispute is the methodology, and the methodology, as McIntyre observed, created a false “hockey stick” shape much like we saw in the Marcott affair:

marcott-A-1000[1]

After McIntyre corrected the methodology used by Marcott, dealing with faulty and missing data, the result looked like this:

 

alkenone-comparison

McIntyre points out this in comments in part 1:

In Marcott’s case, because he took anomalies at 6000BP and there were only a few modern series, his results were an artifact – a phenomenon that is all too common in Team climate science.

So, clearly, the correction McIntyre applied to Marcott’s data made the result better, i.e. more representative of reality.

That’s the same sort of issue that we saw in Goddard’s plot; data was thinning near the endpoint of the present.

Goddard_screenhunter_236-jun-01-15-54

[ Zeke has more on that here: http://rankexploits.com/musings/2014/how-not-to-calculate-temperatures-part-3/ ]

While I would like nothing better than to be able to use raw surface temperature data in its unadulterated “pure” form to derive a national temperature and to chart the climate history of the United States, (and the world) the fact is that because the national USHCN/co-op network and GHCN is in such bad shape and has become largely heterogeneous that is no longer possible with the raw data set as a whole.

These surface networks have had so many changes over time that the number of stations that have been moved, had their time of observation changed, had equipment changes, maintenance issues,or have been encroached upon by micro site biases and/or UHI using the raw data for all stations on a national scale or even a global scale gives you a result that is no longer representative of the actual measurements, there is simply too much polluted data.

A good example of polluted data can be found in Las Vegas Nevada USHCN station:

LasVegas_average_temps

Here, growth of the city and the population has resulted in a clear and undeniable UHI signal at night gaining 10°F since measurements began. It is studied and acknowledged by the “sustainability” department of the city of Las Vegas, as seen in this document. Dr. Roy Spencer in his blog post called it “the poster child for UHI” and wonders why NOAA’s adjustments haven’t removed this problem. It is a valid and compelling question. But at the same time, if we were to use the raw data from Las Vegas we would know it would have been polluted by the UHI signal, so is it representative in a national or global climate presentation?

LasVegas_lows

The same trend is not visible in the daytime Tmax temperature, in fact it appears there has been a slight downward trend since the late 1930′s and early 1940′s:

LasVegas_highs

Source for data: NOAA/NWS Las Vegas, from

http://www.wrh.noaa.gov/vef/climate/LasVegasClimateBook/index.php

The question then becomes: Would it be okay to use this raw temperature data from Las Vegas without any adjustments to correct for the obvious pollution by UHI?

From my perspective the thermometer at Las Vegas has done its job faithfully. It has recorded what actually occurred as the city has grown. It has no inherent bias, the change in surroundings have biased it. The issue however is when you start using stations like this to search for the posited climate signal from global warming. Since the nighttime temperature increase at Las Vegas is almost an order of magnitude larger than the signal posited to exist from carbon dioxide forcing, that AGW signal would clearly be swamped by the UHI signal. How would you find it? If I were searching for a climate signal and was doing it by examining stations rather than throwing out blind automated adjustments I would most certainly remove Las Vegas from the mix as its raw data is unreliable because it has been badly and likely irreparably polluted by UHI.

Now before you get upset and claim that I don’t want to use raw data or as some call it “untampered” or unadjusted data, let me say nothing could be further from the truth. The raw data represents the actual measurements; anything else that has been adjusted is not fully representative of the measurement reality no matter how well-intentioned, accurate, or detailed those adjustments are.

But, at the same time, how do you separate all the other biases that have not been dealt with (like Las Vegas) so you don’t end up creating national temperature averages with imperfect raw data?

That my friends, is the $64,000 question.

To answer that question, we have a demonstration. Over at the blackboard blog, Zeke has plotted something that I believe demonstrates the problem.

Zeke writes:

There is a very simple way to show that Goddard’s approach can produce bogus outcomes. Lets apply it to the entire world’s land area, instead of just the U.S. using GHCN monthly:

Averaged Absolutes

Egads! It appears that the world’s land has warmed 2C over the past century! Its worse than we thought!

Or we could use spatial weighting and anomalies:

 

Gridded Anomalies

Now, I wonder which of these is correct? Goddard keeps insisting that its the first, and evil anomalies just serve to manipulate the data to show warming. But so it goes.

Zeke wonders which is “correct”. Is it Goddard’s method of plotting all the “pure” raw data, or is it Zeke’s method of using gridded anomalies?

My answer is: neither of them are absolutely correct.

Why, you ask?

It is because both contain stations like Las Vegas that have been compromised by changes in their environment, that station itself, the sensors, the maintenance, time of observation changes, data loss, etc. In both cases we are plotting data which is a huge mishmash of station biases that have not been dealt with.

NOAA tries to deal with these issues, but their effort falls short. Part of the reason it falls short is that they are trying to keep every bit of data and adjust it in an attempt to make it useful, and to me that is misguided, as some data is just beyond salvage.

In most cases, the cure from NOAA is worse than the disease, which is why we see things like the past being cooled.

Here is another plot from Zeke just for the USHCN, which shows Goddard’s method “Averaged Absolutes” and the NOAA method of “Gridded Anomalies”:

Goddard and NCDC methods 1895-2013

[note: the Excel code I posted was incorrect for this graph, and was for another graph Zeke produced, so it was removed, apologies – Anthony]

Many people claim that the “Gridded Anomalies” method cools the past, and increases the trend, and in this case they’d be right. There is no denying that.

At the same time, there is no denying that the entire CONUS USHCN raw data set contains all sorts of imperfections, biases, UHI, data dropouts and a whole host of problems that remain uncorrected. It is a Catch-22; on one hand the raw data has issues, on the other, at the bare minimum some sort of infilling and gridding is needed to produce a representative signal for the CONUS, but in producing that, new biases and uncertainty is introduced.

There is no magic bullet that always hits the bullseye.

I’ve known and studied this for years, it isn’t a new revelation. The key point here is that both Goddard and Zeke (and by extension BEST and NOAA) are trying to use the ENTIRE USHCN dataset, warts and all, to derive a national average temperature. Neither method produces a totally accurate representation of national temperature average. Keep that thought.

While both methods have flaws, the issue that Goddard raised has one good point, and an important one; the rate of data dropout in USHCN is increasing.

When data gets lost, they infill with other nearby data, and that’s an acceptable procedure, up to a point. The question is, have we reached a point of no confidence in the data because too much has been lost?

John Goetz asked the same question as Goddard in 2008 at Climate Audit:

How much Estimation is too much Estimation?

It is still an open question, and without a good answer yet.

But at the same time we are seeing more and more data loss, Goddard is claiming “fabrication” of lost temperature data in the final product and at the same advocating using the raw surface temperature data for a national average. From my perspective, you can’t argue for both. If the raw data is becoming less reliable due to data loss, how can we use it by itself to reliably produce a national temperature average?

Clearly with the mess the USHCN and GHCN are in, raw data won’t accurately produce a representative result of the true climate change signal of the nation because the raw data is so horribly polluted with so many other biases. There are easily hundreds of stations in the USHCN that have been compromised like Las Vegas has been, making the raw data, as a whole, mostly useless.

So in summary:

Goddard is right to point out that there is increasing data loss in USHCN and it is being increasingly infilled with data from surrounding stations. While this is not a new finding, it is important to keep tabs on. He’s brought it to the forefront again, and for that I thank him.

Goddard is wrong to say we can use all the raw data to reliably produce a national average temperature because the same data is increasingly lossy and is also full of other biases that are not dealt with. [ added: His method allows for biases to enter that are mostly about station composition, and less about infilling see this post from Zeke]

As a side note, claiming “fabrication” in a nefarious way doesn’t help, and generally turns people off to open debate on the issue because the process of infilling missing data wasn’t designed at the beginning to be have any nefarious motive; it was designed to make the monthly data usable when small data dropouts are seen, like we discussed in part 1 and showed the B-91 form with missing data from volunteer data. By claiming “fabrication”, all it does is put up walls, and frankly if we are going to enact any change to how things get done in climate data, new walls won’t help us.

Biases are common in the U.S. surface temperature network

This is why NOAA/NCDC spends so much time applying infills and adjustments; the surface temperature record is a heterogeneous mess. But in my view, this process of trying to save messed up data is misguided, counter-productive, and causes heated arguments (like the one we are experiencing now) over the validity of such infills and adjustments, especially when many of them seem to operate counter-intuitively.

As seen in the map below, there are thousands of temperature stations in the US co-op and USHCN network in the USA, by our surface stations survey, at least 80% of the USHCN is compromised by micro-site issues in some way, and by extension, that large sample size of the USHCN subset of the co-op network we did should translate to the larger network.

USHCN_COOP_Map

When data drops out of USHCN stations, data from nearby neighbor stations is infilled to make up the missing data, but when 80% or more of your network is compromised by micro-site issues, chances are all you are doing is infilling missing data with compromised data. I explained this problem years ago using a water bowl analogy, showing how the true temperature signal gets “muddy” when data from surrounding stations is used to infill missing data:

bowls-USmap

The real problem is the increasing amount of data dropout in USHCN (and in Co-op and GHCN) may be reaching a point where it is adding a majority of biased signal from nearby problematic stations. Imagine a well sited long period station near Las Vegas out in a rural area that has its missing data infilled using Las Vegas data, you know it will be warmer when that happens.

So, what is the solution?

How do we get an accurate surface temperature for the United States (and the world) when the raw data is full of uncorrected biases and the adjusted data does little more than smear those station biases around when infilling occurs? Some of our friends say a barrage of  statistical fixes are all that is needed, but there is also another, simpler, way.

Dr. Eric Steig, at “Real Climate”, in a response to a comment about Zeke Hausfather’s 2013 paper on UHI shows us a way.

Real Climate comment from Eric Steig (response at bottom)

We did something similar (but even simpler) when it was being insinuated that the temperature trends were suspect, back when all those UEA emails were stolen. One only needs about 30 records, globally spaced, to get the global temperature history. This is because there is a spatial scale (roughly a Rossby radius) over which temperatures are going to be highly correlated for fundamental reasons of atmospheric dynamics.

For those who don’t know what the Rossby radius is, see this definition.

Steig claims 30 station records are all that are needed globally. In a comment some years ago (now probably lost in the vastness of the Internet) we heard Dr. Gavin Schmidt said something similar, saying that about “50 stations” would be all that is needed.

[UPDATE: Commenter Johan finds what may be the quote:

I did find this Gavin Schmidt quote:

“Global weather services gather far more data than we need. To get the structure of the monthly or yearly anomalies over the United States, for example, you’d just need a handful of stations, but there are actually some 1,100 of them. You could throw out 50 percent of the station data or more, and you’d get basically the same answers”

http://earthobservatory.nasa.gov/Features/Interviews/schmidt_20100122.php ]

So if that is the case, and one of the most prominent climate researchers on the planet (and his associate) says we need only somewhere between 30-50 stations globally…why is NOAA spending all this time trying to salvage bad data from hundreds if not thousands of stations in the USHCN, and also in the GHCN?

It is a question nobody at NOAA has ever really been able to answer for me. While it is certainly important to keep these records from all these stations for local climate purposes, but why try to keep them in the national and global dataset when Real Climate Scientists say that just a few dozen good stations will do just fine?

There is precedence for this, the U.S. Climate Reference Network, which has just a fraction of the stations in USHCN and the co-op network:

crn_map

NOAA/NCDC is able to derive a national temperature average from these few stations just fine, and without the need for any adjustments whatsoever. In fact they are already publishing it:

USCRN_avg_temp_Jan2004-April2014

If it were me, I’d throw out most of the the USHCN and co-op stations with problematic records rather than try to salvage them with statistical fixes, and instead, try to locate the best stations with long records, no moves, and minimal site biases and use those as the basis for tracking the climate signal. By doing so not only do we eliminate a whole bunch of make work with questionable/uncertain results, and we end all the complaints data falsification and quibbling over whose method really does find the “holy grail of the climate signal” in the US surface temperature record.

Now you know what Evan Jones and I have been painstakingly doing for the last two years since our preliminary siting paper was published here at WUWT and we took heavy criticism for it. We’ve embraced those criticisms and made the paper even better. We learned back then that adjustments account for about half of the surface temperature trend:

We are in the process of bringing our newest findings to publication. Some people might complain we have taken too long. I say we have one chance to get it right, so we’ve been taking extra care to effectively deal with all criticisms from then, and criticisms we have from within our own team. Of course if I had funding like some people get, we could hire people to help move it along faster instead of relying on free time where we can get it.

The way forward:

It is within our grasp to locate and collate stations in the USA and in the world that have as long of an uninterrupted record and freedom from bias as possible and to make that a new climate data subset. I’d propose calling it the the Un-Biased Global Historical Climate Network or UBGHCN. That may or may not be a good name, but you get the idea.

We’ve found at least this many good stations in the USA that meet the criteria of being reliable and without any need for major adjustments of any kind, including the time-of-observation change (TOB), but some do require the cooling bias correction for MMTS conversion, but that is well known and a static value that doesn’t change with time. Chances are, a similar set of 50 stations could be located in the world. The challenge is metadata, some of which is non-existent publicly, but with crowd sourcing such a project might be do-able, and then we could fulfill Gavin Schmidt and Eric Steig’s vision of a much simpler set of climate stations.

Wouldn’t it be great to have a simpler and known reliable set of stations rather than this mishmash which goes through the statistical blender every month? NOAA could take the lead on this, chances are they won’t. I believe it is possible to do independent of them, and it is a place where climate skeptics can make a powerful contribution which would be far more productive than the arguments over adjustments and data dropout.

 

The climate data they don't want you to find — free, to your inbox.
Join readers who get 5–8 new articles daily — no algorithms, no shadow bans.
5 1 vote
Article Rating
274 Comments
Inline Feedbacks
View all comments
Nick Stokes
June 27, 2014 2:13 pm

John Goetz says: June 27, 2014 at 1:26 pm
“In addition to infilling data for stations that no longer report data, GISS estimates data missing from a station’s record using an averaging algorithm.”

Do you mean USHCN? That’s what we’ve been talking about. AFAIK, GISS doesn’t infill non-reporting stations, nor estimate missing data within station time series. They use a rather complicated gridding method in which such measures would not be needed.

Nick Stokes
June 27, 2014 2:28 pm

richardscourtney says: June 27, 2014 at 6:36 am
“Each team that produces values of global average surface temperature (GASTA) uses a different definition; e.g. the weightings they apply to land and ocean differ and they compute gridding differently.”

That’s not a different definition. They are calculation a spatial integral of temperature anomaly, and using different numerical approximations. There are a huge number of different numerical ways to approximate integrals, and most work well. Using a different method doesn’t mean you are redefining the thing you are calculating.

Evan Jones
Editor
June 27, 2014 2:44 pm

You don’t say what period. It obviously depends on that.
1979 – 2008. A period of unequivocal warming. After all, for bad siting to affect trend, there must be a real, genuine trend to exaggerate in the first place.
Overall, warming from 1950 to date is ~+0.107C/decade (Haddy4). My bottom-line “best guess” is that the “true signal” is closer to 0.07C per decade, exaggerated by poor microsite. That translates to raw CO2 warming with perhaps a bit of negative feedback in play.

richardscourtney
June 27, 2014 2:47 pm

Nick Stokes:
Your post at June 27, 2014 at 2:28 pm says

richardscourtney says: June 27, 2014 at 6:36 am

“Each team that produces values of global average surface temperature (GASTA) uses a different definition; e.g. the weightings they apply to land and ocean differ and they compute gridding differently.”

That’s not a different definition. They are calculation a spatial integral of temperature anomaly, and using different numerical approximations. There are a huge number of different numerical ways to approximate integrals, and most work well. Using a different method doesn’t mean you are redefining the thing you are calculating.

Nice try but no coconut.
Clearly, the different methods provide different results. Indeed, the methods change from month to month and that changes the results.
I repeat my question and I add another.
If the different methods are not analysing different definitions then why do values of global average surface temperature (GASTA) from decades ago alter when the method is changed from month to month: which is the right determination any of the ones before a change or any of those after it? In other words, if the different – and frequently changed – methods are are not assessing different definitions of GASTA then why do the methods not provide the same results?
And my additional question is
What is the definition of global average surface temperature (GASTA).
Richard

Evan Jones
Editor
June 27, 2014 2:52 pm

TOBS, when it is needed, is a big adjustment.
Yes, surely. We find it to be ~0.09C (trend, sic) from 1979 – 2008 (Comparing USHCN Class 1\2 TOBS-biased stations with Class 1\2 non-TOBS-biased). It obviously depends on when the TOBS change occurs during the series: If it’s in the middle of the series (in this case, 1979 – 2008), it has a very large effect, and if it occurs at the end or beginning of the series it has little effect. But is is a big step change, no matter how you slice it. Even averaged out over 30 years.
And it is the uncertainty therein that causes us to drop all stations with TOBS bias (for the current study): Sometimes the best way to deal with a problem is simply to bypass it. (We confirm our accuracy by comparing “TOBS-adjusted data” for the non-TOBS biased stations we include, and the raw+MMTS figures match the TOBS figures very well. So we pruned correctly.)
That is why we filter out all stations with moves and TOBS bias and apply only MMTS adjustment (regrettable but unavoidable). That way we get a near-pure assessment of poorly sited stations vis a vis well sited stations. And the difference is staggering.

Svend Ferdinandsen
June 27, 2014 3:29 pm

The whole problem is that they pretend to make a (artificial) measurement of the temperature, as it would be without mans influence on land use. The meaning is that then CO2 would be the only one to change it.
That is anyway one of the reasons it has not so much to do with temperature.
The result is, we have not a temperature that reflects the reality.
I mean, UHI, land use, aircondtion and what ever is a natural part of the temperature. If UHI raises the temperature, then it is real and should be reported as it is.
If they want a rural average, then make such one out of real rural measurements, but keep the two temperatures apart.
But again, what is rural? It can be influenced by the crop and its harvest, the forrests and pinebeetles and some other changes in the land. No matter what is done, it is not possible to make an average that is not somehow influenced by our use of the land.

Nick Stokes
June 27, 2014 3:29 pm

richardscourtney says: June 27, 2014 at 2:47 pm
“And my additional question is
What is the definition of global average surface temperature (GASTA).”

Here’s my definition:
1. Surface temperature is as measured by weather stations on land, SST over ocean.
2. Anomaly is the local difference between ST and the local average over a prescribed period (eg 1961-90).
3. GASTA is the mean calculated by spatial integral of a time average (say one month) of anomaly.

June 27, 2014 3:33 pm

Calibration studies show that unaspirated land-station temperature sensors produce a systematic warm-bias. That warm bias-error is variable in time and space, non-random, occurs in summer and winter measurements, and is not a constant offset.
Any USHCN – USCRN anomaly comparison, as Zeke H. referenced, in which both temperature data sets were differenced against a common normal, ought to show that bias. However, in the event, it doesn’t.
It is hard to believe that the systematic USHCN warm bias-error would just automatically average away, so that the recent USHCN surface anomaly trend should be within (+/-)0.1 C of the USCRN trend.

Evan Jones
Editor
June 27, 2014 4:11 pm

But there is no real trend from when CRN started . So there would be no trend difference between USHCN and CRN over that period. For bad siting to affect a trend, there must first be a trend to affect.
For the 1979 – 2008 period, there was significant warming, greatly exaggerated by poor microsite.

June 27, 2014 4:16 pm

This is funny. NOAA has this new nClimDiv using a 5k grid.
NOAA TMAX May 1998 to 2014 Trend = .-0.6F/decade
My simple USHCN 5×5 TMAX May 1998 to 2014 Trend = -0.62F
Mine vs NOAA so far this year.
1 Gridded 5×5 -1.02 -.93
2 Gridded 5×5 -2.54 -2.66
3 Gridded 5×5 0.67 1.01 ** Strange
4 Gridded 5×5 -0.11 -0.10
5 Gridded 5×5 -0.62 -0.6
Scary. Because all you have to do to change the trend is change the gridding.
Here is comparing my 5×5 to 1×1
1 Gridded 1×1 -1.17″
1 Gridded 5×5 -1.02″
2 Gridded 1×1 -2.96″
2 Gridded 5×5 -2.54″
3 Gridded 1×1 0.86″
3 Gridded 5×5 0.67″
4 Gridded 1×1 -0.2″
4 Gridded 5×5 -0.11″
5 Gridded 1×1 -0.51″
5 Gridded 5×5 -0.62″

Alexej Buergin
June 27, 2014 4:45 pm

Carrick says: June 27, 2014 at 1:39 pm
Well good. I’ll just say that people often say “absolute temperature” in physics, and it is most common to interpret this as “absolute thermodynamic (temperature) scale”, but the metrological meaning is valid too.
Supposing that you intended to say meteorological my answer would be: I learned about WX at an US university, we used the book by Ahrens, and it says:”Most scientists use a temperature scale called the absolute or Kelvin scale…” Seems unambiguous to me. (But that was 20 years ago and it was the fourth edition, now the ninth is current.)
(By the way: Do you personally know somebody who uses Rankine, or can you find some recent writing which uses Rankine? In non-US countries C and K are as normal as the fact that football is played with the feet.)

Latitude
June 27, 2014 4:56 pm

evanmjones says:
June 27, 2014 at 4:11 pm
For the 1979 – 2008 period, there was significant warming, greatly exaggerated by poor microsite.
====
Wildman, what do you or Anthony consider “significant”?

June 27, 2014 5:33 pm

evanmjones, there is a trend. It’s just approximately zero slope. The systematic and variable warm bias of unaspirated (USHCN) sensors is demonstrated and beyond dispute. The USHCN bias ought to show up in the anomalies. if they have a common normal, the USHCN trend ought to be variably warmer than, i.e., variably and positively offset from, the USCRN trend. But it’s not.
The warm bias is not from bad siting, by the way. It occurs even in perfectly sited sensors that are in excellent repair. The bias is due to thermal loading of the shield. It’s intrinsic to the construction of the original USHCN sensors, including the MMTS sensors, and is endemic in the USHCN record.

Evan Jones
Editor
June 27, 2014 6:13 pm

Wildman, what do you or Anthony consider “significant”?
We find ~0.19C/decade over the 30-year period from 1979 – 2008 (for our precious Class 1\2s). Or at about two degrees per century. (But this is during a positive PDO. Negative PDO years show no trend at all, or slight cooling. So it averages maybe half that.)
That would be statistically significant for the 1979 – 2008 period. This is adjusted by NOAA to 0.324/decade — for the Class 1/2 stations.
The ~0.14C/decade trend difference between the well and poorly sited stations is significant as well: Our chance of obtaining the same results by random chance is calculated by J-NG, using Monte Carlo methodic, at 0.00000. (Yes, that many zeros.)
The reason we beat 95% significance so easily is the large number of stations we use (~400), even after dropping 2 out of 3 (of the full 1221) for TOBS bias, station moves, etc. (And, yes, if a CRS and MMTS are at the same site and have different ratings, we consider this to be a station move.)

Evan Jones
Editor
June 27, 2014 6:23 pm

The warm bias is not from bad siting, by the way. It occurs even in perfectly sited sensors that are in excellent repair.
And yet the well sited stations, on average, have only a little over half the trend as the badly sited stations.
The bias is due to thermal loading of the shield. It’s intrinsic to the construction of the original USHCN sensors, including the MMTS sensors, and is endemic in the USHCN record.
MMTS was adjudged to be far more accurate than CRS. And, yes, well sited CRS shows considerably higher trends than well sited MMTS. So that is part of it. We are compelled to adjust MMTS trends (upward) as a result.
There may be an inherent overall equipment bias (including both CRS and MMTS), but microsite appears to be king. Bad microsite spuriously increases trends by over two thirds.

Editor
June 27, 2014 6:52 pm

Nick Stokes:
Thanks for clarifying part of my point. USHCN does the infilling, and GISS uses the gridding in it’s average global (and regional) temperature calculations.
However, GISS does estimate data. If they pick up a station’s data from USHCN or GHCN, and a month or an entire season are missing here and there, they will create an estimate for that missing month or season. They use an averaging algorithm to do this, and the algorithm looks at all “relevant” data in the station’s record from the time the station went online to the present.
This is big reason why the station records change.

Latitude
June 27, 2014 7:02 pm

Thanks Evan..[got] it……this one begs a question
“MMTS was adjudged to be far more accurate than CRS. And, yes, well sited CRS shows considerably higher trends than well sited MMTS. So that is part of it. We are compelled to adjust MMTS trends (upward) as a result. ”
If MMTS is far more accurate…..and CRS shows a higher trend….why are you adjusting MMTS up?
Shouldn’t you be adjusting CRS down?

Latitude
June 27, 2014 7:03 pm

put a “t” in there……got it……not go it! LOL

Mark Albright
June 27, 2014 11:08 pm

Since we are looking for high quality single station records I nominate Lincoln NE as a candidate with complete data from 1887-2014 near the geographical center of the US. The Lincoln NE time series shows the 1930s to be the warmest decade:
http://www.atmos.washington.edu/marka/lincoln.1887_2013.png
Monthly data source for Lincoln NE:
http://snr.unl.edu/lincolnweather/data/monthly-avg-temperatures.asp

Evan Jones
Editor
June 28, 2014 12:16 am

If MMTS is far more accurate…..and CRS shows a higher trend….why are you adjusting MMTS up?
Shouldn’t you be adjusting CRS down?

That’s what I asked myself at the time. One could. But it makes no difference to trend (sic) if CRS is adjusted down or MMTS is adjusted up. Here’s why:
We are talking trend, here, not absolute measurements. In 1979, all of the stations were CRS. MMTS conversion did not even begin until 1983. So we start with an absolute reading that is too high, and then introduce a downward step change when the CRS is replaced with the MMTS. That produces a spurious negative effect on trend. So the trend must be adjusted upwards to mitigate the downward step-change. If you adjusted the CRS downward, you would also mitigate the change. Just so long as you eliminate that spurious upward data spike at the beginning — either by adjusting what comes after (i.e., MMTS) up or by adjusting what comes before (i.e., CRS) down.
Bear in mind that for purposes of the study, we don’t give a rat’s patoot what the actual readings are. We only care about what the trend of the actual readings are. Therefore it matters not which half of the data is pushed up or down, just so that step change is removed.
If this is not clear, ask again and I will try to do better.
(Even after this adjustment, MMTS stations have lower trends. CRS stations warm too fast and cool too fast. We definitely note this and point it out. But that is a separate issue and is dealt with separately.)

richardscourtney
June 28, 2014 1:49 am

Nick Stokes:
Thankyou for your reply to my post at June 27, 2014 at 2:47 pm which you provide at June 27, 2014 at 3:29 pm.
It (again) ignores my question which was

If the different methods are not analysing different definitions then why do values of global average surface temperature (GASTA) from decades ago alter when the method is changed from month to month: which is the right determination any of the ones before a change or any of those after it? In other words, if the different – and frequently changed – methods are are not assessing different definitions of GASTA then why do the methods not provide the same results?

However, it purports to provide an answer to my other question which was

What is the definition of global average surface temperature (GASTA)?

You say

Here’s my definition:
1. Surface temperature is as measured by weather stations on land, SST over ocean.
2. Anomaly is the local difference between ST and the local average over a prescribed period (eg 1961-90).
3. GASTA is the mean calculated by spatial integral of a time average (say one month) of anomaly.

No, Nick. That is NOT a definition: it is an evasion.
What do you mean by “average”; mean, median, mode, something else, weighted, etc.?

Nick, a definition of a parameter specifies what the parameter is. It does not provide a range of possible things the parameter could be. As I said in my first post to you at June 27, 2014 at 3:40 am which is here.

Every determination of GASTA is determined by a faulty method, and can give you anything.

Your so-called “definition” can give you anything by changing the used type of “average” at any time, and IT DOES so historical determination of GASTA change from month to month.
Richard

richardscourtney
June 28, 2014 1:57 am

Nick Stokes:
I see there is a possible claim of ambiguity in my post at June 28, 2014 at 1:49 am so I write to forestall a possible misunderstanding.
When I wrote

No, Nick. That is NOT a definition: it is an evasion.
What do you mean by “average”; mean, median, mode, something else, weighted, etc.?

I was referring to your use of the word “average” in this statement

3. GASTA is the mean calculated by spatial integral of a time average (say one month) of anomaly.

Richard

Nick Stokes
June 28, 2014 3:05 am

richardscourtney says: June 28, 2014 at 1:57 am
“I was referring to your use of the word “average” in this statement
3. GASTA is the mean calculated by spatial integral of a time average (say one month) of anomaly.”

Average is simple time average – add the days and divide by 31 etc. Same for the anomaly base.

richardscourtney
June 28, 2014 3:44 am

Nick Stokes:
Sincere thanks for your clarification at June 28, 2014 at 3:05 am.
That, of course, returns us to my original question which you have still not addressed; i.e.

If the different methods are not analysing different definitions then why do values of global average surface temperature (GASTA) from decades ago alter when the method is changed from month to month: which is the right determination any of the ones before a change or any of those after it? In other words, if the different – and frequently changed – methods are are not assessing different definitions of GASTA then why do the methods not provide the same results?

You see, Nick, the original measurements don’t change but the processing does. There can only be one valid method to process the data if there is only one definition of GASTA (i.e. global AVERAGE surface temperature anomaly).
There is an ‘elephant’ filling the room. The ‘elephant’ is that there is no agreed definition of GASTA so anybody can compute a value of GASTA as they alone desire and THEY DO; indeed, they each frequently change how they calculate GASTA and, thus, change their determined values of it.
I am shouting, “Look at the elephant!”
You are replying to my shouts by looking out the window of the room and saying you cannot see a nit to pick.
Richard

A C Osborn
June 28, 2014 4:39 am

evanmjones says: June 28, 2014 at 12:16 am
CRS stations warm too fast and cool too fast.
How do you know?
Why isn’t it the MMTS that is wrong?