Much Ado About Very Little
Guest post by Zeke Hausfather and Steve Mosher
E.M. Smith has claimed (see full post here: Summary Report on v1 vs v3 GHCN ) to find numerous differences between GHCN version 1 and version 3, differences that, in his words, constitute “a degree of shift of the input data of roughly the same order of scale as the reputed Global Warming”. His analysis is flawed, however, as the raw data in GHCN v1 and v3 are nearly identical, and trends in the globally gridded raw data for both are effectively the same as those found in the published NCDC and GISTemp land records.
Figure 1: Comparison of station-months of data over time between GHCN v1 and GHCN v3.
First, a little background on the Global Historical Climatology Network (GHCN). GHCN was created in the late 1980s after a large effort by the World Meteorological Organization (WMO) to collect all available temperature data from member countries. Many of these were in the form of logbooks or other non-digital records (this being the 1980s), and many man-hours were required to process them into a digital form.
Meanwhile, the WMO set up a process to automate the submission of data going forward, setting up a network of around 1,200 geographically distributed stations that would provide monthly updates via CLIMAT reports. Periodically NCDC undertakes efforts to collect more historical monthly data not submitted via CLIMAT reports, and more recently has set up a daily product with automated updates from tens of thousands of stations (GHCN-Daily). This structure of GHCN as a periodically updated retroactive compilation with a subset of automatically reporting stations has in the past led to some confusion over “station die-offs”.
GHCN has gone through three major iterations. V1 was released in 1992 and included around 6,000 stations with only mean temperatures available and no adjustments or homogenization. Version 2 was released in 1997 and added in a number of new stations, minimum and maximum temperatures, and manually homogenized data. V3 was released last year and added many new stations (both in the distant past and post-1992, where Version 2 showed a sharp drop-off in available records), and switched the homogenization process to the Menne and Williams Pairwise Homogenization Algorithm (PHA) previously used in USHCN. Figure 1, above, shows the number of stations records available for each month in GHCN v1 and v3.
We can perform a number of tests to see if GHCN v1 and 3 differ. The simplest one is to compare the observations in both data files for the same stations. This is somewhat complicated by the fact that station identity numbers have changed since v1 and v3, and we have been unable to locate translation between the two. We can, however, match stations between the two sets using their latitude and longitude coordinates. This gives us 1,267,763 station-months of data whose stations match between the two sets with a precision of two decimal places.
When we calculate the difference between the two sets and plot the distribution, we get Figure 2, below:
Figure 2: Difference between GHCN v1 and GHCN v3 records matched by station lat/lon.
The vast majority of observations are identical between GHCN v1 and v3. If we exclude identical observations and just look at the distribution of non-zero differences, we get Figure 3:
Figure 3: Difference between GHCN v1 and GHCN v3 records matched by station lat/lon, excluding cases of zero difference.
This shows that while the raw data in GHCN v1 and v3 is not identical (at least via this method of station matching), there is little bias in the mean. Differences between the two might be explained by the resolution of duplicate measurements in the same location (called imods in GHCN version 2), by updates to the data from various national MET offices, or by refinements in station lat/lon over time.
Another way to test if GHCN v1 and GHCN v3 differ is to convert the data of each into anomalies (with baseline years of 1960-1989 chosen to maximize overlap in the common anomaly period), assign each to a 5 by 5 lat/lon grid cell, average anomalies in each grid cell, and create a land-area weighted global temperature estimate. This is similar to the method that NCDC uses in their reconstruction.
Figure 4: Comparison of GHCN v1 and GHCN v3 spatially gridded anomalies. Note that GHCN v1 ends in 1990 because that is the last year of available data.
When we do this for both GHCN v1 and GHCN v3 raw data, we get the figure above. While we would expect some differences simply because GHCN v3 includes a number of stations not included in GHCN v1, the similarities are pretty remarkable. Over the century scale the trends in the two are nearly identical. This differs significantly from the picture painted by E.M. Smith; indeed, instead of the shift in input data being equivalent to 50% of the trend, as he suggests, we see that differences amount to a mere 1.5% difference in trend.
Now, astute skeptics might agree with me that the raw data files are, if not identical, overwhelmingly similar but point out that there is one difference I did not address: GHCN v1 had only raw data with no adjustments, while GHCN v3 has both adjusted and raw versions. Perhaps the warming the E.M. Smith attributed to changes in input data might in fact be due to changes in adjustment method?
This is not the case, as GHCN v3 adjustments have little impact on the global-scale trend vis-à-vis the raw data. We can see this in Figure 5 below, where both GHCN v1 and GHCN v3 are compared to published NCDC and GISTemp land records:
Figure 5: Comparison of GHCN v1 and GHCN v3 spatially gridded anomalies with NCDC and GISTemp published land reconstructions.
If we look at the trends over the 1880-1990 period, we find that both GHCN v1 and GHCN v3 are quite similar, and lie between the trends shown in GISTemp and NCDC records.
1880-1990 trends
GHCN v1 raw: 0.04845 C (0.03661 to 0.06024)
GHCN v3 raw: 0.04919 C (0.03737 to 0.06100)
NCDC adjusted: 0.05394 C (0.04418 to 0.06370)
GISTemp adjusted: 0.04676 C (0.03620 to 0.05731)
This analysis should make it abundantly clear that the change in raw input data (if any) between GHCN version 1 and GHCN version 3 had little to no effect on global temperature trends. The exact cause of Smith’s mistaken conclusion is unknown; however, a review of his code does indicate a few areas that seem problematic. They are:
1. An apparent reliance on station Ids to match stations. Station Ids can differ between versions of GHCN.
2. Use of First Differences. Smith uses first differences, however he has made idiosyncratic changes to the method, especially in cases where there are temporal lacuna in the data. The method which used to be used by NCDC has known issues and biases – detailed by Jeff Id. Smith’s implementation and his method of handling gaps in the data is unproven and may be the cause.
3. It’s unclear from the code which version of GHCN V3 that Smith used.
STATA code and data used in creating the figures in this post can be found here: https://www.dropbox.com/sh/b9rz83cu7ds9lq8/IKUGoHk5qc
Playing around with it is strongly encouraged for those interested.
phi
You’re starting to catch on phi.
On a tangent to this: Thermometer bulbs shrink over time; see the chart at the top of page 98:
http://tinyurl.com/7jngj73
Its contribution might depend on how often thermometers are replaced (?)
Mosher, I used Environment Canada monthly summaries.
Again and again you miss the main point:
“But where the heck did BEST get the -13.79 number in the first place.”
The name is in the last comment. The “code” that downloads it is ftp built in to most browsers. The code that reads it is published in the source code posting. (Not very remarkable code. Fixed format FORTRAN read… following the file layout from the data source.)
Frankly, I think that’s a bit of a Red Herring as any download of v3 ought to be substantially the same and ‘how to read it’ comes from their web site. But yes, it’s nice to have the specifics.
@Amino Acids in Meteorites:
A good point that is often obscured. In the 2009 era version of GIStemp that I analyzed, for example, temperatures are kept AS temperatures right up until the very last “grid / box” anomaly creation stage. So all the RSM and Homogenizing are done on recorded temperatures, not anomalies (though it looks like an obscure kind of anomaly is temporarily created in step 2 IIRC when it does some of the UHI “adjustment” – but it’s based on an assemblage of records that are a new collection with each change of the data set… It picks and chooses out of the available data until it has ‘enough’ records, so change what records are in, you get a different ‘selection’. This can be up to 1200 km away from the target station and may contain ‘homogenized’ data from a further 1200 km away, so limiting things to inside a small grid / box misses that.)
Only at the bitter end are “anomalies” created, but these are NOT between readings on a thermometer, nor readings on any specific thermometers. They are created between two eras of the “grid / box”. One from the “baseline” and one from the “present”. In the current version of GIStemp they have 16,000 grid/boxes, yet not nearly enough thermometers to have even ONE in most grid boxes. So most of the “anomalies” are being created between two values that are both a “polite fiction” where there is no real temperature data. (These grid / boxes are filled with values based on, yes, data from up to 1200 km away that may have been UHI “adjusted” from a further 1200 km based on data that may have been homogenized from up to 1200 km away that is based on ‘corrected and adjusted monthly averages’ based on looking at “nearby ASOS” stations in the QA process up to an unknown distance away…
IMHO, that is why it is important to look at effects from blending data from outside any individual grid / box. Since it is what the climate codes like GIStemp do (or did in 2009-10 when I worked on it) and if you ignore that, you miss something important.
I’ll watch the videos a bit later, they look interesting…
@Phi:
The segments WERE separate in v2, but in v3 the data come ‘pre-homogenized’. (Which likely also means the 1200 km ‘reach’ of GIStemp may have been moved upstream to NCDC, but I’ve not gone through the newer GIStemp to see if they removed their homogenizing or just left it in to rehomogenize)
@richard T. Fowler:
My method of handling the gaps is not in the peer reviewed literature, so is “unproven”. I think it is trivially obvious to do “no bad thing” and fix some problems in the peer reviewed form. But technically, yes, it is “unproven” and it ought to be examined closely (as any change in process or code ought to be closely examined. Bugs and ‘issues’ can be incredibly subtle things…)
@ATheoK:
That’s the ‘grid / box’ issue… and the composition of change issue… Perhaps I ought to have read comments before writing the quick response, look like folks already covered some of it 😉
@mfo:
I have “variable awake time” and Anthony had even odds that I was awake then (as I have been most of the prior week) or not (as today). The world is not time-synchronous anyway, so while I appreciate the sentiment, it’s just how life is. Besides, this lets me get a nice comment thread to read 😉
(Time to put milk and sugar in the tea… back in mo…)
wrong phi.
start with canada which adjusts for tobs
pamela. you might be interested in the test of homogenizstion algorithms
in blind studies. experiment trumps
your speculation
rob
the point about there being no translation table
between version 1 and 3 really concerns Ems comparison. you can’t
simply match on id number. in all cases you
you must check location name the actual
data and then the id. id can change
as suppliers correct or upgrade their
systems. relying on id is a known
recipe for disaster. been there done that.
Mosher, I looked at the daily data for Malahat. There was one E value and 4 M (missing).
The average is 5.825925926 with the E value, 5.95 without. EC used 5.8 for the monthly summary value.
http://www.climate.weatheroffice.gc.ca/climateData/dailydata_e.html?timeframe=2&Prov=BC&StationID=65&mlyRange=1920-01-01|2005-04-01&Year=1992&Month=1&Day=22
But I am still really curious where BEST got -13.79C for Dec 2002 from this:
http://www.climate.weatheroffice.gc.ca/climateData/dailydata_e.html?timeframe=2&Prov=BC&StationID=65&mlyRange=1920-01-01|2005-04-01&Year=2002&Month=12&Day=22
All the means were above zero (and one value was missing).
micheal r
if the differences were distributed in time as you suggest the rends would be different
that is the point of comparing
trends.
sunshinehours1, I am a bit surprised about the value being lower as I have found hundreds of Winter values in BEST that are higher than the Summer values, when they should be at least 5-10 degrees lower, maybe even 20 degrees lower, the data is riddled with those kinds of errors.
Steven, observation should precede experimentation. My speculation should properly begin with observation in classic experimental design. If observed (the 1st stop following speculation) dropout and ENSO patterns show correlation, homogenization (the experiment – the 2nd step) would be better informed. It is okay for you to speculate that temperatures have risen in response to CO2. It is okay for me to speculate otherwise. But what of the 1st and 2nd step in research methods? Have you done the necessary observations first or did you skip to experimentation?
Steven, first, my congratulations to you and
NickZeke for an interesting and meticulously documented post. It is clear, well researched and presented, and eminently checkable, just like science should be done.Heck, contrary to your usual practice, you even actually answered a few questions. However, you seem to be falling back into your bad habits when you say in the comments:
steven mosher says:
June 23, 2012 at 7:22 am
I gotta say, that habit of yours of making cryptic responses that simultaneously don’t answer anything, insult the person asking the question, and pretend to great knowledge on your part, is getting really old.
Phi has posed an interesting question. He has asked you, as is bog-standard scientific practice, simply for a reference for your claim. Either answer the man’s question or admit that you don’t have an reference. The kind of response you have given is both meaningless and damaging to your reputation.
w.
Could we get this by decade: “Figure 3: Difference between GHCN v1 and GHCN v3 records matched by station lat/lon, excluding cases of zero difference.”
I know enough about statistics to know that global sea surface temperatures in all global temperature data sets are modeled and extrapolated from an inadequate number of data sources. Since 70% of the planet’s surface are oceans, the extrapolated data is speculative and readily manipulated
@Mosher
@Smith
question: if you took each raw temperature measurement and plotted it against the time when the measurement was made from the earliest known temperature measurement to the latest known temperature measurement so as to create a complete scatterplot of available raw data, what would it look like?
question: if you computed the trend through the whole of the above described scatterplot using ordinary least squares regression what would the slope of that trend be?
question: if you computed a set of trends through the whole of the above described scatterplot beginning at the beginning with the first month of data only (i.e. a one month trend), and recomputing the trend with one additional months data points added each time again using OLS to create a sequence of slopes of increasing temporal length and then plotted the values of each slope against the date of the month which had been added in in order to calculate the trend at that month, what would that graph look like?
Chas,
“Thermometer bulbs shrink over time”
It’s an interesting point, I addressed it there:
http://rankexploits.com/musings/2012/a-surprising-validation-of-ushcn-adjustments/#comment-95708
E.M.Smith,
“The segments WERE separate in v2, but in v3 the data come ‘pre-homogenized’.”
So we have actually no unhomogenized global temperature because it would require unadjusted aggregate series for it.
Steven Mosher,
“start with canada which adjusts for tobs”
Well, you have a reference which shows that “the most significant adjustment is Tobs” for Canada?
@Paul in Sweden
I’m a volunteer, so not a lot of resources consumed by me 😉
But yes, the amount of global waste over the fantasy of a Global Average Temperature is astounding. That it could instead be used to absolutely fix and solve a large number of significant problems instead is a tragedy. That the “answers” that come from AGW Panic are ‘exactly wrong’ is heart wrenching. (For example, if we did Coal To Liquids or Gas To Liquids we could be free of OPEC oil in 5 to 10 years fairly easily and at lower cost than gasoline here last week. Similar fixes to global “fuel poverty” are just as available. Then the “oil wars” could end… as just ONE example of how the misallocation of resources is a travesty.)
FWIW, the very notion of a Global Average Temperature is based in a fundamental error of Philosophy of Science. It is an intrinsic or intensive property. You simply can not average two temperatures from different things and have any meaning in the result. It is an obscure, but vitally important point; that is consistently ignored by the entire Climate Panic Industry…
http://chiefio.wordpress.com/2011/09/27/gives-us-this-day-our-daily-enthalpy/
http://chiefio.wordpress.com/2011/07/01/intrinsic-extrinsic-intensive-extensive/
Take two pans of water, one at 0 C and the other at 20 C, average temperature is 10 C. Mix them. What is the resultant temperature? You do not know. What are the relative masses of water in the two pans? Is that 0 C frozen or melted? Same problem in climate and temperatures. Average a place with 1 C change from below zero to above zero and 2 feet of snow melt with a desert that has a 1 C change the other way, you get “no change”, yet tons of snow melt heat of fusion moved… It is just wrong to use average temperatures for heat flow problems. But “it’s what they do”…
So I’m occasionally a bit “bummed” that so much time is spent examining “average of temperatures” since it really is an “angels and pins” question; even when I do it. But as that is where the “debate” is mired, it’s the only place to participate. Just somewhere in the back of your mind remember that GAT (or any other average of temperatures) is just meaningless.
@david_in_ct:
I could say “They won’t do that as it will find what I found”, but that would be cheeky… 😉
In defense of what they have done: In forensics you always want to look at things from a different point of view. Watching the magician from the audience front row doesn’t illuminate any “issues” as well as looking from back stage (or in the basement under the stage…)
So it IS important to have a divergent look at things.
My only complaint about this critique, really, is that “part two” is typically to look at things in the same way, then “part three” is to “compare and contrast” followed by “why are the things that are different, different? What is going on?”. That is where learning and wisdom come, and that is where the “perp” is typically caught and the Magician has his trick “found out’. All those are missing here.
No curiosity at all about why a ‘grid / box’ method finds similarity while a FD variable assortment method finds interesting patterns of divergence. Just “He must be wrong because he didn’t do it my way”. It isn’t about which way is “right”, it is about how each method illuminates what can be happening in the climate codes. That this ‘grid / box’ method finds similar data, yet the climate codes smear data all over, and my method that blends data also finds difference; that ought to cause much more curiosity about the nature of the data blending and smearing in the climate codes. Instead it just devolves in a more tribal ‘fling poo’ display. “I’m right, he is in error”. Not very useful, IMHO.
So the “look at it different” is a valid step, but only a first step.
It could be that the data are being carefully manicured in just such a way that the distribution of changes about a mean, ignoring date, gives a surprisingly symmetrical curve. ( I’d expect some kind of small bias in ‘accidental’ non-directed adjustment…) It could be that the ‘grid / box’ averages are so close because the data are being manicured up front to assure that result (but knowing that the actual effect will be different when run through an RSM or Homogenizing blending step). Or it could just be a reasonable and randomly collected set of data that has those properties, but tickles an unexpected ‘bug’ like behaviour when spread around. It could be malice, so a more paranoid forensic mind set ought be applied; or it could be simple error and a less paranoid and more subtle analytical examination ought to be applied. Or, in fairness, it might just be that FD has a different result than the grid / box / baseline method (and that, then turns into an argument over which is more ‘accurate’).
We just don’t know. (And “Trust me” from the data set creation folks is NOT acceptable, not now that Hansen has testified that breaking the law is justified if your goal is “good” enough.)
But even IF it is an artifact of FD vs grid/box/baseline: That, then, just raises the question of how do we benchmark the grid/box/baseline codes in use so as to prove THEY are not out of whack with the warming they find? If 1/2C to 1C of “warming” depends on method, how do you prove the “error bars” are less than 1/10 C when you can not run a benchmark through the codes? Or prove the data unbiased to that precision?
At any rate, taking a different POV is valid, just incomplete.
@Geoff Sherrington
That’s the kind of “devil in the details” stuff that gets ignored in the kind of examination done in this critique. It is the kind of thing that the forensics mindset is dedicated to finding. It is why doing a broad comparison of data sets is an important thing to do, and why past versions ought to all be archived and publicly available (but are not).
@Wayne:
Pixel counting? Talk about dedication 😉 Nicely done, though…
Yes, things always cool the past and warm the present (with the only exception being the occasional warming of the very deep past where the data are thrown away in HadCRUT and GIStemp.)
My question can be recast in the context of your finding as “To what extent does the small induced bias found in a non-spreading grid/box examination become a larger bias when spread via homogenizing, UHI “adjusting” and infilling via the RSM?” The critique here can not answer that question. My examination points out that there’s plenty of opportunity for selected data items to have more impact when spread outside the grid / box. I ask the question “Is that enough to account for the imputed ‘warming’?” Others say “It can not be a problem, trust us.” I’d rather not trust. I’d rather verify…
@Phi:
It is exactly that ‘joining of segments’ as a kind of ‘splice artifact’ that IMHO can be a source of issues. In places like Marble Bar Australia, GIStemp finds a ‘warmer than ever’ trend; yet the record temperature there has never been matched. IMHO that is evidence for a “splice artifacty” behaviour in GHCN/GIStemp interactions. So my approach is to look for the potential for that kind of artifact stimulation in the data. (Segments with a low start point and a high end point). FD, being more sensitive to end states, will react more to those “high ends” and show the rest of the past lower in comparison. It will illustrate the “splice artifact” potential in the data.
The critique here takes deliberate action to suppress that effect (uses a baseline from a semi-random spot in the middle) then simply asserts that the various climate codes can’t have a sensitivity to the splice as they do something similar. I find that inadequate (especially in the face of the known data smearing methods and the specific examples of places like Marble Bar where the artifact effect is clearly present.) So theory runs headlong into existence proof…
No, I’ve not been able (yet) to prove exactly how the artifact effect gets through the codes. (GIStemp is brittle to station change so hard to benchmark) but to simply assert it does not matter because a “toy world” version of the code shows only a small part of it is, well, inadequate.
@all:
I see Steven Moser is more interested in tossing insults than thinking. Spending more time with the Warmers, eh?
BTW, I’ve generally been pretty careful to say I use a variation on a peer reviewed method and to NOT claim that that variation was peer reviewed. Please to do not assert that I lie about that. I’ve said FD is peer reviewed, and it is; and then I point out where I do something different from the classical form. That, too, is true.
Per “Didn’t Get The Memo”:
Steve seems to have not read my statements about wanting the end point sensitivity in the look for splice artifact potential. Oh well. He also seems to have ignore my response in the prior thread where I quoted Jeff Id saying 10,000 segment compares had near zero error but down in the 50 range there was an issue; and where I pointed out that using FD on over 6000 thermometers, by month, gives over 72,000 segments and is well over 10,000. Oh Well.
Steve seems fixated on what finds the perfect Global Average Temperature (ignoring that it is impossible and ignoring that it is meaningless even if found). Not “got the memo” on intrinsic properties I guess. Oh Well.
Here’s a hint Steven: Drop the insult tone. Makes you look vindictive. Contributes nothing. I’d also suggest being careful hanging out with Warmers. It is slowly making your presentation negative and bitter in public.
Personally, I don’t think there will ever be a “right way to go”. Just different methods with different uses and different “issues”, so usable for different things. The world is not a nail, even if you love your hammer…
@Andrew:
I spent a while trying to find “raw” data. It doesn’t exist. There are, as Steve calls it, “First report” data. And it has errors in it.
I noted one where I found 144 C for a station in the USA.
Now Steve is happy to say ~’but we can find that 144 C and change it to something valid’. I think “Hmmm… So we catch the 144 C, but do we catch the 40C that ought to be 38 C?” There is a system that looks for some large number of sigmas off and replaces them with an average of “nearby” ASOS stations (think hot airports, then averaged that will by definition suppress things like large down spikes) but this does nothing to catch whole classes of bad data.
That the electronic instruments seem to regularly “fail high” to insane values, THEN are caught and fixed, implies a probably slow fail to higher than would not be caught, THEN a catastrophic fail high. A periodic calibration is supposed to prevent / catch this, but that would just result in a series of segments with ‘start low end high’ effects that, in smearing / splicing, can induce bias into the result.
So while some folks are happy to just say “Trust me, we fix it good.”, I’m less willing to accept that. The data from the electronic thermometers that “sucked their own exhaust” and heated from the humidity measurement device is still in the record, for example.
@Pouncer:
One clarification:
While I do think all the adjustment and such of individual data items is an issue; I think that the larger issue is the potential for changing which records are in, vs out, of the data set, to have impacts. What I show in the v1 vs v3 is that they DO have impacts, and those can be large.
What the critique here does is say in essence, if we avoid the question of station records in a small area having influence over others we don’t find that problem. Yes, don’t look for it, you won’t find it. Then the assertion is left that the various climate codes are also insensitive to the effect (yet we know that they do data spreading).
IMHO that is the larger issue, not things like tossing out a 144 C and replacing it with an interpolation from the neighbors…
Basically, my major point is just that “The way you look at the data has more effect than the presumed Global Warming”. So is GIStemp right, or wrong? You can not know. It is not possible to benchmark their code and do comparative runs on different data. But you can put comparative sets of data through different systems and get an idea of the general range of sensitivity. That shows a range greater than the Global Warming signal. So ‘choice of method’ can have more ‘warming’ than the GW signal. Choice of data set can have more ‘warming’ than the GW signal. I find that “a problem”. Others like to say “Trust us, we have it right”. This critique says “Here is A way that is less sensitive”. Which is sort of a “so what” kind of moment for me…
@Pamela Gray:
IMHO, the “QA method” is even worse than that. It takes a collection of “nearby” ASOS stations at airports and uses them to replace any data found too far from the expected value. IMHO this will incorporate Airport Heat Islands by default AND via averaging them, prevent any low going extreme values from showing up. (An average always suppresses range).
For homogenizing, the techniques vary, but largely use the same kind of “average a bunch” to get a value. That will always tend to put in mid-values and not what would be there in the real world of actual high or low range values. And what if “the bunch” have “issues”? Like Airport Heat Island? Well, you get that too. (And very large percentages of present GHCN data come from Airports, while by definition none came prior to 1914).
The whole issue of how potentially wrong data is detected and handled is, IMHO, not very good. IIRC the threshold for tossing a “bogus” value was a large fixed temperature in one code ( so low excursions will be tossed more often than high excursions… Inspection of the data shows cold spikes are more extreme than high spikes). In others it is several std deviations. That will miss a whole lot of “somewhat off” errors (like those electronic thermometers that “sucked their own exhaust”).
It’s a modestly large source of error that is largely ignored, IMHO.
(Time for another break… back soon.)
E.M., thanks for your response to my point that Steve Mosher also responded to. I think you’ve done some very good work here.
RTF
Is the data “Fit for purpose”?
First you have to know what their “purpose” is.
The people who want to use this data want to take control of energy. All production, manufacturing, development etc would be with their approval. They want to use it to stop fossil fuel use. They want to use it to stop third world advancement. They want to use it to tell you where you can live, what you can buy and how to live your life. All this based on “the adjusted data”.
So I ask, is it fit for purpose?
steven mosher says: June 23, 2012 at 7:17 am
You can expect some updates to that Sante fe chart in the coming months. I suspect folks who do spectral analysis will be very interested.
I will look forward to your results.
I suggest to separate two hemispheres, South is less volatile, ocean inertia and CPC flywheel effect, North is affected by gyres and more in sympathy with the GMF
Till then this is what I get
http://www.vukcevic.talktalk.net/NH+SH.htm
When done I’ll email you magnetic data, so you can have some fun with it
EM: “Pixel counting? Talk about dedication 😉 Nicely done, though…”
That’s me stepping all of the way back to Newton physically counting areas under the curve when no closed form is available (or in this case no data at all) and is all you have to work with, but that burning question presses you to answer it! It’s just a qick and easy c# program to count each color’s pixels. Pick up any chart and you can either color under the curve, or like this one, color the bars, and presto, you’ve got the approx. multiple integrations. Does work. And if you ever try it, use mspaint to first convert png or jpg to 16 color or 256 color fmt and pick only pure colors so you can easily identify which count goes with which color. A whistle lets me exclude numerous colors that have less than ‘n’ pixels, makes it real easy. Your limits are on the axises already.
But tell me, is that your Figure 3? If so you should have the exact ratio and I’m curiuos just how close this method works. Only if you get the time.
Are we to take away from this that there are no significant adjustments to the GHCN data?
The USHCN adjustments are nearing +0.5C but GHCN to NCDC Land is just 0.004C per decade?
This requires a confirmation because that is what your charts and data is showing.
——————–
Perhaps you could re-try the analysis with the original 1992 GHCN data held at the CDIAC.
http://cdiac.ornl.gov/ftp/ndp041/
probably can just unzip the Temp.data file (22.8 Mb unzipped) and replace it in your programs.
EM Smith says
“the perpetually rosy red arctic in GIStemp output where there is only ice and no thermometers does not lend comfort…”
For those who don’t know what he means, he’s a video that will help you get a start on understanding:
Phi, these bulb contractions do not seem to be as large as those you mention, but 0.1 C in the first 4 years (in what is presumably a top-notch thermometer ) is big enough to explain all of the ‘global warming’ if the thermometers were broken roughly every 5 years.
I have a “Meterological Observers Handbook 1939″ (HMSO) and there is not one mention of calibration cards or any suggestion that thermometers should be sent for recalibration after a period of time. There are paragraphs on methods to avoid breakages (” (your) coat should be held with the left hand” etc) .
I have the 1946 amendment list which just adds “Maximum thermometers with solid stems (not sheathed) are very likely to break above the bulb. They should therefore be gripped above the bulb rather than in the middle when setting”.
It might be that they did not consider recalibrating the thermometers simply because they tended to have a short service life.
Broken mercury/spirit thread failures are ‘upwards’ too. Perhaps there is a case for trimming away data prior to ‘missing values’ and trimming for a few years afterwards ?
Re:
http://wattsupwiththat.com/2012/06/22/comparing-ghcn-v1-and-v3/#comment-1016421
failed link try: http://www.vukcevic.talktalk.net/NH-SH.htm