Guest Post By Walter Dnes:
There have been various comments recently about GISS’ “dancing data”, and it just so happens that as GISS data is updated monthly, I’ve been downloading it monthly since 2008. In addition, I’ve captured some older versions via “The Wayback Machine“. Between those 2 sources, I have 94 monthly downloads between August 2005 and May 2014, but there are somegaps in the 2006 and 2007 downloads. Below is my analysis of the data.
Data notes
- I´ve focused on the data to August 2005, in order to try to make this an apples-to-apples comparison.
- The net adjustments between the August 2005 download and the May 2014 download (i.e. the earliest and latest available data). I originally treated 1910-2005 as one long segment (the shaft of the “hockey-stick”). Later, I broke that portion into 5 separate periods.
- A month-by-month comparison of slopes of various portions of the data, obtained from each download.
- Those of you who wish to work with the data yourselves can download this zip file, which unzips as directory “work”. Please read the file “work/readme.txt” for instructions on how to use the data.
- GISS lists its reasons for adjustments at two webpages:
- The situation with USHCN data, as summarized in Anthony´s recent article , may affect the GISS results, as GISS global anomaly uses data from various sources including USHCN.
In the graph below, the blue dots are the differences in hundredths of a degree C for the same months between GISS data as of May 2014 versus GISS data as of August 2009. GISS provides data as an integer representing hundredths of a degree C. The blue (1880-1909) and red (1910-2005) lines show the slope of the adjustments for the corresponding periods. Hundredths of a degree per year equal degrees per century. The slopes of the GISS adjustments are…
- 1880-1909 -0.520 C degree per century
- 1910-2005 +0.190 C degree per century
The next graph is similar to the above, except that the analysis is more granular, i.e. 1910-2005 is broken up into 5 smaller periods. The slopes of the GISS adjustments are…
- 1880-1909 -0.520 C degree per century
- 1910-1919 +0.732 C degree per century
- 1920-1939 +0.222 C degree per century
- 1940-1949 -1.129 C degree per century
- 1950-1979 +0.283 C degree per century
- 1980-2005 +0.110 C degree per century
The next graph shows the slopes (not adjustments) for the 6 periods listed above on a month-by-month basis, from the 94 monthly downloads in my possession.
- 1880-1909; dark blue;
- From August 2005 through December 2009, the GISS data showed a slope of -0.1 C degree/century for 1880-1909.
- From January 2010 through October 2011, the GISS data showed a slope between +0.05 and +0.1 C degree/century for 1880-1909.
- From November 2011 through November 2012, the GISS data showed a slope around zero for 1880-1909.
- From December 2012 through latest (May 2014), the GISS data showed a slope around -0.6 to -0.65 C degree per/century for 1880-1909.
- 1910-1919; pink;
- From August 2005 through December 2008, the GISS data showed a slope of 0.7 C degree/century for 1910-1919.
- From January 2009 through December 2011, the GISS data showed a slope between +0.55 and +0.6 C degree/century for 1910-1919.
- From January 2012 through November 2012, the GISS data showed a slope bouncing around between +0.6 and +0.9 C degree/century for 1910-1919.
- From December 2012 through latest (May 2014), the GISS data showed a slope around 1.4 to 1.5 C degree per/century for 1910-1919.
- 1920-1939; orange;
- From August 2005 through December 2005, the GISS data showed a slope between +1.15 and +1.2 C degree/century for 1920-1939.
- From May 2006 through November 2011, the GISS data showed a slope of +1.3 C degree/century for 1920-1939.
- From December 2011 through November 2012, the GISS data showed a slope around +1.25 C degree/century for 1880-1909.
- From December 2012 through latest (May 2014), the GISS data showed a slope around +1.4 C degree per/century for 1880-1909.
- 1940-1949; green;
- From August 2005 through December 2005, the GISS data showed a slope between -1.25 and -1.3 C degree/century for 1940-1949.
- From May 2006 through December 2009, the GISS data showed a slope between -1.65 and -1.7 C degree/century for 1940-1949.
- From January 2010 through November 2011, the GISS data showed a slope around -1.6 C degree/century for 1940-1949.
- From December 2011 through November 2012, the GISS data showed a slope bouncing around between -1.6 to -1.7 C degree/century for 1940-1949.
- From December 2012 through latest (May 2014), the GISS data showed a slope bouncing around between -2.35 to -2.45 C degree per/century for 1940-1949.
- 1950-1979; purple;
- From August 2005 through October 2011, the GISS data showed a slope between +0.1 and +0.15 C degree/century for 1950-1979.
- From November 2011 through November 2012, the GISS data showed a slope bouncing around between +0.2 and +0.3 C degree/century for 1950-1979.
- From December 2012 through latest (May 2014), the GISS data showed a slope around +0.4 C degree per/century for 1950-1979.
- 1980-2005; brown;
- From August 2005 through November 2012, the GISS data showed a slope of +1.65 C degree/century for 1980-2005.
- From December 2012 through latest (May 2014), the GISS data showed a slope around +1.75 to +1.8 C degree per/century for 1980-2005.
- 1910-2005; red;
- This is a grand summary. From August 2005 through December 2005, the GISS data showed a slope of +0.6 C degree/century for 1910-2005.
- From May 2006 through December 2011, the GISS data showed a slope of +0.65 C degree/century for 1910-2005.
- From January 2012 through November 2012, the GISS data showed a slope bouncing around +0.65 to +0.7 C degree/century for 1910-2005.
- From December 2012 through latest (May 2014), the GISS data showed a slope of +0.8 C degree per/century for 1980-2005.
In 7 years (December 2005 to December 2012), the rate of temperature rise for 1910-2005 has been adjusted up from +0.6 to +0.8 degree per century, an increase of approximately 30%.
Commentary
- It would be interesting to see what the data looked like further back in time. Does anyone have GISS versions that predate 2005? Can someone inquire with GISS to see if they have copies (digital or paper) going further back? Have there been any versions published in scientific papers prior to 2005?
- Given how much the data has changed in the past 9 years, what might it be like 9 years from now? Can we trust it enough to make multi-billion dollar economic decisions based on it? I find it reminiscent of George Orwell’s “1984” where;
“Winston Smith works as a clerk in the Records Department of the Ministry of Truth, where his job is to rewrite historical documents so they match the constantly changing current party line.”
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.



Paraphrasing Nick Stokes says at July 3, 2014 at 5:22 pm
I don’t see Clinton having sex with interns here. But the fact is that he doesn’t do much Lewinskis at all now. Since the end of his presidency, he has become more handy, with Hillary’s impending run. That’s the main reason for the change
Walter,
Thanks for the analysis. You are showing the local trends as discontinuous lines, but that’s only fair to GISS if they really have step-wise adjustment changes at the transition times. If they don’t, would it not would be fairer to present this as connected trend lines? Basically, you’d change from a fit on 6×2=12 variables to 6×1+1=7 variables.
Frank
The earliest version of the US National Temperatures from the NCDC are in this report on Page 25 Climate Assessment. A Decadal Review 1981-1990. 1934, 1921, 1953, 1954 were warmer than the 1980s.
The actual data numbers have long been removed from the net by the NCDC. It is just a pdf and someone would have to digitize the data.
http://www1.ncdc.noaa.gov/pub/data/cmb/bams-sotc/climate-assessment-1981-1990.pdf
Also, a huge pdf 56 MB, Trends ’93 A compendium of data on Global Change
Contiguous United States Temperatures from 1900-1991 are on Page 689. Just annual and seasonal anomalies in C in a table. You can download the report on this page.
http://www.osti.gov/scitech/biblio/10106351
Lots and Lots of data in this report some of which can still be found here – US temps have been replaced of course by the NCDC in the database.
http://cdiac.ornl.gov/ftp/trends93/
“My question is: where are the politicians with the guts to acknowledge that the whole CO2 story has reached the end of the line, and that no more money should be wasted on this fairy story?”
When the practical consequences of ‘climate change’ policies start to hit the ordinary voter in his or her pocket in a significant and easily seen manner, politicians will gain the balls to face down the pro-AGW brigade. We have already seen this in the UK, where rising domestic energy prices are now a political hot potato. Politicians are still managing to hold two conflicting ideas in their heads, namely being against ‘fuel poverty’ and being pro-AGW at the one and same time, despite their policies to combat the latter actively causing the former. But the fact that they now have two ideas in there when before they only had one is itself progress. Slowly they will cotton on there are votes in reversing AGW policies and once one party breaks ranks others will follow. UKIP kind of have in the UK, but it’ll take one of the traditional parties to follow suit before the dam breaks.
“The bottom line is that there is no “temperature data,” prior to satellite measurements”
Thank you, Phil.
Do they have an explanation of why they are doing this?
============
they have a rationalization. human beings have an infinite capacity to rationalize any decision, good or bad, such that it is justified.
So this claim is cherry-picked, “In 7 years (December 2005 to December 2012), the rate of temperature rise for 1910-2005 has been adjusted up from +0.6 to +0.8 degree per century, an increase of approximately 30%.” with 1910 being picked, presumably because it’s at the bottom of the adjustment curve. Despite the cherry-picking, I still find the method troubling, and I agree with this claim, “Given how much the data has changed in the past 9 years, what might it be like 9 years from now? Can we trust it enough to make multi-billion dollar economic decisions based on it?”
Here’s the thing about adjustments. You need to label your units. Correctly.
If these are temperatures, then the temperature in Central Park should be the same, whether it’s 2005 or 2014.
That’s not happening in this data. The claim is because the data is “adjusted.” OK. There are some plausible reasons for that, but my simple mind needs an analog. Do we have any other adjusted data sources? Yes, the news always talks about “inflation adjusted dollars.” So maybe adjusted temperatures are like adjusted dollars? If so, there’s something missing. When the news quotes inflation adjusted dollars, they tell me the frame of reference, “buying power of a 2005 dollar,” for example. And the Minneapolis Fed gives me an inflation calculator so I can go to their web site and convert between 2014 dollars and 2005 dollars.
GISS data is labeled in degrees C. Not “2014 adjusted degrees C,” and they provide no way to compare between years. Giving the benefit of the doubt and assuming that all of the adjustment methods used are “above boards,” I’d still have to say that their implementation is incomplete. They need to label their units as adjusted degrees, and they owe the public a conversion tool to convert between years the same way the Minneapolis fed has done for inflation adjustments.
Also, if these adjustments are valid, once a temperature is measured and adjusted once for a set of stations at a specified time, the proportionate relationship between those stations should stay the same through all future adjustments until the end of time. It’s not apparent to me whether that is happening, but I suspect not.
w.r.t that last paragraph, I downloaded the data and see that it doesn’t include individual stations. With exceptions for closely ranked years, the same statement should also hold true for points in time. If month1/year1 is markedly warmer than month2/year2 when they were collected in 2005, that should also be true in 2014. Barring rounding differences, if we happen to find that month2/year2 is markedly cooler in 2005 and warmer in 2014, for any pair of months/years, then I think that should be a huge red flag, as it indicates that the size of an “adjusted degree” is not being held constant, even within a particular dataset.
If an “adjusted degree” is bigger for one station than it is for another, even during a particular data sweep, than I don’t see that how it conveys any useful information. On the other hand, if the inter-time-period ranking of months stays the same for all pairs at all times (except for rounding differences), that would offer some confidence in the adjustments.
If I get around to it, I may run this test, but it might be a while… Don’t hold your breath waiting for me.
The people involved, the people who help Obummer keep repeating his lies, “hottest year ever,” blah blah blah, have jobs requiring them to check their personal integrity at the door. Stokes you seem to be one of them. This is the ugliest thing in America these days, and most Dems are unaware that their chosen Liar-In-Chief stoops this low. The Dem donors who fund this wretched nonsense have no shame, cannot have any pride left.
The universities have become cesspools of mendacity. I took Econ 101 from a Communist at U of Michigan. His Teaching Assistant propounded the lies from the Club of Rome, Malthusians. I scoffed openly at this in class, he flunked me on the first hourly despite my having answered every question correctly, I dropped the class. This was in 1979! Things have not improved, seems actually far worse now. “Climate Science” indeed, not very scientific.
This blog is doing God’s work. Expose it all to the light, watch it wither away like in the vampire movies, and the sooner the better…
You extremists just can’t see the proper role for public servants adjusting the published temperatures on a daily basis so I don’t feel the hot or cold extremes so much. It’s called sustainability and trust them, they’re from the Gummint and they’re here to help.
This is a critical area as it underpins the whole of CAGW. Articles I have read over the past few years suggest every dataset has been altered in a similar fashion. The suspicion is that most of the alterations are fraudulent.
It suggests the organisations looking after this data are ‘not fit for purpose’ and the functions should be removed from these bodies.
At the end of the day it is just a bunch of data. Most governments have statistical departments which could take over this function. Using professional statisticians would also overcome some of the other criticisms on how the data is handled. Any adjustments should be requested via a committee and the reasons for any accepted published on the web in case someone wants to check or appeal against their inclusion.
There is another aspect of this. If any changes are fraudulent then in effect the data is being altered to provide more funding. Alternatively it may be regarded as using government funds to create propaganda. I would expect this to be illegal and making them open to a criminal prosecution.
These kind of articles come and go and are ignored by all apart from skeptics. What I would like to see is a coordinated effort to produce a document on how these datasets have been altered and a campaign to remove the functions from the current bodies. A campaign forever until things are changed. These datasets are so fundamental to CAGW that destroying their credibility would be a massive blow. I think that this is the one battle that if won would win the war.
Bill_W says:
July 4, 2014 at 5:06 am
John Harmon,
I’m not so sure the Congressional record can’t be changed. After all, congress-folk are allowed to record a speech at a later date and insert it into the record when they were not even present the day the debate took place. I hope that the actual date they made the recording is prominently displayed but I have not checked
==========================================================
If you can stomach watching and listening, congressional types nearly always end their spiel with:
“I reserve the right to REVISE and extend my remarks”. A convenient out for the times they get called out being for or against an issue. Just another way of saying “I was for it before I was against it.”
Thanks. Dancing data makes all analyses based on any one version suspect.
Changing the time of temperature observations at a given weather station makes a huge difference to the reading. Surely that is obvious. Further, the work trend has been for people to stay at work all day, rather than return home at lunch time to eat which used to happen. So older manual readings tend to be taken closer to the hottest part of the day, and newer manual readings tend to be taken early in the morning before going to work. Most recently automated stations presumably give you a full temperature profile throughout the day.
So there’s no way that you can get a sensible trend from readings at the same place with different “Time of Observation”, which forces you to adjust it and guarantees that the adjusted temperature is going to change with time as the knowledge of how to do it improves. Thus, to cope with the change in time you would expect to increase the trend over time for the majority of stations, though a minority may reduce, and you should have the data (“Time of Observation” in the paper record) available to do this. Exactly what adjustment you should then make is going to be a matter of doing analyses and developing expertise which would then let you create rules, which you might hope would get better over time.
The other factor which needs adjustment is the well-known UHI (“Urban Heat Island”) effect. This time you need to identify changes in individual station readings not reflected in readings in the same area (which need to include urban and rural stations), and mostly this ought to result in reducing the more recent temperatures for urban stations. The process need not be done by explicitly identifying urban stations – just to make adjustments for any station which suddenly starts to get out of line with the other stations in the region ought to be able to cope with any UHI effects and similar changes in either direction.
There are research papers and code available comparing “urban” with “rural” trends for the USA which determine whether there are any UHI distortions left in the data set, and the GISS temperatures come out pretty well on this analysis taking any one of four definitions of what “urban” might actually mean!
Peter:
re your post at July 4, 2014 at 10:20 am.
Your excuses for the ‘adjustments’ to GISS data do not wash.
If you were right then each past datum would require being ‘adjusted’ ONCE. But past data are changed almost every month.
The frequent GISS changes cause this.
GISS data may be propaganda but it certainly is NOT scientific information.
Richard
From Phil.
“””””…..The bottom line is that there is no “temperature data,” prior to satellite measurements (which suffer from their own issues) and the USCRN network, which only has about 10 years worth of data, IIRC, BEST, GISS, HADCRUT, etc. etc. notwithstanding. All of the temperature indices are essentially models (which may reflect the modelers’ biases and preconceptions more than the historical weather.)……”””””
And the top line is that almost concurrent with the satellite data era, is the era of those oceanic buoys, that have been measuring simultaneously, both the near surface (-1 metre) oceanic water Temperature, and the near surface/lower troposphere (+3 metre) oceanic air Temperatures, on which Prof. John Christy, et al, reported about Jan 2001 that those Temps are not identical (why would they be). And further more, they aren’t correlated, so all that 150 years of ocean water temps obtained haphazardly at sea, as proxies for oceanic lower troposphere air Temperatures, to mingle with the other 29% of the earth surface, that is solid ground, instead of water; is just plain junk. Total rubbish.
So lacking correlation, the correct air temps, are not recoverable from the phony baloney ersatz water Temperatures.
So I’m in agreement. No global Temperature data, pre circa 1979/80; so actually anything approximating global Temperatures have been getting measured for about one earth climate “interval” time (30-35 years). Wait till we have maybe five under our belt.
the bottom line is always the same – unless they (i.e. any of the dataset holders) can provide the genuine real RAW data, along with each and every subsequent change/adjustment and the documented reason for each and every adjustment – we, the ‘consumers’ – have absolutely no idea what we are getting.
Now, if this was a food type consumer product 0 we could take it or leave it if we don’t like. However, this is, in effect, a WORLD consumed ‘product’ (although force fed product might be a better description) and we are not allowed to question it!
I don’t have a problem with that in itself – if mugs want to use such data with blind trust – so be it. However, when said data is used and upheld as scientifically valid – that changes the game – it must be reproducible and proven as valid. Frankly, to this day, I don’t think we have reliable data – and certainly not without questions as to its history or validity! Ergo, in almost any other scientific endeavour, the ‘results’ or ‘conclusions’ based on such data would be thrown out or at best held in very low esteem (think wagonload of salt here!)………
Walter Dnes (July 3, 2014 at 5:46 pm) “I can’t see any over-riding reason why the endpoints must join up.”
Think of it as a single trend with multiple linear segments, not a set of unrelated linear trends.
Taken separately, each trend line you show is technically correct for its period. But the period segments are typically very short and trends are dubious over short periods. More importantly, there’s a strong short cooling segment in which the start point is way above the previous end point, and the end point is way below the next start point. ie, your composite trend has large jumps in it which a trend line (even a composite trend line) shouldn’t have – not if it’s a “trend”. The trend for that particular segment is also heavily dependent on its start and end points – a modest change in end point will make a large change in trend. Lastly, the sum of all the trends over their selected periods is not equal to the trend over the total period, so the picture they give is misleading.
It’s no big deal to calculate multi-segment trends that join up, if you have a general-purpose optimising algorithm for multiple variables. You simply do a least-squares optimisation with the temperatures at the segment ends as the variables. That’s n+1 variables for n segments. Even better, add in time as a variable at each internal segment end (ie, allow the join points to move horizontally), that’s another n-1 variables. I am horribly busy over the next few days so might not be able to do the calc for a while.
Maybe instead of “Think of it as a single trend with multiple linear segments” try “Think of it as a curve-fit with multiple linear segments”.
Richard Courtney says that the GISS guys should get only one shot at adjusting the data – once they have had that – no further changes. This would be a pretty stupid approach. Here’s how it would work….
The GISS team have already adjusted readings for Time of Observation but now realise that the new electronic mini-max thermometers installed in stations on average give a maximum temperature 0.4C lower and a minimum temperature 0.3C higher than the previous manual readings. But Richard’s rule says that they have already had their shot at adjusting, so its not permissible to make the new change.
And more scrutiny of the data has brought to light that a few guys were cheating in the 1920’s – they went on holiday without arranging a standin and decided the easiest approach was to record the same temperature for a week. Sorry guys – you’ve had your one adjustment and the adjusted temperatures are now fixed in perpetuity.
See http://reason.com/archives/2014/07/03/did-federal-climate-scientists-fudge-tem for further cases – despite the title whoever wrote that piece had done more than take a superficial view saying “adjustments are always bad – particularly adjustments in the direction we don’t like”.
http://eric.worrall.name/giss/ is a 3d navigable “lunar lander” construction based on the temperature data supplied by Walter Dnes in the post above. Tested in Safari and Chrome, *might* work in IE10.
Only click the link if you have a high spec computer running the latest OS – it uses advanced 3D / html code, which stretches the computer to the limit.
The terrain from left to right is oldest GIS snapshots to newest snapshots on the right. From front to back is oldest temperature anomalies (1880s) to newest anomalies at the back.
You can navigate with the arrow keys, up = forwards, down = backwards, left and right.
Enjoy 🙂
People really ought to look into R. R is great for examining data. For example, it took me all of ten minutes (including the time to familiarize myself with some unfamiliar commands) to generate 94 images showing the changes between each iteration of the data set. I’ve uploaded that collection of images here. It takes only five or so lines of code to repeat the process.
I’m not drawing any conclusions on this topic. I haven’t spent the time. I just would like to see people move away from things like Excel.
Kev-in-Uk says the GISS team should document each change and the effect on all the data from the previous change. The implication seems to be that he thinks they take the previous set of adjusted changes and make a further set of changes on that. Of course that isn’t what happens – each time they would go back and reprocess the raw data.
What happens is that some new anomaly comes to light in the temperature data set, leading to a requirement for a change to the adjustment process. A new version of the adjustment computer code would be written. Once this has been run against the raw data, the new set of temperature data set version would be subject to regression testing to check that the new code version has not re-introduced any of old identified problems and that it does indeed fix the newly identified anomaly. At that point the new version of data can be released with a note saying why the processing has changed.
All this should be standard IT practice – nothing special about temperature data sets here.
In fact it will be a two-step process – the raw station data is processed in the manner described above, then a separate step is run which aggregates the results with suitable weighted averaging to provide data for a region, set of grid points or globally.
Now what is the important thing? Is it important to know what errors were present in every single one of the past versions of data and to work out exactly how correction of these errors has changed every single station reading? This would be a total waste of time.
The important thing is to be able to show that the new set of data contains none of the original flaws or biases that the data analysts have spent so much time trying to identify. The real deliverable for openness is the new version of code used to make adjustments on the raw data. An expert reading this code could determine the method of adjustment, whether it resolved all the issues identified in the past, and whether it was likely to introduce any systematic biases of its own.
You would hope that over time the frequency and the effect of new adjustments would go down, and this seems to be what is happening to the GISTEMP data set.
There are also ways you can use to validate whether the adjusted data is as good as you can make it….
Someone else can write a completely different set of code using different techniques for resolving known anomalies in the same raw data, and this is what was done with the BEST project. BEST found that the adjustment techniques used by the NOAA were pretty good.
You can compare with the satellite data, available only since 1979. However, this has problems of its own. While weather stations take readings from one place at one point in time, satellite sensors in the various microwave bands read the average temperature over a particular (and usually fairly broad) range of heights in the atmosphere. To get the “lower troposphere” temperature you have to add and subtract temperatures for different bands, then make allowances for orbital changes. Further, somehow you have to back-calibrate all sensors from old satellites against the better sensors on the newer satellites, when not all old satellite sensors did not overlap in time with the new sensors. Some new satellite sensors have had to be disregarded because they were clearly faulty. All in all it makes ground station adjustments look easy which why the RSS and UAH processing often give very different answers from the same set of available sensor readings.
Some fairly significant problems were found with these satellite data sets in the early days, and it is still not clear how accurate they are. The resulting temperatures swing higher on the peaks and lower on the troughs of global temperature than the ground station readings. But it does provide a completely independent set of readings against which ground station readings can be compared.
Peter:
At July 5, 2014 at 1:44 am you say
NO! ABSOLUTELY NOT! Only a charlatan could think such a thing!
The asserted “total waste of time” provides uncorrected error to every analysis which uses the data set. No scientist can find that acceptable.
So, it IS “important to know what errors were present in every single one of the past versions of data”. And that can only be determined in falsifiable manner by knowing “exactly how correction of these errors has changed every single station reading”
You follow that with this
But that is an admission of “original flaws” in the previous data, and it says the data analysts have identified those “flaws”. Why not publish the “flaws” and their effect if they truly are identified?
Richard
Ooops.
I wrote
The asserted “total waste of time” provides uncorrected error to every analysis which uses the data set. No scientist can find that acceptable.
I intended to write
The asserted “total waste of time” provides unquantified error to every analysis which uses the data set. No scientist can find that acceptable.
Sorry.
Richard