NEW UPDATES POSTED BELOW
As regular readers know, I have more photographs and charts of weather stations on my computer than I have pictures of my family. A sad commentary to be sure, but necessary for what I do here.
Steve Goddard points out this NASA GISS graph of the Annual Mean Temperature data at Godthab Nuuk Lufthavn (Nuuk Airport) in Greenland. It has an odd discontinuity:
Source data is here
The interesting thing about that end discontinuity is that is is an artifact of incomplete data. In the link to source data above, GISS provides the Annual Mean Temperature (metANN) in the data, before the year 2010 is even complete:
Yet, GISS plots it here and displays it to the public anyway. You can’t plot an annual value before the year is finished. This is flat wrong.
But even more interesting is what you get when you plot and compare the GISS “raw” and “homogenized” data sets for Nuuk, my plot is below:
Looking at the data from 1900 to 2008, where there are no missing years of data, we see no trend whatsoever. When we plot the homogenized data, we see a positive artificial trend of 0.74°C from 1900 to 2007, about 0.7°C per century.
When you look at the GISS plotted map of trends with 250KM smoothing, using that homogenized data and GISS standard 1951-1980 baseline, you can see Nuuk is assigned an orange block of 0.5 to 1C trend.
Source for map here
So, it seems clear, that at least for Nuuk, Greenland, their GISS assigned temperature trend is artificial in the scheme of things. Given that Nuuk is at an airport, and that it has gone through steady growth, the adjustment applied by GISS is in my opinion, inverted.
The Wikipedia entry for Nuuk states:
With 15,469 inhabitants as of 2010, Nuuk is the fastest-growing town in Greenland, with migrants from the smaller towns and settlements reinforcing the trend. Together with Tasiilaq, it is the only town in the Sermersooq municipality exhibiting stable growth patterns over the last two decades. The population increased by over a quarter relative to the 1990 levels, and by nearly 16 percent relative to the 2000 levels.
Nuuk population growth dynamics in the last two decades. Source: Statistics Greenland
Instead of adjusting the past downwards, as we see GISS do with this station, the population increase would suggest that if adjustments must be applied, they logically should cool the present. After all, with the addition of modern aviation and additional population, the expenditure of energy in the region and the changing of natural surface cover increases.
The Nuuk airport is small, but modern, here’s a view on approach:

Closer views reveal what very well could be the Stevenson Screen of the GHCN weather station:

Here’s another view:

The Stevenson Screen appears to be elevated so that it does not get covered with snow, which of course is a big problem in places like this. I’m hoping readers can help crowdsource additional photos and/or verification of the weather station placement.
[UPDATE: Crowdsourcing worked, the weather station placement is confirmed, this is clearly a Stevenson Screen in the photo below:

Thanks to WUWT reader “DD More” for finding this photo that definitively places the weather station. ]
Back to the data. One of the curiosities I noted in the GISS record here, was that in recent times, there are a fair number of months of data missing.
I picked one month to look at, January 2008, at Weather Underground, to see if it had data. I was surprised to find just a short patch of data graphed around January 20th, 2008:
But even more surprising, was that when I looked at the data for that period, all the other data for the month, wind speed, wind direction, and barometric pressure, were intact and plotted for the entire month:
I checked the next missing month on WU, Feb 2008, and noticed a similar curiosity; a speck of temperature and dew point data for one day:
But like January 2008, the other values for other sensors were intact and plotted for the whole month:
This is odd, especially for an airport where aviation safety is of prime importance. I just couldn’t imagine they’d leave a faulty sensor in place for two months.
When I switched the Weather Underground page to display days, rather than the month summary, I was surprised to find that there was apparently no faulty temperature sensor at all, and that the temperature data and METAR reports were fully intact. Here’s January 2nd, 2008 from Weather Underground, which showed up as having missing temperature in the monthly WU report for January, but as you can see there’s daily data:
But like we saw on the monthly presentation, the temperature data was not plotted for that day, but the other sensors were:
I did spot checks of other dates in January and February of 2008, and found the same thing: the daily METAR reports were there, but the data was not plotted on graphs in Weather Underground.The Nuuk data and plots for the next few months at Weather Underground have similar problems, as you can see here:
But they gradually get better. Strange. It acts like a sensor malfunction, but the METAR data is there for those months and seems reasonably correct.
Since WU makes these web page reports “on the fly” from stored METAR reports in a database, to me this implies some sort of data formatting problem that prevents the graph from plotting it. It also prevents the WU monthly summary from displaying the correct monthly high, low, and average temperatures. Clearly what they have for January 2008 is wrong as I found many temperatures lower than the monthly minimum of 19 °F they report for January 2008, for example, 8°F on January 17th, 2008.
So what’s going on here?
- There’s no sensor failure.
- We have intact hourly METAR reports (the World Meteorological Organization standard for reporting hourly weather data for January and February 2008.
- We have erroneous/incomplete presentations of monthly data both on Weather Underground and NASA GISS for the two months of Jan Feb 2008 I examined.
What could be the cause?
WUWT readers may recall these stories where I example the impacts of failure of the METAR reporting system:
GISS & METAR – dial “M” for missing minus signs: it’s worse than we thought
Dial “M” for mangled – Wikipedia and Environment Canada caught with temperature data errors.
These had to do with missing “M’s” (for minus temperatures) in the coded reports, causing cold temperatures like -25°C becoming warm temperatures of +25°C, which can really screw up monthly average temperatures with one single bad report.
Following my hunch that I’m seeing another variation of the same METAR coding problem, I decided to have a look at that patch of graphed data that appeared on WU on January 19th-20th 2008 to see what was different about it compared to the rest of the month.
I looked at the METAR data for formatting issues, and ran samples of the data from the times it plotted correctly on WU graphs, versus the times it did not. I ran the METAR reports through two different online METAR decoders:
http://www.wx-now.com/Weather/MetarDecode.aspx
http://heras-gilsanz.com/manuel/METAR-Decoder.html
Nothing stood out from the tests with the decoder I did. The only thing that I can see is that some of the METAR reports seem to have extra characters, like /// or 9999, like these samples, the first one didn’t plot data on WU, but the second one did an hour later on January 19th:
METAR BGGH 191950Z VRB05G28KT 2000 -SN DRSN SCT014 BKN018 BKN024 M01/M04 Q0989 METAR BGGH 192050Z 10007KT 050V190 9999 SCT040 BKN053 BKN060 M00/M06 Q0988
I ran both of these (and many others from other days in Jan/Feb) through decoders, and they decoded correctly. However, that still leaves the question of why Weather Underground’s METAR decoder for graph plotting isn’t decoding them correctly for most of Jan/Feb 2008, and leaves the broader question of why GISS data is missing for these months too.
Now here’s the really interesting part.
We have missing data in Weather Underground and in GISS, for January and February of 2008, but in the case of GISS, they use the CLIMAT reports (yes, no ‘e”) to gather GHCN data for inclusion into GISS, and final collation into their data set for adjustment and public dissemination.
The last time I raised this issue with GISS I was told that the METAR reports didn’t effect GISS at all because they never got numerically connected to the CLIMAT reports. I beg to differ this time.
Here’s where we can look up the CLIMAT reports, at OGIMET:
http://www.ogimet.com/gclimat.phtml.en
Here’s what the one for January 2008 looks like:
Note the Nuuk airport is not listed in January 2008
Here’s the decoded report for the same month, also missing Nuuk airport:
Here’s February 2008, also missing Nuuk, but now with another airport added, Mittarfik:
And finally March 2008, where Nuuk appears, highlighted in yellow:
The -8.1°C value of the CLIMAT report agrees with the Weather Underground report, all the METAR reports are there for March, but the WU plotting program still can’t resolve the METAR report data except on certain days.
I can’t say for certain why two months of CLIMAT data is missing from OGIMET, why the same two months of data is missing from GISS, or why Weather Underground can only graph a few hours of data on each of those months, but I have a pretty good idea of what might be going on. I think the WMO created METAR reporting format itself is at fault.
What is METAR you ask? Well in my opinion, a government invented mess.
When I was a private pilot (which I had to give up due to worsening hearing loss – tower controllers talk like auctioneers on the radio and one day I got the active runway backwards and found myself head-on to traffic. I decided then I was a danger to myself and others.) I learned to read SA reports from airports all over the country. SA reports were manually coded teletype reports sent hourly worldwide so that pilots could know what the weather was in airport destinations. They were also used by the NWS to plot synoptic weather maps. Some readers may remember Alden Weatherfax maps hung up at FAA Flight service stations which were filled with hundreds of plotted airport station SA (surface aviation) reports.
The SA reports were easy to visually decode right off the teletype printout:
From page 115 of the book “Weather” By Paul E. Lehr, R. Will Burnett, Herbert S. Zim, Harry McNaught – click for source image
Note that in the example above, temperature and dewpoint are clearly delineated by slashes. Also, when a minus temperature occurs, such as -10 degrees Fahrenheit, it was reported as “-10″, not with an “M”.
The SA method originated with airmen and teletype machines in the 1920′s and lasted well into the 1990′s. But like anything these days, government stepped in and decided it could do it better. You can thank the United Nations, the French, and the World Meteorological Organization (WMO) for this one. SA reports were finally replaced by METAR in 1996, well into the computer age, even though it was designed in 1968.
From Wikipedia’s section on METAR
METAR reports typically come from airports or permanent weather observation stations. Reports are typically generated once an hour; if conditions change significantly, however, they can be updated in special reports called SPECIs. Some reports are encoded by automated airport weather stations located at airports, military bases, and other sites. Some locations still use augmented observations, which are recorded by digital sensors, encoded via software, and then reviewed by certified weather observers or forecasters prior to being transmitted. Observations may also be taken by trained observers or forecasters who manually observe and encode their observations prior to transmission.
History
The METAR format was introduced 1 January 1968 internationally and has been modified a number of times since. North American countries continued to use a Surface Aviation Observation (SAO) for current weather conditions until 1 June 1996, when this report was replaced with an approved variant of the METAR agreed upon in a 1989 Geneva agreement. The World Meteorological Organization‘s (WMO) publication No. 782 “Aerodrome Reports and Forecasts” contains the base METAR code as adopted by the WMO member countries.[1]
Naming
The name METAR is commonly believed to have its origins in the French phrase message d’observation météorologique pour l’aviation régulière (“Aviation routine weather observation message” or “report”) and would therefore be a contraction of MÉTéorologique Aviation Régulière. The United States Federal Aviation Administration (FAA) lays down the definition in its publication the Aeronautical Information Manual as aviation routine weather report[2] while the international authority for the code form, the WMO, holds the definition to be aerodrome routine meteorological report. The National Oceanic and Atmospheric Administration (part of the United States Department of Commerce) and the United Kingdom‘s Met Office both employ the definition used by the FAA. METAR is also known as Meteorological Terminal Aviation Routine Weather Report or Meteorological Aviation Report.
I’ve always thought METAR coding was a step backwards.
Here is what I think is happening
METAR was designed at a time when teletype machines ran newsrooms and airport control towers worldwide. At that time, the NOAA weatherwire used 5 bit BAUDOT code (rather than 8 bit ASCII) transmitting at 56.8 bits per second on a current loop circuit. METAR was designed with one thing in mind: economy of data transmission.
The variable format created worked great at reducing the number of characters sent over a teletype, but that strength for that technology is now a weakness for today’s technology.
The major weakness with METAR these days is the variable length and variable positioning of the format. If data is missing, the sending operator can just leave out the data field altogether. Humans trained in METAR decoding can interpret this, computer however are only as good as the programming they are endowed with.
Characters that change position or type, fields that shift or that may be there one transmission and not the next, combined with human errors in coding can make for a pretty complex decoding problem. The problem can be so complex, based on permutations of the coding, that it takes some pretty intensive coding to handle all the possibilities.
Of course in 1968, missed characters or fields on a teletype paper report did little more than aggravate somebody trying to read it. In today’s technological world, METAR reports never make it to paper, they get transmitted from computer to computer. Coding on one computer system can easily differ from another, mainly due to things like METAR decoding being a task for an individual programmer, rather than a well tested off the shelf universally accepted format and software package. I’ve seen all sorts of METAR decoding programs, from ancient COBOL, to FORTRAN, LISP, BASIC, PASCAL, and C. Each program was done differently, each was done to a spec written in 1968 for teletype transmission, and each computer may run a different OS, have a program written in a different language, and different programmer. Making all that work to handle the nuances of human coded reports that may contain all sorts of variances and errors, with shifting fields presence and placement, is a tall order.
That being said, NOAA made a free METAR decoder package available many years ago at this URL:
http://www.nws.noaa.gov/oso/metardcd.shtml
That has disappeared now, but a private company is distributing the same package here:
This caveat on the limulus web page should be a red flag to any programmer:
Known Issues
- Horrible function naming.
- Will probably break your program.
- Horrible variable naming.
I’ve never used either package, but I can say this: errors have a way of staying around for years. If there was an error in this code that originated at NWS, it may or may not be fixed in the various custom applications that are based on it.
Clearly Weather Underground has issues with some portion of it’s code that is supposed to plot METAR data, coincidentally, many of the same months where their plotting routine fails, we also have missing CLIMAT reports.
And on this page at the National Center for Atmospheric Research (NCAR/UCAR), we have reports of the METAR package failing in odd ways, discarding reports:
>On Tue, 2 Mar 2004, David Larson wrote: > >> >>Robb, >> >>I've been chasing down a problem that seems to cause perfectly good >>reports to be discarded by the perl metar decoder. There is a comment >>in the 2.4.4 decoder that reads "reports appended together wrongly", the >>code in this area takes the first line as the report to process, and >>discards the next line. >>
Here’s another:
On Fri, 12 Sep 2003, Unidata Support wrote: > > ------- Forwarded Message > > >To: support-decoders@xxxxxxxxxxxxxxxx > >From: David Larson <davidl@xxxxxxxxxxxxxxxxxx> > >Subject: perl metar decoder -- parsing visibility / altimeter wrong > >Organization: UCAR/Unidata > >Keywords: 200309122047.h8CKldLd027998 > > > The decoder seems to mistake the altimeter value for visibility in many > non-US METARs.
So my point is this. METAR is fragile, and at the world’s premier climate research center, it seems to have problems that suggest it doesn’t handle worldwide reports yet it appears to be the backbone for all airport based temperature reports in the world, which get collated to GHCN for climate purposes. I think the people that handle these systems need to reevaluate and test their METAR code.
Even better, let’s dump METAR in favor of a more modern format, that doesn’t have variable fields and variable field placements requiring the code to not only decode the numbers/letters, but also the configuration of the report itself.
In today’s high speed data age, saving a few characters from the holdover of teletype days is a wasted effort.
The broader point is; our reporting system for climate data is a mess of entropy on so many levels.
UPDATE:
The missing data can be found at the homepage of DMI, the Danish Meteorological Institute http://www.dmi.dk/dmi/index/gronland/vejrarkiv-gl.htm “vælg by” means “choose city” choose Nuuk to get the numbers monthly back to january 2000.
Thanks for that. The January 2008 data is available, and plotted below at DMI’s web page.
So now the question is, if the data is available, and a prestigious organization like DMI can decode it, plot it, and create a monthly average for it, why can’t NCDC’s Peterson (who is the curator of GHCN) decode it and present it? Why can’t Gavin at GISS do some QC to find and fix missing data?
Is the answer simply “they don’t care enough?” It sure seems so.
UPDATE: 10/4/10 Dr. Jeff Masters of Weather Underground weighs in. He’s fixed the problem on his end:
Hello Anthony, thanks for bringing this up; this was indeed an error in
our METAR parsing, which was based on old NOAA code:
/* Organization: W/OSO242 – GRAPHICS AND DISPLAY SECTION */
/* Date: 14 Sep 1994 */
/* Language: C/370 */
What was happening was our graphing software was using an older version of
this code, which apparently had a bug in its handling of temperatures
below freezing. The graphs for all of our METAR stations were thus
refusing to plot any temperatures below freezing. We were using a newer
version of the code, which does not have the bug, to generate our hourly
obs tables, which had the correct below freezing temperatures. Other
organizations using the old buggy METAR decoding software may also be
having problems correctly handling below freezing temperatures at METAR
stations. This would not be the case for stations sending data using the
synoptic code; our data plots for Nuuk using the the WMO ID 04250 data,
instead of the BGGH METAR data, were fine. In any case, we’ve fixed the
graphing bug, thanks for the help.
REPLY: Hello Dr. Masters.
It is my pleasure to help, and thank you for responding here. One wonders if similar problems exist in parsing METAR for generation of CLIMAT reports. – Anthony
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.


















Thank you for a very clear, detailed, and informative article.
Enlightening, succinct and alarming (not new by Anthony – in addition to the enlightening, succinct & inspiring posts at WUWT!). I suggest a massive sharing of this post to afford the world’s population with another explicit example of the need for real time actual transparency and honest peer review of papers and “documentation” upon which our futures are being based.
Damn good work Anthony/WUWT!
I’ve enjoyed this one a lot, Anthony. However, please don’t let everyone think that the recent (stupid) discontinuity is the only one for Nuuk. There is a truly real one, occurring at September 1922, of about 1.1 deg C, and it happened in the space of a month or so. It endured at a constant level for several decades. In the period before the witching month another stable period had also endured for decades. You don’t believe this 😉 ! Well I could demonstrate it easily if only I knew how to post graphics (GIFs) to the thread. The step change occurs at all southern Greenland sites, and in Iceland too. If you’d like to see this, and if you wish, many other plots of related discontinuities, for example in well known indices like PDO and others, I could send them by email (what address?). Believe me, it’s not my imagination.
Robin – who’s done a large amount of work on climate data over the last 16 years.
david says:
Anthony: “So, it seems clear, that at least for Nuuk, Greenland, their GISS assigned temperature trend is artificial in the scheme of things. Given that Nuuk is at an airport, and that it has gone through steady growth, the adjustment applied by GISS is in my opinion, inverted.”
Is this aspect more then just programing error?
I don’t know if it’s programming ‘error’ or ‘by design’ but it is a consistent part of GIStemp. I’ve benchmarked the first 2 steps of GIStemp and the data are ‘warmed’ by the process fairly consistently.
This is one of my older postings. Full of long boring tables of numbers, but it lets you see in detail form what GIStemp did to the character of the data for some particular countries. It consistently makes changes of the form that would induce a warming trend.
http://chiefio.wordpress.com/2009/11/05/gistemp-sneak-peek-benchmarks-step1/
Step1 is the “homogenize” step and, IMHO, does most of the ‘screwing up’. It is then followed by STEP2 that does UHI correction – backwards 1/2 the time, and finishes the job. Step3 hides this mess via making the broken grid/box anomalies where it compares “apples to frogs” as the thermometer in a grid/box in the baseline is almost always NOT the one in the box today, so an anomaly comparing them is exactly what again?… BTW, with 8000 boxes (until recently) and about 1200 thermometers in GHCN in 2010 most of the boxes have NO thermometer in them. GIStemp makes it’s anomalies by comparing an old thermometer to a ghost thermometer in many cases. Perhaps most cases. It “makes up” a grid / box value from “nearby” thermometers up to 1200 km away. THEN it makes the grid / box anomaly. Sometimes even ‘ghost to ghost’… Such are the ways of ‘climate science’… When folks bleat that it’s done with anomalies and trot out examples of A thermometer being compared to itself: That is how it OUGHT to be done, but not how it IS done. Most of GIStemp breakage happens using straight temperatures and long before it makes the grid/box anomalies, and those are done ‘apples to frogs’ anyway. (Apples to ghost frogs? Just the nature of the analogy needed to describe it speaks volumes…)
And here we find out the ghost frogs have deformed and missing limbs to begin with…
Is there any graphical representation of flow of climate information, available anywhere?
For example:
Airports -> METAR -> GHCN -> CRU
-> GISS
This kind of incredible analysis alone should be sufficient to debunk the AGW claims.
More awesome work by Mr Watts. I am truly awed by his diligence and tenacity.
Now for my question: How do we turn all of this work into actionable items, agents for change? To whom do we write? How do we format such in-depth analysis for people (politicians I presume) who are used to one-liners and sound bites?
I had an opportunity to attend a conference last week where Robert Glennon of ASU spoke about his new book: “Unquenchable: America’s Water Crisis and What To Do About It”.
It was a very interesting presentation, little focus on climate change per se, but a lot about population growth, water use and responses to it. Haven’t read the book, but if it’s like the presentation, it’s probably a good read.
You might invite him to do a post.
Why is the station data even being used anymore for climate study?
There are so many drawbacks for reaching any valid conclusions with global averages that use station data. The number of devices that are “assumed” to be calibrated and function correctly over a range of temperatures is absurd. Each device matters because each one is used to represent such a vast area.
I strongly argue that satellite data should be the only source of data used in the era that it is available. It is a single (or pair) device that is measuring for all points. That is an inherently more reliable system to collecting useful data.
Perhaps one of the most frustrating things in the study of the climate is the difference in the sources of temperature. A standard should be set and everyone should use that. The difficulty is as usual…. politics.
John Kehr
The Inconvenient Skeptic
Well, this Software Engineering cloud-loving pilot also thinks XML is the way to go.
Awesome analysis, Anthony. The only way I can remember METARS is Rod Machado’s:
Should Tina Walk Vera’s Rabbit Without Checking The Dog’s Appetite
Station Time Wind Visibility RVR Weather Clouds Temp Dewpoint Altimeter
Nice to know GISS is at the forefront of technology and 2010 is the warmest ever…
GISS does not agree with anything but Hansen’s predictions.
Why is it used at all? It’s alarming, like Hansen.
What is really alarming is the false state of temperature it generates, and that derived from faulty data interpretation plus bad siting issues.
What we have in GISS, Nuuk and the METAR system is data reporting that is worse than previously imagined.
Anthony et. al.
GISS is not the only government climate outfit with automated quality control issues. Acording to Canada’s Auditor General, Environment Canada has similar problems as outlined in the following report – http://www2.canada.com/story.html?id=3430304 -.. This was briefly picked up by Canada’s MSM. It didn’t get legs because, I think, the Canadian public said ” Ho Hum, what else is new?”
For quality world climate reports, color Canada blank.
Another outstanding research effort, Anthony. Fascinating analysis, excellent detective work.
I notice that one of the pictures of the airport shows a couple of Dash 7’s parked on the tarmac, in front of the alleged Stevenson screen. The engines put out a fair amount of heat, and the props will push that back at the screen quite effectively.
REPLY: I don’t know for certain that is the weather station there, NCDC doesn’t know the exact location either. that’s why I’m asking readers for help. – Anthony
Further to the Audior Generals report a quote, ” Human quality control of climate data ceased as of April 1, 2008. Automated quality control is essentially non-existent. There is no program in place to prevent erroneous data from entering the national climate archive”.
RSS Monthly for September is out:
http://www.remss.com/data/msu/monthly_time_series/RSS_Monthly_MSU_AMSU_Channel_TLT_Anomalies_Land_and_Ocean_v03_2.txt
More like: “Lies, damned lies, and climate science.”
I have codes to extract data from metar reports as part of my work. yes, they are an utter mess. human readable but difficult to get a computer to do (it’s hard to find EVERY case where the data comes through in a funny format)
Another GISS bites the dust….
Another GISS bites the dust….
And another one’s gone….and another one’s gone….
Excellent sleuth work, Anthony!
-Chris
Norfolk, VA, USA
Instead of adjusting the past downwards, as we see GISS do with this station, the population increase would suggest that if adjustments must be applied, they logically should cool the present.
It always burns me to see 1930s era temperature readings “adjusted downward to account for modern day UHI rather than the other way round. I am reminded of this as Los Angeles last week set a new record 113 F that broke the thermometer. Any UHI adjustment is worse than a guess.
GISS (should I use two ZZs to end the acronym?)
I propose that one fine day when Hansen and all of his other co-conspirators there are sacked…or sacked, arrested and tried (lol), that Anthony be named its new director.
He seems to have a better understanding of proper temperature measurements than the turkeys currently running it.
Now folks don’t take me the wrong way, I know there are plenty of extremely bright people working there no doubt, its just that their leader is a former scientist who is now nothing but an activist.
And this tone, set by Hansen, from the top, shows in the GISS sloppiness at the poles.
-Chris
Norfolk, VA, USA
Thanks, Anthony and Steve, for yet another example of why it’s worse than we thought.
When do you guys sleep?
“geoff says:
October 3, 2010 at 9:38 am
I have codes to extract data from metar reports as part of my work. yes, they are an utter mess. human readable but difficult to get a computer to do (it’s hard to find EVERY case where the data comes through in a funny format)”
That the US did WX one way and the other countries did it another way was not good for international aviation. Thus came METAR, and it was something like a political compromise (including all the defects of both sides, of course).
At the time Richard Collins of “Flying” mag complained that it was no good because too MUCH info had to be transmitted by the then limited connections. Remember that a transponder still only uses numbers 1 to 7 to save 1 bit.
Good catch in tracing down all the possible sources. The chief difficuly and problem lies with GHCN. GISS, starts with the assumption that GHCN data is complete. They account for a few exceptions:
1. They use USHCN
2. The pull antarctice data in.
3. they correct a a couple individual stations when they have been supplied the missing data.
They dont go checking for more data or different data sources. That’s the job of NOAA.
For their annual figure they adopt a rule that says you can calculate an annual figure with missing months. A good check would be to determine if this rule biases the record hi or low. that would be a good job for somebody to pick up and plow through thousands of stations for hundreds of months. Or you can download gisstemp, change the rule and see if it matters. ( since I dump any year that has less than 12 months of data, I’ll suggest that result doesnt change much. )