The January Leading Indicator

GISS Data – Image Credit: Walter Dnes

By Walter Dnes – Edited by Just The Facts

Investopedia defines “Leading Indicator” thusly…

A measurable economic factor that changes before the economy starts to follow a particular pattern or trend. Leading indicators are used to predict changes in the economy, but are not always accurate.

Economics is not the only area where a leading indicator is nice to have. A leading indicator that could predict in February, whether this calendar year’s temperature anomaly will be warmer or colder than the previous calendar year’s anomaly would also be nice to have. I believe that I’ve stumbled across exactly that. Using data from 1979 onwards, the rule goes like so…

  1. If this year’s January anomaly is warmer than last year’s January anomaly, then this year’s annual anomaly will likely be warmer than last year’s annual anomaly.
  2. If this year’s January anomaly is colder than last year’s January anomaly, then this year’s annual anomaly will likely be colder than last year’s annual anomaly.

This is a “qualitative” forecast. It doesn’t forecast a number, but rather a boundary, i.e. greater than or less than a specific number. I don’t have an explanation for why it works. Think of it as the climatological equivalent of “technical analysis”; i.e. event X is usually followed by event Y, leaving to others to figure out the underlying “fundamentals”, i.e. physical theory. I’ve named it the “January Leading Indicator”, abbreviated as “JLI” (which some people will probably pronounce as “July”). The JLI has been tested on the following 6 data sets, GISS, HadCRUT3, HadCRUT4, UAH5.6, RSS and NOAA

In this post I will reference this zipped GISS monthly anomaly text file and this spreadsheet. Note that one of the tabs in the spreadsheet is labelled “documentation”. Please read that tab first if you download the spreadsheet and have any questions about it.

The claim of the JLI would arouse skepticism anywhere, and doubly so in a forum full of skeptics. So let’s first look at one data set, and count the hits and misses manually, to verify the algorithm. The GISS text file has to be reformatted before importing into a spreadsheet, but it is optimal for direct viewing by humans. The data contained within the GISS text file is abstracted below.

Note: GISS numbers are the temperature anomaly, multiplied by 100, and shown as integers. Divide by 100 to get the actual anomaly. E.g. “43” represents an anomaly of 43/100=0.43 Celsius degrees. “7” represents an anomaly of 7/100=0.07 Celsius degrees.

  • The first 2 columns on the left of the GISS text file are year and January anomaly * 100.
  • The column after “Dec” (labelled “J-D”) is the January-December anomaly * 100

The verification process is as follows:

  • Count all the years where the current year’s January anomaly is warmer than the previous year’s January anomaly. Add a 1 in the Counter column for each such year.
  • For each such year, we count all where the year’s annual anomaly is warmer than the previous year’s annual anomaly and add a 1 in the Hit column for each such year.
Jan(current) > Jan(previous) J-D(current) > J-D(previous)
Year Counter Compare Hit Compare Comment
1980 1 25 > 10 1 23 > 12
1981 1 52 > 25 1 28 > 23
1983 1 49 > 4 1 27 > 9
1986 1 25 > 19 1 15 > 8
1987 1 30 > 25 1 29 > 15
1988 1 53 > 30 1 35 > 29
1990 1 35 > 11 1 39 > 24
1991 1 38 > 35 0 38 < 39 Fail
1992 1 42 > 38 0 19 < 38 Fail
1995 1 49 > 27 1 43 > 29
1997 1 31 > 25 1 46 > 33
1998 1 60 > 31 1 62 > 46
2001 1 42 > 23 1 53 > 41
2002 1 72 > 42 1 62 > 53
2003 1 73 > 72 0 61 < 62 Fail
2005 1 69 > 57 1 66 > 52
2007 1 94 > 53 1 63 > 60
2009 1 57 > 23 1 60 > 49
2010 1 66 > 57 1 67 > 60
2013 1 63 > 39 1 61 > 58
Predicted 20 > previous year Actual 17 > previous year

Of 20 candidates flagged (Jan(current) > Jan(previous)), 17 are correct (i.e. J-D(current) > J-D(previous)). That’s 85% accuracy for the qualitative annual anomaly forecast on the GISS data set where the current January is warmer than the previous January.

And now for the years where January is colder than the previous January. The procedure is virtually identical, except that we count all where the year’s annual anomaly is colder than the previous year’s annual anomaly and add a 1 in the Hit column for each such year.

Jan(current) < Jan(previous) J-D(current) < J-D(previous)
Year Counter Compare Hit Compare Comment
1982 1 4 < 52 1 9 < 28
1984 1 26 < 49 1 12 < 27
1985 1 19 < 26 1 8 < 12
1989 1 11 < 53 1 24 < 35
1993 1 34 < 42 0 21 > 19 Fail
1994 1 27 < 34 0 29 > 21 Fail
1996 1 25 < 49 1 33 < 43
1999 1 48 < 60 1 41 < 62
2000 1 23 < 48 1 41 < 41 0.406 < 0.407
2004 1 57 < 73 1 52 < 61
2006 1 53 < 69 1 60 < 66
2008 1 23 < 94 1 49 < 63
2011 1 46 < 66 1 55 < 67
2012 1 39 < 46 0 58 > 55 Fail
Predicted 14 < previous year Actual 11 < previous year

Of 14 candidates flagged (Jan(current) < Jan(previous)), 11 are correct (i.e. J-D(current) < J-D(previous)). That’s 79% accuracy for the qualitative annual anomaly forecast on the GISS data set where the current January is colder than the previous January. Note that the 1999 annual anomaly is 0.407, and the 2000 annual anomaly is 0.406, when calculated to 3 decimal places. The GISS text file only shows 2 (implied) decimal places.

The scatter graph at this head of this article compares the January and annual GISS anomalies for visual reference.

Now for a verification comparison amongst the various data sets, from the spreadsheet referenced above. First, all years during the satellite era, which were forecast to be warmer than the previous year

Data set Had3 Had4 GISS UAH5.6 RSS NOAA
Ann > previous 16 15 17 18 18 15
Jan > previous 19 18 20 21 20 18
Accuracy 0.84 0.83 0.85 0.86 0.90 0.83

Next, all years during the satellite era, which were forecast to be colder than the previous year

Data set Had3 Had4 GISS UAH5.6 RSS NOAA
Ann < previous 11 11 11 11 11 11
Jan < previous 15 16 14 13 14 16
Accuracy 0.73 0.69 0.79 0.85 0.79 0.69

The following are scatter graph comparing the January and annual anomalies for the other 5 data sets:

HadCRUT3

HadCRUT3 Data – Walter Dnes

HadCRUT4

HadCRUT4 Data – Walter Dnes

UAH 5.6

UAH 5.6 Data – Walter Dnes

RSS

RSS Data – Walter Dnes

NOAA

NOAA Data – Walter Dnes

The forecast methodology had problems during the Pinatubo years, 1991 and 1992. And 1993 also had problems, because the algorithm compares with the previous year, in this case Pinatubo-influenced 1992. The breakdowns were…

  • For 1991 all 6 data sets were forecast to be above their 1990 values. The 2 satellite data sets (UAH and RSS) were above their 1990 values, but the 4 surface-based data sets were below their 1990 values
  • For 1992 the 4 surface-based data sets (HadCRUT3, HadCRUT4, GISS, and NCDC/NOAA) were forecast to be above their 1991 values, but were below
  • The 1993 forecast was a total bust. All 6 data sets were forecast to be below their 1992 values, but all finished the year above

In summary, during the 3 years 1991/1992/1993, there were 6*3=18 over/under forecasts, of which 14 were wrong. In plain English, if a Pinatubo-like volcano dumps a lot of sulfur dioxide (SO2) into the stratosphere, the JLI will not be usable for the next 2 or 3 years, i.e.:

“The most significant climate impacts from volcanic injections into the stratosphere come from the conversion of sulfur dioxide to sulfuric acid, which condenses rapidly in the stratosphere to form fine sulfate aerosols. The aerosols increase the reflection of radiation from the Sun back into space, cooling the Earth’s lower atmosphere or troposphere. Several eruptions during the past century have caused a decline in the average temperature at the Earth’s surface of up to half a degree (Fahrenheit scale) for periods of one to three years. The climactic eruption of Mount Pinatubo on June 15, 1991, was one of the largest eruptions of the twentieth century and injected a 20-million ton (metric scale) sulfur dioxide cloud into the stratosphere at an altitude of more than 20 miles. The Pinatubo cloud was the largest sulfur dioxide cloud ever observed in the stratosphere since the beginning of such observations by satellites in 1978. It caused what is believed to be the largest aerosol disturbance of the stratosphere in the twentieth century, though probably smaller than the disturbances from eruptions of Krakatau in 1883 and Tambora in 1815. Consequently, it was a standout in its climate impact and cooled the Earth’s surface for three years following the eruption, by as much as 1.3 degrees at the height of the impact.” USGS

For comparison, here are the scores with the Pinatubo-affected years (1991/1992/1993) removed. First, where the years were forecast to be warmer than the previous year

Data set Had3 Had4 GISS UAH5.6 RSS NOAA
Ann > previous 16 15 17 17 17 15
Jan > previous 17 16 18 20 19 16
Accuracy 0.94 0.94 0.94 0.85 0.89 0.94

And for years where the anomaly was forecast to be below the previous year

Data set Had3 Had4 GISS UAH5.6 RSS NOAA
Ann < previous 11 11 11 10 10 11
Jan < previous 14 15 13 11 12 15
Accuracy 0.79 0.73 0.85 0.91 0.83 0.73

Given the existence of January and annual data values, it’s possible to do linear regressions and even quantitative forecasts for the current calendar year’s annual anomaly. With the slope and y-intercept available, one merely has to wait for the January data to arrive in February and run the basic “y = mx + b” equation. The correlation is approximately 0.79 for the surface data sets, and 0.87 for the satellite data sets, after excluding the Pinatubo-affected years (1991 and 1992).

There will probably be a follow-up article a month from now, when all the January data is in, and forecasts can be made using the JLI. Note that data downloaded in February will be used. NOAA and GISS use a missing-data algorithm which results in minor changes for most monthly anomalies, every month, all the way back to day 1, i.e. January 1880. The monthly changes are generally small, but in borderline cases, the changes may affect rankings and over/under comparisons.

The discovery of the JLI was a fluke based on a hunch. One can only wonder what other connections could be discovered with serious “data-mining” efforts.

0 0 votes
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

165 Comments
Inline Feedbacks
View all comments
Editor
February 1, 2014 9:13 pm

Willis Eschenbach says:
> February 1, 2014 at 8:21 pm
> As to the question you raise above about different
> months, David, there’s not a whole lot of difference. When
> I do the analysis on the whole 132 years of the GISS LOTI
> dataset, I get the following results:
In addition to less coverage back in the past, the people behind GISS have had more opportunity to “adjust” the data from longer ago. See http://wattsupwiththat.com/2011/01/13/tale-of-the-global-warming-tiger/ and http://wattsupwiththat.com/2011/01/16/the-past-is-not-what-it-used-to-be-gw-tiger-tale/ Not to mention that the mid-1940’s warm period has been “disappeared” just like the MWP. I selected 1979-to-2013 in order to be able to do an apples-to-apples comparison between land and satellite data.
> Finally, the author hasn’t adjusted for the fact that the
> data has a trend … and that means that on average,
> both the January to January and the year to year data
> both will have a positive value.
I provided separate numbers for forecasts of warmer years versus forecasts of colder years. The warmer-forecast-years do have higher percentage. That alone should be enough to infer a warming trend.

Editor
February 1, 2014 9:15 pm

Recall from above that the average of all of the months of “leading indicators” in the GISS LOTI data is 0.61. That is to say, 61% of the time, if a given month is warmer (colder) than the same month a year previous, 61% of the time that year’s average (12 months starting with and including the given month) is warmer (colder) than the previous 12 months. Recall also that for January the relevant figure was 70%.
What was missing, and needed, were the details of the “null hypothesis”, which is that this a random event. To see what that value would be, I just finished running the Monte Carlo analysis.
I took the GISS LOTI data, detrended it, and calculated the AR and MA coefficients (0.93 and -0.48, typical values for global temperature datasets). Then I generated random ARIMA proxy temperature datasets with those parameters, and set them to the trend of the GISS dataset … here is a group of the randomly generated proxies:

Just kidding, the bottom left one is the actual GISS LOTI data. I do this when I’m doing a Monte Carlo analysis, to make sure I’m actually comparing apples to apples.
So … what did I find from that? Well, the so-called “leading indicator”, which isn’t leading, agrees with the annual results some 66 percent of the time, with a standard deviation of ± 4%. This means that 95% of the “leading indicator” results for the proxy temperature datasets fell between 58% and 74%.
And this, as I suspected, means that at 70%, the author’s “leading indicator” is not doing any better than random chance … as as such, it is useless as a prognostication device.
w.

Nick Stokes
February 1, 2014 9:43 pm

I’ve put here a table of the correlation coefficients for the six land/ocean indices I deal with. Each is over the whole length of data. The correlations are between each month and the annual (calendar) average. Naturally they improve as you advance in the year; mid-year months are a more representative sample.
I don’t think the result is worthless. As said, Jan is a leading indicator. It has limited use as a predictor.

Editor
February 1, 2014 9:44 pm

Willis Eschenbach says:
> February 1, 2014 at 9:15 pm
> And this, as I suspected, means that at 70%, the author’s
> “leading indicator” is not doing any better than random
> chance … as as such, it is useless as a prognostication
> device.
For the satellite era, the GISS numbers I get are 85% (17 of 20 when forecast warmer than previous year) and 79% (11 of 14 when forecast cooler than previous year). That’s with the Pinatubo years included. The numbers look even better with Pinatubo years eliminated.

February 1, 2014 10:42 pm

Mt Sinabung eruptions, if they continue and strengthen, could make this a bust year and 2015 too. The bright side is that a Sinabung cooling effect could dampen any Summer-Fall El Nino, thus spoiling Trenberth’s hoped-for El Nino warming.

Editor
February 1, 2014 10:49 pm

walterdnes says:
February 1, 2014 at 9:44 pm

For the satellite era, the GISS numbers I get are 85% (17 of 20 when forecast warmer than previous year) and 79% (11 of 14 when forecast cooler than previous year). That’s with the Pinatubo years included. The numbers look even better with Pinatubo years eliminated.

Thanks for the quick response, Walter.
I’m sorry, but that is special pleading. You need to use the full dataset, not just the section that might be favorable to your theory. Yes, if you throw out the data that gives poor results, your results will get stronger and strong … consider what that means. It means nothing.
In any case, using a much shorter subset of the data greatly widens the variations in the results. Recall from my analysis above that the results from using each of the 12 months as a “leading indication” for the year that starts with that month were:

Jan, 70%
Feb, 62%
Mar, 57%
Apr, 61%
May, 56%
Jun, 58%
Jul, 55%
Aug, 65%
Sep, 55%
Oct, 59%
Nov, 70%
Dec, 67%
AVERAGE, 61%
95% CI 50% to 72%

If, on the other hand, we use only the satellite era data we get

Jan, 83%
Feb, 59%
Mar, 62%
Apr, 59%
May, 50%
Jun, 62%
Jul, 44%
Aug, 50%
Sep, 44%
Oct, 47%
Nov, 71%
Dec, 68%
AVERAGE, 58%
95% CI 35% to 81% 

Note that while the average is not much different, the spread is wider. Now, my Monte Carlo analysis for the full GISS LOTI gave me a 95% CI from 54% to 71%. Hang on, let me recalculate the results …
OK. As you’d expect, the average is about the same, but now the confidence interval has widened, to from 43% to 77%.
So once again, the results that you are finding are not at all surprising. Instead, we find them in random pseudo-data. They are a consequence of three things. The first is that the temperature data is highly autocorrelated. The lag-1 autocorrelation term AR of the total GISS LOTI dataset is > 0.9. This means that once a trend is started, it tends to persist … which in turn means that January is more likely to resemble the following year.
The second thing that increases correlation is any multi-year overall trend from any reason. If the temperatures drop for a few years, January will more similar to the yearly average. Given that the world has been warming for the last few centuries …
Finally, you’ve done something which is an absolute no-no in the forecasting world. This is to include the predictor data in the response. Since January is a part of the yearly average, if there were no other factors (no trend, no autocorrelation), we’d expect the January trend to agree with the yearly trend some 53% of the time.
The difficulty is that these three factors conspire together give results from random “red noise” pseudo-data which are indistinguishable from the results we find in the GISS LOTI data.
As a result, you have not been able to falsify the null hypothesis. You have not shown that your results are different from what we find in random red-noise pseudo-data.
Please take this in the supportive sense in which it is offered. You need to learn to do a Monte Carlo analysis in order to see if your results from this (or any other) “indicator” are doing any better than random chance.
Best regards, and thanks for all the work,
w.

February 1, 2014 11:10 pm

What’s the point? Isn’t this just suggesting the existence of multiyear trends, which is already well known.
You can get 75+% success in weather forecasting as well by simply saying “the weather tomorrow will be like the weather today”.

Editor
February 1, 2014 11:15 pm

Joel O’Bryan says:
> February 1, 2014 at 10:42 pm
> Mt Sinabung eruptions, if they continue and strengthen,
> could make this a bust year and 2015 too. The bright side
> is that a Sinabung cooling effect could dampen any
> Summer-Fall El Nino, thus spoiling Trenberth’s hoped-for
> El Nino warming.
Volcano activity is an excuse only when a volcano *EXPLODES VERTICALLY*, pumping a lot of SO2, etc, into the *STRATOSPHERE*, like Pinatubo, Tambora, etc. Spewing lava (Etna, Mauna Loa, etc) causes localized damage, and maybe some forest-fire-equivalant smoke from burning buildings and vegetation, but doesn’t cool the planet noticeably.

Greg Goodman
February 1, 2014 11:57 pm

“The 1993 forecast was a total bust. All 6 data sets were forecast to be below their 1992 values, but all finished the year above”
Which pretty much shows the extent of even largest stratospheric eruption of 20th was very limited in duration. This is what was found by my volcano stack analysis that superimposed six major eruptions and looked at the evolution of the degree.day (growing days) integral.
In these graphs a straight downward slope indicates cooler temps, rather than actual cooling. Flat portions even if lower are where the temp has recovered to pre-eruption levels. In the tropics where the degree.days integral comes back to the same level means that temps actually got warmer and made up for the loss of growth days during the post eruption years. ie climate responded to the loss of solar input in some way and self corrected. Not just to restore temperature to previous levels but to compensate for the lost growth caused by making it warmer for an equivalent period.
http://climategrog.wordpress.com/?attachment_id=285
Follow the links for similar plots of SST and NH , SH comparisons.

Greg Goodman
February 2, 2014 12:10 am

Willis : “I’m sorry, but that is special pleading. You need to use the full dataset, not just the section that might be favorable to your theory. Yes, if you throw out the data that gives poor results, your results will get stronger and strong … consider what that means. It means nothing.”
That would be a reasonable comment if he was arbitrarily removing an “inconvenient ” sections. However, there is a good, accepted, physical reason why that section should have a different behaviour. Removing it is a legitimate step.
If you were investigation the variation of diurnal temperature variation against solar elevation, it would be legitimate to remove a day that had a solar eclipse at 2pm.
Other than that, a lot of this is about auto-correlation as you rightly say.

Editor
February 2, 2014 12:11 am

Willis Eschenbach says:
> February 1, 2014 at 10:49 pm
> If, on the other hand, we use only the satellite era data we get
>
> Jan, 83%
> Feb, 59%
> Mar, 62%
> Apr, 59%
> May, 50%
> Jun, 62%
> Jul, 44%
> Aug, 50%
> Sep, 44%
> Oct, 47%
> Nov, 71%
> Dec, 68%
> AVERAGE, 58%
> 95% CI 35% to 81%
OK, maybe it’s been a freaky/flukey 1/3rd of a century, but it’s nice to know that my calculations agree with yours about GISS having approx 80% correlation for January versus the entire year (I get similar numbers for the other data sets). The 80%+ correlation is the whole point of the article. I don’t have a physical explanation for why that is, versus the lower numbers for other months. I get that you’re saying it could be entirely due to chance. Maybe it is. But I’ll stick my neck out this month and make forecasts. A year from now you may be laughing at me.
As I mentioned previously, I used satellite-era data to enable an apples-to-apples comparison between the surface-based data sets, and the satellite-based data sets.

Greg Goodman
February 2, 2014 12:16 am

wbrozek: “4. According to Bob Tisdale, effects of El Nino or La Nina often show themselves in January so in those cases, it would be obvious why the rest of the year follows.”
That does not explain it , it is just saying the same thing. It is another ‘predicitve’ observation that warm Jan is often followed by warm year. Call it El Ninjo or whatever, when it happens, it’s just giving a name to the same observation.
What this means is that the auto-correlation is longer than AR(1) in monthly data, there would seem to be some annual AR1 as well.

Greg Goodman
February 2, 2014 12:35 am

Willis: “And this, as I suspected, means that at 70%, the author’s “leading indicator” is not doing any better than random chance … as as such, it is useless as a prognostication device.”
What you are doing is a valid attempt at assessing the effect and the need to test a null is a very good point.
However, your data is not “random” . You have constructed pseudo data with a similar statistical structure based on an analysis of the data and find the “predictor” works similarly. This demonstrates that at least a large part of the effect is due to the auto-correlation structure.
This does not mean the predictor is useless, it means it will retain its, rather limited predictive ability as long as the data retains its auto-regressive nature. That is probably a reasonable expectation (and as long as there are no major volcanoes).

Hoser
February 2, 2014 12:53 am

It’s just diagnostic testing. http://en.wikipedia.org/wiki/Likelihood_ratios_in_diagnostic_testing
Use this calculator to see how well the test works. http://www.medcalc.org/calc/diagnostic_test.php
You fill in a 2×2 grid like this:
Cold Year Warm Year
Cold Jan 11 3
Warm Jan 3 17
Results:
a c
b d
Positive Predictive Value (Warm)
a /(a + c ) = 78.57 % (*) 95% CI: 49.21 % to 95.09 %
Negative Predictive Value (Cold)
d /(b + d ) = 85.00 % (*) 95% CI: 62.08 % to 96.62 %
More likely to be right when the test indicates cold..
Nobody needs to get all cranky. Interesting test. But more than likely the FDA would prefer at least 1000 tests. Oops! There goes the share price.

daddylonglegs
February 2, 2014 2:10 am

Nick Stokes on February 1, 2014 at 9:43 pm
I’ve put here a table of the correlation coefficients for the six land/ocean indices I deal with. Each is over the whole length of data. The correlations are between each month and the annual (calendar) average. Naturally they improve as you advance in the year; mid-year months are a more representative sample.
I don’t think the result is worthless. As said, Jan is a leading indicator. It has limited use as a predictor.

Thanks – I was just about to ask what about the other months.

Berényi Péter
February 2, 2014 2:31 am

Is it not possible, that any month could be used as a “leading indicator” for the average temperature of the 12 month period starting with it? Just because things tend to change slowly, perhaps.

lemiere jacques
February 2, 2014 2:39 am

january anomaly is used to calculate yearly anomaly.
let s imagine the anomaly is random on a monthly basis and Zero in average.
if you pick up a year with a given anomaly in january the anomaly of the year will be this anomaly on average on a statistical point of view.
I just want to say autocorrelation.
Less the 40 points in your graph…

February 2, 2014 3:05 am

This is completely unsurprising: there is substantial autocorrelation in the anomalies, so any point measurement is a reasonable predictor of the coming year, barring “funny things” happening, such as major volcanoes. As Nick Stokes points out, measurements in the middle of the year (June/July) would be expected to have the best correlation with the calendar year, and it looks like they do.
The fact that January is a reasonable predictor tells you that the correlation time scale is not “short” compared with one year; the fact that June/July do better tells you that the correlation time scale is not “very long” compared with one year. But we already knew that.

A C Osborn
February 2, 2014 3:17 am

Jonathan Jones says: February 2, 2014 at 3:05 am
“the fact that June/July do better tells you that the correlation time scale is not “very long” compared with one year”
I don’t know who’s data you are looking at, but it certainly isn’t the OP’s.
> Jan, 83%
Versus
> Jun, 62%
> Jul, 44%

February 2, 2014 3:36 am

A C OSborn,
I was using the Nick Stokes numbers at http://www.moyhu.org.s3.amazonaws.com/misc/janlead.txt

son of mulder
February 2, 2014 4:03 am

As January is being used as an indicator and is part of the annual average being predicted and compared to, hence introducing a bias, what happens if you use the preceding December as an indicator ie if Dec 2001 is warmer than Dec 2000 will 2002 be warmer than 2001? This would also have the advantage of “predicting” a year before it starts.

Steve from Rockwood
February 2, 2014 5:11 am

Looking at Willis’ numbers the winter months give higher predictive coefficients than the summer months, suggesting average winter temperatures are a better predictor of annual temperatures. The fact that even the summer temperature predictive coefficients are so high is likely a result of length of time it takes for the trends to turn around (longer trends producing a higher correlation than shorter ones). If you compare the change in coefficients from a min / max point of view the difference between summer and winter becomes even greater (so does the variance).
I liked the article Walter and the comments even more.

Editor
February 2, 2014 5:12 am

I suspect this JTI works because January is the month in which we see how much the Northern Hemisphere has cooled from the previous summer warmth and overall the NH has tended to warm more than the Southern Hemisphere, having therefore a greater effect on the global average anomaly.
GISS Graph of Hemispheric temperature change

Martin 457
February 2, 2014 5:56 am

Layman perspective:
Isn’t this how the climate models failed? Not by ‘flipping coins’, but, ‘rolling loaded dice’.