Digital Signal Processing analysis of global temperature data time series suggests global cooling ahead

This DSP engineer is often tasked with extracting spurious signals from noisy data. He submits this interesting result of applying these techniques to the HadCRUT temperature anomaly data. Digital Signal Processing analysis suggests cooling ahead in the immediate future with no significant probability of a positive anomaly exceeding .5°C between 2023 and 2113. See figures 13 and 14. Code and data is made available for replication. – Anthony

Guest essay by Jeffery S. Patterson, DSP Design Architect, Agilent Technologies

Harmonic Decomposition of the Modern Temperature Anomaly Record

Abstract: The observed temperature anomaly since 1900 can be well modeled with a simple harmonic decomposition of the temperature record based on a fundamental period of 170.7 years. The goodness-of-fit of the resulting model significantly exceeds the expected fit to a stochastic AR sequence matching the general characteristic of the modern temperature record.

Data

I’ve used the monthly Hadcrut3 temperature anomaly data available from http://woodfortrees.org/data/hadcrut3vgl/every as plotted in Figure 1.

clip_image002

Figure 1 – Hadcrut3 Temperature Record 1850-Present

To remove seasonal variations while avoiding spectral smearing and aliasing effects, the data was box-car averaged over a 12-month period and decimated by 12 to obtain the average annual temperature plotted in Figure 2.

clip_image004

Figure 2 – Monthly data decimated to yearly average

A Power Spectral Density (PSD) plot of the decimated data reveals harmonically related spectral peaks.

clip_image006

Figure 3 – PSD of annual temperature anomaly in dB

To eliminate the possibility that these are FFT (Fast Fourier Transform) artifacts while avoiding the spectral leakage associated with data windowing, we use a technique is called record periodization. The data is regressed about a line connecting the record endpoints, dropping the last point in the resulting residual. This process eliminates the endpoint discontinuity while preserving the position of the spectral peaks (although it does extenuate the amplitudes at higher frequencies and modifies the phase of the spectral components). The PSD of the residual is plotted in Figure 4.

clip_image008

Figure 4 – PSD of the periodized record

Since the spectral peaking is still present we conclude these are not record-length artifacts. The peaks are harmonically related, with odd harmonics dominating until the eighth. Since spectral resolution increases with frequency, we use the eighth harmonic of the periodized PSD to estimate the fundamental. The following Mathematica (Mma) code finds the 5th peak (8th harmonic) and estimates the fundamental.

wpkY1=Abs[ArgMax[{psdY,w>.25},w]]/8

0.036811

The units are radian frequency across the Nyquist band, mapped to ±p (the plots are zoomed to 0 < w < 1 to show the area of interest). To convert to years, invert wpkY1 and multiply by 2p, which yields a fundamental period of 170.7 years.

From inspection of the PSD we form the harmonic model (note all of the radian frequencies are harmonically related to the fundamental):

(*Define the 5th order harmonic model used in curve fit*)

model=AY1*Sin[wpkY1 t+phiY1]+AY2*Sin[2*wpkY1* t+phiY2]+AY3*

Sin[3*wpkY1* t+phiY3]+AY4*Sin[4*wpkY1* t+phiY4]+AY5*

Sin[5*wpkY1* t+phiY5]];

vars= {AY1,phiY1,AY2,phiY2,AY3,phiY3,AY4,phiY4,AY5,phiY5 }

and fit the model to the original (unperiodized) data to find the unknown amplitudes, AYx, and phases, phiYx.

fitParms1=FindFit[yearly,model,vars,t]

fit1=Table[model/.fitParms1,{t,0,112}];

residualY1= yearly- fit1;{AY1→-0.328464,phiY1→1.44861,AY2→-

0.194251,phiY2→3.03246,AY3→0.132514,phiY3→2.26587,AY4→0.0624932,

phiY4→-3.42662,AY5→-0.0116186,phiY5→-1.36245,AY8→0.0563983,phiY8→

1.97142,wpkY1→0.036811}

The fit is shown in Figure 5 and the residual error in Figure 6.

clip_image010

Figure 5 – Harmonic model fit to annual data

clip_image012 clip_image014

Figure 6 – Residual Error Figure 7 – PSD of the residual error

The residual is nearly white, as evidenced by Figure 7, justifying use of the Hodric-Prescott filter on the decimated data. This filter is designed to separate cyclical, non-stationary components from data. Figure 8 shows an excellent fit with a smoothing factor of 15.

clip_image016

Figure 8 – Model vs. HP Filtered data (smoothing factor=3)

Stochastic Analysis

The objection that this is simple curve fitting can be rightly raised. After all, harmonic decomposition is a highly constrained form of Fourier analysis, which is itself a curve fitting exercise that yields the harmonic coefficients (where the fundamental is the sample rate) which recreate the sequence exactly in the sample domain. That does not mean however, that any periodicity found by Fourier analysis (or by implication, harmonic decomposition) are not present in the record. Nor, as will be shown below, is it true that harmonic decomposition on an arbitrary sequence would be expected to yield the goodness-of-fit achieved here.

The 113 sample record examined above is not long enough to attribute statistical significance to the fundamental 170.7 year period, although others have found significance in the 57-year (here 56.9 year) third harmonic. We can however, estimate the probability that the results are a statistical fluke.

To do so, we use the data record to estimate an AR process.

procY=ARProcess[{a1,a2,a3,a4,a5},v];

procParamsY = FindProcessParameters[yearlyTD["States"],procY]

estProcY= procY /. procParamsY

WeakStationarity[estProcY]

{a1→0.713,a2→0.0647,a3→0.0629,a4→0.181,a5→0.0845,v→0.0124391}

As can be seen in Figure 9 below, the process estimate yields a reasonable match to observed power spectral density and covariance function.

clip_image018 clip_image020

Figure 9 – PSD of estimated AR process (red) vs. data Figure 9b – Correlation function (model in blue)

clip_image022 clip_image024

Figure 10 – 500 trial spaghetti plot Figure 10b – Three paths chosen at random

As shown in 10b, the AR process produces sequences which in general character match the temperature record. Next we perform a fifth-order harmonic decomposition on all 500 paths, taking the variance of the residual as a goodness-of-fit metric. Of the 500 trials, harmonic decomposition failed to converge 74 times, meaning that no periodicity could be found which reduced the variance of the residual (this alone disproves the hypothesis that any arbitrary AR sequences can be decomposed). To these failed trials we assigned the variance of the original sequence. The scattergram of results are plotted in Figure 11 along with a dashed line representing the variance of the model residual found above.

clip_image026

Figure 11 – Variance of residual; fifth order HC (Harmonic Coefficients), residual 5HC on climate record shown in red

We see that the fifth-order fit to the actual climate record produces an unusually good result. Of the 500 trials, 99.4% resulted in residual variance exceeding that achieved on the actual temperature data. Only 1.8% of the trials came within 10% and 5.2% within 20%. We can estimate the probability of achieving this result by chance by examining the cumulative distribution of the results plotted in Figure 12.

clip_image028

Figure 12 – CDF (Cumulative Distribution Function) of trial variances

The CDF estimates the probability of achieving these results by chance at ~8.1%.

Forecast

Even if we accept the premise of statistical significance, without knowledge of the underlying mechanism producing the periodicity, forecasting becomes a suspect endeavor. If for example, the harmonics are being generated by a stable non-linear climatic response to some celestial cycle, we would expect the model to have skill in forecasting future climate trends. On the other hand, if the periodicities are internally generated by the climate itself (e.g. feedback involving transport delays), we would expect both the fundamental frequency and importantly, the phase of the harmonics to evolve with time making accurate forecasts impossible.

Nevertheless, having come thus far, who could resist a peek into the future?

We assume the periodicity is externally forced and the climate response remains constant. We are interested in modeling the remaining variance so we fit a stochastic model to the residual. Empirically, we found that again, a 5th order AR (autoregressive) process matches the residual well.

tDataY=TemporalData[residualY1-Mean[residualY1],Automatic];

yearTD=TemporalData[residualY1,{ DateRange[{1900},{2012},"Year"]}]

procY=ARProcess[{a1,a2,a3,a4,a5},v];

procParamsY = FindProcessParameters[yearTD["States"],procY]

estProcY= procY /. procParamsY

WeakStationarity[estProcY]

clip_image030 clip_image032

A 100-path, 100-year run combining the paths of the AR model with the harmonic model derived above is shown in Figure 13.

clip_image034

Figure 13 – Projected global mean temperature anomaly (centered 1950-1965 mean)

clip_image036

Figure 14 – Survivability at 10 (Purple), 25 (Orange), 50 (Red), 75 (Blue) and 100 (Green) years

The survivability plots predict no significant probability of a positive anomaly exceeding .5°C between 2023 and 2113.

Discussion

With a roughly one-in-twelve chance that the model obtained above is the manifestation of a statistical fluke, these results are not definitive. They do however show that a reasonable hypothesis for the observed record can be established independent of any significant contribution from greenhouse gases or other anthropogenic effects.

0 0 votes
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

313 Comments
Inline Feedbacks
View all comments
WillR
September 11, 2013 5:11 am

After figure 3:
To eliminate the possibility that these are FFT (Fast Fourier Transform) artifacts while avoiding the spectral leakage associated with data windowing, we use a technique [that] is called record periodization.
Note the word [that] — it appears to be missing…

Nick Stokes
September 11, 2013 5:11 am

How can you establish a 170 year periodicity from 113 years of data?

joshv
September 11, 2013 5:12 am

Though I am extremely skeptical of model projections in general, I see no reason to believe the projections of this model any less than the projections of GCMs.

Eliza
September 11, 2013 5:13 am

Unfortunately Hadcrut etc not reliable and even this would be showing cooling haha. BTW BBC is shouting NH ice lowest volume this PAST winter on its running red tape news please someone advise them.

Eliza
September 11, 2013 5:15 am

Try same with CET(reliable rural station) you will probably get a much more pronounced cooling in the future

Jonathan Abbott
September 11, 2013 5:15 am

Not being a statistician I understand about one term in ten but I think I get the gist. It all looks very impressive but appears to boil down to more curve fitting. At least the author admits to fundamental reasons why the forecast made could be entirely wrong.
Could anyone supply a summary of the the process used in plain English?

September 11, 2013 5:17 am

http://wattsupwiththat.files.wordpress.com/2013/09/clip_image034.png
Can you see? The warming will come back and will haunt us again!! 😀

Gary
September 11, 2013 5:18 am

Evaluating the math and the assumptions are way beyond me, but I would be more likely to accept the likelihood of this exercise being reasonable if a proven example of the method was demonstrated — something in signal processing which is well-understood both theoretically and experimentally.

September 11, 2013 5:22 am

As an electrical engineer, I’ve been wondering about using this approach for sometime. Mr. Patterson has done all the right things, so I don’t doubt the numbers per se. We don’t know is if the forecast portion is valid, but now that we have a model we can at least determine that going forward. What is interesting is the relative simplicity of the final result when compared with the climate models.

September 11, 2013 5:25 am

Jonathan Abbott says: Could anyone supply a summary of the the process used in plain English?
……………………
It is.

September 11, 2013 5:28 am

Nick Stokes said How can you establish a 170 year periodicity from 113 years of data?
Because they lie in the same harmonic series (in a 3/2 ratio). The other harmonic periodicity noted (56.9) was in 3:1 ratio. (That’s what I gathered anyway.)

WillR
September 11, 2013 5:30 am

As Geofff says — it is in plain English. 😉
Plausible? Why not? I am doing similar work with other climate data using a similar technique. You can indeed spot these periodic form with less than the (normally required) full data stream.

Swiss Bob
September 11, 2013 5:36 am

The maths is way over my head but if I understand the gist correctly you are saying that using standard DSP analysis techniques you can state with greater than 90% confidence that global temperature is cyclical and is heading down.
Unlike the IPCC you at least provide the maths!

David L. Hagen
September 11, 2013 5:38 am

Compliments Jeffrey on a thorough naive evaluation of the observed temperature data updating us on the latest techniques, with estimates of the probabilities involved.
For comparison:
Armstrong & Kesten Green evaluate a naive no-change model and several simple models. See Scientific Forecasting eg for Public Policy
The near term projections are similar to a linear rise from the Little Ice Age plus a 50-60 year PDO model. See Syun Akasofu’s work provokes journal resignation
Nicola Scafetta developed an astronomical climate model using harmonic analysis to identify natural cycles plus an anthropogenic contribution. Scafetta recently published the invited review:
Solar and Planetary Oscillation Control on Climate Change: Hind-cast, Forecast and a Comparison with the CMIP5 GCMs
All of these appear to give global temperature predictions with higher near term probability than the current IPCC global climate models.
Jeffrey, It would be interesting to see who well your analysis techniques perform to hindcast/forecast from half the data to the other half.

Mike Lewis
September 11, 2013 5:41 am

If the same method was applied to data just through 1980, would it accurately predict the warming in the 80’s and 90’s?

Editor
September 11, 2013 5:43 am

A few comments:
1) decimated by 12
I know that “decimate” came from the Roman practice of inspiring soldiers after battles where the commanders felt were uninspired. They lined up the soldiers and killed every tenth man, leaving the remaining 90% inspired to win the next battle. At first blush, this sound like torturing the data until it confesses or perhaps this is “dodecimated” – killing every 12th measurement. 🙂
2) The obvious counter is that the relatively recent surge in CO2 levels and (possible) concomitant heating may be messing up any nice periodicities.
3) Would these techniques return similar results to the 2,450 year record described in http://wattsupwiththat.com/2011/12/07/in-china-there-are-no-hockey-sticks/ ?
And a free, bonus rant:
Ever since hurricane Katerina (and weak defenses) devastated New Orleans, new reporters have been saying that and other disasters have “decimated” the area. Bugs the heck out of me. I keep retorting that the reporter is a stupid idiot, but they keep on talking as though I’m not in the same room as said idiot.

johnmarshall
September 11, 2013 5:44 am

A lot of work and for that many thanks. BUT, there is always a but with temperature, the data set used is altered to reflect current theories that temperatures must increase with rising CO2 levels. Raw temperature data is notoriously variable even if all collections are from correctly sited and using the same calibrated instruments there will be temperature differences even a few yards apart. So what is the real temperature of the earth’s surface? we really do not know because temperature can only be accurate if the item being measured is at temperature equilibrium which the earth never is.

Chuck Nolan
September 11, 2013 5:46 am

WillR says:
September 11, 2013 at 5:11 am
After figure 3:
To eliminate the possibility that these are FFT (Fast Fourier Transform) artifacts while avoiding the spectral leakage associated with data windowing, we use a technique [that] is called record periodization.
Note the word [that] — it appears to be missing…
———————————————-
Don’t add words, delete the word “is”.
cn

IanE
September 11, 2013 5:52 am

‘With four parameters I can fit an elephant, and with five I can make him wiggle his trunk.’ John von Neumann.

The Ghost Of Big Jim Cooley
September 11, 2013 5:55 am

Eliza, what the BBC are doing (http://www.bbc.co.uk/news/science-environment-23964372) is to shift to ‘winter’ ice VOLUME. They don’t mention that it has recovered (http://psc.apl.washington.edu/wordpress/wp-content/uploads/schweiger/ice_volume/BPIOMASIceVolumeAnomalyCurrentV2_CY.png). But then this is the BBC now. As Jeremy Paxman said, they gave up the pretence of being impartial some time ago. This is what warmist organisations and individuals are doing now. As we find things not happening as planned, focus will be shifted to areas that somehow impart the feeling that we’re still doomed. It’s only when the global temperature actually starts to properly go into freefall that we will be able to prosecute warmists for their absurdity and ignorance. Just be patient (although none of us welcome cooler world) the Earth will show up the nonsense of AGW without sceptics having to do or say anything.

Editor
September 11, 2013 5:56 am

Nick Stokes says:
September 11, 2013 at 5:11 am
> How can you establish a 170 year periodicity from 113 years of data?
From the post:

The 113 sample record examined above is not long enough to attribute statistical significance to the fundamental 170.7 year period, although others have found significance in the 57-year (here 56.9 year) third harmonic.

So he found a 56.9 year cycle and extended it looking for the fundamental.
Here’s a mental exercise:
Draw a one cycle sine wave, marvel that it’s continuously differentiable and the results are more sinusoids. Get off that tangent.
Trim the ends so that only 113/170 of the cycle remains. Observe that “Hey, it looks like a clipped sine wave.” Apply appropriate techniques to determine the phase, amplitude, and period.
Back in my EE days, we barely had decent enough computers to do DSP, except in the CompSci dept, and I was focused on operating system and ARPAnet code. So I don’t know beans about DSP. Barring bothering to learn anything about it, it would be easy enough to take the digital data for that 113/170 th of a sine wave and convolve with with various test sines looking for the best match. So I’m sure there are mathematical techniques beyong Fourier transforms that can handle those fractional sines.

Sam The First
September 11, 2013 5:57 am

Menahwile in the broadsheet papers here in the UK, notably The Independent and The Guardian which are read by liberal / left opinion formers (inc teachers at all levels), the comments to this article below demonstrate that none of their readers is taking any notice of the long pause in warming and the overall cyclical record.
Will someone with the time and expertise please add a comment or two to explain what is really going on?
http://www.independent.co.uk/news/science/woolly-mammoths-became-extinct-because-of-climate-change–not-hunting-8808684.html

September 11, 2013 6:00 am

Hadcrut 3 is not a balanced sample of weather stations. It may give you some indication of periodicity but with bias.
Just take a random sample, of weather stations, balanced by latitude and 70/30 @sea/inland,
http://blogs.24.com/henryp/2013/02/21/henrys-pool-tables-on-global-warmingcooling/
pay particular attention to maxima, and plot:
http://blogs.24.com/henryp/2012/10/02/best-sine-wave-fit-for-the-drop-in-global-maximum-temperatures/
I don’t understand why people want to do so difficult.

Pamela Gray
September 11, 2013 6:04 am

Comment 1: While data massaging here is extreme, it does demonstrate the futility in ordinary least squares statistics so adored by AGWing scientists.
Comment 2: We know from the outset that the signal is representative of a complicated internal system (regardless of external possibilities) in its own right, with teleconnected yet varying unbalanced short and long term internal oscillations (ie decidedly non sine-waveish and very discordant). Therefore the assumption should always be that the variations are entirely internal, quite resistant to demonstrating affects of small external stimulus, and nearly impossible to predict going forward.

Pamela Gray
September 11, 2013 6:05 am

HenryP, quit calling your sample random. It is not.

1 2 3 13