Riding a Pseudocycle

Guest Post by Willis Eschenbach

Loehle and Scafetta recently posted a piece on decomposing the HadCRUT3 temperature record into a couple of component cycles plus a trend. I disagreed with their analysis on a variety of grounds. In the process, I was reminded of work I had done a few years ago using what is called “Periodicity Analysis” (PDF).

A couple of centuries ago, a gentleman named Fourier showed that any signal could be uniquely decomposed into a number of sine waves with different periods. Fourier analysis has been a mainstay analytical tool since that time. It allows us to detect any underlying regular sinusoidal cycles in a chaotic signal.

Figure 1. Joseph Fourier, looking like the world’s happiest mathematician

While Fourier analysis is very useful, it has a few shortcomings. First, it can only extract sinusoidal signals. Second, although it has good resolution as short timescales, it has poor resolution at the longer timescales. For many kinds of cyclical analysis, I prefer periodicity analysis.

So how does periodicity analysis work? The citation above gives a very technical description of the process, and it’s where I learned how to do periodicity analysis. Let me attempt to give a simpler description, although I recommend the citation for mathematicians.

Periodicity analysis breaks down a signal into cycles, but not sinusoidal cycles. It does so by directly averaging the data itself, so that it shows the actual cycles rather than theoretical cycles.

For example, suppose that we want to find the actual cycle of length two in a given dataset. We can do it by numbering the data points in order, and then dividing them into odd- and even-numbered data points. If we average all of the odd data points, and we average all of the even data, it will give us the average cycle of length two in the data. Here is what we get when we apply that procedure to the HadCRUT3 dataset:

Figure 2. Periodicity in the HadCRUT3 global surface temperature dataset, with a cycle length of 2. The cycle has been extended to be as long as the original dataset.

As you might imagine for a cycle of length 2, it is a simple zigzag. The amplitude is quite small, only plus/minus a hundredth of a degree. So we can conclude that there is only a tiny cycle of length two in the HadCRUT3.

Next, here is the same analysis, but with a cycle length of four. To do the analysis, we number the dataset in order with a cycle of four, i.e. “1, 2, 3, 4, 1, 2, 3, 4, 1, 2, 3, 4 …”

Then we average all the “ones” together, and all of the twos and the threes and the fours. When we plot these out, we see the following pattern:

Figure 3. Periodicity in the HadCRUT3 global surface temperature dataset, with a cycle length of 4. The cycle has been extended to be as long as the original dataset.

As I mentioned above, we are not reducing the dataset to sinusoidal (sine wave shaped) cycles. Instead, we are determining the actual cycles in the dataset. This becomes more evident when we look at say the twenty year cycle:

Figure 4. Periodicity in the HadCRUT3 dataset, with a cycle length of 20. The cycle has been extended to be as long as the original dataset.

Note that the actual 20 year cycle is not sinusoidal. Instead, it rises quite sharply, and then decays slowly.

Now, as you can see from the three examples above, the amplitudes of the various length cycles are quite different. If we set the mean (average) of the original data to zero, we can measure the power in the cyclical underlying signals as the sum of the absolute values of the signal data. It is useful to compare this power value to the total power in the original signal. If we do this at all possible frequencies, we get a graph of the strength of each of the underlying cycles.

For example, suppose we are looking at a simple sine wave with a period of 24 years. Figure 5 shows the sine wave, along with periodicity analysis in blue showing the power in each of the various length cycles:

Figure 5. A sine wave, along with the periodicity analysis of all cycles up to half the length of the dataset.

Looking at Figure 5, we can see one clear difference between Fourier analysis and periodicity analysis — the periodicity analysis shows peaks at 24, 48, and 72 years, while a Fourier analysis of the same data would only show the 24-year cycle. Of course, the apparent 48 and 72 year peaks are merely a result of the 24 year cycle. Note also that the shortest length peak (24 years) is sharper than the longest length (72-year) peak. This is because there are fewer data points to measure and average when we are dealing with longer time spans, so the sharp peaks tend to broaden with increasing cycle length.

To move to a more interesting example relevant to the Loehle/Scafetta paper, consider the barycentric cycle of the sun. The sun rotates around the center of mass of the solar system. As it rotates, it speeds up and slows down because of the varying pull of the planets. What are the underlying cycles?

We can use periodicity analysis to find the cycles that have the most effect on the barycentric velocity. Figure 6 shows the process, step by step:

Figure 6. Periodicity analysis of the annual barycentric velocity data. 

The top row shows the barycentric data on the left, along with the amount of power in cycles of various lengths on the right in blue. The periodicity diagram at the top right shows that the overwhelming majority of the power in the barycentric data comes from a ~20 year cycle. It also demonstrates what we saw above, the spreading of the peaks of the signal at longer time periods because of the decreasing amount of data.

The second row left panel shows the signal that is left once we subtract out the 20-year cycle from the barycentric data. The periodicity diagram on the second row right shows that after we remove the 20-year cycle, the maximum amount of power is in the 83 year cycle. So as before, we remove that 83-year cycle.

Once that is done, the third row right panel shows that there is a clear 19-year cycle (visible as peaks at 19, 38, 57, and 76 years. This cycle may be a result of the fact that the “20-year cycle” is actually slightly less than 20 years). When that 19-year cycle is removed, there is a 13-year cycle visible at 13, 26, 39 years etc. And once that 13-year cycle is removed … well, there’s not much left at all.

The bottom left panel shows the original barycentric data in black, and the reconstruction made by adding just these four cycles of different lengths is shown in blue. As you can see, these four cycles are sufficient to reconstruct the barycentric data quite closely. This shows that we’ve done a valid deconstruction of the original data.

Now, what does all of this have to do with the Loehle/Scafetta paper? Well, two things. First, in the discussion on that thread I had said that I thought that the 60 year cycle that Loehle/Scafetta said was in the barycentric data was very weak. As the analysis above shows, the barycentric data does not have any kind of strong 60-year underlying cycle. Loehle/Scafetta claimed that there were ~ 20-year and ~ 60-year cycles in both the solar barycentric data and the surface temperature data. I find no such 60-year cycle in the barycentric data.

However, that’s not what I set out to investigate. I started all of this because I thought that the analysis of random red-noise datasets might show spurious cycles. So I made up some random red-noise datasets the same length as the HadCRUT3 annual temperature records (158 years), and I checked to see if they contained what look like cycles.

A “red-noise” dataset is one which is “auto-correlated”. In a temperature dataset, auto-correlated means that todays temperature depends in part on yesterday’s temperature. One kind of red-noise data is created by what are called “ARMA” processes. “AR” stands for “auto-regressive”, and “MA” stands for “moving average”. This kind of random noise is very similar observational datasets such as the HadCRUT3 dataset.

So, I made up a couple dozen random ARMA “pseudo-temperature” datasets using the AR and MA values calculated from the HadCRUT3 dataset, and I ran a periodicity analysis on each of the pseudo-temperature datasets to see what kinds of cycles they contained. Figure 6 shows eight of the two dozen random pseudo-temperature datasets in black, along with the corresponding periodicity analysis of the power in various cycles in blue to the right of the graph of the dataset:

Figure 6. Pseudo-temperature datasets (black lines) and their associated periodicity (blue circles). All pseudo-temperature datasets have been detrended.

Note that all of these pseudo-temperature datasets have some kind of apparent underlying cycles, as shown by the peaks in the periodicity analyses in blue on the right. But because they are purely random data, these are only pseudo-cycles, not real underlying cycles. Despite being clearly visible in the data and in the periodicity analyses, the cycles are an artifact of the auto-correlation of the datasets.

So for example random set 1 shows a strong cycle of about 42 years. Random set 6 shows two strong cycles, of about 38 and 65 years. Random set 17 shows a strong ~ 45-year cycle, and a weaker cycle around 20 years or so. We see this same pattern in all eight of the pseudo-temperature datasets, with random set 20 having cycles at 22 and 44 years, and random set 21 having a 60-year cycle and weak smaller cycles.

That is the main problem with the Loehle/Scafetta paper. While they do in fact find cycles in the HadCRUT3 data, the cycles are neither stronger nor more apparent than the cycles in the random datasets above. In other words, there is no indication at all that the HadCRUT3 dataset has any kind of significant multi-decadal cycles.

How do I know that?

Well, one of the datasets shown in Figure 6 above is actually not a random dataset. It is the HadCRUT3 surface temperature dataset itself … and it is indistinguishable from the truly random datasets in terms of its underlying cycles. All of them have visible cycles, it’s true, in some cases strong cycles … but they don’t mean anything.

w.

APPENDIX:

I did the work in the R computer language. Here’s the code, giving the “periods” function which does the periodicity function calculations. I’m not that fluent in R, it’s about the eighth computer language I’ve learned, so it might be kinda klutzy.

#FUNCTIONS

PI=4*atan(1) # value of pi

dsin=function(x) sin(PI*x/180) # sine function for degrees

regb =function(x) {lm(x~c(1:length(x)))[[1]][[1]]} #gives the intercept of the trend line

regm =function(x) {lm(x~c(1:length(x)))[[1]][[2]]} #gives the slope of the trend line

detrend = function(x){ #detrends a line

x-(regm(x)*c(1:length(x))+regb(x))

}

meanbyrow=function(modline,x){ #returns a full length repetition of the underlying cycle means

rep(tapply(x,modline,mean),length.out=length(x))

}

countbyrow=function(modline,x){ #returns a full length repetition of the underlying cycle number of datapoints N

rep(tapply(x,modline,length),length.out=length(x))

}

sdbyrow=function(modline,x){ #returns a full length repetition of the underlying cycle standard deviations

rep(tapply(x,modline,sd),length.out=length(x))

}

normmatrix=function(x) sum(abs(x)) #returns the norm of the dataset, which is proportional to the power in the signal

# Function “periods” (below) is the main function that calculates the percentage of power in each of the cycles. It takes as input the data being analyzed (inputx). It displays the strength of each cycle. It returns a list of the power of the cycles (vals), along with the means (means), numner of datapoints N (count), and standard deviations (sds).

# There’s probably an easier way to do this, I’ve used a brute force method. It’s slow on big datasets

periods=function(inputx,detrendit=TRUE,doplot=TRUE,val_lim=1/2) {

x=inputx

if (detrendit==TRUE) x=detrend(as.vector(inputx))

xlen=length(x)

modmatrix=matrix(NA, xlen,xlen)

modmatrix=matrix(mod((col(modmatrix)-1),row(modmatrix)),xlen,xlen)

countmatrix=aperm(apply(modmatrix,1,countbyrow,x))

meanmatrix=aperm(apply(modmatrix,1,meanbyrow,x))

sdmatrix=aperm(apply(modmatrix,1,sdbyrow,x))

xpower=normmatrix(x)

powerlist=apply(meanmatrix,1,normmatrix)/xpower

plotlist=powerlist[1:(length(powerlist)*val_lim)]

if (doplot) plot(plotlist,ylim=c(0,1),ylab=”% of total power”,xlab=”Cycle Length (yrs)”,col=”blue”)

invisible(list(vals=powerlist,means=meanmatrix,count=countmatrix,sds=sdmatrix))

}

# /////////////////////////// END OF FUNCTIONS

# TEST

# each row in the values returned represents a different period length.

myreturn=periods(c(1,2,1,4,1,2,1,8,1,2,2,4,1,2,1,8,6,5))

myreturn$vals

myreturn$means

myreturn$sds

myreturn$count

#ARIMA pseudotemps

# note that they are standardized to a mean of zero and a standard deviation of 0.2546, which is the standard deviation of the HadCRUT3 dataset.

# each row is a pseudotemperature record

instances=24 # number of records

instlength=158 # length of each record

rand1=matrix(arima.sim(list(order=c(1,0,1), ar=.9673,ma=-.4591),

n=instances*instlength),instlength,instances) #create pseudotemps

pseudotemps =(rand1-mean(rand1))*.2546/sd(rand1)

# Periodicity analysis of simple sine wave

par(mfrow=c(1,2),mai=c(.8,.8,.2,.2)*.8,mgp=c(2,1,0)) # split window

sintest=dsin((0:157)*15)# sine function

plotx=sintest

plot(detrend(plotx)~c(1850:2007),type=”l”,ylab= “24 year sine wave”,xlab=”Year”)

myperiod=periods(plotx)

0 0 votes
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

416 Comments
Inline Feedbacks
View all comments
August 9, 2011 5:40 am

Geoff Sharp says:
August 9, 2011 at 12:26 am
One date also does not convey the depth of the downturn.
Doesn’t matter as I have solar data. I just need to know which of the wiggles you consider grand minima.
If you cannot find fault with the data provided I will have to assume your correlation statements are without supporting evidence.
You have this backwards. It is you trying to convince the world, not me. Provide the data in full 3000BC-3000AD as I originally suggested by annotating my graph or providing the tables.

sky
August 9, 2011 4:07 pm

Willis,
The deterioration of communication and analytic insight in the ongoing discussion that I found upon returning from my trip is dismaying. I’ll look past the personally-directed remarks to sharpen the focus on signal structure.
If the sample acf you compute from the HADCRUT3 series doesn’t show very significant NEGATIVE correlation at multidecadal lags, then obviously the series trend and mean were not properly removed before that computation. This basic blunder introduces a systematic POSITIVE BIAS in the results. Conclusions about signal structure drawn from such grossly faulty estimates are mathematically baseless.
Because you’re an extraordinary guy, Willis, I’ll take extraordinary means to clear the communication channel and attempt to clarify this and other isuuse, whenever I find the free time. Pressing matters will occupy me today.

sky
August 10, 2011 7:51 pm

Willis,
My (unprovoked) comments on WUWT are always about the technical issue, rather than the person. I write formally, with logical thinking foremost. Please note the if-then construction of my statement about HADCRUT3 at multidecadal lags. Failure to properly remove the mean and trend is a frequently made basic blunder, of which I did not accuse you .
Since WUWT is an open forum, rather than a private exchange, I do keep the comprehension of the general reader in mind in choosing my examples. That is the reason I chose random ocean waves as an example of a real, oscillatory process that is aperiodic. I wasn’t lecturing you at all.
Although ACF estimation algorithms differ in results, my choice for HADCRUT3 is #3. But that is NOT an example of red-noise, commonly modeled in discrete time by x(n) = r(1)*x(n-1) + g(n), where r(1) is the value of acf at lag 1 and g(n) is gaussian white noise. Red noise has an exponentially decaying NON-negative acf (aside from sampling fluctuations) characteristic of diffusion-type processes. Once you get into higher-order ARIMA processes with oscillatory acfs , you have real, albeit random, cycles instead of mere wandering. Cycles are not an artifact of autocorrelation , they are its expression.
Because this is an important topic, I want clear up. with your indulgence, other issues as well in the coming days.
Cheers.

sky
August 11, 2011 5:18 pm

Willis,
Can’t really help you with R code, because I’m not adept at it. The signal analysis and time-series modeling software I use is the property of our research and consulting firm and the results often belong to clients. I can try to explain, however, an analytic concept that many Ph.D. scientists in various fields have a difficult time in wrapping their heads around. That concept is the linear superposition of many band-limited random-phase oscillations covering a broad range of frequencies. That’s the structure manifested by most most real-world signals not produced by periodic astronomical forces (such as tides). Think of swell from a distant fetch. The apparent wave height and period varies from wave to wave within a wave group, and there’s can be abrupt transitions of phase btween wave groups. Now, instead of wave periods being confined to generally less than 20sec as in the ocean, imagine entirely separate processes that produce swell-like waves at widely seperated frequency bands to complete the conceptual picture.
To get away from the information limitations of short duration temperature records, we have to turn to proxy data. The GISP2 del18-O isotope data, whose acquisition benefitted greatly from a signal-savvy instrumentation engineer on the team, provides perhaps the best available proxy indication of multidecadal, quasi-centennial and much longer random temperature variations.
Power spectrum analysis of the entire Holocene portion of GISP2 data provides frequency resolution two orders of magnitude greater than is available from most instrument records.
It reveals a rich superposition of fairly narrow-band multidecadal and quasi-centennial oscillations, which form a very complex, unpredictable interference pattern in time. The most powerful component in this range, however, has a pronounced spectral peak at ~62 yrs. The Lohle non-dendro series similarly shows, inter alia, a prominent peak at ~59yrs. These oscillations, which cry out for a physical explanation, are not the impersistent wanderings of red noise or the the oscillations of low order ARIMA processes.
I’ll have time again tomorrow to remark upon the pitfalls of modelling. After a long break, pigskins are flying in the air tonight.
Cheers.

sky
August 12, 2011 5:20 pm

The GISP2 isotope data I refer to is not the core profile data that you link
to, but the bidecadal series available from NCDC paleoclimate link. Simple
plotting of that data shows the complex superposition pattern of MANY
oscillations, that I speak of. To obviate the need for accurate
detrending, always a problematic task, I spectrum analyzed the bidecadal
ROC, which applies an analytically well-known high-pass filtering to the
data. Due to sampling uncertainty and mild non-stationarities, there are
the expected differences, of course, in the power densities of the
bi-sectioned data. The pronounced ~62yr peak in the spectral estimate for
entire Holocene, shows up closer to ~65yrs in the early Holocene portion
and there are other differences as well. A ~44-yr oscillation is more
precisely common to both sections. But that’s what real-world random
signals that are not very narrow-band do: they vary from section to
section. And thats what makes thier predicatiability over long time
horizons virtually nil. Pronounced peaks in the power spectrum are not at
all imcompatible with unpredicatabilty in complex cases.
Since you are championing the idea on this post that real-world temperature variations are explained by red-noise aor ARIMA processes, it should be your burden to provide compelling evidence. My problem is that I have not seen anything that would convince anyone competent in signal analysis that such is the case. I’ve already explained why no code will be forthcoming from me. I’m further handicapped by lack of web-surfing skills. I don’t even know how to provide a link, let alone post a graph for everyone to see. That’s why in a disupte at CA with Koutsoyannis about the adequacy of the red-noise model he champions, I could only post the numerical values of the sample acf for a representative set of vetted USA stations. But I’d be happy to e-mail you any results that I’m free to divulge. Hope that helps mutual understanding.

Joachim Seifert
August 12, 2011 7:44 pm

Dear Willis,
let me send you 2 pages from my book (ISBN 978-3-86805-604-4) to show a historical
comparison of the “Net Solar-irradiation Gain (NSG)” and the “Hadley-staircase, the GMT
(Global Mean Temp). I will sent them separately to you, since someone has to paste the
page into here:
……. 2 graphs here……
Comment: The “Hadley Staircase” -GMT- is the previous HadCRUT2 curve, used
until 2005, from which the staircase form can be more easily detected than from
HadCRUT3.
If you subject HadCRUT3 to your mathematical approach , you are blurring the
staircase form, thus not very helpful. The staircase steps are detected and talked
about in the L&S-paper as 60 year cycles (one step surface plus one step height)!
But, always a but in life, L&S include 2 mistakes 1. The first step (1850-1910) is longer in flat
surface, due to crossing the NSG-line (60 years are one flat surface plus one step height),
thus is longer than 60 years, and the 2. and most important, there cannot be an additional
step upward (as L&S predict) and therefore also no 0.66 C man made additional warming
in the 21.Century, because we reached the NSG top limit and the NSG constitutes
nothing less than the “mysterious ” heat for global warming since the LIA… From year
2000 onwards, the GMT temps will stay flat as “plateau” , as you can also see…..
Both graphs show: We have reached the top of the NSG-cycle [790 year
Earth orbit multicentennial cycle, so far widely unknown, due to the IPCC silence
about it (grounds for my ongoing AR4-error complaint this month). If you like, I will
copy my complaint to you….., if you want….
Furthermore, the Net Solar-irradiation Gain (which has risen since middle of the LIA
by 2.17 W/m^2) will peak at the year 2045 and will decrease from thereon into the
coming LIA, 395 years further ahead……
The booklet calculates transparently all mentioned figures of this natural astronomic
cycle, no simulations, no assumptions, only straightforward calculations,
for everyone, and impossible to refute….
Please give it a thought, if I can get the IPCC into looking at the Earth’s orbit, then
a great leap forward has been made……..
Regards Your JSei.
PS to Geoff: My reply to your question at the end of the L&S blog..

sky
August 13, 2011 2:24 pm

Willis Eschenbach says:
August 12, 2011 at 6:16 pm
I’ll try posting some graphs to illustrate the issue next week, when I return from yet another assignment. In the meantime, I prepared some remarks on the pitfalls of discrete-time modeling. They may help to bridge the gap between what I’m saying and what your numerical simulations show.
The commonly used model in red-noise simulations produces an overlay of
white noise. To obtain pure red noise, as in Gauss-Markov processes, the
last term in the recursive equation I gave should contain the factor (1 –
r(1)) to rein in the overlay of gaussian white noise. The generating
equation then is an exponential filter, with white noise input. After the
initial transient settles down, it will produce a series whose power
density is the circular Cauchy distribution. Even this, however, is not
quite the analytic f^(-2) spectrum of white noise integration in continuous
time that consitutes the red noise in some physical systems.
Success in discrete-time simulations meeting various analytic expectations
depends strongly upon how well the random number generator produces truly
uncorrelated numbers, without any cyclical components. Many library
routines are deficient in that respect, producing rythmic ripples in the
sample acf as seen in Christian P’s first graph.
While band-limited random signals can be obtained by band-passing true
white noise, this requires mastery of digital filter design, involving the
complex-valued discrete shift-operator z. The approach is entirely AR,
without any IMA. The coefficients in a power series of various orders in z
are used to control the bandwidth and roll-off rate. It isn’t just any
high-order AR process that produces narrow-band series. BTW, Burg’s MEM
spectral estimation algorithm determines those coefficients optimally from
a sample acf for any chosen order of analysis.
A simplified approach to emulating narrow-band random signals is to
superimpose a handful of sinusoids with INCOMMENSURABLE periods unevenly
clustered around the target peak frequency. Each sinusoid is assigned a
random phase drawn from a UNIFORM distribution in the range 0 to 2pi, which
remains CONSTANT throughout the simulation. It varies only from realization
to realization in an ensemble of simulations. Assigning smaller amplitudes
to the flanks of the cluster helps emulate roll-off and varying the spread
in the cluster regulates bandwidth.
Hope this helps. Let’s enjoy the weekend.

sky
August 17, 2011 5:50 pm

Let’s see if this link works: http://s1188.photobucket.com/albums/z410/skygram/

sky
August 17, 2011 5:56 pm

Can’t take any more time today to fix the size problem in the Spectral Graphs. Hope to fix that and provide a succinct write up in the next day or two.

sky
August 18, 2011 5:26 pm

I have time today only for a brief note on the power spectrum estimates shown by the above link . They were obtained from the bidecadal GISP2 isotope series of only the first 408 data points, covering the age range from 10-8150BP. This truncation was necessary to avoid a severe non-stationarity in the data from the preceding century. The truncated series is weak-sense stationary and is virtually trendless. The sample acf through lag 50 (1000yrs) of both the data series and of the bidecadal ROC (first difference) series was computed by a robustly unbiased algorithm. The cosine transform was then applied and the resulting raw power density estimates were “hanned.”
This classic estimation procedure was chosen not only because of the flexibility it provides for handling any record length, but because it avoids the substitution of the CIRCULAR acf implicit in psd estimates produced from raw FFT periodograma either by decimation in frequency or in time algorithms. The present estimates, which have ~15 degrees of freedom, show peaks and valleys that are appreciably less sharp than would be obtained by either of the FFT algorithms, and far less sharp than could be obtained using Burg’s algorithm. This constitutes a very conservative approach to identifying signal structure, giving the maximum benefit of doubt to smoothly varying academic models.
Since, at the lowest frequencies, first differencing is a close approximation to continuous differentiation, which has a power transfer proportional to f^2, it’s apparent at a glance at the first graph that the physical process that produced the GISP2 data is NOT the integration of white noise. The latter would produce a sample psd that is FLAT at the lowest frequencies. Instead, we get a sample psd that RISES nearly linearly with frequency and showws significant peaks and valleys at the higher frequencies.
Hope to find more time tomorrow to discuss other salient differences wrt the red-noise model provided by Gauss-Markov proceeses.

sky
August 19, 2011 4:48 pm

Having seen that the GISP2 data is not at all the integrated white noise of
the Wiener process (aka “drunkard’s walk”), let’s examine its resemblance
to the red noise model of Gauss-Markov processes. The comparison is based
on the data sample value of r(1) =.197. This low value tells us mmediately
that the data is very noisy or the signalis very wide-band–or both.
Indeed the red noise model spectrum shown in the second graph differs
little from the flat spectrum of white noise. But the always monotonically
declining red noise model denies any possibility of significant spectral
peaks and valleys, which are amply evident in the GISP2 spectrum at widely
separated frequencies 3, 12, 33, with a double peak at 44 and 47. These
spectral features indicate the presence of band-limited signal components
whose power density rises above the noise.
Because the spectral densities have been normalized by their total
variances, the plotted FRACTIONAL power density values add up to unity in
each case. Thus adding the ordinates from frequency 0 through 8 tells us
that 30% of the total variance of GISP data is due to multi-centennial and
quasi-millennial oscillations; correspondingly the red noise model would
allow only 21% in that range. Frequencies 9-16 show 18% and 30-36 show
12%, while 42-50 shows 14%. Thus the four spectral bands account for 74%
of the GISP2 total variance. While other spectrum analysis schemes may
increase the peakedness and change the peak frequency somewhat, the latter
result should not change materially.
It is not noise alone, but the very wide frequency range covered by various
signal components that produces the low absolute value of r(1) for the
GISP2 data. First differencing narrows the effective signal bandwidth
considerably, producing the sample value r(1) = -0.47 for the ROC series.
Such large NEGATIVE correlation, which is virtually unattainable by differencing red noise, provides further confirmation of oscillatory
signal components in the multi-decadal range. Whatever broad-brush resemblence red noise may have to real-world signals manifesting low r(1) values, that resemblence disappears as r(1)increases to the positive levels often shown by vetted station data.
To be sure, low-order ARIMA models are capable of producing non-monotic spectra and oscillating behavior that qualitatively resembles instrument measurements. But close QUANTITATIVE agreement is the province of high-order AR or random phase models.
I’ll try to wrap up my comments tommorrow.

sky
August 20, 2011 2:25 pm

The spectral signature of GISP2 Holocene data is quite instructive. It
reveals the presence of oscillatory signal components in widely separated
freuency bands from the quasi millennial to the multi-decadal, which are
obscured by considerable noise. These signal components make any concept of
linear trend over quasi-centennial time-scales or less quite chimeral. The
apparent linear trend will oscillate in response to the trans-centennial
and multidecadal oscillations.
What Lohle&Scaffeta did wrong was to idealize the multidecadal oscillations
as STRICTLY periodic cycles and treat the apparent trend as a SECULAR one,
rather than the product of trans-centennial oscillations. This entirely
deterministic model is grossly unrealistic, because it ignores the very
limited predictability of random oscillations.
I’ll check back for any questions next Tuesday.

1 15 16 17