Guest post by Richard Saumarez
Correlations are often used to relate changes in climate variables in order to establish a potential relationship between them. These variables may have been processed in some way before they are correlated, possibly by filtering them. There has been some debate, which may have shed more heat than light, about the validity of combining these processes and whether they interact to make the conclusions drawn from them invalid. The object of this post is to explain the processes of correlating and filtering signals, the relationship between them and show that the results are predictable provided one takes care.
The importance of the Fourier Transform/Series.
Fourier analysis is of central importance in filtering and correlation because it allows efficient computation and gives theoretical insight into how these processes are related. The Fourier Transform is an analytical operation that allows a function that exists between the limits of – and + infinity to be expressed in terms of (complex) frequency and is a continuous function of frequency. However, a Fourier Transform of a real-World signal, which is sampled over a specific length of time – a record -, is not calculable. It can be approximated by a Fourier series, normally calculated through the discrete Fourier Transform algorithm, in which the signal is represented as the sum of a series of sine/cosine waves whose frequencies are an exact multiple of the fundamental frequency (=1./length of the record). Although this may seem to be splitting hairs, the differences between the Fourier Transform and the series are important. Fortunately, many of the relationships for continuous signals, for example a voltage wave form, are applicable to signals that are samples in time, which is the way that a signal is represented in a computer. The essential idea about the Fourier transform is that it takes a signal that is dependent on time, t, and represents it in a different domain, that of complex frequency, w. An operation performed on the time domain signal has an equivalent operation in the frequency domain. It is often simpler, far more efficient, and more informative to take a signal, convert it into the frequency domain, perform an operation in the frequency domain and then convert the result back into the time domain.
Some of these relationships are shown in figure 1.
Figure 1. The relationship between the input and output of a system in the time and frequency domains and their correlation functions. These are related through their (discrete) Fourier Transforms.
If a signal, x(t) passes through a linear system, typically a filter, the output, y(t) can be calculated from the input and the impulse response of the filter h(t), which is, mathematically, its response to an infinite amplitude spike that lasts for an infinitesimal time. The process by which this is calculated is called a convolution, which is often represented as “*”, so that:
y(t)=x(t)*h(t)
Looking at figure 1, this is shown in the blue upper panel. The symbol t, representing time, has a suffix, k, that indicates that this is a sampled signal at t0, t1, t2 …… Arrows represent the Discrete Fourier Transform (DFT) that convert the signal from the time to the frequency domain the inverse transform (DFT-1) that converts the signal from the frequency to the time domain. In the frequency domain, convolution is equivalent to multiplication. We can take a signal and transform it from x(tk) to X(wn). If we know the structure of the filter, we can calculate the DFT, H(wn), of its impulse response. We can write, using the symbol F as the forward transform and F-1 as the inverse transform, the following relationships to get the filter output:
X(w)=F[x(t)]
H(w)=F[h(t)]
Y(w)=X(w)H(w)
y(t)=F-1[Y(w)]
What we are doing is taking a specific frequency component of the input signal, modifying it by the frequency response of the filter to get the output at that frequency. The importance of the relationships shown above is that we can convert the frequency response of a filter, which is how filters are specified, into its effect on a period of a time domain signal, which is usually what we are interested in. These deceptively simple relationships allow the effects of a system on a signal to be calculated interchangeably in the time and frequency domains.
Looking at the lower panel in Figure 1, there is a relationship between the (discrete) Fourier Transform and the correlation functions of the inputs and outputs. The autocorrelation functions, which are the signals correlated with themselves are obtained by multiplying the transform by a modified form, the complex conjugate, written as X(w)*, (see below), which gives the signal power spectrum and taking the inverse transform. The cross correlation function is obtained by multiplying the DFT of the input by the complex conjugate of the output to obtain the cross-power spectrum, Gxy(w) and taking the inverse transform, i.e.:
Gxy(w)=X(w)Y*(w)= X(w)X*(w)H*(w)
Rxy(t)=F-1[Gxy(w)]
Therefore there is an intimate relationship between time domain signals representing the input and output of a system and the correlation functions of those signals. They are related through their (discrete) Fourier Transforms.
We now have to look in greater detail at the DFT, what we mean by a frequency component and what a cross-correlation function represents.
Figure 2 shows reconstruction of a waveform, shown in the bottom trace in bold by adding increasingly higher frequency components of its Fourier series. The black trace is the sum of the harmonics up to that point and the red trace is the cosine wave at each harmonic. It is clear that as the harmonics are summed, its value approaches the true waveform and when all the harmonics are used, the reconstruction is exact.
Up to this point, I have rather simplistically represented a Fourier component as being a cosine wave. If you compare harmonics 8 and 24 in figure 2, the peak of every third oscillation of harmonic 24 coincides with the peak in harmonic 8. In less contrived signals this does not generally occur.
Figure 2 A wave form, shown in bold, is constructed by summing its Fourier components shown in red. The black traces show the sum the number of Fourier components up to that harmonic.
The Importance of Phase
Each Fourier component has two values at each frequency. A sine and cosine waves are generated by a rotation of a point about an origin (Figure 3). If it starts on the y axis, the projection of the point is a sine wave and its projection on the x axis is a cosine wave. When the Fourier coefficients are calculated, the contribution of both a sine and a cosine wave to the signal at that frequency are determined. This gives two values, the amplitude of the cosine wave and its phase. The red point on the circle is at an arbitrary point and so its projection becomes a cosine wave that is shifted by a phase angle, usually written as j. Therefore the Fourier component at each frequency has two components, amplitude and phase and can be regarded as a vector.
Earlier, I glibly said that convolution is performed by multiplying the transform of the input by the transform of the impulse response (this is true since they are complex). This is equivalent to multiplying their amplitudes and adding their phases. In correlation, rather than multiplying X(w) and Y(w), we use Y(-w), the transform represented in negative frequency, the complex conjugate. This is equivalent to multiplying their amplitudes and subtracting their phases. Understanding the phase relationships between signals is essential in correlation[i].
Figure 3 The Fourier component is calculated as a sine and cosine coefficient, which may be converted to amplitude, A, and phase angle j. The DFT decomposes the time domain signal into amplitude and phases at each frequency component. The complex conjugate at is shown in blue.
Signal shape is critically determined by phase. Figure 4 shows two signals, an impulse and a square wave shown in black. I have taken their DFT, randomised the phases, while keeping their amplitudes the same, and then reconstructed the signals, shown in red.
Figure 4. The effect of phase manipulation. The black traces are the raw signal, and the red trace is the signal with a randomised phase spectrum but an identical amplitude spectrum
This demonstrates that phase has very important role in determining signal shape. There is a classical demonstration, which space doesn’t allow here, of taking broad band noise and imposing either the phase spectrum of a deterministic signal, while keeping the amplitude spectrum of the noise unaltered or doing the reverse: imposing the amplitude spectrum of the deterministic signal and keeping the phase of the noise unaltered. The modified spectrum is then inverse-transformed into the time domain. The phase manipulated signal has a high correlation with the deterministic signal, while the amplitude manipulated signal has a random correlation with deterministic signal, so underlining the importance of phase in determining signal shape.
Figure 5 The phase spectra of delayed impulses.
A very important concept is that phase represents delay in a signal. A pure time delay is a linear change in phase with frequency as shown in figure 5. The amplitude of the signal is unaltered, but in the case of a delay, there is increasing negative phase with frequency. However, any system that changes the shape of the input signal as it is passed through it, as is usually the case, will not have a linear phase spectrum. This is a particularly important concept when related to correlation.
We are now in a position to understand the cross-correlation function. Looking at the formula for correlation shown in figure 1, the CCF is:
which can be normalised to give the coefficient:
, where N is the number of points in the record.
This rather formidable looking equation is actually quite straightforward. If k is zero, this is simply the standard formula for calculating the correlation coefficient and x is simply correlated with y. If k is one, the y signal is shifted by one sample and the process is repeated. We repeat this for a wide range of k. Therefore the function rxy is the correlation between signals two signals at different levels of shift, k and this tells one something about the relationship between the input and output.
We have a signal x(t) which has been passed through a physical system, with specific characteristics, which results in an output y(t) and we are trying to deduce the characteristics of the system, h(t). Since, from Figure1, the DFT, Y(w) of the output is the product of the DFTs of the input X(w) and the impulse response, H(w), could we not simply divide the DFT of the output by the DFT of the input to get the response? In principle, we can, providing the data is exact.
However most real world measurements contain noise, which is added to the inputs and outputs, or even worse other deterministic signals, and this renders the process somewhat error prone and the results of such a calculation are shown below (figure 6), illustrating the problem:
Figure 6. Left: Input (black) and output (red) signals for a system. Right: the calculated impulse response with 2.5% full scale amplitude noise added to the input and output (black) compared with the true impulse response (red). The low pass filtered response is shown in green.
This calculation illustrates another very important concept: linear physical systems store and dissipate energy. For example, a first order system, which could be the voltage output of a resistor/capacitor network or the displacement of a mechanical spring/damper system, absorbs energy transients and then releases the energy slowly, resulting in the negative exponential impulse response shown in figure 6. The variables which fully define the first order system are its gain and the time constant. The phase spectrum of the impulse response in distinctly non-linear. Attempts to measure another variable, for example delay, which implies a linear phase response, is doomed to failure because it doesn’t really mean anything. For example, if one is looking at the relationship between CO2 and temperature, this is likely to be a complex process that is not defined by delay alone and therefore the response of the system should be identified rather than a physically meaningless variable.
Noise and Correlation
Correlation techniques are used to reduce the effects of noise in the signal. They depend on the noise being independent of, and uncorrelated with, the underlying signal. As explained above, correlation is performed by shifting the signal in time, multiplying the signal by itself (auto-correlation) or with another signal (cross-correlation), summing the result, and performing this at every possible value of shift.
Figure 7. Broad band noise ( black) with the autocorrelation function( red) superimposed.
In broad band noise, each point is, by definition uncorrelated with its neighbours. Therefore, in the auto-correlation function, when there is no shift, there will be perfect correlation between it and its non-shifted self. For all other values of shift, the correlation is, ideally, zero, as shown in figure 7.
The auto correlation function of a cosine wave is obtained in the same manner. When it is unshifted, there will be a perfect match and the correlation will be 1. When shifted by ¼ of its period, the correlation will be zero, be -1 when shifted by ½ a period and zero when shifted by ¾ of period.
The ACF of a cosine wave is a cosine wave of the same frequency with an amplitude of the square of the original wave. However if there is noise in the signal, the value of the correlation will be reduced.
Figure 8 shows the ACF of broadband noise with two sine wave embedded in it. This indicates recovery of two deterministic signals that have serial correlation and are not correlated with the noise. This is a basis for spectral identification in the presence of noise.
Figure 8 The ACF of a signal containing two sine waves of different frequencies embedded in noise. The ACF (red) is the sum of two cosine waves with the same frequencies.
A very important feature of the ACF is that if destroys phase information. Referring to Figure 1, the DFT of the ACF is X(w) (or Y(w)) multiplied by its complex conjugate, which has the same amplitude and negative phase. Thus when they are multiplied together, the amplitudes are squared and the phases are added together making the resultant phase zero. This is the “power spectrum” and is the ACF is its inverse DFT. Therefore the ACF is composed entirely of cosine waves and is symmetrical about a shift of zero.
However, the cross-correlation function, which is the inverse DFT of the cross-power spectrum contains phase. By multiplying the complex conjugate of the output by the input in the frequency domain, one is extracting the phase difference and the delays at each frequency between the input and the output and the cross-correlation function reflects this relationship. If the power spectrum, e.g.: the DFT, of rxx(t) is Gxx(w) and the cross-power spectrum of rxy(t) is Gxy(w), then:
H(w)=Gxy(w)/Gxx(w)
Figure 9 shows the same data as used in figure 6 to calculate the impulse response and the error is very much reduced because signal correlation is a procedure that separates the signal from noise.
Figure 9. Estimate of the impulse response using the data in figure 6 via cross-correlation (black) and pre-filtering the data (green).
These calculations are based on a single record of the input and output. When available, one uses multiple records and calculates the estimated response E[H(w)] from the averaged power spectra:
E[H(w)]=< Gxy(w)>/<Gxx(w)>
Where < x> means the average of x. This leads to a better estimate of the impulse response. It is possible to average because correlation changes the variable from the time domain to relative shift between the signals so aligning them.
One simple check that can be performed to check that one is getting reasonable results, assuming that one has enough individual records, is to calculate the coherence spectrum. This is effectively the correlation between the input and output at each frequency component in the spectrum. If this is significantly less than 1.0, it is likely that there is another input, which hasn’t been represented, or the system is non-linear.
One of the major problems in applying signal processing methods to climate data is that there is only one, relatively short, record and therefore averaging cannot be applied to improve estimates.
Improving resolution by record segmentation and filtering.
One can improve estimates of the response if one has a model of what the signal represents. If one is dealing with a short term process, in other words the output varies quickly and transiently in response to a short term change in input, one can estimate the length of record that is required to capture that response and hence the frequency range of interest. This enables one to segment the record into shorter sections. Each segment has the same sampling frequency, therefore the highest frequency is preserved. By shortening the length of each record we have thrown away low frequencies because the lowest frequency is 1/(record length). However, we have created more records containing high frequencies, which can be averaged to obtain a better estimate of the response.
The other strategy is filtering. This, again, involves assumptions about the nature of signal. Figure 10 shows the same data as in figures 7 & 8 after low-pass filtering. The ACF is no longer an impulse but is expanded about t=0. However the ACF of the deterministic signal is recovered with higher accuracy.
This can be done here because the signal in question has a very small, low frequency, bandwidth and is not affected by the filtering (figure 11). The effects of the filter are easily calculable. If it has a frequency response of A(w), the input and output signals become X(w)A(w) and Y(w)A(w). The cross correlation spectrum is therefore simply:
Gxy(w)= A(w)X(w)A*(w)Y*(w)=A2(w)X(w)Y*(w)
Figure 10 The ACFs using the same data as in Figure 6. Note that the ACF of noise is no longer an impulse at t=0 and that there has been a considerable improvement in the ACF of the two sine waves as they now represent a higher fraction of the total power in the signal.
A2(w)is the autocorrelationfunction of the filter, which has no phase shift and will not affect the phase of the cross-power spectrum, provided the same filter is used on the input and output. This is because the phase reversal of the complex conjugate of the filter in the output cancels out that applied to the input, so the timing relationships between the input and output will not be affected. Provided the ACF spectrum is in the pass band of the filter, it will be preserved. In figure 9, the estimated impulse responses are shown using filtered (green) and non-filtered data. If one wishes to characterise the response, by assuming it is a first order system (which this is), one can fit an exponential to the data so getting its gain and time constant. The filtered result gives marginally better estimates but one has to design the filter rather carefully, appreciate that filtering modifies the impulse response and correct the results for this.
Thus, it is possible to filter signals and perform correlations, provided that the frequency band being filtered does not overlap to much the system response, as illustrated in figure 11, and one is careful.
Figure 11. The signal spectrum is composed of the true signal and noise spectra. A good filter response (green) will attenuate (grey) some of the noise but preserve the true signal, while a bad filter will modify the true signal spectrum and hence the ACF.
In practice, however there is likely to be an overlap between the noise and signal spectra. If the system response is filtered, the correlation functions will be filtered and become widened and oscillatory. In this case, the results won’t mean much and almost certainly will not be what you think they are! There are more advanced statistical methods of determining which part of the spectra contain deterministic signal but, in the case of climate data, the short length of the modern record and the poor quality of the historical record makes this very difficult.
Degrees of Freedom.
Suppose we have two signals and we want to determine if they have different means. They both have a normal distribution and the same variance. Can we test the difference in means by performing a “Student’s t” test? This will almost certainly be wrong, because in most simple statistical tests, there is an assumption that each observation is independent. In figure 7, the ACF is an impulse and nominally zero elsewhere, showing that each point is independent of each other. If the signal is filtered, the points are no longer independent because we have convolved the signal with the impulse response of the filter, as shown in figure 10. Looking at figure 1, the time domain convolution is given by:
This is similar to the correlation formula, except that the impulse response is reversed in time. It shows that the output at any point is a weighted sum of the inputs that have preceded it and are therefore no longer independent. Therefore in applying statistical tests to signal data, one has to measure the dependence of each sample on others by using the autocorrelation of the signal to calculate the number of independent samples or “degrees of freedom”.
Conclusion.
Correlation and filtering are highly interdependent through a set of mathematical relationships. The application of these principles is often limited because of signal quality and the “art” of signal processing is to try to get the best understanding of a physical system in the light of these constraints. The examples shown here are very simple, giving well defined results but real world signal processing may be messier, require much more statistical characterisation and give results that may be limited statistically by inadequate data.
One always has to ask what is the goal of processing a signal and does this make any sense physically? For example, as discussed earlier, cross correlation is often used interchangeably with “delay and it is only meaningful if one believes that phase response of the system in question has a linear phase response with frequency. If one is estimating something that is not meaningful, additional signal processing will not be helpful.
Rather, if one has a model of the system, one can then calculate the parameters of the model and, having done this, one should look carefully at the model to see if it accounts for the measured signals. Ideally, this should be tested with fresh data if it is available, or one can segment the raw data and use one half to create the model and test it with the remaining data.
Modern scripting programs such as “R” allow one to perform many signal processing calculations very easily. The use of these programs lies in not applying them blindly to data but in deciding how to use them appropriately. Speaking from bitter experience, it is very easy to make mistakes in signal processing and it is difficult to recognise them. These mistakes fall into three categories, programming errors, procedural errors in handling the signals and not understanding the theory as well as one should. While modern scripting languages are robust, and may largely eliminate straight programming errors, they most certainly do not protect one from making the others!
[i] I have used the electrical engineering convention in this post, i.e.: the basis function of the Fourier Transform is a negative complex exponential.
Related articles
- Fourier series before Fourier (johndcook.com)
- Retrotechtacular: The Fourier Series (hackaday.com)
- Sampling Rate (tamino.wordpress.com)
cd: “Can I just add, and perhaps what Dr Foster is alluding to this,…”
No , it was he who introduced the DFT (of the straight time series) in order to “school” me. When asked whether he had a done any test of stationarity he went and hid behind a wall.
He’s a sham.
cd : “The other point to note is that the signal itself may not be stationary. ”
Indeed, and that is why I’d investigated d/dt (the subject was Arctic ice extent) . The std ADF test indicated that rate of change could be regarded as stationary and the time series not.
I guess that may have _something_ to do with Grant Forst going to hide behind his control of what gets seen/deleted on his “Open Mind” pseudo-science fiasco.
Greg
Sorry, I’m sort of joining the conversation late.
I think when looking at time series I was always of the opinion that you remove drift with a spline then work on the residuals. I can see this might be a problem where the dependency is defined in the trend you’re removing (e.g. CO2 vs Temperature).
As for determining whether the signal is stationary – excuse my ignorance but I’m not sure what std ADF is – a wavelet transform, or less efficiently, a moving window DFT (or even better the moving window cosine transform of the autocorrelation – reduced clutter in the spectra) would show if there is possible drift in the signal. Anyways, if you have found that, you could apply an attenuation to both signals – I can see Richard’s head in his hands.
But still, and as I said in my original comment, I don’t think that all this data processing is getting us anywhere. In this field statistical analysis takes priority. You just need look how statistical solutions are proposed for experimental problems. Surely we need to start at observations and check if they are reliable. For example, take a look at the global temperature record. You’re talking about a number of sources of error:
Experimental error:
1) Land use changes
2) Instrumental failure
3) Instrumental changes
Processing error:
4) Choice of interpolation method for gridding
5) Choice of projection system for gridding
7) Summary statistic
And all for a global measure that changes on the order of a few tenths of a degree. Then you add all this processing and analyses. This is nuts. We should be trying to get a handle on the principle sources of error, principally 1-3 and then 4-7 (as listed above).
Synonym for Duration is Length. 🙂 Therefore, “duration of time” and “length of time” are both acceptable phrases, scientific journal or otherwise.
Ah but Tom, domains have different metrics. And in the time domain, length has no meaning. Now let that be an end to it, otherwise the pedant police will close WUWT down ;-).
FrankK says:
April 9, 2013 at 4:51 pm (Edit)
How typical of a grammar nazi … you don’t attack the message, you simply deny it and then claim the messenger is unbalanced … and you mis-spell my name.
w.
KevinK says:
April 9, 2013 at 6:37 pm
I maintain it is not. I can tell you exactly how things are going to unfold in the next several decades.
The global temperature metric is going to continue to trend upward at about 0.6 degC/century, as it has for the past century. At the same time, it is going to exhibit a superimposed cycle of about 60-70 years fundamental period which, in the near term of roughly 2 more decades, is going to cause a gradual decline. If you want to know what is going to happen, take the data in the first plot at roughly 1945, raise it up, and splice it onto the record now.
It is all perfectly natural, and following a long term pattern which was in place long before CO2 in the atmosphere grew significantly in the last century.
For CO2, the rate of change is going to continue to track the temperature anomaly, and so the accumulation will continue to decelerate for the next two decades, which should produce a marked deviation from the anthropogenic CO2 emissions, which will continue accelerating.
And then, perhaps at last when these portents become obvious and undeniable, we can rid ourselves of this Chicken Little silliness once and for all.
Greg Goodman says:
April 9, 2013 at 11:56 pm
“So “will yield a flat spectrum” was a totally speculative remark that you can not back up. “
What??? That’s cracked. It’s completely backed up. It is a flat spectrum. It is trivially true. You can apprehend it at a glance. Look at dCO2/dt and T – they are affinely similar. Ergo, flat spectrum. Fini. You are arguing with me over whether 2 + 2 = 4, when I have plainly laid out two sticks, and another two sticks, and shown you that together they are four.
Greg Goodman says:
April 10, 2013 at 12:18 am
“It also requires that the mean remains constant over time.”
Such processes often have stationary increments, and those can be analyzed using Fourier methods. This is fairly standard in the field.
thisisgettingtiresome says:
April 9, 2013 at 11:38 pm
Indeed, just multiply it by the speed of light, and you’ve got a length. Physicists do this as a matter of course to simplify the notation when dealing with relativistic effects.
peterg says:
April 10, 2013 at 2:12 am
Yes. The whole debate could be settled by people who were familiar with these methods.
RCSaumarez says:
April 10, 2013 at 2:52 am
“Transform of an integrator is 1/s and therefore one would not expect to see a flat cross-correlation function if you are correct.”
Of course not. That is why I specifically advocated using the derivative of CO2 in your correlation analysis.
FrankK says:
April 9, 2013 at 4:51 pm (Edit)
Frank, the use of “length of time” has brought you static for being nitpicky. To make things worse, you are also wrong about length not being a term for quantity of time. It is a perfectly good idiom used for centuries in one form or another. Would you for example reject the usage – “He spoke at great length on the topic.” “It took a very long time.” “Long ago and far away.” “At length, the fruit ripened.” “He lengthened his stay by two weeks”. And there are other idiomatic uses such as: “He went to great lengths to solve the problem”. In this instance we are measuring how much “effort” was made, or how “hard” one worked on the problem.
I heard this kind of stuff from one my less-gifted professors, probably as you did – drop it for your own sake.
Bart says:
April 10, 2013 at 9:47 am
“Such processes often have stationary increments, and those can be analyzed using Fourier methods. This is fairly standard in the field.”
Let me further elucidate. We are not looking for series with a stationary individual representation. We are looking for an input/output relationship which is Lebesque integrable. It is not necessary that the series we are correlating represent stationary processes to apply the Fourier analysis. Only, as-stated, that the I/O be L2. When we seek the impulse response of the system with temperature as input and the derivative of CO2 as output, then the response is L2 – it is all-pass, flat spectrum, with an impulse response of a scaled impulse.
Bart: saysGreg Goodman says:
“So “will yield a flat spectrum” was a totally speculative remark that you can not back up. “
Bart: says “What??? That’s cracked. It’s completely backed up. It is a flat spectrum. It is trivially true. You can apprehend it at a glance. Look at dCO2/dt and T – they are affinely similar. Ergo, flat spectrum. Fini. You are arguing with me over whether 2 + 2 = 4”
So , it’s cracked, trivially ture , affinely similar, flat, Fini etc.
Yet , obvious as you seem to think it is, you are unable to back up your claim that is as simple as 2 + 2 = 4 with a plot of this supposedly “flat cross-correlation function “.
In very rough terms I think d/dt(CO2) is as was shown by Allen MacRae, your _may_ be correct (roughly). Now let’s see it.
No more waffle. You made several categorical statements, even when challenged, now please back it up if you have something specific or admit you were speaking a bit loosely and don’t have anything specific to back up your rather over-assertive comments.
I repeat that I suspect you may be largely, roughly, vaguely correct. I would like to have proof of what you say. (and I don’t count your thice posted and crappy WFT.org running mean) .
You spoke with certainty,several times, about “flat cross-correlation function “, let’s see one.
My guess from the data is that it is far from “flat” and also that you have not produced one so far.
I suspect that you don’t know how but would be delighted to be proved wrong on that point.
Bart says “… then the response is L2 – it is all-pass, flat spectrum, …”
can you demonstrate that rather than just asserting it?
Greg Goodman says:
April 10, 2013 at 10:25 am
“can you demonstrate that rather than just asserting it?”
I am not just asserting it. It is an equivalence relationship. The series are affinely related if and only if the frequency response is constant in magnitude and phase across frequency.
Bart says:
April 10, 2013 at 10:58 am
Greg Goodman says:
April 10, 2013 at 10:25 am
Allow me to preempt further discussion by simply laying out how the argument will progress from here.
Greg: Ah, so I see you can’t show it.
Bart: I did show it. It is an equivalence relationship. The series are affinely related if and only if the frequency response is constant in magnitude and phase across frequency.
Repeat ad infinitum.
@ur momisugly bart.
The differential equation you present is:
dCO2/dt=k(t-t0)
on integration, we get:
CO2 = Integral[k(t-t0)] +constant.
Hence, if you treat the system as a co2 producer/absorber as a function of temperature, the transfer function is 1/s which is of course a hyperbola in amplitude with a 1/-j = j =pi/2 phase shift.
If you feel the need to differentiate the CO2 signal numerically, you are on shifting sands. Differentiation is a filter with frequency response:
D[exp(-jwt] = -jw.exp(-jwt). ie a linear increase in magnitude with frequency and a -pi/2 phase shift. If you apply this filter to an integrator you will end up with balderdash.
If you cannot see that if a linear function of one variable is equal to the derivate of a second variable is equivalent to the second variable being equal to the intergral of the first variable, it would seem that you have some way to go. I suggest to read the last paragraph of my post and reflect.
RCSaumarez says:
April 10, 2013 at 11:21 am
“If you apply this filter to an integrator you will end up with balderdash.”
It is getting very difficult to restrain myself. Do they teach calculus anymore in schools?
If you apply a differentiator to the output of an integrator, you get a unity gain, all pass, zero phase filter. The one is the inverse of the other. That is why, in some quarters – particularly British, the integral is referred to as the “anti-derivative”.
The response of an integrator is 1/s. The response of a differentiator is “s”. Now, follow this calculation closely, because it is oh so hard: (1/s) * s = 1.
Differentiation amplifies at high frequency, which is why noise would kill an ideal, continuous time differentiator. But, the response of a discrete time differentiator is 2*j*sin(w/2), and only goes up to the Nyquist frequency, and so the noise response is bounded. Filtering by, e.g., annual averaging also reduces the amplification of the noise, but the derivative is still well represented within the passband of the averaging filter.
“If you cannot see that if a linear function of one variable is equal to the derivate of a second variable is equivalent to the second variable being equal to the integral of the first variable, it would seem that you have some way to go.”
What does that have to do with anything? Fine, do your correlation on CO2 versus the integral of the temperature series as input. It doesn’t matter. The relationship is the same all-pass transfer function.
Sheesh.
RCS: : “If you feel the need to differentiate the CO2 signal numerically, you are on shifting sands. Differentiation is a filter with frequency response:
D[exp(-jwt] = -jw.exp(-jwt). ie a linear increase in magnitude with frequency and a -pi/2 phase shift.
If you apply this filter to an integrator you will end up with balderdash.”
What Bart refers to as “increaments” and is otherwise known as “first difference” and is also the simple numerical approximation of the d/dt or rate of change, is frequently recommended in ecomometrics as a means of rendering a time series stationary, precisely for this kind of analysis.
This process is indeed a crude high pass filter and this the reason that it can render a non stationary process, roughly stationary.
Now, I’m not saying what is done in econometrics is a reference for what is valid signal processing, but are you saying that such a process will necessarily produce “balderdash” ?
If I followed your article correctly, auto(cross)-correlation functions remove phase anyway, so deriving the power spectrum will still be valid , though it may be necessary to take account of the attenuation.
In case of confusion, the response of a discrete time differentiator is 2*j*sin(w*T/2)/T, where T is the sampling period. I am used to working with systems where this is all normalized.
RCSaumarez:
Did the bad and good filters get switched in Fig. 11?
Bart: I did show it. It is an equivalence relationship. The series are affinely related if and only if the frequency response is constant in magnitude and phase across frequency.
So you have stated a theoretical relationship. So if you can now prove that (T-To) and d/dt (CO2) are “affinely related” we can expect flat cross-correlation, the two being equivalent. So far you have not shown that either is true for the data in question other than a lash up plot on WTF.org.
I think it must be clear to all by this stage that you do not have more than that, which is a shame. Like I said I would like to have proof of such a relationship, without which ” the debate rages on”.
Bart , could you just clarify whether you are the same Bart that did the rather impresive spectral analysis on SSN that was reproduced on Tallbloke’s talkshop?
Greg Goodman says:
April 10, 2013 at 1:00 pm
“So you have stated a theoretical relationship”
Theoretical, perhaps, but not with the connotation of being conjectural. To the degree that the axioms of mathematics hold true, it is a definite relationship.
“So if you can now prove that (T-To) and d/dt (CO2) are “affinely related” we can expect flat cross-correlation, the two being equivalent. So far you have not shown that either is true for the data in question other than a lash up plot on WTF.org.”
The WTF plots show it quite plainly. It matches in every significant detail, with just minor deviations due to measurement error and leakage through the non-causal averaging (WTF automatically adjusts the data for phase lag) of CO2 measurements (we have to introduce a zero at the annual cycle, or it tends to overwhelm, and the 12 month average does that). The dips and curves occur in the same location and at the same magnitude, and the longer term trend is the same. It’s about as good a match as I’ve ever seen in test data in the lab.
You can demand some R^2 figure or other, but such measures are generally useful for teasing out relationships in borderline cases. In such a case as this, where the relationship is blatantly obvious, it amounts to gilding the lily. I am loathe to divert even a minute of time from other tasks to “prove” the blatantly obvious, and would prefer to leave that to others so inclined. It is your choice whether to accept it or not. But, it really is rather blatantly obvious.
Greg Goodman says:
April 10, 2013 at 1:08 pm
Likely enough though, no offense to Roger, I would not generally want to be associated with some of the fringe stuff which he likes to post to stimulate ideas and conversation.
@ur momisuglyBart
Thanks, for your comments.
If I understand you correctly. which some might find difficult, you have differentiated a low pass fileered signal. In which case you have attenuated the deternistic components and amplified the noise. In which case you will end up with a practically random signal.
Thank you, I learnt a certain amount of signal processing theory during my PhD..
@ur momisugly Greg Goodman. I am well aware of the difference between numerical approximation “first differnce” and the true derivative,
I note that you appear to be incapable of integrating a first order, sperable differential equationn
and appreciating its significance in the frequecy domain.
Again, I suggest that you read the last paragraph og my post and reflect.
Bart: “But, it really is rather blatantly obvious.”
OK, so eye-balling the graph is all you’ve got. We could have got there a lot quicker, but at least it’s clear now. Thanks.
“Likely enough though …”
I was asking whether you were to author of this rather good analysis.
http://tallbloke.wordpress.com/2011/07/31/bart-modeling-the-historical-sunspot-record-from-planetary-periods/
You didn’t seem to realise what I was referring to in your reply, so I guess that’s a no. There’s plenty of Bart’s around , I was just wondering, though I was having some difficulty reconciling your posts here with the degree of expertise shown in that article.
@ur momisugly Bart,
Having looked at ypur last comment in detail, you are talking complete rubbish. You appear to be unable to do elementary calculus and your derivation of a the frequency response of a differebtiator is just plain wrong
Frankly, I have seen your idiotic, uneducated posts on other threads and and, as many others woukd wish, a period of silence from you would be desirable while you reflect on your lack of mathematical skill.
RCSaumarez ” I am well aware of the difference between numerical approximation “first differnce” and the true derivative,”
I was not suggesting you weren’t. I appreciated your well informed and well written article. In view of your apparent expertise in this area I was asking a question.
To do the kind of frequency analysis being discussed it is necessary AFAIK, to ensure the data is at least approximately stationary. There are several common techniques. Taking first differences, “detrending” with linear or other low frequency function or windowing. All have their short-comings.
You mention the phase shift introduced by differentiation, however, it would seem from your article that autocorrelation removes phase.
Am I correct therefore in thinking that a valid power density spectrum can be derived by using differencing to make the data stationary, compensating if necessary, for the attenuation of lower frequencies?