A mathematician's response to BEST

Doug Keenan in 2009

Doug Keenan, who readers may remember doggedly pursued and won some tree ring data that Queens University held back, was asked to comment of the BEST papers by the Economist. He posted up the full correspondence, including his critiques. There’s some interesting things in there. Since Dr. Muller and BEST want full transparency, in that interest, I’m making this available here. Start from the bottom up to maintain the timeline. h/t to Bishop Hill

He writes:

The Economist asked me to comment on four research papers from the Berkeley Earth Surface Temperature (BEST) project. The four papers, which have not been published, are as follows.

Below is some of the correspondence that we had. (Note: my comments were written under time pressure, and are unpolished.)

From: D.J. Keenan

To: Richard Muller [BEST Scientific Director]; Charlotte Wickham [BEST Statistical Scientist]

Cc: James Astill; Elizabeth Muller

Sent: 17 October 2011, 17:16

Subject: BEST papers

Attach: Roe_FeedbacksRev_08.pdf; Cowpertwait & Metcalfe, 2009, sect 2-6-3.pdf; EmailtoDKeenan12Aug2011.pdf

Charlotte and Richard,

James Astill, Energy & Environment Editor of The Economist, asked Liz Muller if it would be okay to show me your BEST papers, and Liz agreed. Thus far, I have looked at two of the papers.

  • Decadal Variations in the Global Atmospheric Land Temperatures
  • Influence of Urban Heating on the Global Temperature Land Average Using Rural Sites Identified from MODIS Classifications

Following are some comments on those.

In the first paper, various series are compared and analyzed. The series, however, have sometimes been smoothed via a moving average. Smoothed time series cannot be used in most statistical analyses. For some comments on this, which require only a little statistical background, see these blog posts by Matt Briggs (who is a statistician):

Do not smooth times series, you hockey puck!

Do NOT smooth time series before computing forecast skill

Here is a quote from those (formatting in original).

Unless the data is measured with error, you never, ever, for no reason, under no threat, SMOOTH the series! And if for some bizarre reason you do smooth it, you absolutely on pain of death do NOT use the smoothed series as input for other analyses! If the data is measured with error, you might attempt to model it (which means smooth it) in an attempt to estimate the measurement error, but even in these rare cases you have to have an outside (the learned word is “exogenous”) estimate of that error, that is, one not based on your current data.

If, in a moment of insanity, you do smooth time series data and you do use it as input to other analyses, you dramatically increase the probability of fooling yourself! This is because smoothing induces spurious signals—signals that look real to other analytical methods.

This problem seems to invalidate much of the statistical analysis in your paper.

There is another, larger, problem with your papers. In statistical analyses, an inference is not drawn directly from data. Rather, a statistical model is fit to the data, and inferences are drawn from the model. We sometimes see statements such as “the data are significantly increasing”, but this is loose phrasing. Strictly, data cannot be significantly increasing, only the trend in a statistical model can be.

A statistical model should be plausible on both statistical and scientific grounds. Statistical grounds typically involve comparing the model with other plausible models or comparing the observed values with the corresponding values that are predicted from the model. Discussion of scientific grounds is largely omitted from texts in statistics (because the texts are instructing in statistics), but it is nonetheless crucial that a model be scientifically plausible. If statistical and scientific grounds for a model are not given in an analysis and are not clear from the context, then inferences drawn from the model should be regarded as unfounded.

The statistical model adopted in most analyses of climatic time series is a straight line (usually trending upward) with noise (i.e. residuals) that are AR(1). AR(1) is short for “first-order autoregressive”, which means, roughly, that this year (only) has a direct effect on next year; for example, if this year is extremely cold, then next year will have a tendency to be cooler than average.

That model—a straight line with AR(1) noise—is the model adopted by the IPCC (see AR4: §I.3.A). It is also the model that was adopted by the U.S. Climate Change Science Program (which reports to Congress) in its analysis of “Statistical Issues Regarding Trends”. Etc. An AR(1)-based model has additionally been adopted for several climatic time series other than global surface temperatures. For instance, it has been adopted for the Pacific Decadal Oscillation, studied in your work: see the review paper by Roe [2008], attached.

Although an AR(1)-based model has been widely adopted, it nonetheless has serious problems. The problems are actually so basic that they are discussed in some recent introductory (undergraduate) texts on time series—for example, in Time Series Analysis and Its Applications (third edition, 2011) by R.H. Shumway & D.S. Stoffer (see Example 2.5; set exercises 3.33 and 5.3 elaborate).

In Australia, the government commissioned the Garnaut Review to report on climate change. The Garnaut Review asked specialists in the analysis of time series to analyze the global temperature series. The report from those specialists considered and, like Shumway & Stoffer, effectively rejected the AR(1)-based statistical model. Statistical analysis shows that the model is too simplistic to cope with the complexity in the series of global temperatures.

Additionally, some leading climatologists have strongly argued on scientific grounds that the AR(1)-based model is unrealistic and too simplistic [Foster et al., GRL, 2008].

To summarize, most research on global warming relies on a statistical model that should not be used. This invalidates much of the analysis done on global warming. I published an op-ed piece in the Wall Street Journal to explain these issues, in plain English, this year.

The largest center for global-warming research in the UK is the Hadley Centre. The Hadley Centre employs a statistician, Doug McNeall. After my op-ed piece appeared, Doug McNeall and I had an e-mail discussion about it. A copy of one of his messages is attached. In the message, he states that the statistical model—a straight line with AR(1) noise—is “simply inadequate”. (He still believes that the world is warming, primarily due to computer simulations of the global climate system.)

Although the AR(1)-based model is known to be inadequate, no one knows what statistical model should be used. There have been various papers in the peer-reviewed literature that suggest possible resolutions, but so far no alternative model has found much acceptance.

When I heard about the Berkeley Earth Surface Temperature project, I got the impression that it was going to address the statistical issues. So I was extremely curious to see what statistical model would be adopted. I assumed that strong statistical expertise would be brought to the project, and I was trusting that, at a minimum, there would be a big improvement on the AR(1)-based model. Indeed, I said this in an interview with The Register last June.

BEST did not adopt the AR(1)-based model; nor, however, did it adopt a model that deals with some of the complexity that AR(1) fails to capture. Instead, BEST chose a model that is much more simplistic than even AR(1), a model which allows essentially no structure in the time series. In particular, the model that BEST adopted assumes that this year has no effect on next year. That assumption is clearly invalid on climatological grounds. It is also easily seen to be invalid on statistical grounds. Hence the conclusions of the statistical analysis done by BEST are unfounded.

All this occurred even though understanding the crucial question—what statistical model should be used?—requires only an introductory level of understanding in time series. The question is so basic that it is discussed by the introductory text of Shumway & Stoffer, cited above. Another text that does similarly is Introductory Time Series with R by P.S.P. Cowpertwait & A.V. Metcalfe (2009); a section from that text is attached. (The section argues that, from a statistical perspective, a pure AR(4) model is appropriate for global temperatures.) Neither Shumway & Stoffer nor Cowpertwait & Metcalfe have an agenda on global warming, to my knowledge. Rather, they are just writing introductory texts on time series and giving students practical examples; each text includes the series of global temperatures as one of those examples.

There are also textbooks that are devoted to the statistical analysis of climatic data and that discuss time-series modeling in detail. My bookshelf includes the following.

Climate Time Series Analysis (Mudelsee, 2010)

Statistical Analysis in Climate Research (von Storch & Zwiers, 2003)

Statistical Methods in the Atmospheric Sciences (Wilks, 2005)

Univariate Time Series in Geosciences (Gilgen, 2006)

Considering the second paper, on Urban Heat Islands, the conclusion there is that there has been some urban cooling. That conclusion contradicts over a century of research as well as common experience. It is almost certainly incorrect. And if such an unexpected conclusion is correct, then every feasible effort should be made to show the reader that it must be correct.

I suggest an alternative explanation. First note that the stations that your analysis describes as “very rural” are in fact simply “places that are not dominated by the built environment”. In other words, there might well be, and probably is, substantial urbanization at those stations. Second, note that Roy Spencer has presented evidence that the effects of urbanization on temperature grow logarithmically with population size.

The Global Average Urban Heat Island Effect in 2000 Estimated from Station Temperatures and Population Density Data

Putting those two notes together, we might expect that the UHI effect will be larger at the sites classified as “very rural” than at the sites classified as urban. And that is indeed what your analysis shows. Of course, if this alternative explanation is correct, then we cannot draw any inferences about the size of UHI effects on the average temperature measurements, using the approach taken in your paper.

There are other, smaller, problems with your paper. In particular, the Discussion section states the following.

We observe the opposite of an urban heating effect over the period 1950 to 2010, with a slope of -0.19 ± 0.19 °C/100yr. This is not statistically consistent with prior estimates, but it does verify that the effect is very small….

If the two estimates are not consistent, then they contradict each other. In other words, at least one of them must be wrong. Hence one estimate cannot be used “verify” an inference drawn from the other. This has nothing to do with statistics. It is logic.

Sincerely, Doug

* * * * * * * * * * * *

Douglas J. Keenan

http://www.informath.org


From: Richard Muller

To: James Astill

Cc: Elizabeth Muller

Sent: 17 October 2011, 23:33

Subject: Re: BEST papers

Dear James,

You’ve received a copy of an email that DJ Keenan wrote to me and Charlotte. He raises lots of issues that require addressing, some that reflect misunderstanding, and some of which just reflect disagreements among experts in the field of statistics. Since these issues are bound to arise again and again, we are preparing an FAQ that we will put on our web site.

Keenan states that he had not yet read our long paper on statistical methods. I think if he reads this he is more likely to appreciate the sophistication and care that we took in the analysis. David Brillinger, our chief advisor on statistics, warned us that by avoiding the jargon of statistics, we would mislead statisticians to think we had a naive approach. But we decided to write in a more casual style, specifically to be able to reach the wider world of geophysicists and climate scientists who don’t understand the jargon. Again, if Keenan reads the methods paper, he will have a deeper appreciation of what we have done.

It is also important to recognize that we are not creating a new field of science, but are adding to one that has a long history. In the past I’ve discovered that if you avoid using the methods of the past, the key scientists in the field don’t understand what you have done. As my favorite example, I cite a paper I wrote in which I did data were unevenly spaced in time, so I did a Lomb periodogram; the paper was rejected by referees who argued that I was using an “obscure” approach and should have simply done the traditional interpolation followed by Blackman-Tukey analysis. In the future I did it their way, always being careful however to also do a Lomb analysis to make sure there were no differences.

His initial comment is on the smoothing of data. There are certainly statisticians who vigorously oppose this approach, but there have been top statisticians who support it. Included in that list are David Brillinger, and his mentor, the great John Tukey. Tukey revolutionize the field of data analysis for science and his methods dominate many fields of physical science.

Tukey argued that smoothing was a version of “pre-whitening”, a valuable way to remove from the data behavior that was real but not of primary interest. Another of his methods was sequential analysis, in which the low frequency variations were identified, fit using a maximum likelihood method, and then subtracted from the data using a filter prior to the analysis of the frequencies of interest. He showed that this pre-whitening would lead to a more robust result. This is effectively what we did in the Decadal variations paper. The long time scale changes were not the focus of our study, so we did a maximum-likelihood fit, removed them, and examined the residuals.

Keenan quotes: “If, in a moment of insanity, you do smooth time series data and you do use it as input to other analyses, you dramatically increase the probability of fooling yourself! This is because smoothing induces spurious signals—signals that look real to other analytical methods.” Then he draws a conclusion that does not follow from this quote; he says: “This problem seems to invalidate much of the statistical analysis in your paper.”

He is, of course, being illogical. Just because smoothing can increase the probability of our fooling ourselves doesn’t mean that we did. There is real value to smoothing data, and yes, you have to beware of the traps, but if you are then there is a real advantage to doing that. I wrote about this in detail in my technical book on the subject, “Ice Ages and Astronomical Causes.” Much of this book is devoted to pointing out the traps and pitfalls that others in the field fell into.

Keenan goes on to say, “In statistical analyses, an inference is not drawn directly from data. Rather, a statistical model is fit to the data, and inferences are drawn from the model.” I agree wholeheartedly! He may be confused because we adopted the language of physics and geophysics rather than that of statistics. He goes on to say that “This invalidates much of the analysis done on global warming.” If we are to move ahead, it does no good simply to denigrate most of the previous work. So we do our work with more care, using valid statistical methods, but write our papers in such a way that the prior workers in the field will understand what we say. Our hope, in part, is to advance the methods of the field.

Unfortunately, Keenan’s conclusion is that there has been virtually no valid work in the climate field, that what is needed is a better model, and he does not know what that model should be. He says, “To summarize, most research on global warming relies on a statistical model that should not be used. This invalidates much of the analysis done on global warming. I published an op-ed piece in the Wall Street Journal to explain these issues, in plain English, this year.”

Here is his quote basically concluding that no analysis of global warming is valid under his statistical standards: “Although the AR(1)-based model is known to be inadequate, no one knows what statistical model should be used. There have been various papers in the peer-reviewed literature that suggest possible resolutions, but so far no alternative model has found much acceptance.”

What he is saying is that statistical methods are unable to be used to show that there is global warming or cooling or anything else. That is a very strong conclusion, and it reflects, in my mind, his exaggerated pedantry for statistical methods. He can and will criticize every paper published in the past and the future on the same grounds. We might as well give up in our attempts to evaluate global warming until we find a “model” that Keenan will approve — but he offers no help in doing that.

In fact, a quick survey of his website shows that his list of publications consists almost exclusively of analysis that shows other papers are wrong. I strongly suspect that Keenan would have rejected any model we had used.

He gives some specific complaints. He quotes our paper, where we say, “We observe the opposite of an urban heating effect over the period 1950 to 2010, with a slope of -0.19 ± 0.19 °C/100yr. This is not statistically consistent with prior estimates, but it does verify that the effect is very small….”

He then complains,

If the two estimates are not consistent, then they contradict each other. In other words, at least one of them must be wrong. Hence one estimate cannot be used “verify” an inference drawn from the other. This has nothing to do with statistics. It is logic.

He is misinterpreting our statement. Our conclusion is based on our analysis. We believe it is correct. The fact that it is inconsistent with prior estimates does imply that one is wrong. Of course, we think it is the prior estimates. We do not believe that the prior estimates were more than back-of-the-envelope “guestimates”, and so there is no “statistical” contradiction.

He complains,

Considering the second paper, on Urban Heat Islands, the conclusion there is that there has been some urban cooling. That conclusion contradicts over a century of research as well as common experience. It is almost certainly incorrect. And if such an unexpected conclusion is correct, then every feasible effort should be made to show the reader that it must be correct.

He is drawing a strong a conclusion for an effect that is only significant to one standard deviation! He never would have let us claim that -0.19 ± 0.19 °C/100yr indicates urban cooling. I am surprised that a statistician would argue that such a statistically insignificant effect indicates cooling.

Please be careful whom you share this email with. We are truly interested in winning over the other analysts in the field, and I worry that if they were to read portions of this email out of context that they might interpret it the wrong way.

Rich


From: D.J. Keenan

To: James Astill

Sent: 18 October, 2011 17:53

Subject: Re: BEST papers

James,

On the most crucial point, it seems that Rich and I are in agreement. Here is a quote from his reply.

Keenan goes on to say, “In statistical analyses, an inference is not drawn directly from data. Rather, a statistical model is fit to the data, and inferences are drawn from the model.” I agree wholeheartedly!

And so the question is this: was the statistical model that was adopted for their analysis a reasonable choice? If not, then–since their conclusions are based upon that model–their conclusions must be unfounded.

In fact, the statistical model that they adopted has been rejected by essentially everyone. In particular, it has been rejected by both the IPCC and the CCSP, as cited in my previous message. I know of no work that presents argumentation to support their choice of model: they have just adopted the model without any attempt at justification, which is clearly wrong.

(It has been known for decades that the statistical model that they adopted should not be used. Although the statistical problems with the model were clear, for a long time, no one knew the physical reason. Then in 1976, Klaus Hasselmann published a paper that explained the reason. The paper is famous and has since been cited more than 1000 times.)

We could have a discussion about what statistical model should be adopted. It is certain, though, that the model BEST adopted should be rejected. Ergo, their conclusions are unfounded.

Regarding smoothing, the situation here requires only little statistics to understand. Consider the example given by Matt Briggs at

Do NOT smooth time series before computing forecast skill

We take two series, each entirely random. We compute the correlation of the two series: that will tend to be around 0. Then we smooth each series, and we compute the correlation of the two smoothed series: that will tend to be greater than before. The more we smooth the two series, the greater the correlation. Yet we started out with purely random series. This is not a matter of opinion; it is factual. Yet the BEST work computes the correlation of smoothed series.

The reply uses rhetorical techniques to avoid that, stating “Just because smoothing can increase the probability of our fooling ourselves doesn’t mean that we did”. The statement is true, but it does not rebut the above point.

Considering the UHI paper, my message included the following.

There are other, smaller, problems with your paper. In particular, the Discussion section states the following.

We observe the opposite of an urban heating effect over the period 1950 to 2010, with a slope of -0.19 ± 0.19 °C/100yr. This is not statistically consistent with prior estimates, but it does verify that the effect is very small….

If the two estimates are not consistent, then they contradict each other. In other words, at least one of them must be wrong. Hence one estimate cannot be used “verify” an inference drawn from the other. This has nothing to do with statistics. It is logic.

The reply claims “The fact that [their paper’s conclusion] is inconsistent with prior estimates does imply that one is wrong”. The claim is obviously absurd.

The reply also criticizes me for “drawing a strong a conclusion for an effect that is only significant to one standard deviation”. I did not draw that conclusion, their paper suggested it: saying that the effect was “opposite in sign to that expected if the urban heat island effect was adding anomalous warming” and that “natural explanations might require some recent form of “urban cooling””, and then describing possible causes, such as “For example, if an asphalt surface is replaced by concrete, we might expect the solar absorption to decrease, leading to a net cooling effect”.

Note that the reply does not address the alternative explanation that my message proposed for their UHI results. That explanation, which is based on the analysis of Roy Spencer (cited in my message), implies that we cannot draw any inferences about the size of UHI effects on the average temperature measurements, using the approach taken in their paper.

I has a quick look at their Methods paper. It affects none of my criticisms.

Rich also cites his book on the causes of the ice ages. Kindly read my op-ed piece in the Wall Street Journal, and especially consider the discussion of Figures 6 and 7. His book claims to analyze the data in Figure 6: the book’s purpose is to propose a mechanism to explain why the similarity of the two lines is so weak. In fact, to understand the mechanism, it is only necessary to do a simple subtraction–as my piece explains. In short, the analysis is his book is extraordinarily incompetent–and it takes only an understanding of subtraction to see this.

This person who did the data analysis in that book is the person in charge of data analysis at BEST. The data analysis in the BEST papers would not pass in a third-year undergraduate course in statistical time series.

Lastly, a general comment on the surface temperature records might be appropriate. We have satellite records for the last few decades, and they closely agree with the surface records. We also have good evidence that the world was cooler 100-150 years ago than it is today. Primarily for those reasons, I think that the surface temperature records–from NASA, NOAA, Hadley/CRU, and now BEST–are probably roughly right.

Cheers, Doug


From: James Astill

To: D.J. Keenan

Sent: 18 October 2011, 17:57

Subject: Re: BEST papers

Dear Doug

Many thanks. Are you saying that, though you mistrust the BEST methodology to a great degree, you agree with their most important conclusion, re the surface temperature record?

best

James

James Astill

Energy & Environment Editor


From: D.J. Keenan

To: James Astill

Sent: 18 October 2011, 18:41

Subject: Re: BEST papers

James,

Yes, I agree that the BEST surface temperature record is very probably roughly right, at least over the last 120 years or so. This is for the general shape of their curve, not their estimates of uncertainties.

Cheers, Doug


From: D.J. Keenan

To: James Astill

Sent: 20 October, 2011 13:11

Subject: Re: BEST papers

James,

Someone just sent me the BEST press release, and asked for my comments on it. The press release begins with the following statement.

Global warming is real, according to a major study released today. Despite issues raised by climate change skeptics, the Berkeley Earth Surface Temperature study finds reliable evidence of a rise in the average world land temperature of approximately 1°C since the mid-1950s.

The second sentence may be true. The first sentence, however, is not implied by the second sentence, nor does it follow from the analyses in the research papers.

Demonstrating that “global warming is real” requires much more than demonstrating that average world land temperature rose by 1°C since the mid-1950s. As an illustration, the temperature in 2010 was higher than the temperature in 2009, but that on its own does not provide evidence for global warming: the increase in temperatures could obviously be due to random fluctuations. Similarly, the increase in temperatures since the mid 1950s could be due to random fluctuations.

In order to demonstrate that the increase in temperatures since the mid 1950s is not due to random fluctuations, it is necessary to do valid statistical analysis of the temperatures. The BEST team has not done such.

I want to emphasize something. Suppose someone says “2+2=5”. Then it is not merely my opinion that what they have said is wrong; rather, what they have said is wrong. Similarly, it is not merely my opinion that the BEST statistical analysis is seriously invalid; rather, the BEST statistical analysis is seriously invalid.

Cheers, Doug


From: James Astill

To: D.J. Keenan

Sent: 20 October 2011, 13:19

Subject: Re: BEST papers

Dear Doug

Many thanks for all your thoughts on this. It’ll be interesting to see how the BEST papers fare in the review process. Please keep in touch.

best

james

James Astill

Energy & Environment Editor


A story about BEST was published in the October 22nd edition of The Economist. The story, authored by James Astill, makes no mention of the above points. It is subheaded “A new analysis of the temperature record leaves little room for the doubters. The world is warming”. Its opening sentence is “For those who question whether global warming is really happening, it is necessary to believe that the instrumental temperature record is wrong”.


www.informath.org/apprise/a5700.htm  was last updated on 2011-10-21.
Get notified when a new post is published.
Subscribe today!
0 0 votes
Article Rating
114 Comments
Inline Feedbacks
View all comments
Richard Hill
October 21, 2011 9:26 pm

It is puzzling that there is no mention of LTP(Long Term Persistence), Koutsoyannis and the Hurst coefficient in this discussion of climatic time series data. Maybe its there but i couldnt find it.

October 21, 2011 9:35 pm

A point to consider when looking at temperature data beginning in the 1950s. Above, someone made the comment that truly urban areas have little increase in temperatures due to UHI, because they are already built up.
I disagree, for the reason that air conditioning consumes considerable electricity, and all that electricity eventually becomes heat due to Second Law of thermodynamics. The 1950s and onward was the period when air conditioning became more and more prevalent across the western world, especially in urban areas. Urban buildings and residences were converted to air conditioning in that time period (typically the late 1950s and early 1960s).
This was definitely the case in US cities such as Houston, Dallas, San Antonio, and many others. The effect of the air conditioners and the heat from their condensers may not have had much effect on daytime high temperatures, but was probably more noticeable in the evening minimum temperatures.

wayne
October 22, 2011 12:12 am

Doug, you are the first person I have read putting the correct emphasis on Dr. Spencer’s urban denstity study. Your whole article is right on the mark. Thanks.

October 22, 2011 1:21 am

We defined a site as “very-rural” if the MOD500 map showed no urban regions within one tenth of a degree in latitude or longitude of the site.
Chicago’s latitude and longitude are listed as 41-52-55 and 87-37-40. O’Hare International Airport’s latitude and longitude are 41-58-41 and 87-54-28. That would make O’Hare “very rural” by the definition stated.

slow to follow
October 22, 2011 1:38 am

Roger Sowell – IIRR the average power density of urban areas is about four or five times that claimed for CO2 forcing. Might have remembered that wrong – please check.

October 22, 2011 2:17 am

I looked at paper 2 about the positions of thermometers. They talk about being unable to find out exactly where many of the stations were because the positions were given only to a tenth of a degree of Lat. or Long..
When I was navigating deepsea using my sextant, we calculated to seconds of a degree, and all positions were given in degrees, minutes and seconds.
Does no-one even use minutes any more?
Or do they think that 59 degrees 50 minutes ( written as 59.50 perhaps ) means 59 and a half degrees?
Or am I looking for mistakes where none exist?

John Marshall
October 22, 2011 2:33 am

An excellent assessment of the BEST work. The more you read the more one thinks that peer review will fail.

Nick Stokes
October 22, 2011 3:23 am

I looked up the study that Doug Keenan said invalidated the use of AR(1). I didn’t find that stated – they mentioned some variants. But I did notice the abstract, which said:

Are global temperatures on a warming trend? It is difficult to be certain about trends when there is so much variation in the data and very high correlation from year to year. We investigate the question using statistical time series methods. Our analysis shows that the upward movement over the last 130-160 years is persistent and not explained by the high correlation, so it is best described as a trend. The warming trend becomes steeper after the mid-1970s, but there is no significant evidence for a break in trend in the late 1990s. Viewed from the perspective of 30 or 50 years ago, the temperatures recorded in most of the last decade lie above the confidence band of forecasts produced by a model that does not allow for a warming trend.

AlexS
October 22, 2011 3:28 am

“This may be definitional, but showing the average temperature has risen proves a real warming. What it does NOT prove is *anthropogenic* global warming, or *permanent, irreversable* global warming. Gotta watch those adjectives.”
Proves a real warming of that stations(not Earth) and only in this circunstances:
-Only with method currently in use to measure “average”
-That there were no changes in stations.
-If the rise of average temperature is above error.

Editor
October 22, 2011 5:43 am

I have yet to read the BEST papers (I want to give them proper time and consideration), but I’m not surprised by the headlines and news stories. The devil is in the detail.
This assessment on the other hand goes right to the heart of the whole issue, and I am very grateful to Doug for writing it. This is one non-mathematician, who now ‘gets’ it much better.
The email to James Astill suggests it was not meant for Doug Kennan’s eyes – especially:
“Please be careful whom you share this email with. And now it is on a blog for all to see.

phlogiston
October 22, 2011 6:37 am

I am with Richard Muller on the question of smoothing. Keenen’s assertion that smoothing is incompatible with subsequent modeling seems a bit exaggerated.
There have been several shrill proclamations here at WUWT that any kind of smoothing renders time series data completely unanalysable and devoid of any predictive value. This must be nonsense. Richard Muller has provided a valuable service to WUWT by explaining why statistically smoothing does not empty data of any meaning.
I work in micro-tomography where reconstructed images are sometimes noisy and require singificant smoothing to allow real structure to emerge from noise. It would make no sense to assert that such images were devoid of real information following smoothing.
This spurious assertion that smoothing invalidates timeseries data has obstructed the discussion of a number of interesting hypotheses here on WUWT. It should do so no longer.

October 22, 2011 6:47 am

Slightly amending Legatus (October 21, 2011 at 6:35 pm) – thanks!
False Flag
Allie “We are fellow skeptics like you! Watts’ concern is important!!”
Neutralize “Our results show that Watts’ work, though a salutory check, is actually nothing to worry about!”
Destroy “MEDIA MEDIA MEDIA!!! Even skeptics now see that warming has been true and records are trustworthy!”
So, standard Communist tactics all along, eh? Berkeley a Marxist bastion, eh?
This tactic was also used by the Inquisition. Inquisitors worked in pairs, one had the brutal touch, the other had the soft appealing touch. Evidently it’s thought to work.

DirkH
October 22, 2011 8:34 am

phlogiston says:
October 22, 2011 at 6:37 am
“There have been several shrill proclamations here at WUWT that any kind of smoothing renders time series data completely unanalysable and devoid of any predictive value. This must be nonsense.”
On a perfectly normal day the temperature can easily vary between 0 deg C and 20 deg C, right now, here in Germany. That makes the IR backradiation vary by about 25% as it varies with the 4th power of the actual temperature, not the smoothed one.
Do you still think a model that manages to reproduce the *smoothed* time series has any predictive value, or any resemblance to reality? 😉

October 22, 2011 9:20 am

Since this post is about statistics and climate, can anybody help me make sense of the NCDC disparity? (see link below) NCDC reports the temperature trend for the 48 contiguous states is 1.2 degrees F per century. However, the mean of the individual states’ trends is 0.78 degrees F per century, and the area-weighted average for the 48 states is 0.74 degrees F per century.
Should not the trend of the entire 48 states should be very close to the average of each state’s trend?
http://sowellslawblog.blogspot.com/2011/09/us-long-term-temperature-trend-from.html

Dave, UK
October 22, 2011 9:32 am

More “lying by omission,” this time by James Astill. I wonder if he is proud of himself for that.

Mikko Stenlund
October 22, 2011 9:43 am

What exactly makes Keenan a Mathematician?
A Google search and a visit to his website do not point to him having a degree in Mathematics; he certainly is no Professor as someone has falsely claimed here. Second, he has posted links to seven papers on his website, none of which have anything to do with Mathematics.

JPeden
October 22, 2011 10:02 am

phlogiston says:
October 22, 2011 at 6:37 am
This spurious assertion that smoothing invalidates timeseries data has obstructed the discussion of a number of interesting hypotheses here on WUWT. It should do so no longer.
Then get a load of what Best’s “smoothing” seems to have done according to the screen shots linked below and the time-space course presented showing the existence of the surface stations, also alleged to cover the whole world. I’m thinking that the changing reality and just making things up problems shown below are partly what Keenan is talking about here and in his WSJ op ed. piece [AR 1 assumption, temporally and spatially?], but I could be wrong!
Latitude says:
October 21, 2011 at 4:48 pm
James Sexton says:
October 21, 2011 at 3:10 pm
Sadly, it is clear the BEST project isn’t interested in science, truth, or even plausible extrapolations. Have you guys seen the global anomaly video put out by BEST?
http://www.berkeleyearth.org/movies.php There’s absolutely no way they can legitimately invent coverage where it never took place. But, they do it anyway. Here’s a couple of screen shots.
http://suyts.wordpress.com/2011/10/21/is-that-the-best-they-can-do/
Guys, I’d not waste much time on these people and just go straight to laughing at them.
=================================================
Strange…..
You picked one of their temp maps – 1891 – and they show the SW cool….
1891 was a famous drought and heat wave in the SW
Is this another case of inventing the past cooler….to make the present warmer?

phlogiston
October 22, 2011 11:03 am

DirkH says:
October 22, 2011 at 8:34 am
On a perfectly normal day the temperature can easily vary between 0 deg C and 20 deg C, right now, here in Germany. That makes the IR backradiation vary by about 25% as it varies with the 4th power of the actual temperature, not the smoothed one.
Do you still think a model that manages to reproduce the *smoothed* time series has any predictive value, or any resemblance to reality? 😉
I guess it depends what you are trying to predict. The discussions I had in mind were of multidecadal trends of global and ocean temperature for instance.

phlogiston
October 22, 2011 11:19 am

This whole BEST thing seems a straw man. Not that many here at WUWT ever asserted that the whole of recorded global tmperature increase in the 20th century was an artefact of UHI and other fabrications. The rise is generally accepted. The substance of the debate and research is – what is the reason for the rise – is it anthropogenic or is it cyclical? Or in other words, is it more plausible and intelligent to propose temperature stasis or oscillation as the normal state or null hypothesis?
In fact the BEST global temperature curve going back to 1800:
http://www.bbc.co.uk/news/science-environment-15373071
reveals some interesting oscillations. There is an appearence of ocsillations of increasing wavelength, suggesting something like an interferance of phase shift effect, that would merit further research.

Gofigure
October 22, 2011 12:39 pm

While the claim that temperature has increased since 1950 may be accurate, that should be no surprise – if it is the case. But, so what? Recall that the period from about 1940 to 1975 was cooling. (Even Obama’s science adviser, (a warmist) should have to concur on that cooling period because, in the 70s Holdren was busily searching for ways to warm the planet to forestall the oncoming ice age.)

October 22, 2011 12:48 pm

Understand that it is alleged that due to increased green house gases in the atmosphere, heat is trapped that cannot escape from earth. So if an increase in green house gases is to blame for the warming, it should be minimum temperatures (that occur during the night) that must show the increase (of modern warming). In that case, the observed trend should be that minimum temperatures should be rising faster than maxima and mean temperatures. That is what would prove a causal link.
What I have discovered so far from my (silly?) carefully chosen sample of 15 weather stations is that the overall increase of maxima, means and minima was 0.036, 0.012 and 0.004 degrees C respectively per annum over the past 35 years. So the ratio is 9:3:1. Assuming that my sample is representative of all those stations listed, I have to conclude that it was the maximum temps (that occur during the day) that pushed up the average temps. and the minima. So either the sun shone more brightly or there were less clouds. Or, even, perhaps the air just simply became cleaner (less dust? Are there records on that?).
I also noted that the warming on the NH is totally different to that of the SH. There is virtually no warming in the SH as seen by the means and minima whereas in the NH, the ratio of the increase in maxima, means and minima is about 1:1:1, amazingly.
Again, if it were an increase in CO2 or GHG’s that is doing the warming, you would expect to see the exactly the same results for NH and SH because these gases should be distributed evenly in the whole of the NH and SH hemisphere. So, even here, we again must conclude that it never was the increase in CO2 that is doing it. The only logical explanation I can think of is the difference in the rate by which the earth is greening. In South America we still had massive de-forestation over this period whereas Australia and Southern Africa have large deserts. Obviously, the NH has most of the landmass and here everyone seems to be planting trees and gardens. A recent investigation by the Helsinki university found that 45 countries were more green then previously out of a sample of 70.
Paradoxically, the increase in greenery is partly due to human intervention, partly due to more heat coming available (increase in maxima!) and partly due to the extra CO2 that we put in the air which appears to be acting as a fertilizer/ accelerator for growth.
For my data, see:
http://www.letterdash.com/HenryP/henrys-pool-table-on-global-warming
(make a copy for yourself of the tables)
Now, if we could have the 3 plots Maxima, Means and Minima for the BEST figures? That would help.

October 22, 2011 3:13 pm

I have been revising the maps on my site by increasing the resolution to 3 mile square grids, the increased detail shown compared to the 30 mile square gridded maps now on site are able to resolve the natural and man made sheltered areas that are responsible for the UHI effects.
from the sample maps shown at the link below there is an area greater than 10-15 miles in diameter out from the population centers that clearly show the UHI effects, so the 7 mile limit used by BEST IMO is lame at best.
Excerpt;It becomes easy to see not all of the warmer and cooler spots are due to cities alone, most are due to sheltering from weather due to surface textures that were preexisting before human occupation. People tended to settle in sheltered areas along water ways, so the natural heat islands have over the years, been human enhanced by urban growth. A fact of life not mentioned in the research literature?
Valleys in slow wind flows patterns can be over 10 degrees warmer than on windy days, like in the Dakotas in this screen shot, finished maps will be masked to block the random noise out side of the borders. Click to expand view and zoom in for more detail.
http://research.aerology.com/project-progress/map-detail/

Septic Matthew
October 22, 2011 6:30 pm

DirkH says:
October 22, 2011 at 8:34 am
You can find examples that illustrate the necessity of smoothing and the liabilities of smoothing. Muller’s response to Keenan is essentially correct: it is necessary to use smoothing and judgment (i.e. explicit tests of various kinds of known problems) together.

JJ
October 22, 2011 7:31 pm

phlogiston says:
October 22, 2011 at 6:37 am
I am with Richard Muller on the question of smoothing.

No you aren’t. You are with yourself on the question of smoothing, as you do not understand
“Keenen’s assertion that smoothing is incompatible with subsequent modeling seems a bit exaggerated.”
Keenan did not make that assertion.
“There have been several shrill proclamations here at WUWT that any kind of smoothing renders time series data completely unanalysable and devoid of any predictive value.”
Not on this thread. In fact, at the time you posted that accusation, there weren’t any.
“This must be nonsense. Richard Muller has provided a valuable service to WUWT by explaining why statistically smoothing does not empty data of any meaning.”
Richard Muller has given no such explanation of why his use of smoothing does not invalidate the statistics he computes.
“I work in micro-tomography where reconstructed images are sometimes noisy and require singificant smoothing to allow real structure to emerge from noise. It would make no sense to assert that such images were devoid of real information following smoothing.”
How very nice, and completely irrelevant to this thread. Keenan made no assertion that smoothing renders any data ‘devoid of real information’. Nor did he say anything about microtomographic images. What he did say, which you dont comprehend, is that smoothing time series data should not be performed before statistical analysis of those data, specifically the computation of correlation statistics.
“This spurious assertion that smoothing invalidates timeseries data …”
Does not appear anywhere but in your ignorant interpretation of what Keenan said. Keenan said that smoothing data invalidates certain statistics and staistical inferences calculated from those data. Muller did nothing but wave his hands and make appeals to authority – always fallacious but egregiously so when the authority is dead and cannot have given explicit support to Muller’s claims.

phlogiston
October 23, 2011 1:50 am

JJ
Well I’m no statistician and no doubt exaggerated my case about smoothing, so your rebuttal is fair up to a point. However Keenan’s views on smoothing of time series are not the only views in the field – in his reply Muller cites the influential work of John Tukey in justifying pre-smoothening (“whitening”) in certain circumstances.
I was not referring to comments on this post but other threads in the last year or so, where interesting empirical correlations between astrophysical parameters and climate, or correlations involving multidecadal oceanic oscillations, have been dismissed perhaps too hastily on the grounds of over-demanding statistical technicalities, considering all the unknowns in the system.