IPCC Climate: A Product of Lies, Damn Lies and Statistics Built On Inadequate Data

Guest essay by Dr. Tim Ball

“If you torture the data enough, nature will always confess” – Ronald Coase.

Facts are stubborn things, but statistics are more pliable. – Anonymous.

Climatology is the study of average weather over time or in a region. It is very different than Climate Science, which is the study by specialists of individual components of the complex system that is weather. Each part is usually studied independent of the entire system and even how it interacts or influences the larger system. A supposed link between the parts is the use of statistics. Climatology has suffered from a pronounced form of the average versus the discrete problem from the early 1980s when computer modelers began to dominate the science. Climatology was doomed to failure from then on, only accelerated by its hijacking for a political agenda.  I witnessed a good example early at a conference in Edmonton on Prairie Climate predictions and the implications for agriculture.

It was dominated by the keynote speaker, a climate modeler, Michael Schlesinger. His presentation compared five major global models and their results. He claimed that because they all showed warming they were valid. Of course they did because they were programmed to that general result. The problem is they varied enormously over vast regions. For example, one showed North America cooling another showed it warming. The audience was looking for information adequate for planning and became agitated, especially in the question period. It peaked when someone asked about the accuracy of his warmer and drier prediction for Alberta. The answer was 50%. The person replied that is useless, my Minister needs 95%. The shouting intensified.

Eventually a man threw his shoe on the stage.  When the room went silent he said, “I didn’t have a towel”. We learned he had a voice box and the shoe was the only way he could get attention. He asked permission to go on stage where he explained his qualifications and put a formula on the blackboard. He asked Schlesinger if this was the formula he used as the basis for his model of the atmosphere. Schlesinger said yes. The man then proceeded to eliminate variables asking Schlesinger if they were omitted in his work. After a few eliminations he said one was probably enough, but you have no formula left and you certainly don’t have a model. It has been that way ever since with the computer models.

Climate is an average, and in the early days averages were the only statistic determined. In most weather offices the climatologist’s job was to produce monthly and annual averages. The subject of climatology was of no interest or concern. The top people were forecasters who were meteorologists with only learning in physics of the atmosphere. Even now few know the difference between a meteorologist and a climatologist. When I sought my PhD essentially only two centers of climatology existed, Reid Bryson’s center in Wisconsin and Hubert Lamb’s Climatic Research Unit (CRU) at East Anglia. Lamb set up there because the national weather office wasn’t interested in climatology. People ridiculed my PhD being in the Geography Department at  the University of London, but university departments weren’t doing such work. Geography accommodated it because of its chorologic objectives. (The study of the causal relationships between geographic phenomena in a region.)

Disraeli’s admonition of lies, damn lies and statistics was exemplified by the work of the IPCC and its supporters. I realized years ago that the more sophisticated the statistical technique the more likely the data was inadequate. In climate the data was inadequate from the start as Lamb pointed out when he formed the CRU.  He wrote in his autobiography “…it was clear that the first and greatest need was to establish the facts of the past record of the natural climate in times before any side effects of human activities could well be important.” It is even worse today. Proof of the inadequacy is the increasing use of more bizarre statistical techniques. Now they invent data such as in parameterization. Now they use output of one statistical contrivance or model as real data in another model.

The climate debate cannot be separated from environmental politics. Global warming became the central theme of the claim humans are destroying the planet promoted by the Club of Rome. Their book, Limits to Growth did two major things both removing understanding and creating a false sense of authority and accuracy. First, was the simplistic application of statistics beyond an average in the form of a straight-line trend analysis: Second, predictions were given awesome, but unjustified status, as the output of computer models. They wanted to show we were heading for disaster and selected the statistics and process to that end. This became the method and philosophy of the IPCC. Initially, we had climate averages. Then in the 1970s, with the cooling from 1940, trends became the fashion. Of course, the cooling trend did not last and was replaced in the 1980s by an equally simplistic warming trend. Now they are trying to ignore another cooling trend.

One problem developed with switching from average to trend. People trying to reconstruct historic averages needed a period in the modern record for comparison. The 30-year Normal was created with 30 chosen because it is a statistically significant sample, n, in any population N.  The first one was the period 1931-1960, because it was believed to have the best instrumental data sets.  They keep changing the 30-year period, which only adds to the confusion. It is also problematic because the number of stations has reduced significantly. How valid are the studies done using earlier “Normal periods”?

Unfortunately, people started using the Normal for the wrong purposes. Now it is used as the average weather overall. It is only the average weather for a 30-year period. Actually it is inappropriate for climate because most changes occur over longer periods.

But there is another simple statistical measure they effectively ignore. People, like farmers, who use climate data in their work know that a most important statistic is variation. Climatology was aware of this decades ago as it became aware of changing variability, especially of mid-latitude weather, with changes in upper level winds. It was what Lamb was working on and Leroux continued.

Now, as the global trend swings from warming to cooling these winds switched from zonal to meridional flow causing dramatic increases in variability of temperature and precipitation. The IPCC, cursed with the tunnel vision of political objectives and limited by their terms of reference did not accommodate natural variability.  They can only claim, incorrectly, that the change is proof of their failed projections.

Edward Wegman in his analysis of the “hockey stick” issue for the Barton Congressional committee identified a bigger problem in climate science when he wrote:

“We know that there is no evidence that Dr. Mann or any of the authors in paleoclimatology studies have had significant interactions with mainstream statisticians.

This identifies the problem that has long plagued the use of statistics, especially in the Social Sciences, namely the use of statistics without knowledge or understanding.

Many used a book referred to as SPSS, (it is still available) the acronym for Statistical Packages for the Social Sciences. I know of people simply plugging in numbers and getting totally irrelevant results. One misapplication of statistics undermined the career of an English Geomorphologist who completely misapplied a Trend Surfaces analysis.

IPCC projections fail for many inappropriate statistics and statistical methods. Of course, it took a statistician to identify the corrupted use of statistics to show how they fooled the world into disastrous policies, but that only underlines the problem with statistics as the two opening quotes attest.

There is another germane quote by mathematician and philosopher A.N. Whitehead about the use, or misuse, of statistics in climate science.

There is no more common error than to assume that, because prolonged and accurate mathematical calculations have been made, the application of the result to some fact of nature is absolutely certain.

_______________

Other quotes about statistics reveal a common understanding of their limitations and worse, their application. Here are a few;

He uses statistics as a drunken man uses lampposts – for support rather than for illumination. – Andrew Lang.

One more fagot (bundle) of these adamantine bandages is the new science of statistics. – Ralph Waldo Emerson

Then there is the man who drowned crossing a stream with an average depth of six inches. – W E Gates.

Satan delights equally in statistics and in quoting scripture. – H G Wells

A statistical analysis, properly conducted, is a delicate dissection of uncertainties, a surgery of suppositions. – M J Moroney.

Statistics are the modern equivalent of the number of angels on the head of a pin – but then they probably have a statistical estimate for that. – Tim Ball

__._,_.___

0 0 votes
Article Rating
81 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments
Marcos
October 2, 2013 11:11 am

the NOAA web site used to have a page discussing the use of ‘normals’ that specifically stated that they were were never intended to be used to gauge climate change. for the life of me, i cant find that page anymore…

Gonzo
October 2, 2013 11:16 am

[He uses statistics as a drunken man uses lampposts – for support rather than for illumination. – Andrew Lang.] I couldn’t help but think of tamino et al LOL

October 2, 2013 11:31 am

This part of an exchange I had with a modeler on my blog.
Let’s go back to basics. What can a model teach us? Or in other words, what is the knowledge gained by running a simulation. Knowledge is gained from analysis of information. Where does information come from? From the resolution of contingency (if that last sentence doesn’t make sense, consult any introductory text on information theory). What contingency then can running a simulation resolve? It’s output trajectory, unknown before the run. On what does that result depend? The model’s equation plus initial state. Since this information is internal to the model, a model can only teach us about itself! Models can create contingency but can never resolve it, and therefor cannot provide information gain if the metric is information about the real world. Models can only provide a hypothesis but information gain (knowledge) cannot occur unless the hypothesis is tested by comparing the models predictions to the observed data, something the IPCC refuses to acknowledge. They claim models do not make predictions. But a model is only as good as its predictive power.

October 2, 2013 11:48 am

Old Scottish saying:
Facts are chiels that winna’ ding.
Approximate translation: Facts are things that cannot be argued with.
Now, how can we get a), IPCC and b), MSM to accept this?
I’m not holding my breath.

October 2, 2013 11:59 am

Curiously, I just posted a list of four statistical “sins” committed in AR5 with regard to figure 1.4 (and in AR4’s SPM as well), on CA in the thread Steve just opened on statistical problems with this figure. They are a bit more specific than the general essay above:
* Cherrypicking (presenting single model runs without telling us how those particular model runs were selected or constructed, permitting them to be selected out of an ENSEMBLE of runs from slightly variable initial conditions per model).
* Data Dredging in the worst possible way and concealing the fact in a plateful of spaghetti.
* Applying the concept of the statistical mean of an ensemble of models as if it has any meaning whatsoever, especially when the models individually fail or do very poorly on an ordinary hypothesis test compared to the actual data, once you remove the illusion of the plate of spaghetti and display the results, one model at a time, compared to the actual data.
* Applying the concept of the statistical variance/standard deviation to the results from an enemble of models as if it has any meaning whatsoever. This is the only rational basis for statements such as “95%” confidence or “high confidence” and is based on the central limit theorem, which does not apply to ensembles of model predictions because the general circulation models used in the ensemble are not randomly selected independent and identically distributed samples drawn from a probability distribution of correct general circulation models.
Steve also added the fact that they cherrypicked and shifted the start date for figure comparison to maximize the overlap of their spaghetti with the actual climate data, certainly compared to the earlier AR5 draft. I’d add another feature of 1.4 I pointed out back in the originally leaked draft — the error bars shown on the current temperatures are utterly meaningless — somebody just made them up at plus or minus 0.1C. HADCRUT4 doesn’t even claim to be accurate to better than 0.15C across the contemporary time frame. As usual, the graphic adds an error bar onto individual points that are what they are (conveying a false sense of precision beyond that represented by the data itself) and omits any sort of error estimate or analysis of fluctuations in the individual GCMs represented in the spaghetti. Finally, one should really look AT the fluctuations in the spaghetti — in particular, how does the autocorrelation of the GAST produced by the GCMs compare to the observed climate autocorrelation. Hmmm, not so well, not so well.
I’d add further that performing the most cursory of fits of the form T(t) = a t + b sin(ct) + d to the 165 year HADCRUT4 data, one obtains a result with vastly smaller a (the linear rate of warming), b around 0.2 C, c around 1/70 inverse years, and d set by the need to optimize the slope and oscillation relative to the data. This curve explains almost all of the visible variation in the data remarkably accurately, including the rapid warming of both the early and late 20th century and the “pause/cooling” in the middle and end of the 20th century, as WELL as a similar cycle visible even in the 19th century data with its presumably large uncertainties. In fact, I think I could pretty easily make up a stochastic function such as:
T(t) = a t + b sin(ct) + d + F(t)
where F(t) is trendless, exponentially correlated or power correlated noise and produce temperature curves that are ALMOST indistinguishable from the HADCRUT4 observed 165 year temperature series, and could do even better if I cherrypicked a different start and (say) threw out all data before 1880 or whatever as being too imprecisely known.
If this curve (which is pure numerology — I have no idea what the physical basis of a and d are beyond mumblings about Milankovitch cycles that add up to “I don’t really know”, might GUESS that c is the inverse of the PDO period with b an empirical amplitude, and of course F(t) is noise because, well, the climate system is empirically noisy as all hell in ways we cannot predict or understand) has any meaning at all, it is that making egregious claims about the “unprecedented” nature of the warming trend in the late 20th century is ridiculous — no trend explainable by a four parameter sinusoidal fit is “unprecedented”, and as Dick Lindzen has shown, no audience that is shown early and late 20th century temperature variations at the same scale and asked to pick which one is with and which without CO_2 can do this unless they are enormously familiar with the data and can pick out individual FINGERPRINTS in the data such as the late 20th ENSO/Pinatubo bobble. The curves are qualitatively and quantitatively identical to within this sort of feature specific noise. What isn’t appreciated is that the curve actually extends decently back into the 19th century as well, at least as far as HADCRUT4 is concerned.
Personally, I’d like to see AR5 criticism be a little less general and a little more specific, and think that one can pick figure 1.4 completely to pieces in a way that is most embarrassing to the IPCC. I’d start by simply separating out the contributing GCMs, one at a time, from the spaghetti. Look at the one in orange near the very top! It doesn’t come within MILES of the empirical curve. To hell with it, it fails an INDIVIDUAL hypothesis test regardless of how you massage starting points and so on. Remove it from the contributing ensemble as a falsified model. Look at another. Yes, it dips down as low as the empirical curve — for less than 10% of its overall values — and it has UTTERLY incorrect autocorrelation, showing wild spikes of warming that really ARE unprecedented in the data. To hell with it — remove it from the contributing ensemble.
In the end, you might have a handful of GCMs that survive this winnowing not because they are CORRECT, but because they at least spend SOME significant fraction of their time at or below the actual temperature record and have at least approximately the right autocorrelation and heating/cooling fluctuation range. Better yet, run entire Monte Carlo ensembles of runs per model and see what fraction of the models have EXACTLY the right mean and autocorrelation behavior. If the answer is “less than 5%” chuck the model. Even fewer of the GCMs would make the cut here.
When you are done, don’t present a pile of spaghetti and don’t data dredge a conclusion. The non-data dredged conclusion one SHOULD draw from considering the GCMs as if they WERE iid samples would be “throw them all out”, because one has to apply MUCH MORE STRINGENT STATISTICAL REQUIREMENTS FOR SIGNIFICANCE if you have many jars of jellybeans you are trying to correlate with acne, and commit even more grevious sins that mere data dredging if one PIECES TOGETHER regions where individual models HAPPEN to descend to barely reach the data in order to be able to claim “the models embrace the data”.
Of course, if one does this one ends up with the conclusion that the rate of warming is highly exaggerated in AR4 and AR5 alike by any of the ensembles of GCMs. You might even end up with a rate of warming that is surprisingly close to a, the slope of the linear trend in my numerological fit. That’s order of a half degree per century, completely independent of any presumed cause such as CO_2.
rgb
[Thank you. An “iid sample” ? Mod]

D Nash
October 2, 2013 12:04 pm

The result of statistical corruption then leads to other forms of corruption:
“The surest way to corrupt a youth is to instruct him to hold in higher esteem those who think alike than those who think differently.”
― Friedrich Nietzsche
“One of the saddest lessons of history is this: If we’ve been bamboozled long enough, we tend to reject any evidence of the bamboozle. We’re no longer interested in finding out the truth. The bamboozle has captured us. It’s simply too painful to acknowledge, even to ourselves, that we’ve been taken. Once you give a charlatan power over you, you almost never get it back.”
― Carl Sagan

October 2, 2013 12:14 pm

statistics is fine to study climate
as long as you keep to probability theory
and taking representative samples
http://blogs.24.com/henryp/2013/02/21/henrys-pool-tables-on-global-warmingcooling/

MattB
October 2, 2013 12:34 pm

Facts are stubborn things, but statistics are more pliable. – Anonymous
FYI, that one is generally attributed to MarkTwain, and I can see that 🙂

October 2, 2013 12:36 pm

Even when the models compare well to observed data, there is still the possibility of no relevance between the two. I deal with this on a regular basis where modelers “tune” their calculations adjusting inputs that really fixed, and ignoring those that are critical.
As an example: A building energy model was created to match the building energy consumption characteristics to past bills. A mechanical contractor had approached the owner with a energy retrofit project that he claimed would save 30% of the building air conditioning energy use. He “tuned” the model to match annual energy use within 2%, calling it good and presented the output.
It took me the better part of 3 weeks to get the input files in order to review the output files.
Monthly energy use was off as much as 200% of the bills. Higher in the summer, lower in the winter.
They had adjusted the building air conditioning operations from the actual 60 hours/week to 94 hours/week.
They had not taken into account a data center that consumed 30% of the total electrical usage. (This was purposely done to provide a larger energy pool to claim savings from.)
Motor KW draw was nameplate, not actual. (Actual was ~50% less than nameplate.)
Chiller efficiency at 0.9KW/ton cooling was worse than actual. (more energy consumption.)
These issues were brought up and the modeler went back for another round. Again I had to ask repeatedly for the input files.
The final showed half the data center load accounted for and the operating hours were corrected,.
However the chiller efficiency was adjusted upwards to 1.4KW/ton cooling. The reason given was to more accurately reflect “actual” usage. (Actual was around 0.68KW/ton.)
The occupant count was increased almost two-fold, including floor plug loads.
By this time the owner saw the writing on the wall and stopped the project.
The contractor was furious that I had deemed it necessary to actually question his model.
The sad thing was that the contractor didn’t know (or acknowledge?) the real reason for poor energy performance was the building envelope. Caulking was failed, and the air conditioning motorized dampers were failed wide open at night, allowing stack effect to “push” the heat out of the building every night.
Had these issues been identified upfront, there would have been no reason to do the modeling in the first place.
This was one instance where I managed to get ahead of the problem. All too often I get brought in afterwards, when the damage has already been done. Those projects make for great meetings…:)
I could write for months on this subject….

October 2, 2013 12:36 pm

There’s some interesting historical details that perhaps may mesh with this story. Reid Bryson was a friend of Hubert Lamb. I believe it was in the 1970s, when global cooling and a coming ice age were the fashion statement of the day, Reid Bryson was claimed to have speculated on the cooling potential that might occur if every human kicked up something like a spoonful of dirt into the atmosphere. Bryson, himself, never accepted the CAGW meme, but he unwittingly gave an inspiration to Thomas Wigley. Hubert Lamb was not particularly computer literate and hired Wigley to transfer climate records onto the computer. Wigley had other ideas and Lamb grew to regret having hired him at the CRU. Wigley, who eventually migrated over to the NCAR, was one of the original AGW mavens. A problem always existed with the global warming meme in that global temperatures dropped from the 1950s till the late 1970s, in contradiction to the greenhouse theory. Thomas Wigley, so to speak, ‘borrowed’ Reid Bryson’s theories on atmospheric aerosols and used that to explain away that drop, claiming that when clean air legislation came on line in the 70s the cooling effect of the aerosols ceased which allowed AGW to light off. The IPCC apparently attempted to resurrect that claim about a year ago to explain the current lull in warming. Lots of twists and turns to this story, eh?

October 2, 2013 12:40 pm

I have been in the water resources business for 30 years, much of which has dealt with groundwater. That segment of the industry has been hijacked by groundwater modelers in much the same way. Groundwater models are no more in touch with reality than climate models but they are the cornerstone of decisions for groundwater remediations. In the absence of an IPCC, EPA is the mover and shaker of the model-driven basis of groundwater reality in the U.S., and if you think they are draconian with climate policy, you should spend some time working on a CERCLA (Superfund) Site and dealing with that branch of the agency. If a model says something is a certain way, then that’s it – no arguments. The premier journal of groundwater science “Ground Water” has become a paper forum for modelers to throw computer codes around and there are virtually no actual studies of groundwater in the real world.
I have represented companies and groups of companies which are paying tens of millions of dollars attempting to remediate groundwater which has no potential of impacting a single receptor and which, through natural processes, remediates itself before any receptor could be exposed to any contamination. I have produced hard time series and spatial distribution data to show that there is no need for the expenditure. But the agency runs a model which says that under a narrow range of specific pre-determined conditions, starting from an impossible initial condition, there is a chance that someone might be exposed. What’s more, they use the highest concentration ever measured at the source of a release (where no-one actually uses the groundwater) as the exposure concentration regardless how far away that exposure point is from the source and ignoring how much measured attenuation of concentration has occurred in the intervening distance/time, and assume that if a million people drink two litres of that most-contaminated water every day for 70 full years (no vacations or moving to a new house is allowed) then there is a chance that one of those people will develop a cancer – and they purport to know this despite the fact that of those million persons, 250,000 will actually develop cancers from other reasons. In other words, if ANYONE happens to spill a few gallons of gasoline onto the ground and it is measureable in groundwater, EPA takes the highest concentration detected, extrapolates THAT concentrations to remote locations and concludes that instead of 250,000 cancers per million people, there MIGHT be 250,001 cancers. Believe me, the transport assumptions are IMPOSSIBLE.
The agency then calculates a number which should be the remediation goal and forces the company to attempt to meet that goal by expensive actions which invariable fail (that is correct – they do not work). EPA KNOWS that NONE of its enforced groundwater remediations have ever succeeded in achieving its impossible goals in the absence of the wonderful ability of nature to take care of itself and do most of the actual work of contaminant elimination. Yet the agency continues to required multiple tens of millions of dollars per year in useless remediations because a model says it will be possible.
Now here’s a familiar scenario for you. Who actually runs the models? It is the contractors who will then be awarded the remedial design contract and construction oversight and operations and maintenance work. In every case in which I have been confronted with a computer model (and that’s about 100 to 125 out of the many many hundreds of cases on which I have worked), I have been able to refute the conclusions of the models by applying a very simple trick. Force everyone to look at the actual field data.
In groundwater modeling, there is a simple precept which everyone except the modelers and EPA accept: in order for a model to be remotely accurate, you need to generate enough data to eliminate assumptions for the input parameters. But by the time you have that much data, you no longer need a model because you HAVE a data-supported answer.

October 2, 2013 12:45 pm

Hi
If the models are the best of the science can produce (and I do accept that the model constructors are experts in the numerical modelling, but may not be experts in much else) and their product does not work, perhaps what some call ‘pseudo science’ should be looked into.
For some years now, I have looked into changes of the Earth’s magnetic field and noticed easily observed ‘apparent correlation’ with the temperature records.
For time being the science indicates that the only way these changes may make an impact is through the cosmic rays nucleation process, but it doesn’t support the Svensmark’s hypothesis since the Earth’s and solar field appear to have negative ‘correlation’:
http://www.vukcevic.talktalk.net/LFC9.htm
There are other physical processes that come into play but sadly, as the current science indicates none has sufficient power available to move the oceanic temperatures.
However, the geomagnetic field as measured on the surface is often indication what is happening further down in the Earth’s interior.
Changes in the interior are also reflected on the surface through the tectonic activity, which contains enough energy to affect efficiency of the ocean currents, the main transporters of energy from the equatorial regions pole-wards. .
Thus next step was to look at the tectonic records for the last 100+ years (relatively good records available) in the areas of the three climate indices (AMO, PDO & ENSO).
An odd ‘oscillation’ was noted in the North Atlantic, but when these events were integrated over period of time ( k ) the picture became far more encouraging.
The process was repeated for the areas of the North and Equatorial Pacific and results are presented here:
http://www.vukcevic.talktalk.net/APS.htm
As it can be clearly seen the ‘forcing formula’ is the same one in all three cases. The k factor was determined by trial and error, it has same value for both Pacific indices, but it is less effective in the Atlantic than in the Pacific, Pacific being more tectonically active.
Finally, it is intriguing that a single (albeit regional) variable can within reason model past 100+ years temperature records, but for the future, ‘the time will tell’.
Note: if anyone is keen to know tectonic data, they are available on the web but it takes time and effort to collate, so do not expect readily made handout.
(the above is an exception, normally do not do long posts)

GlynnMhor
October 2, 2013 12:54 pm

How to ‘prove’ causality with graphs:
http://www.smbc-comics.com/?id=3129

Lloyd Martin Hendaye
October 2, 2013 1:00 pm

Normal curves are like mastectomies: Fine with a cosmetic overburden, yet nothing underneath.

heyseuss
October 2, 2013 1:23 pm

Nice essay, Dr. Ball.

Gary Pearse
October 2, 2013 1:23 pm

“Edward Wegman in his analysis of the “hockey stick” issue for the Barton Congressional committee identified a bigger problem in climate science when he wrote:
“We know that there is no evidence that Dr. Mann or any of the authors in paleoclimatology studies have had significant interactions with mainstream statisticians.”
Unfortunately Bradley pole axed Wegman’s effectiveness by charging him with plagiarism of some lines in his report and that is effectively about all that remains of Wegman’s contribution. Attacking the man is all that’s needed to neutralize his work these days. If Mann used the Tiljander series (Finnish lake sediment cores) upside down, that doesn’t matter. He attacked his accusers and has been since exonerated, even decorated and a dozen studies have been done since supporting the Hockey Stick. That’s how it’s done in post normal times.

stan stendera
October 2, 2013 1:25 pm

Brad and Tom G: You should both write books!!!! Among the all time great comments on WUWT, and that’s saying something.

October 2, 2013 1:29 pm

Climatology is the study of average weather over time or in a region. It is very different than Climate Science, which is the study by specialists of individual components of the complex system that is weather.

=================================================================
I read that and thought of the medical field or “The Healing Arts” as I heard it called once. A nutritionist is not a MD but what MD would say that nutrition has nothing to do with health and vise/versa. “Health foods”? “Herbal remedies”? They may have their place. (Slippery elm can relieve the symptoms of a simple sore throat.) The problem enters when one “field” thinks theirs is the only valid one.
That’s what’s going on here. Some think tree rings (one tree ring?) and CO2 from fossil fuels are the only valid things to consider.

chris y
October 2, 2013 1:37 pm

stan stendera says:
October 2, 2013 at 1:25 pm
“Brad and Tom G: You should both write books!!!! Among the all time great comments on WUWT, and that’s saying something.”
I agree. These are absolutely excellent comments.
I encourage both of you to consider putting together a post for Anthony. Your writing styles are similar to Willis Eschenbach, whose posts I enjoy immensely!

October 2, 2013 1:39 pm

rgb
Always interesting hearing your take on models. Recently I’ve been keeping tabs on Bishop Hill, especially the recent Nic Lewis interactions with Julia Slingo of the Met Office.
My take on all this has been pretty much constant: where has IR forcing been characterised in a realistic environment, how have losses been measured and how effective is IR heating at very low power densities. IR is used to for example make bacteria inactive, but the power densities are in the KW/m2.
Anyway some commenters notably AlecM (posts on WUWT as AlecMM) suggested looking at Hansen_etal_1981 (if you Googled this it would come up). AlecM believes that Hansen has miscalculated the greenhouse effect and that more focus should be on LTE in the atmosphere (sorry for the paraphrase). What struck me was that Hansen attributes 90% of warming to the 15 micron band -“downwelling radiation” – in essence that Co2 is the main driver of climate. There is a reference for this pointing to an earlier paper, which then relates to Manabe and then finally to the first recent era estimation of atmosphere heating due to greenhouse gases – by Moller.
The paper I have from 1961 shows the logic of the calculation. Basically the idea is that slabs of the atmosphere absorb and radiate depending on composition (some gases absorp and thermalise). But for 15 microns it looks very much like the surface is the main contributor as this freqency passes through the slabs largely unaffected. The net radiation (the difference between upwelling and downwelling from the atmosphere) is used to adjust the blackbody emission of the Earth. The reduction in radiation will result in an increase in BB emission to balance it.
Now in a vacuum this is fine. I have experience in the space industry testing plasma thrusters and thermal properties are a big part of it. The temperature of surfaces can have repercussions on operating temperature of cables and components and temperature can only be determined by radiative processes to space.
The problem is that this modelling assumes that all the downwelling radiation is turned into real heat. That none of it is lost. And that this occurs in an atmosphere, with surface conduction, evaporation, convection not to mention scattering and thermal losses.
And this idea has been perpetuated for 50 years or more.
All the models assume that W/m2 is instantaneously converted 100% into real heat.
It’s a small detail buried in what is a good attempt to model an atmosphere, using tested radiative properties of gases. But the issue is that there is no correction for “effective forcing” due to this downwelling radiation in an atmosphere interface. This would have been picked up if the effect had been studied and characterised.
It was interesting then that in Julia Slingo’s reply she mentioned that scientists are still trying to isolate radiative forcing from C02 alone. Not eliminating it using known characteristics. They are still trying to look for it. And use the whole Earth climate system with all the other effects thrown in.
That is a real problem.

chris y
October 2, 2013 1:40 pm

Here is a timeless quote regarding statistics-
“If you need statistics to prove something, and you have no proof except statistical proof, then what you have proved probably isn’t true. Statistical evidence is the lowest form of evidence there is. What a depressing conclusion.”
William Briggs, statistician (to the stars!), October 5, 2011

October 2, 2013 1:41 pm

Statistical methods are tools. Nothing more. In the hands of a skilled craftsman they can help build wonders like medical discoveries or pricing models. But most users are not skilled craftsmen, as Drs Ball and Brown point out. To repeat an old saw (carpentry pun intended), if all you have is a hammer, even screws look like nails.
But statistical incompetence is no excuse for mendacity as practiced by IPCC AR5 in SPM figure 1-4. Rewrote previous IPCC pronouncements by arbitrarily revising previous starting data. Included unrealistic scenario B1 to lower the lower bound. Cherry picked individual model runs instead of using CMIP3 ensemble averages. Who knows what else.

YEP
October 2, 2013 1:51 pm

First, SPSS is not a book, it’s a statistical software package. Perhaps the fact that it’s easy to use creates the temptation to use it without proper understanding. But SPSS is certainly not the problem; it’s GIGO, enabled by lack of understaning on many sides, not least graduate supervisors and peer reviewers.
Second, the use of statistics in the social sciences is not uniformly poor. Of course there are plenty of papers written and published based on simplistic analyses and poor methodologies, but any graduate program worth its salt should provide a proper grounding in fundamental statistical theory and practice. In my own field of economics, econometrics can boast very high standards of sophistication and has been at the forefront of the development of statistical theory and practice.
Thirdly (mainly in response to comments), there is nothing wrong epistemologically with using models, so please let’s stop the generalized bashing. The real world is too complex, stochastic and chaotic to be analyzed in its totality, so simplifications are needed to further any understanding. A good analogy is a road map: it’s an imperfect and simplistic model, but it gets you there, and it’s a lot more useful and practical than carrying a full-size replica of the earth in your car. As long as we don’t treat models or their output as truth or fact, not only do they have their uses, but the advancement of knowledge becomes impossible without them. So by all means criticize those who use model outputs as inputs to other investigations, or claim them as proofs for hypotheses, or ignore facts incompatible with the models’ forecasts, or talk of “settled science”. But dismissing whole fields of inquiry by saying “it’s only models” is absurd. A model’s usefulness is in simplifying, highlighting the essential aspects of a set of phenomena, and using logically coherent methods to deduce testable hypotheses from actual data and well-understood causal mechanisms. If the testable hypotheses fail, then you go back and reexamine assumptions and causal links. Failure is as important as success (or more so) in illuminating the assumptions and causal mechanisms used and promoting understanding.
Feel free to apply any of the above to the IPCC and its “settled science”. But the generalized assault on the social sciences and the use of models misunderstands the most basic aspects of the philosophy of science, and is getting a little tiresome.

Jordan
October 2, 2013 1:55 pm

“The 30-year Normal was created with 30 chosen because it is a statistically significant sample, n, in any population N. ”
The above statement is only true on the condition of independent samples. This means sample errors follow a theoretical pattern of cancellation when aggregated. Put another way, we need to be confident that sample errors are zero mean random variables with a predictable variance.
If this condition is met, we can accurately calculate confidence limits for the sample. If these conditions are not met, confidence limits must be greater (greater than the estimates we would produce using methods which assume zero mean random error).
Proper, well-designed, sampling schemes are necessary to be able to meet the above condition, and to make claims about the result of statistical analysis.
Proper, well designed sampling schemes require prior understanding of the characteristics of the population. If you don’t have this, you are off to a bad start, and statistical methods become more difficult to apply (although this requires equal justification and understanding of the population).
Even then, there can be issues with stationarity (the assumption that a sample taken at one time is a representation of the system behaviour at another.
Climate analysis and climate modelling has no reliable measure of the “climate population”, or understanding of the fundamental properties (meaning statistical behaviour). It therefore lacks a dependable basis for design of an adequate sampling system. Without this, it cannot make any reliable claims based on statistics.
If we say N=statistically significant, we fall into a trap. We concede too much ground to the amazing claims of the doomsayers, when we really need to be sticking to fist principles. Sticking to first principles allows us to say: “you haven’t a clue”.

October 2, 2013 1:56 pm

Marcos says:
October 2, 2013 at 11:11 am
the NOAA web site used to have a page discussing the use of ‘normals’ that specifically stated that they were were never intended to be used to gauge climate change. for the life of me, i cant find that page anymore…
====================================
Wayback machine at archive.org might help? I wouldn’t know where to start to look for the page you are after.

October 2, 2013 2:19 pm

“The 30-year Normal was created with 30 chosen because it is a statistically significant sample, n, in any population N.”
Wrong. there is no such thing as a statistically significant sample.
the samples required is dependent upon the variance of the thing you are measuring and your desired accuracy.
here is something simple that even ball might get
http://statswithcats.wordpress.com/2010/07/11/30-samples-standard-suggestion-or-superstition/

Mac the Knife
October 2, 2013 2:43 pm

With regards to the UN-IPCC AR5 report:
“The whole aim of practical politics is to keep the populace alarmed (and hence clamorous to be led to safety) by menacing it with an endless series of hobgoblins, all of them imaginary. ” H.L. Mencken

October 2, 2013 3:02 pm

stan stendera says:
“Brad and Tom G: You should both write books!!!!”
I second that request. Both subjects are familiar to me. Building energy “efficiency” assessments where the perp walks away with the folding stuff while the costs go up. Groundwater transport requirements to protect an ephemeral watercourse 50m away, and the permeability rate is less than 2mm per day. I have done a fair amount of modelling. Models have now become a pandemic worse than swine flu.

Reply to  Martin Clark
October 2, 2013 4:03 pm

I am halfway done…. Going to cover a lot more than modeling errors.

Jquip
October 2, 2013 3:35 pm

Robert Brown @ 11:59 — “Applying the concept of the statistical variance/standard deviation to the results from an enemble of models as if it has any meaning whatsoever.”
You did most of the heavy lifting for a point I wished to raise, so I’ll pick on you just a bit here. Though with nothing personal intended, consider it a guitar jam. Is there meaning to the ensemble var/sd? Well, yes, absolutely. This is an Economics Toolkit issue. Where *an* investor is always an idiot, even if they are an expert idiot. And the var/sd speaks directly to the amount of conformity in the individual opinions of the idiots in the ensemble. The mean of the distribution directly speaks only to the opinion of a constructed Statistical Idiot.
To speak to any more than that requires demonstration that the subset ensemble is reflective of a larger population. In Economics the Market is *the* ensemble of idiots. And a panel of selected expert idiots, if selected without bias, can be extended from a subset ensemble to a population ensemble. Because the collected opinions of *all* idiots is directly causal to the Market outcomes. And I don’t think it terribly likely that you’ll find Climatologists or Catastro-fans signing onto the explicit statement that the Climate ensembles causally manifest outcomes in reality. Though I’m quite sure you’ll find a random fellow here and there that will sign onto it.
But the models are not reflective of a larger population. The Statistical Idiot they construct is not an abstract notion of the distribution of the opinions in the subset ensemble being representative of a population ensemble. It is a single, and solitary object. A fictional, but constructed, Committee Chairman that speaks as one voice for the ensemble of idiots used to construct it. Which is fine for what it’s worth. But how accurate are the Delphic prognostications of the Statistical Idiot? And what metric do we use to justify investing on the basis of those Delphic prognostications?
If we’re interested in sciency things such as laws and theories, then a Statistical Idiot applied to a single reality isn’t going to get us there. It requires the explicit statement that ‘Science,’ as a body of knowledge, is not yet competent enough to produce a law or theory on the given subject. And it requires the explicit statement that we reject that *any specific* idiot is actually correct. Either known to be correct, or *possibly but unknowably* correct. If we held any candle of hope that there was such an idiot in the ensemble, then we would not use the ensemble. And it requires the final explicit statement that, while we are self-certain that every idiot in the ensemble is wrong in material ways, that blunting individual wrongness by constructing a Statistical Idiot is useful.
For those not familiar with gambling, this is how the Bookie at the horse races determines what odds to give on each race. The gamblers are the ensemble of idiots that are not reflective of the population of dogs or ponies in the given race. Nor are they a subset of the population of race contestants. Nor can their constructed Statistical Idiot be causal to the outcome of the race; unless there are criminal elements involved. The Bookie merely uses the Statistical Idiot vying for each dog or pony as a manner to ensure he stays in business. To adjust his odds of profit and loss for himself.
But if there is only one Statistical Idiot, then a Bookie is out of business. His payout must go to zero. Unless he has better knowledge that the Statistical Idiot about how often that dog or pony wins. That is, absent empirical knowledge — though still statistically aggregated — he cannot permit any payout greater than zero. If he wants to stay in business. And here, in respect to climatology, the ensembles are the gamblers, reality is the pony, and the Bookie is each consumer of the pair. Which, in consideration to the recent IPCC report and its purposes, the Bookie is any Govenment or Actuary.
So is there worth in the Statistical Idiot constructed by the model ensembles? Absolutely. It is a direct and explicit reflection of the consensus in the Consensus Science of Climastrology.

DaveS
October 2, 2013 4:25 pm

I have an alternative vision of what a drunken man might do with a lamppost. And perhaps it’s an equally appropriate metaphor for how some climate change alarmists treat statistics.

Fernando (in Brazil)
October 2, 2013 5:02 pm

Me too.
I am now.
I am skeptical sarcastic.
=================
Sincerely, I apologize to the Rev.
The indignation with such stupidity will lead me to be even more stupid.
And do not want to cause any embarrassment to my longtime friends.
There are no words. To describe. the incredible. 5 IPCC.
========================
Sorry.
The crazy liars ever crossed the line.

October 2, 2013 5:03 pm

It is clear after the AR5 SPM that the IPCC forecasts based on climate models are completely useless as a basis for policy. This was a last chance for the IPCC contributors and editors to acknowledge frankly that their models have no skill in forecasting and begin to change their CAGW paradigm The IPCC scientists are so far out on a limb that for psychological and professional and funding reasons they cannot scramble back .Further discussion of the IPCC science is a waste of time – a new forecasting paradigm is required. Such a different approach is outlined in a series of posts at http://climatesense.norpag.blogspotr.com
Here are some quotes from the latest post.
“b) A Simple Rational Approach to Climate Forecasting based on Common Sense and Quasi Repetitive- Quasi Cyclic Patterns.
How then can we predict the future of a constantly changing climate?
When,about ten years ago ,I began to look into the CAGW – CO2 based scare, some simple observations immediately presented themselves. These seem to have escaped the notice of the Climate Establishment. ( See the Post 5/14/13 Climate Forecasting for Britain’s Seven Alarmist Scientists and for UK Politicians.)
a) Night is colder than day.
b) Winter is colder than summer.
c) It is cooler in the shade and under clouds than in the sun
d) Temperatures vary more widely in deserts and hot humid days are more uncomfortable than dry hot days – humidity (enthalpy) might be an important factor. We use Sun Screen against UV rays – can this be a clue?
e) Being a Geologist I knew that the various Milankovitch cycles were seen repeatedly in the Geologic record and were the main climate drivers controlling the Quaternary Ice Ages.
f) I also considered whether the current climate was unusually hot or cold. Some modest knowledge of history brought to mind frost fairs on the Thames and the Little Ice Age and the Maunder Minimum without sunspots during the 17th century . The 300 years of Viking settlements in Greenland during the Medieval Warm Period and viniculture in Britain suggested a warmer world in earlier times than at present while the colder Dark Ages separate the MWP from the Roman Climate optimum.
g) I noted that CO2 was about 0.0375% of the Atmosphere and thought ,correctly as it turns out, that it was highly unlikely that such a little tail should wag such a big dog.
I concluded ,as might any person of reasonable common sense and average intelligence given these simple observations that solar activity and our orbital relations to the sun were the main climate drivers. More specific temperature drivers were the number of hours of sunshine,the amount of cloud cover,the humidity and the height of the sun in the sky at midday and at Midsummer . It seemed that the present day was likely not much or very little outside the range of climate variability for the last 2000 years and that no government action or policy was required or would be useful with regard to postulated anthropogenic CO2 driven climate change.
These conclusions based on about 15 minutes of anyone’s considered thought are, at once , much nearer the truth and certainly would be much more useful as a Guide to Policymakers than the output of the millions of man hours of time and effort that have been spent on IPCC – Met Office models and the Global Warming impact studies and the emission control policies based on them.”
For a forecast of the coming cooling – very likely until 2035 and possible until 2650 check on the link above.

Gail Combs
October 2, 2013 5:43 pm

YEP says:
October 2, 2013 at 1:51 pm
First, SPSS is not a book, it’s a statistical software package. Perhaps the fact that it’s easy to use creates the temptation to use it without proper understanding…..
Feel free to apply any of the above to the IPCC and its “settled science”. But the generalized assault on the social sciences and the use of models misunderstands the most basic aspects of the philosophy of science, and is getting a little tiresome.
>>>>>>>>>>>>>>>>>>>>>
In my field, Quality Engineering/Chemistry, statistical software packages are routinely ‘used without proper understanding.’ It was one of my major gripes with the Six Sigma program as ‘taught’ in the companies I worked for. Heck my state had to pass a law making it mandatory kids are taught the multiplication tables!
As for “the generalized assault on the social sciences and the use of models misunderstands the most basic aspects of the philosophy of science” that is what happens when you intentionally use bad ‘science’ and suborned scientific societies to shove bad political programs down peoples throats.
If and when the general public finds out they have been intentionally misled by ‘scientists’ the backlash against science could get rather nasty and that is the real crime committed by ‘Climate Scientists’ (Aside from all the human deaths they have caused that is.)

Gail Combs
October 2, 2013 5:56 pm

Martin Clark says: @ October 2, 2013 at 3:02 pm
stan stendera says:
“Brad and Tom G: You should both write books!!!!”
I second that request…. I have done a fair amount of modelling. Models have now become a pandemic worse than swine flu.
>>>>>>>>>>>>>>>
I rather see them write short essays like they just did so they can be plastered across the internet and in letters to the editor. (The seven second rule. )
As far as models and modelers go… well now we know what all those would be horse traders do for a living in the modern age.
Yes models are useful but only when used for a better ‘Dig Here’ guess. Without validation with real world data they are of no more use than fairy tales… Well actually of less use.

October 2, 2013 5:56 pm

No wonder Mann wants to sue everyone. His integrity has been defined.

observa
October 2, 2013 6:01 pm

We don’t care about no steenking statistical underpinnings, just the HEADLINES-
http://www.adelaidenow.com.au/news/breaking-news/september-hottest-on-record/story-fni6ul2m-1226732022605
bearing in mind whitefellas only rolled up seriously in Gondwanaland in 1788, Adelaide itself was only founded in 1836-
http://en.wikipedia.org/wiki/Adelaide
and a reasonable network of Stevenson Screens was only rolled out throughout Australia around 1910, but never let the paucity of an historical data record get in the way of a ripping good headline here scary folk.
The priceless thing about the self appointed climatology club is they’ll happily wave away any early settler temp records and heat wave horror stories and then in the next breath tell you how important it is to recognise the thousands of years of aboriginal settlement and their Dreamtime stories.

kuhnkat
October 2, 2013 6:39 pm

iid sample
In probability theory and statistics, a sequence or other collection of random variables is independent and identically distributed (i.i.d.) if each random variable has the same probability distribution as the others and all are mutually independent.[1]
http://en.wikipedia.org/wiki/Independent_and_identically-distributed_random_variables

RACookPE1978
Editor
October 2, 2013 6:40 pm

What is the “real data” that these much-vaunted global Models supposedly generate? I have NEVER seen a plot or graphic of their output winds, temperatures, ice areas, oceans and land areas EVER.
nor have we been told what the original (0,0) conditions are for their fixed variables: % land, % ice, locations of the land and sea masses, original currents, original temperature and solar radiation simulations, original land and water albedo simulations, etc. We are only told of the one-line output.
For example, summer and winter land albedoes are going to change (going to get darker! and cause the land to absorb more energy over longer periods of the year) as the increased CO2 DOES increases growth of every plant, tree, shrub, grass, and plankton on the planet by 12 to 27% fro 1970 through today’s more lush growth.
Is this 1% – 2% or 4% decrease in land albedo included as a function of time? .
We are told that “land use” changes (cutting trees) is some 15% of the man-caused increase in CO2 – but where is that estimate justified by data worldwide?

observa
October 2, 2013 7:20 pm

Your average HS student with internet skills could put these catastrophist chicken littles into perspective starting here-
http://en.wikipedia.org/wiki/List_of_disasters_in_Australia_by_death_toll#cite_note-39
Now bear in mind that’s for a whole continent full of just over 23 mill people nowadays-
http://www.abs.gov.au/ausstats/abs%40.nsf/94713ad445ff1425ca25682000192af2/1647509ef7e25faaca2568a900154b63?OpenDocument
But looking down that Wiki tragedy list including those various killer heat waves/bushfires and natural disasters you need to bear in mind in 1910 our popn was only 4.5mill and climbed to 7 mill at the end of WW2-
http://www.populstat.info/Oceania/australc.htm
Now to really put all that climate related death and the chicken littles into perspective here kiddies-
http://www.bitre.gov.au/publications/ongoing/rda/files/RDA_Aug_2013.pdf
note that during the 12 months ended August 2013 there were 1,265 road deaths nationally so be VERY VERY afraid when mum and dad want you to hop in the car to get to school!

r murphy
October 2, 2013 7:44 pm

I am always enlightened after reading one of Dr. Ball’s essays. His down home, common sense commentary on the climate field is plain refreshing, thanks.

policycritic
October 2, 2013 7:50 pm

Thanks, Dr. Ball.

kuhnkat
October 2, 2013 7:51 pm

Moshpup,
do you ever get tired of making incorrect assertions??
“Wrong. there is no such thing as a statistically significant sample.”
Even eHow has got you:
http://www.ehow.com/how_6967270_calculate-statistically-significant-sample-size.html
HAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHAHA
Maybe you could just hire this guy for help:
http://www.statisticallysignificantconsulting.com/SampleSize.htm
or
http://www.amnavigator.com/blog/2010/10/27/how-to-calculate-statistically-significant-sample-size/

wayne
October 2, 2013 7:59 pm

Read all about it!
THE NEW AND IMPROVED IPCC AR5 REPORT
Propaganda with raw fear at its very finest.
Don’t wait, get your copy while they’re bright red and hot!
Use to scare the daylights out of your fellow comrades with ease!
Doubts? Just ask the environmentalist on your block for instructions.
IMPORTANT: Never let mere data or observations lead you astray,
all must first pass through an official UN.IPCC statistical filter.
Enjoy!
/sarc off

wayne
October 2, 2013 8:05 pm

Now why did I put /sarc off at the end of that? That’s not right.

October 2, 2013 8:43 pm

Kuhnkat,
There is no general statistically significant sample size. It is subject to and conditional upon the variance of the data and the question you are asking.
“30” is a rule of thumb, but you should know from polling for example that 30 people would not work to predict a presidential election. now would it
To repeat. 30 is not a magic number. Sometimes you can have fewer samples and sometimes you need more. There is no such thing as a canonical sample size for statistical significance. It depends. its conditional. Notcarved in stone, not in dropped from heaven. it depends.
Ask yourself what the sample sizes are for particle physics.. you wont get 99.99% from 30 samples
Ask yourself about 6 sigma practices.
The bottomline is that 30 years was not selected because of the reason Ball asserts. In fact, you can find a discussion about this in the climategate mails.

October 2, 2013 9:07 pm

As they say in many parts of the Middle East, and this part of Canuckistan, “Shoe-kran, Dr. Tim!”

Steve Obeda
October 2, 2013 9:13 pm

I have three things to say, and I will say them. Now.
A drunk wanders down the middle of the road. On average, he is fine.
Even as an undergrad, it became apparent to me that the social sciences were for the statistical lightweights.
Climatologists seem to want everyone else to stfu because they’re not members of the high priesthood of climatology. Well, to a great extent this is just applied statistics. No great mystery.

RACookPE1978
Editor
October 2, 2013 9:20 pm

So, educate me here please:
The “average of the average of all of the world’s published (not measured or actual but “published”) average temperature anomalies” over 30 years is NOT an independently sampled statistical “number” by any means, right? So, though each year’s “average temperature anomoly” could be assigned an error band, why would somebody want to periodically “rest” the client for yet another 30 years? Are they not continuously trying to flat-line a ever-rising periodic wave ?
By definition, the global “average temperature anomaly” IS going to be changing over time, so understandably, we have to have some reference point. Should not that single reference point be assigned, then fixed?
Given that there is a very visible 55 year ACI variation over time (or 60, or 65, and now some writers are claiming 88 and 100 year periods!) all on top of a undeniable 800 year long-term rise and fall from the Roman Warming to the Dark Ages to the Medieval Warming Period to the Little Ice Age to today’s Modern Warming Pause …. Should not the climate community at least recognize that their temperature trends are NOT simple linear models, but have to include a period cycle that MAY LIKELY BE influenced by a linear CO2 increase on top the original trend?
Yet they seem zealously and emotionally fixated on projecting a simple linear trend out for 100 years, as this 30 year ploy of accepting their own models shows.

October 2, 2013 9:57 pm

I like Cork Hayden’s comment : The average American has one breast and one ball .
From my perspective I see “climate scientists” following the path of “social scientists” learning ever more esoteric statistics and arguing endlessly over data sets .
But never having learned the most basic physics and its classical analytical approach . It appears to me that there are career “climate scientists” on both sides of the “debate” who have never even learned how to calculate the temperature of a radiantly heated colored ball .
Until that becomes an absolute requirement for any undergraduate presuming to enter the field , along with the return to the teaching of the analytical approach of comparable branches of applied physics , this decade upon decade of near total stagnation will continue .
Another of Cork Hayden’s aphorisms : If it were science , there would be 1 model instead of 30 ( now 73 ) .

Brian H
October 3, 2013 2:03 am

Dr. Ball,
Your title reminded me of an analogous article, in the Atlantic: Lies, Damned Lies and Medical Science. When big bucks are available, human ingenuity devotes itself to acquisition.

SteveP
October 3, 2013 2:38 am

Some great quotes on statistics. My favorite statistics quote is by Homer Simpson:
‘Kent Brockman: Mr. Simpson, how do you respond to the charges that petty vandalism such as graffiti is down eighty percent, while heavy sack beatings are up a shocking nine hundred percent?
Homer Simpson: Aw, you can come up with statistics to prove anything, Kent. Forty percent of all people know that. ‘

oMan
October 3, 2013 2:47 am

Dr. Ball: thanks, a wonderful essay. Better yet, it prompts some outstanding comments and mini-essays. In particular I want to echo the praise from “chris y” and “stan stendera”–
“chris y on October 2, 2013 at 1:37 pm
stan stendera says:
October 2, 2013 at 1:25 pm
“Brad and Tom G: You should both write books!!!! Among the all time great comments on WUWT, and that’s saying something.”
I agree. These are absolutely excellent comments.
I encourage both of you to consider putting together a post for Anthony. Your writing styles are similar to Willis Eschenbach, whose posts I enjoy immensely!”

RC Saumarez
October 3, 2013 4:17 am

@rgb
You have outlined very eloquently the problems with statistics and climate modelling. I am simply astonished by the errors made in basic probability in justifying models. The problem is when you have trashed the obvious non-performers and you are left with something that approximates real data. Is this simply a fortuitous choice of initial conditions and parameters or does the model really work? I cannot see any way of testing this except through prospective testing. In other words, there is a hypothesis that looks as though it might be right but it has not been tested.
Most of the models seem to be no more accurate than a first order Taylor’s expansion. Unfortunately for the IPCC, if the reference point is moved backwards, so giving a wider range of estimates that include current data, the second order term is probably significant and how you extract the second derivative of the temperature signal with any confidence is beyond me.
The point about auto-correlation is very important, since many models give wildly different ACFs to real data. The ACF is important in model identification and seems to be roundly ignored, which it should not be.

Paul Vaughan
October 3, 2013 5:02 am

RC Saumarez (October 3, 2013 at 4:17 am) wrote:
“I cannot see any way of testing this […]”
Step back from the univariate view to see other variables, some of which are rigidly constrained. If you really cannot see this, you definitely need jolting perspective change. Step One: Sober Up. Global average temperatures alone do not determine large scale terrestrial flows, which are constrained by the law of conservation of angular momentum in earth rotation summaries. A specification of the spatiotemporal distribution of temperature gradients is needed. When attractors in well-constrained flow data are taken into account, the set of permissible ensembles is reduced by an order of magnitude (or an order of idiocy in YEP’s translation). This can be proven geometrically. This is the natural extension of the work of Lamb & Leroux mentioned by Tim Ball.
Universal Fluctuations in Correlated Systems
http://arxiv.org/abs/cond-mat/9912255
“Self similarity is an important feature of the natural world. It arises in strongly correlated many body systems when fluctuations over all scales from a microscopic length a to a diverging correlation length ξ lead to the appearence of “anomalous dimension” [1] and fractal properties. However, even in an ideal world the divergence of of ξ must ultimately be cut off by a macroscopic length L, allowing the definition of a range of scales between a and L, over which the anomalous behaviour can occur. Such systems are found, for example, in critical phenomena, in Self-Organized Criticality [2,3] or in turbulent flow problems. By analogy with fluid mechanics we shall call these finite size critical systems “inertial systems” and the range of scales between a and L the “inertial range”. One of the anomalous statistical properties of inertial systems is that, whatever their size, they can never be divided into mesoscopic regions that are statistically independent. As a result they do not satisfy the basic criterion of the central limit theorem and one should not necessarily expect global, or spatially averaged quantities to have Gaussian fluctuations about the mean value.”
That’s what makes earth orientation parameters such indispensable climate indicators.
Macroscopic control parameter for avalanche models for bursty transport
http://arxiv.org/abs/0806.1133
“It is increasingly recognized that a large group of physical systems can be characterized as driven, dissipative, out-of-equilibrium and having a conservation law or laws (see the comprehensive treatments of [1, 2]). They usually have many degrees of freedom (d.o.f.), or excited modes, and long range correlations leading to scaling or multiscaling. Two examples are fully developed turbulence (see e.g. [3, 4]) and Self-Organized Criticality (SOC, [5, 6, 7]).
[…]
Our focus here is then to establish the macroscopic similarities and differences between turbulence and SOC in the most general sense. A central idea in physics is that complex and otherwise intractable behavior may be quantified by a few measurable macroscopic control parameters. In fluid turbulence, the Reynolds number RE expresses the ratio of driving to dissipation and parameterizes the transition from laminar to turbulent flow. Control parameters such as the Reynolds number can be obtained from dimensional analysis (see e.g.[3, 56]), without reference to the detailed dynamics. From this perspective the level of complexity resulting from the detailed dynamics is simply characterized by the number N of excited, coupled d.o.f. (or energy carrying modes) in the system. The transition from laminar to turbulent flow then corresponds to an (explosive) increase in N. The nature of this transition, the value of the RE at which it occurs, and the rate at which N grows with RE all depend on specific system phenomenology. Dimensional arguments, along with the assumptions of steady state and energy conservation, are however, sufficient to give the result that N always grows with RE (as in [57], see also [3].
We anticipate that an analogous control parameter for complexity, RA, will exist for the wider group of systems discussed above. Interestingly it is now known that such a control parameter that expresses the ratio of driving to dissipation does indeed exist for SOC. In this paper we will give a prescription to obtain RA generally from dimensional analysis, that is, without reference to the range of detailed and rich phenomenology that any given system will also exhibit.”

Editor
October 3, 2013 6:05 am

Thanks, Tim. I enjoyed that.
Regards

Solomon Green
October 3, 2013 6:09 am

Steven Mosher says,
“The bottomline is that 30 years was not selected because of the reason Ball asserts. In fact, you can find a discussion about this in the climategate mails.”
Could Mr. Mosher tell us why 30 years was selected, please. Otherwise some might suspect that it was because at the time that it chosen it provided the steepest incline in global temperatures.

Mindert Eiting
October 3, 2013 6:24 am

I have no problems with models if these are used for testing simple hypotheses. Let a certain sensitivity be the hypothesis. Do a thousand model runs with random initial conditions. Note per run the slope of linear regression for the past 17 years. Compute the mean and variance of all obtained slopes. If the real slope deviates more than 2 sigma from the model mean, the hypothesis should be rejected. Results can be reported on one page. It is rather suspect that the IPCC needs 2000 pages for telling their story.

October 3, 2013 6:47 am

Guest essayist by Dr. Tim Ball said,
. . . First, was the simplistic application of statistics beyond an average in the form of a straight-line trend analysis: Second, predictions were given awesome, but unjustified status, as the output of computer models. They wanted to show we were heading for disaster and selected the statistics and process to that end. This became the method and philosophy of the IPCC. . . .
. . .
There is another germane quote by mathematician and philosopher A.N. Whitehead about the use, or misuse, of statistics in climate science.

There is no more common error than to assume that, because prolonged and accurate mathematical calculations have been made, the application of the result to some fact of nature is absolutely certain.
. . .

– – – – – – – –
Why did the IPCC choose models as the locus of their arguments? Was their decision an objective one or subjective one?
The corpus of A. N. Whitehead’s philosophy of science was influential to a following generation of scientists via his extensive treatment on the conception of objectivity.
From a philosophy of science basis, was the IPCC’s choice explicitly justified in peer reviewed literature or was it based on an inherited, and presumed to be correct, paradigm of science?
The later justification, if found to be the case, is a central strategic critique point for skeptics.
John

Jeff Alberts
October 3, 2013 7:10 am

This identifies the problem that has long plagued the use of statistics, especially in the Social Sciences, namely the use of statistics without knowledge or understanding.

I disagree. I think the Team understood exactly what they were doing, especially Mann. They just (rightly) assumed the general public’s eyes would glaze over when the “errors” were made public.

more soylent green!
October 3, 2013 8:02 am

A couple of things I want to emphasize”

He claimed that because they all showed warming they were valid. Of course they did because they were programmed to that general result.

Absolutely spot-on. Computer programs output the results they do because they were created to do so. Computer models do not output facts. Computer models to not output data.

Second, predictions were given awesome, but unjustified status, as the output of computer models.

This highlights one of the many, many instances of double-talk and dissembling on the part of the alarmists. They scream “the sky is falling, the sky is falling” because of the model results. In the next breath they tell you the models used by the IPCC don’t make predictions.
Repeat after me, “Computer models do not output facts. Computer models to not output data.”

Unfortunately, people started using the Normal for the wrong purposes. Now it is used as the average weather overall. It is only the average weather for a 30-year period. Actually it is inappropriate for climate because most changes occur over longer periods.

This is a pet peeve of mine. I sometimes talk back to the TV weatherman when he/she says today’s weather is warmer/colder/wetter/drier/etc, than normal. “The high today will be 86, which is above normal.” What they mean to say is the high is above the 30 year average for today’s date. To get a real idea of what normal is, look at the record high and record low for a date. That range defines the normal range, which can be quite variable.

Alan Robertson
October 3, 2013 8:18 am

That was a very interesting read, Dr. Ball. Thank you.

RC Saumarez
October 3, 2013 8:46 am

@Paul Vaughn
Thank you for jolting me out of my complacency. I am sure that all you have quoted is highly relevant.
I simply said that if you ditch the obviously non-performing models you are left with ones that seem to conform to observations. By selecting this subgroup of models one has formed a hypothesis. Remembering the statistical adage that you cannot test a hypothesis with the data used to construct it, these models have to be tested with fresh data, in other words seeing how they predict the future.

richardscourtney
October 3, 2013 9:04 am

RC Saumarez says at October 3, 2013 at 8:46 am

I simply said that if you ditch the obviously non-performing models you are left with ones that seem to conform to observations. By selecting this subgroup of models one has formed a hypothesis. Remembering the statistical adage that you cannot test a hypothesis with the data used to construct it, these models have to be tested with fresh data, in other words seeing how they predict the future.

Yes! That is right and it needs to be understood. But many don’t understand it.
For those who do not know why it is true, see
http://en.wikipedia.org/wiki/Texas_sharpshooter_fallacy
Richard

Bill Taylor
October 3, 2013 9:16 am

the climate is simply the average weather over the previous 30 years…….just like a baseball players batting average is his PAST actual record at the plate…..his average exerts NO control on any level over what happens in his next at bat….a player that has never hit one home run in his life COULD hit one his next time up, and the all time leader could strike out his next time up…….those claiming the climate exerts any control over the weather are showing incredible IGNORANCE.

aequitas45
October 3, 2013 10:12 am

Vukcevic says:
“For some years now, I have looked into changes of the Earth’s magnetic field and noticed easily observed ‘apparent correlation’ with the temperature records.
For time being the science indicates that the only way these changes may make an impact is through the cosmic rays nucleation process, but it doesn’t support the Svensmark’s hypothesis since the Earth’s and solar field appear to have negative ‘correlation’:”
I have seen the same anti correlation and interpret it as another case of two opposite acting forces (the solar and geomagnetic field) that contribute to the stability of the earths climate through a governoring effect on Svensmark’s hypothesis of cosmic ray nucleation process. This is actually necessary to match the temperature anomaly record.

more soylent green!
October 3, 2013 10:15 am

Does everybody know Barrack Obama never signed the Kyoto Treaty? It’s true, it’s true.

October 3, 2013 10:15 am

I plead guilty to using SPSS for my MSc in 1985 (i am getting old!) and getting it wrong.

Sun Spot
October 3, 2013 10:57 am

Dr. Ball, if you can divulge the name “Eventually a man threw his shoe on the stage. who was the man that went on the stage ?

DaveGinOly
October 3, 2013 1:21 pm

My theory of climate change/global warming – painting houses yellow creates global warming.
I can test the theory by creating a computer model that adds heat to the atmosphere every time a house is painted yellow. Run the simulation, with large numbers of simulated houses painted with simulated yellow paint, and – the simulated atmosphere gets warmer!
Not very surprising results. Its pretty obvious that if you create a simulation in which heat is driven by a factor in the simulation, an increase in that factor will yield more heat – even if that factor is yellow paint. And, naturally, you can substitute CO2 for yellow paint; it will work just as well. In the end, the simulation proves nothing.

rgbatduke
October 3, 2013 2:15 pm

Steven Mosher says,
“The bottomline is that 30 years was not selected because of the reason Ball asserts. In fact, you can find a discussion about this in the climategate mails.”
Could Mr. Mosher tell us why 30 years was selected, please. Otherwise some might suspect that it was because at the time that it chosen it provided the steepest incline in global temperatures.

Well, except for the interval between 1910 and 1940. In fact, the incline in GASTA from 1910-1940 almost exactly matches the incline in GAST from 1980 to 2010. Dick Lindzen is fond of presenting e.g. HADCRUT4:
http://www.woodfortrees.org/plot/hadcrut4gl/offset:288
broken into 1900-1950 and 1950-2000 without numbers but on identical vertical anomaly scales side by side and asking people to identify which one is which, which incline is associated with CO_2 and which one isn’t. It ain’t easy.
I also recommend looking at:
http://www.woodfortrees.org/plot/hadcrut4gl/detrend:0.7
Note that this subtracts a trend of 0.7/165 = 0.43 degrees per century from the anomaly data. What’s left looks like a sinusoid with a period approximately equal to the period of the PDO, that is SO CLOSE to being flat — maybe there is a bare 0.2 C that could be attributed to the entire “anthropogenic era” of increased CO_2 in the second half of the 20th century. Could be, or not, because the mid-1800s exceeded this trend line by just as much without CO_2, and we don’t understand this graph or its underlying causality.
rgb

Paul Vaughan
October 3, 2013 7:43 pm

@RC Saumarez (October 3, 2013 at 8:46 am)
My concern is that the modelers never draw attention to multivariate model output. The FIRST thing I want to see is their global AAM (atmospheric angular momentum) output. WHY DO THEY HIDE IT? Because they know that if I analyze it I can PROVE BEYOND ALL SHADOW OF A DOUBT that their model output is TOTAL BS TOTAL BS TOTAL BS in the most egregious sense possible. Like the sharpest razor blade possible, cutting effortlessly through something that senses nothing
Demand that they report global AAM for all models. I dare them to….
Best Regards!

TonG(ologist)
October 3, 2013 8:32 pm

Thanks to all for your kind and generous comments and your request for “more of the same”. If you want to stray a little off topic I do have a blog which addresses timely topics in the news pertaining to geology. I don’t want to post the link here without permission from Anthony and the moderators, as it would be a shameless plug to drag the audience from here.
Back on track, however, I have been considering a more concentrated blog about the reckless waste inflicted by EPA on the American taxpayer and consumer since the enactment of CERCLA in 1980. It would be an insider’s perspective with ample examples (intentional word play there). I have refrained until now because until this past January, I was still representing clients in that regulatory arena and couldn’t risk it. Now however, I am one of those lying, untrustworthy, evil, polluting, “frackers”.(Those who know my comments also know how hard it is for me to write it that way). Now that I don’t deal with CERCLA any longer, and before I lose my insider’s edge I might begin that blog now.
Cheerio.
Tom

October 4, 2013 1:53 am

I find it strange that within the study of and use of statistics, for the purpose of analyzing a dataset, that we do not have an agreed methodology.
Not being versed in stats, I can not comment well, however, there are many parallels with other professions which I do work within: software and hardware design.
In both of these skill areas, a number of competed paradigms can be used, some favorites become displaced by ‘new kids on the block’ for clear and understandable reasons, they do a better job at achieving your intended outcome.
By now I would have expected, perhaps half a dozen or so, methods or techniques would be vying to be ‘the best’ way to analyse time series datasets.
However, we have endless, heated debate, not about which method, but how to implement a method for what is ‘standard data’.
Is it not possible to expect a few good statisticians to get together and hammer out a few strong methods that people could agree on because they will be able to ‘see’ the value of each method?
Or am I misunderstanding the maturity level of applied statistics today?
As an ex software engineer/hardware designer I would like to be able to find the most appropriate method to use to find a trend in time series data.
It is disconcerting that I can not.

markx
October 4, 2013 1:59 am

TonG(ologist) says: October 3, 2013 at 8:32 pm
Back on track, however, I have been considering a more concentrated blog about the reckless waste inflicted by EPA on the American taxpayer and consumer since the enactment of CERCLA in 1980. …… Now that I don’t deal with CERCLA any longer, and before I lose my insider’s edge I might begin that blog now.
Anything which brings a public focus onto the regulatory monster that the EPA has become would be a wonderful thing.

markx
October 4, 2013 2:09 am

DaveGinOly says: October 3, 2013 at 1:21 pm
“…Its pretty obvious that if you create a simulation in which heat is driven by a factor in the simulation, an increase in that factor will yield more heat …”
I personally believe you are exactly correct Dave, and that it is that simple:
If one of your parameter changes is that heat is added to the system, and there are no major negative feedbacks, the system must warm, no matter how many hoops your computer program jumps through.

rgbatduke
October 4, 2013 7:52 am

Anyway some commenters notably AlecM (posts on WUWT as AlecMM) suggested looking at Hansen_etal_1981 (if you Googled this it would come up). AlecM believes that Hansen has miscalculated the greenhouse effect and that more focus should be on LTE in the atmosphere (sorry for the paraphrase). What struck me was that Hansen attributes 90% of warming to the 15 micron band -”downwelling radiation” – in essence that Co2 is the main driver of climate. There is a reference for this pointing to an earlier paper, which then relates to Manabe and then finally to the first recent era estimation of atmosphere heating due to greenhouse gases – by Moller.
I think I’ve already read the paper in question, and indeed consider its figure 7 to ge the direct precursor to figure 1.4 in AR5. I spent a quiet, happy day a couple of years ago attaching the UAH LTT onto the Hansen’s graph to note that the entire warming of the 1980s is solidly within his estimate of “natural variation” (in spite of the fact that it curves downward for no reasons he could possibly justify given the mean warming without CO_2 of the previous two centuries) and is already solidly distinguishable from his three climate sensitivity predictions.
Note well that he doesn’t actually claim that the CO_2 is the main driver of the climate — the whole point of his “climate sensitivity” estimates which were and continue to be the “catastrophic” part of the problem is that he assumes that the warming due to CO_2 alone will be multiplied by a factor of 2 to 5 by water vapor. IIRC, he describes this in some detail in the paper, and back in 1981, at least, he was still somewhat uncertain about its feedback except that it would definitely be positive, hence he posted the three curves that would produce anywhere from 1.4 to 5.6 C of warming by 2100. 1.4 was “CO_2 alone”.
In 1981 it was basically impossible to compute anything but nearly trivial single slab models — 1D spatiotemporal average reductions of the Earth and atmosphere and oceans. The Cray 1 owned by NCAR at the time had 8 MB of core (1 million 64 bit words) and operated at a whopping 80 MHz, and on a good day could run at around 100 MFLOPS The computer I’m typing this reply on, an aged laptop with an Intel CoreDuo with 4 GB of RAM and running 2.5 GHz, has more effective computational power than all the supercomputers on Earth put together at that time. Hell, my CELL PHONE has more compute power, by far, than any supercomputer of that day. It would have been considered a munition with export restrictions at the time lest the Evil Empire use it to design nuclear bombs. (There’s an app for that:-).
However, that was then, this is now. Grant Petty pointed out in another conversation I’ve been participating in (that I’ve just retired from, as it was with the PSI whackos and going nowhere as usual) that anyone can go grab the source for e.g. CAM 3.x, a genuine open source GCM (written, sadly, in Fortran90) and perhaps more usefully, access and read its documentation. To the extent that CAM 3.x is typical, it is no longer the case that GCMs are in any sense “simple” single slab models. The partition the atmosphere up into many layers, over an (IMO unbelievably stupid) regular but weakly adaptive latitude/longitude decomposition of the Earth’s surface into cells. CAM 3 at least does a single slab ocean (basically only considering the surface, no real deep ocean heat transport) but it very definitely has a radiation physics engine that is a lot more sophisticated than Hansen’s 1981 paper. It also devotes a considerable amount of code to handling e.g. the water cycle, to aerosols, etc.
The point is that most of the criticisms of GCMs (that I’ve read on e.g. WUWT) are levelled at not the actual structure of the GCM, but at a sort of “toy reduction” of the GCM into a 1 dimensional single slab model, as if this were the mid-1950s and computers were a piece of paper and a pencil (and maybe a mechanical adding machine off to the side or a slide rule to help with the heavy lifting). They overestimate this, underestimate that. I’m guilty of this myself — we naturally try to simplify a complex thing so that we can discuss it at all, because believe me — I’ve LOOKED at the CAM 3.1 code and it is a poorly documented mess in my rather professional opinion (I teach programming among other things, and have written a few zillion lines of code on my own at this point in my life, starting with a plastic mechanical adder that worked with marbles used to activate plastic marble-track flip-flops that one hand programmed to add two numbers or do a simple binary multiplication back in 1965. I was 10. It was cool. (The CAM 3.1 TOPLEVEL documentation I refer to above is actually pretty good though. It’s just too far removed from the code itself.)
GCMs are not really susceptible to this sort of reductionist consideration, although we try. In the end, there are thousands of lines of code, many megabytes of initialization data, hundreds of control variables, hundreds of separate physical process computations controlled by those variables and coupled by a dynamical engine that has to manage everything from bulk mass transport to latent heat to parcel-to-parcel radiation.
Every single bit of this is subject to a rather staggering list of possible errors.
The code could have bugs. In fact, by the time you have this many lines of code, the code DOES have bugs, I promise — probability is near unity that there are non-fatal bugs. Having written code with deep, functional bugs before, they are by far the most difficult thing to expunge, because the only way to know they are there is by comparing the result from the computation to a result you know some other way. If the bug produces otherwise “sensible” (but incorrect) results you’ll literally have no reason to think it is there. I’m talking things as simple as typing N instead of M in a single line of code in a module where both make sense and are used, both have values with identical ranges, and where using N instead of M produces a perfectly sensible result. I had just this sort of bug in a huge program I wrote 20 years ago in Fortran and the only way I knew it was there is because the output had to satisfy a certain scaling and although it had exactly the right form (was sensible) it failed the scaling test. I was lucky — the program could have otherwise run for years producing wrong answers.
The code is implemented on a ludicrous grid. I’m serious, who in their right mind would perform NUMERICAL quadratures on a spherical manifold on a spherical polar grid? Did I mention that the bug above in my own code was in a module designed to perform quadratures (spherical harmonic projections) on a spherical polar grid? It seems so easy, so sensible — spherical polar coordinates are the natural analytic representation of a sphere, after all — until you hit that pesky Jacobean, the map problem. If you divide latitude and longitude into equal interval cells that are say 5×5 degrees they are nearly square cells 350 miles to the side at the equator and thin slices 350 miles north to south that you can STEP across near the poles. Nearly all quadrature methods rely on breaking a continuum up into some sort of a regular grid so one can do nice, simple polynomial fits across them, and while there are kludges — renormalizing the longitude grid size at some latitude, creating a complex map to account for the ever varying area of the cells, the ever varying VOLUME of the PARCELS you are trying to transport between the cells — all I can say is “eeeewww”. Been there, done that, not the right way.
The “right way” is to use a symmetric, uniformly scalable icosahedral tesselation of the sphere:
http://i.stack.imgur.com/Cyt2B.png
or better yet:
http://www.mathworks.com/matlabcentral/fx_files/37004/3/screen_shot.jpg
By being uniformly renormalizable/scalable, one can systematically run the code at different spatial resolutions without ANY tweaking and actually determine of the result is spatially convergent! The price you pay, of course, is that you have to hand re-code and test all of your quadrature and ODE solution and transport routines, because they are all DESIGNED FOR A RECTANGULAR PLANAR GRID! Adaptive quadrature on a sphere in polar coordinates, in contrast, just doesn’t work BECAUSE one cannot really properly handle the cell shape distortion — the map problem where antarctica is rendered huge when the world is presented as a large rectangle.
I frankly doubt that anybody in climate science has any idea how much error is introduced by this. CAM 3.0 contains a comment that when they DID renormalize the cell sizes near the poles, the code sped up and maybe got more accurate (understandable because by the time your grid is barely adequate for the equator, you’re parsing cells you could throw a softball over at the poles). At the very least it costs compute time, hence precision.
And then there is the physics. CAM 3.1 integrates time in hour long timesteps. Each hour, a staggering amount of physics per cell has to be updated on the basis of mass and energy transport. In each case the physical process is known but has to be approximated/averaged and renormalized to correspond to the (distorted, non-uniform) cells. The averaging ALREADY “integrates” over all shorter time/length scale processes, with the catch being that we do not know and cannot measure whether or not we are doing that integral correctly (the only way to do so is to use the GCM code at ever finer length/time granularity and demonstrate some sort of convergence using NON-approximated code). All of these processes are coupled by the coarse-grained computation, but have to be computed SEPARABLY in the step computations (otherwise you’d have to have radiative code in the middle of your latent heat code in the middle of your mass transport code in the middle of your albedo code) so that all sorts of dynamically coupled processes are again necessarily handled semi-heuristically as if they are independent and separate at the finest grained timescale used.
I’m not saying they are doing this wrong, only that it is essentially impossible to be certain that they are doing it right any other way besides COMPARING THE PREDICTIONS OF THE CODE TO REALITY!
This is the fundamental rule of predictive modeling in all venues. When you write a complex program to predict the future, you can claim anything you like about how fabulous, accurate, complete, well-written your code is. It can match the recent past (the training set) perfectly well, it can hindcast the remote past (where it is known) perfectly well (although AFAIK, NO GCM can hindcast the secular variations of any significant portion of the climate history across human historic, let alone geological timescales) but if it does not actually predict the future, it doesn’t work and your assertions of fabulosity and accuracy just aren’t to be taken seriously.
This is true one model at a time. The mean of three models that fail to predict the future one model at a time is not expected to do better than any of the three models — indeed, it CANNOT do better than the BEST model (think about it!) and one is better off using the best model, and better off still working on improving the models and not asserting that they work when they manifestly don’t.
This SHOULDN’T be all that big a deal. If your code doesn’t work, you work on it until it does. Again, a simple, obvious rule of computer programming. Your code can pass all sorts of internal tests and still be broken. Another simple rule of computer programming, and the reason that real code goes through alpha and beta test phases. In computerspeak, for all that it is “version 3”, cam 3.x is IMO no better than beta code — it is past alpha in that “it runs”, but is very definitely at the stage where one has to actually test it against reality, which is what beta testing IS. Does your game actually work when the consumer gets their hands on it, or does it lock up when you jerk the joystick and press fire three times within 1 second? My wife has a genius for revealing hidden flaws in supposedly release-ready computer code — a lot of times problems are revealed by somebody who uses it a way the designers didn’t anticipate and hence activate a completely untested pathway (one that often shows, when resolving the bug, that even in the supposedly TESTED pathways it isn’t working right, producing sensible but wrong answers or activity).
If I had a ten year grant for (say) a million dollars a year for salary, staff, computers, indirect costs, I’d personally tackle a complete white-room rewrite of e.g. CAM 3, taking its documentation and references only and re-implementing the whole thing in C or C++ on top of a scalable truncated icosahedral grid, starting by inventing and perfecting adaptive quadrature and interpolation routines, continuing by building a scalable, tessellated world map (needed to get height above sea level, surface albedo, nature of the surface — forest, rock, desert, ocean, ice — etc. right) and then slowly, gently, add physics on top of it, probably redoing the dynamical engine (which is likely to be necessary on top of the new grid). It would actually be fun, I think and would very likely take the rest of my professional career to accomplish. I might even throw in a few things I’ve learned and coded that can do EMPIRICAL nonlinear function approximation — neural networks, other sorts of Bayesian predictors — that I’d bet a substantial amount of money could easily be trained to beat the physics-based code. I’ve fantasized about building an empirically trained climate predictor (neural network) on an icosahedral grid for some time now. A network would discover things like the linear trend and simple nonlinear functional relationships that modify it quite quickly (and without ever “identifying” it as an actual linear function) — the really interesting thing would be seeing if it could actually predict trial set behavior outside of the range exemplified within the training set, the “challenge” of ALL predictive models. Interpolation is easy, extrapolation difficult.
rgb

October 4, 2013 8:45 am

Now here’s come good news. National Groundwater Assoc. reports that 94% of the EPA has been furloughed. Wouldn’t it be lovely if this worked out to be like a baseball strike? After living without them long enough, we could all learn that we could live without them.

RACookPE1978
Editor
October 4, 2013 10:10 am

rgbatduke says:
October 4, 2013 at 7:52 am
THANK YOU! (Can I find a bigger font? 8<) )
More seriously, should not the rock, ice, granite, copper, bronze, brass, iron, silver, and final gold standards be the "test" against which we measure the value of yesterday's and today's Global Circulation Models before we accept tomorrow's Global Circulation Models. (Granted, building a valid model has to be done – and by that I mean NOT rebuilding tomorrow's models using code copied yet agian from the 1980's.
Today's GCM began from the early computer models during Colorado's attempts to model the aerosol and sulfuric acid particles coming from extremely localized point sources inside local regions: Specifically the nickel smelter in Canada's north of Lake Erie and the smog/air particles getting trapped by the convection inversions in the Los Angeles basin. The first target was to force its shutdown during the "acid rain" scares, the second was a legitimate and very important attempt to analyze the LA Basin for smog control. The problem of CAGW starts with those "local" problems because those calculations (valid at medium latitudes of simple grid square approximations) was the foundation to "grow" those same approximations into a unified global-spanning approximation. Their error in Colorado was convincing a very, very willing group of other "enablers" that the small details of particle and vapor differential equations that can be properly integrated up from meter x meter x meter volumes into kilometer x kilometer x kilometer cubes – which can (sort of) work most of the time – into a 50,000 kilometer sphere valid not only over a single day, then week, then month, then year and multi-decade analysis of all temperatures and ground conditions.
Like the "models" of glacier ice flow: What is correct and evident and simple and measurable when looking at a meter of the "ideal" rock-ice "average" coefficients and slopes and densities and friction values is simply "integrated" over an "average" entire valley! The hills and cliffs and varying rock conditions and shapes and protuberances inside that valley (not to mention the changing slope fro top to bottom of a 100 km long valley or icecap!) are "smoothed over by one value. The equation is (almost) a valid approximation of an "ideal" glacier: but it fails because that computer model cannot predict the performance and speed and shape and melt rate and mass gain rate of any single glacier correctly for any single specific glacier in the real world.
So, how do you "test" a CGM?
The first standard should be the "rock in space": The GCM MUST be able to predict in 3D latitude and longitude at a simple 100 x 100 km grid the temperature of the entire actual moon's surface (no atmosphere, no ice, no water, no phase changes, solid simple rock types of only simple albedo that don't change over time.) They DO have to get right the heat lags and delays and surface exchange rates, the emissivities and outward long wave radiation, the rotation of the sphere, the changes in solar energy over a year's orbit, the spherical coordinate and grid problem, the basic radiation/absorption problems of albedo and emissivity. Throw in the back-of-moon (rough, mountains and crater texture, and the "front-of-moon problem of smooth "seas" and isolated white craters to prove that the various land masses and ocean areas can be modeled accurately in a heat transfer model. We have the Apollo surface temperature data spanning many decades: Why don't they show us the model that has "simple" lunar temperatures over every day of that entire period? If the GCM is "stable" then it MUST be able to "stabilize" to a continuous plot of the hour by hour temperature changes of the moon over that entire period of 5 decades.
Next: The "ice in space" – Add the ice and vapor changes: Accurately model a comet passing by the sun: Phase changes, deposit and emission changes of mass flows. various albedo black and white surfaces that change with time. Long term radiation change problems spanning centuries: See if the "ice standard" DOES predict the different comet glows over time.
Then go the "Copper and Brass and Bronze" models: Mercury, Venus, Mars. Do the GCM simulate over centuries the "no water vapor, no phase changes, no surface terrain or ice problems of those "simple" atmospheres?
THEN try the earth's simple one-layer or two layer atmosphere. This begins to test the actual oceans and ice and land mass changes of albedo over the year.

rgbatduke
October 5, 2013 2:59 pm

Yes, that does sound like a plan of sorts, doesn’t it. The closest I’ve heard of following this sort of thing is the water world test of four GCMs earlier this year. They all four converged to different answers. Of course we don’t know what the right answer is, but we can be certain that at least three of the four were wrong. But I do not pretend to know what steps modelers have taken to verify their models. One hopes they’ve done precisely what you suggest, very early on. As you say, if you cannot do the simple problems, how can you expect to do well on the difficult ones, and the Earth is difficult.
rgb