This comment from rgbatduke, who is Robert G. Brown at the Duke University Physics Department on the No significant warming for 17 years 4 months thread. It has gained quite a bit of attention because it speaks clearly to truth. So that all readers can benefit, I’m elevating it to a full post
Saying that we need to wait for a certain interval in order to conclude that “the models are wrong” is dangerous and incorrect for two reasons. First — and this is a point that is stunningly ignored — there are a lot of different models out there, all supposedly built on top of physics, and yet no two of them give anywhere near the same results!
This is reflected in the graphs Monckton publishes above, where the AR5 trend line is the average over all of these models and in spite of the number of contributors the variance of the models is huge. It is also clearly evident if one publishes a “spaghetti graph” of the individual model projections (as Roy Spencer recently did in another thread) — it looks like the frayed end of a rope, not like a coherent spread around some physics supported result.
Note the implicit swindle in this graph — by forming a mean and standard deviation over model projections and then using the mean as a “most likely” projection and the variance as representative of the range of the error, one is treating the differences between the models as if they are uncorrelated random variates causing >deviation around a true mean!.
Say what?
This is such a horrendous abuse of statistics that it is difficult to know how to begin to address it. One simply wishes to bitch-slap whoever it was that assembled the graph and ensure that they never work or publish in the field of science or statistics ever again. One cannot generate an ensemble of independent and identically distributed models that have different code. One might, possibly, generate a single model that generates an ensemble of predictions by using uniform deviates (random numbers) to seed
“noise” (representing uncertainty) in the inputs.
What I’m trying to say is that the variance and mean of the “ensemble” of models is completely meaningless, statistically because the inputs do not possess the most basic properties required for a meaningful interpretation. They are not independent, their differences are not based on a random distribution of errors, there is no reason whatsoever to believe that the errors or differences are unbiased (given that the only way humans can generate unbiased anything is through the use of e.g. dice or other objectively random instruments).
So why buy into this nonsense by doing linear fits to a function — global temperature — that has never in its entire history been linear, although of course it has always been approximately smooth so one can always do a Taylor series expansion in some sufficiently small interval and get a linear term that — by the nature of Taylor series fits to nonlinear functions — is guaranteed to fail if extrapolated as higher order nonlinear terms kick in and ultimately dominate? Why even pay lip service to the notion that or
for a linear fit, or for a Kolmogorov-Smirnov comparison of the real temperature record and the extrapolated model prediction, has some meaning? It has none.
Let me repeat this. It has no meaning! It is indefensible within the theory and practice of statistical analysis. You might as well use a ouija board as the basis of claims about the future climate history as the ensemble average of different computational physical models that do not differ by truly random variations and are subject to all sorts of omitted variable, selected variable, implementation, and initialization bias. The board might give you the right answer, might not, but good luck justifying the answer it gives on some sort of rational basis.
Let’s invert this process and actually apply statistical analysis to the distribution of model results Re: the claim that they all correctly implement well-known physics. For example, if I attempt to do an a priori computation of the quantum structure of, say, a carbon atom, I might begin by solving a single electron model, treating the electron-electron interaction using the probability distribution from the single electron model to generate a spherically symmetric “density” of electrons around the nucleus, and then performing a self-consistent field theory iteration (resolving the single electron model for the new potential) until it converges. (This is known as the Hartree approximation.)
Somebody else could say “Wait, this ignore the Pauli exclusion principle” and the requirement that the electron wavefunction be fully antisymmetric. One could then make the (still single electron) model more complicated and construct a Slater determinant to use as a fully antisymmetric representation of the electron wavefunctions, generate the density, perform the self-consistent field computation to convergence. (This is Hartree-Fock.)
A third party could then note that this still underestimates what is called the “correlation energy” of the system, because treating the electron cloud as a continuous distribution through when electrons move ignores the fact thatindividual electrons strongly repel and hence do not like to get near one another. Both of the former approaches underestimate the size of the electron hole, and hence they make the atom “too small” and “too tightly bound”. A variety of schema are proposed to overcome this problem — using a semi-empirical local density functional being probably the most successful.
A fourth party might then observe that the Universe is really relativistic, and that by ignoring relativity theory and doing a classical computation we introduce an error into all of the above (although it might be included in the semi-empirical LDF approach heuristically).
In the end, one might well have an “ensemble” of models, all of which are based on physics. In fact, the differences are also based on physics — the physicsomitted from one try to another, or the means used to approximate and try to include physics we cannot include in a first-principles computation (note how I sneaked a semi-empirical note in with the LDF, although one can derive some density functionals from first principles (e.g. Thomas-Fermi approximation), they usually don’t do particularly well because they aren’t valid across the full range of densities observed in actual atoms). Note well, doing the precise computation is not an option. We cannot solve the many body atomic state problem in quantum theory exactly any more than we can solve the many body problem exactly in classical theory or the set of open, nonlinear, coupled, damped, driven chaotic Navier-Stokes equations in a non-inertial reference frame that represent the climate system.
Note well that solving for the exact, fully correlated nonlinear many electron wavefunction of the humble carbon atom — or the far more complex Uranium atom — is trivially simple (in computational terms) compared to the climate problem. We can’t compute either one, but we can come a damn sight closer to consistently approximating the solution to the former compared to the latter.
So, should we take the mean of the ensemble of “physics based” models for the quantum electronic structure of atomic carbon and treat it as the best predictionof carbon’s quantum structure? Only if we are very stupid or insane or want to sell something. If you read what I said carefully (and you may not have — eyes tend to glaze over when one reviews a year or so of graduate quantum theory applied to electronics in a few paragraphs, even though I left out perturbation theory, Feynman diagrams, and ever so much more:-) you will note that I cheated — I run in a semi-empirical method.
Which of these is going to be the winner? LDF, of course. Why? Because theparameters are adjusted to give the best fit to the actual empirical spectrum of Carbon. All of the others are going to underestimate the correlation hole, and their errors will be systematically deviant from the correct spectrum. Their mean will be systematically deviant, and by weighting Hartree (the dumbest reasonable “physics based approach”) the same as LDF in the “ensemble” average, you guarantee that the error in this “mean” will be significant.
Suppose one did not know (as, at one time, we did not know) which of the models gave the best result. Suppose that nobody had actually measured the spectrum of Carbon, so its empirical quantum structure was unknown. Would the ensemble mean be reasonable then? Of course not. I presented the models in the wayphysics itself predicts improvement — adding back details that ought to be important that are omitted in Hartree. One cannot be certain that adding back these details will actually improve things, by the way, because it is always possible that the corrections are not monotonic (and eventually, at higher orders in perturbation theory, they most certainly are not!) Still, nobody would pretend that the average of a theory with an improved theory is “likely” to be better than the improved theory itself, because that would make no sense. Nor would anyone claim that diagrammatic perturbation theory results (for which there is a clear a priori derived justification) are necessarily going to beat semi-heuristic methods like LDF because in fact they often do not.
What one would do in the real world is measure the spectrum of Carbon, compare it to the predictions of the models, and then hand out the ribbons to the winners! Not the other way around. And since none of the winners is going to be exact — indeed, for decades and decades of work, none of the winners was even particularly close to observed/measured spectra in spite of using supercomputers (admittedly, supercomputers that were slower than your cell phone is today) to do the computations — one would then return to the drawing board and code entry console to try to do better.
Can we apply this sort of thoughtful reasoning the spaghetti snarl of GCMs and their highly divergent results? You bet we can! First of all, we could stop pretending that “ensemble” mean and variance have any meaning whatsoever bynot computing them. Why compute a number that has no meaning? Second, we could take the actual climate record from some “epoch starting point” — one that does not matter in the long run, and we’ll have to continue the comparison for the long run because in any short run from any starting point noise of a variety of sorts will obscure systematic errors — and we can just compare reality to the models. We can then sort out the models by putting (say) all but the top five or so into a “failed” bin and stop including them in any sort of analysis or policy decisioning whatsoever unless or until they start to actually agree with reality.
Then real scientists might contemplate sitting down with those five winners and meditate upon what makes them winners — what makes them come out the closest to reality — and see if they could figure out ways of making them work even better. For example, if they are egregiously high and diverging from the empirical data, one might consider adding previously omitted physics, semi-empirical or heuristic corrections, or adjusting input parameters to improve the fit.
Then comes the hard part. Waiting. The climate is not as simple as a Carbon atom. The latter’s spectrum never changes, it is a fixed target. The former is never the same. Either one’s dynamical model is never the same and mirrors the variation of reality or one has to conclude that the problem is unsolved and the implementation of the physics is wrong, however “well-known” that physics is. So one has to wait and see if one’s model, adjusted and improved to better fit the past up to the present, actually has any predictive value.
Worst of all, one cannot easily use statistics to determine when or if one’s predictions are failing, because damn, climate is nonlinear, non-Markovian, chaotic, and is apparently influenced in nontrivial ways by a world-sized bucket of competing, occasionally cancelling, poorly understood factors. Soot. Aerosols. GHGs. Clouds. Ice. Decadal oscillations. Defects spun off from the chaotic process that cause global, persistent changes in atmospheric circulation on a local basis (e.g. blocking highs that sit out on the Atlantic for half a year) that have a huge impact on annual or monthly temperatures and rainfall and so on. Orbital factors. Solar factors. Changes in the composition of the troposphere, the stratosphere, the thermosphere. Volcanoes. Land use changes. Algae blooms.
And somewhere, that damn butterfly. Somebody needs to squash the damn thing, because trying to ensemble average a small sample from a chaotic system is so stupid that I cannot begin to describe it. Everything works just fine as long as you average over an interval short enough that you are bound to a given attractor, oscillating away, things look predictable and then — damn, you change attractors.Everything changes! All the precious parameters you empirically tuned to balance out this and that for the old attractor suddenly require new values to work.
This is why it is actually wrong-headed to acquiesce in the notion that any sort of p-value or Rsquared derived from an AR5 mean has any meaning. It gives up the high ground (even though one is using it for a good purpose, trying to argue that this “ensemble” fails elementary statistical tests. But statistical testing is a shaky enough theory as it is, open to data dredging and horrendous error alike, and that’s when it really is governed by underlying IID processes (see “Green Jelly Beans Cause Acne”). One cannot naively apply a criterion like rejection if p < 0.05, and all that means under the best of circumstances is that the current observations are improbable given the null hypothesis at 19 to 1. People win and lose bets at this level all the time. One time in 20, in fact. We make a lot of bets!
So I would recommend — modestly — that skeptics try very hard not to buy into this and redirect all such discussions to questions such as why the models are in such terrible disagreement with each other, even when applied to identical toy problems that are far simpler than the actual Earth, and why we aren’t using empirical evidence (as it accumulates) to reject failing models and concentrate on the ones that come closest to working, while also not using the models that are obviously not working in any sort of “average” claim for future warming. Maybe they could hire themselves a Bayesian or two and get them to recompute the AR curves, I dunno.
It would take me, in my comparative ignorance, around five minutes to throw out all but the best 10% of the GCMs (which are still diverging from the empirical data, but arguably are well within the expected fluctuation range on the DATA side), sort the remainder into top-half models that should probably be kept around and possibly improved, and bottom half models whose continued use I would defund as a waste of time. That wouldn’t make them actually disappear, of course, only mothball them. If the future climate ever magically popped back up to agree with them, it is a matter of a few seconds to retrieve them from the archives and put them back into use.
Of course if one does this, the GCM predicted climate sensitivity plunges from the totally statistically fraudulent 2.5 C/century to a far more plausible and stillpossibly wrong ~1 C/century, which — surprise — more or less continues the post-LIA warming trend with a small possible anthropogenic contribution. This large a change would bring out pitchforks and torches as people realize just how badly they’ve been used by a small group of scientists and politicians, how much they are the victims of indefensible abuse of statistics to average in the terrible with the merely poor as if they are all equally likely to be true with randomly distributed differences.
rgb
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.
@ur momisugly [rgb]
What he said!
Here is one other big issue with models on comments sections everywhere. When you point out a failure Warmists will often say phrases like “oh, but the models predicted it”. In other words they just pick one model run from ANY paper to back up their claim. They have so much crap out there that they can always back up any claim. Winters to be warmer, winters to be colder. Earth to spin faster, Earth to spin slower and so on……………. This is what lot’s of funding can achieve, that’s why I pay no attention to claims about the mountains of ‘evidence’. LOL.
I’ve always thought that averaging the models was the same as saying;
I have 6 production lines that make cars. Each line has a defect.
1. Makes cars with no wheels.
2. Makes cars with no engine.
3. Makes cars with no gearbox.
4. Makes cars with no seats.
5. Makes cars with no windows.
6. Makes cars with no brakes.
But “on average” they make good cars.
I thank the good Doctor for explaining why I’ve always thought that. 😉
Am I right in summarizing this thread as?:
The climate models have failed. Observations don’t match reality. Stop beating this poor, dead horse. The jig is up. The party is over. The fat lady is inhaling. The rats are scampering. It has flatlined. The final whistle has been blown.
Good night all.
Ooops. Correction:
“The climate models have failed. Projections don’t match reality. Stop beating this poor, dead horse. The jig is up. The party is over. The fat lady is inhaling. The rats are scampering. It has flatlined. The final whistle has been blown. “
You are correct, Jimbo!
Good night. Sleep well.
*****************
Hey! There is MAGIC working around here! #[:)]
(thanks Anthony or Moderator — for the “that’ll do” enhancement!)
[de nada. — mod.]
Tim Ball says:
June 19, 2013 at 9:54 am
Is the problem wider than the ensemble of 23 models that Brown discusses? As I understand, each of the 23 model results are themselves averages. Every time the model is run starting from the same point the results are different so they do several runs and produce an average. I also understand the number of runs is probably not statistically significant because it takes so long to do a single run.
—-
Thank you, Dr. Ball, I found your post(s) on this topic incredibly illuminating. I had no idea of the incredible amount of time, effort and resources, Climate Scientist\Modelers go through to ultimately reach their intended, predetermined results. I don’t think I’ve ever seen such an incredibly ridiculous, worthless, futile human endeavor.
The immediate thought that comes to mind is, “Why bother?”
Of course the answer is, “We need pseudo-facts to sell the pseudo-science to the unwashed, low-intelligence voters.”
I sense a certain distress in the author’s words in that piece. Wonder if he’s followed the goings on at Climate Audit – upside down Mann, HS extractors, Gergis etc – will need chill pills after that.
rgb is right about the incoherence of the IPCC’s models. You could make exactly the same refutation of Mike Mann’s “Paleo” papers that blend many proxies. Anyone familiar with scotch whiskey will know that blending several “Single Malts” makes no sense.
If “Tree Rings” are the best data it won’t help to blend them with varves lakes or ice cores.
The ‘average of the ensemble’ is akin to having one foot in boiling water and one foot in freezing water and then declaring that on average you feel just fine !
Jimbo says:
June 19, 2013 at 5:11 am
“For Warmists who say that Robert Brown doesn’t know much about computing or models see an excerpt from his about page.”
http://www.phy.duke.edu/~rgb/About/about.php
As a humble member of the Duke university physics department it was my intention to be “low maintenance” even though I had a bunch of Macs, PCs and Sun machines to look after. Whenever I got into serious trouble it was “rgb” who got my chestnuts out of the fire.
If there are Warmists who question “rgb”s competence they will have to work very hard to convince me that they have a clue.
paddylol says:June 19, 2013 at 10:29 am
Mosher, where art thou?
Mosher? Mosher? Mosher?
covering all contingencies?
19 June: Bloomberg: Alessandro Vitelli: EU Carbon Market Needs Immediate Changes, Policy Exchange Says
An independent advisory institution could be established to review the system every two to three years and recommend changes to the market, Newey said.
The specific circumstances in which the market could be adjusted would include “when macroeconomic conditions change significantly from what they were when the cap was set, if the climate science changes, or if there is progress on an international climate deal that would require the EU to take on more ambition, or less,” Newey said.
Policy Exchange was founded by Michael Gove, Francis Maude and Nicholas Boles, who are all now ministers in the U.K.’s coalition government…
http://www.bloomberg.com/news/2013-06-18/eu-carbon-market-needs-immediate-changes-policy-exchange-says.html
Time to stop arguing about climate change: World Bank
LONDON, June 19 (Reuters) – The world should stop arguing about whether humans are causing climate change and start taking action to stop dangerous temperature rises, the president of the World Bank said on Wednesday…
http://www.pointcarbon.com/news/reutersnews/1.2425075?&ref=searchlist
Reblogged this on thewordpressghost and commented:
Friends,
Maybe someday, I will write a great comment upon the problem of Global Warming. And maybe then, my comment will rise to the level of a sticky post.
But, whether you wait for me to write a great comment, or you go read this great comment (post), remember global warming has been going on for thousands of years.
Profiting off of the fear of climate change (global warming) is a very recent marketing strategy.
Enjoy,
Ghost.
I’ve been a Robert Brown fan for years having become familiar with him from the compute angle. This is one bright fellow. Climate science can’t progress without more physicists like him wading in. Rather than displaying prowess in physics and statistics, climate scientists too often seem to be bench warmers.
Nick wrote: “It’s hardly in the fine print – it’s prominent in the introduction. It seems like a very sensible discussion.”
Try reading the section on sea level rise (p13-14) in SPM for AR4 WG1. The scientists responsible for sea level rise refused to make an predictions for acceleration of ice flow from ice sheets because they believed they had no sound basis for estimating that acceleration. Note how clearly the authors responsible for the SPM explained this key caveat associated with their projections, both in Table SPM.3 and in the bullet point at the bottom of page 14. Now look at Figure SPM.5 showing projected climate change with one standard deviation ranges. There is NO mention of the caveats about their use of an “ensemble of opportunity and they show an estimate of uncertainty in these projections when the the introduction to Chapter 10 specifically says the ensemble is NOT suitable for this purpose: ” statistical interpretation of the model spread is therefore problematic”. Figure SPM.3 references Figures 10.4 and 10.29, which also do inappropriate statistical analysis. FIgure 10.29 adds in the uncertainty due the carbon cycle (how accurately can we predict GHG accumulation in the atmosphere from emission scenarios) and the SPM doesn’t mention this caveat either.
Stainforth’s ensembles (referenced above) have shown how dramatically projections of warming change – five-fold – when parameters controlling precipitation/clouds are randomly varied within a range consistent with laboratory measurements. Parameters associated with thermal diffusion of heat (below the mixed layer, between atmospheric grid cells, and between surface and air) weren’t varied in his scenarios, so five-fold uncertainty is only a start to estimating parameter uncertainty for a single model. Then one needs to add uncertainty associated with chaotic behavior, the carbon cycle, aerosols, and model differences. If the IPCC correctly accounted for all of these uncertainties, everyone would realize that the range of possible futures associated with various emissions scenarios is too wide to be useful. Instead, they report results from a ensemble of less than twenty models with single values for each parameter and allow policymakers to believe that a few runs from each model represent the range of possible futures associated with a particular emission scenario. Then they ignore even that uncertainty and tell us how little we can emit to limit warming to +2 degC above pre-industrial temperature (an unknown temperature during the LIA.)
Nick also wrote: “People do use ensemble averages. That’s basically what the word ensemble means. What we still haven’t found is anything that remotely matches the rhetoric of this post.”
However, one doesn’t perform a statistical analysis (mean, std, etc) of an “ensemble of opportunity”. The national models in the IPCC’s “ensemble of opportunity” were not chosen to cover the full range of possible futures consistent with our understanding of the physics of the atmosphere and ocean. They were chosen and considered to be equally valid (“model democracy”) for political reasons. All of these models evolved when climate sensitivity was assumed to be 2-4 degC and with full knowledge of the 20th century warming record. As scientists “optimized” these models without the computing power to properly explore parameter space, these models naturally converged on the “consensus” projections we have today (high sensitivity models are strongly cooled by aerosol and low sensitivity models are not). Stainforth has shown that hundreds of different models provide equally valid representations of current climate and radically different projections of the future. Observational estimates of climate sensitivity now are centered around 2 degC. We might have found, but haven’t found, systematic errors in the temperature record (UHI, for example) that reduced 20th century warming by 50%. Under those circumstances, “optimization” of the parameters in the IPCC’s models could easily have produced very different projections.
(Even when you appear to be wrong, your “trolling” improves the quality of the science discussed here.)
In Nicks Defense
I must say that while using the general circulation models as global climate models is a generally bad idea. The idea behind using multi model means and using an ensemble of a model set is not new. In hurricane forecasting it is used frequently because the overall skill in predicting the storms path is actually slightly better in the multi model mean. Additionally NOAA is working towards developing a more robust seasonal predictive capability based on a multi model mean approach. You can read about it here
http://www.cpc.ncep.noaa.gov/products/ctb/MMEWhitePaperCPO_revised.pdf
So in a day to day use of the models in weather forecasting these ideas are not new and not considered a bad practice.
That being said, the GCM’s modified for temperature prediction over decades has a proven track record of no skill at predicting the temperature so using a mean of all those failed attempts is still a failed attempt.
Much ado about nothing. RGB is correct, that a multi-model mean has no physical meaning. It has no stasticial significance with respect to measurement, and no predictive value.
However, even though it has no physical meaning or significance, it is still a useful tool. It is a visual tool that allows you to show a trend. If many models have upwards trajectories, the mean will illustrate that. It’s just a summary. When you look at a large number of different lines, sometimes you would like to know what is the general trend.
In this case, the multi-model mean, the general trend, is showing that most models are predicting temperature increases (with the usual caveat about means that they can be unduly affected by outliers). That is a valid and interesting fact, even if it has no physical meaning.
TomRude says:
June 19, 2013 at 12:57 pm
Mike MacCracken the Director of the Climate Institute shares his knowledge -or lack of- on Yahoo climatesceptic group…
“The emerging answer seems to be that a cold Arctic (and so strong west to east jet stream) is favored when there is a need to transport a lot of heat to the Arctic”
Que? Transporting a lot of heat to the Arctic makes it colder? That’s, ahem, highly original.
Frank:’However, one doesn’t perform a statistical analysis (mean, std, etc) of an “ensemble of opportunity”.’
Though not for Frank specifically, but for everyone. An rather late on in things.
It’s worth noting that rgbatduke produced a good example of things. It’s also worth noting that it requires some rather fundamental understandings of the scientific process that the reader is not only aware of, but unwilling to budge about. This is all before we begin speaking about secondary details such as mathematical illiteracy. And it paves no easy road for the skeptic of a process to be put in the position of having to prove the process is valid or not. It is the onus of those that put forward the process in the first place that carry the burden of establishing its worth, correctness, and limitations. So while rather late in the game, here is a simple thought experiment for proponents of ensemble means to answer satisfactorily:
——–
Mrs. Jones teaches math to a class of 23 students. They are all taught from the same source material and have the same basic grounding in how to approach and solve the math problems they’re presented. Each student is unique, has their own unique thought process, and unique understanding of the concepts involved in solving math problems; despite that they share the same fundamental teaching from Mrs. Jones.
Last Tuesday Mrs. Jones gave her students a pop-quiz to test their knowledge and growth in learning math; the subject she is teaching them. Each student dutifully answers all the questions and turns in their papers. A couple students produce the same answer while every other student produces a unique answer that no other student produced.[1] And when grading the papers, to her horror, Mrs. Jones discovers that she has lost the answer key. And despite being a teach of mathematics, she is uncertain as to the correct answer to some of the questions.
But if she admits that she has lost the answer key, her job will be in jeopardy. So she chooses to take the average value of the answers to each question and use that as the value of the correct answer. On question #16 there is a single student, Jimmy, whose answer is the same as the average value of the answers from all 23 students.
To the proponents of ensemble means and any others who do not state in uncertain terms that it is mathematical illiteracy:
1) Prove that the correct answer is the answer that Jimmy produced.
Failing this:
2) Prove that no other student produced an answer that was ‘closer’ to the correct one than Jimmy’s.
——–
Those that assert that there is nothing amiss with ensemble means should have no difficulty correctly proving one of these two questions. And if they cannot produce such a proof then there is nothing further to discuss with them. They either need to find that proof, or steadfastly hold an irrational belief in square circles.
[1] Of note here, science being science, we can firmly state that at most one student has the correct answer, and that at least 22 of them have the wrong one. Though these constraints are hardly necessary for the purposes of the problem posed.
Galloping Camel says;
‘If “Tree Rings” are the best data it won’t help to blend them with varves lakes or ice cores.’
If tree rings ARE the best data, it is high time we looked for different data.
Why did other climate scientists think it a good idea to accept a method that loosely records only its micro climate in a highly generalised fashion for its three month growing season with its subdued night time record.?
tonyb
Frank says: June 19, 2013 at 10:27 pm
“Now look at Figure SPM.5 showing projected climate change with one standard deviation ranges.”
Well, at last at the bottom of the second thread, someone has actually pointed to something tangible that might be statistically problematic. But this isn’t a multi-model issue. Both SPM 5 and Fig 10.4 are explicit about the sd:
” Shading denotes the ±1 standard deviation range of individual model annual averages.”
To me, that means the spread of the individual series – something one can calculate for any time series. By contrast, they refer to the center line as “are multi-model global averages”. And Fig 10.29 says:
“For comparison, results are shown for the individual models (red dots) of the multi-model AOGCM ensemble for B1, A1B and A2, with a mean and 5 to 95% range (red line and circle) from a fi tted normal distribution.”
So they are explicit about where they are getting the range, and again it seems to be a fit to an individual model. On carbon they say
“The 5 to 95% ranges (vertical lines) and medians (circles) are shown from probabilistic methods”
and give a whole stack of references.
So I still don’t think we’ve matches RGB’s bitchslap rhetoric to any actual deeds.
Jquip: Your ensemble is interesting, but you will find Stainforth’s more relevant to this post. His ensemble tried systematically varying all of the model parameters that control clouds and precipitation within the limits established by laboratory experiments. These parameters interact in surprising ways that can’t be understood by varying/optimizing one parameter at a time.
http://media.cigionline.org/geoeng/2005%20-%20Stainforth%20et%20al%20-%20Uncertainty%20in%20predictions%20of%20the%20climate%20response%20to%20rising%20GHGs.pdf
In subsequent papers, Stainforth has tried (and failed) to identify a small subset of his ensemble of models which provide the best representation of the earth’s current climate.
Gail Combs wrote:
“This Graph shows world bank funding for COAL fired plants in China, India and elsewhere went from $936 billion in 2009 to $4,270 billion in 2010.”
It should be:
$936 million in 2009 to $4270 million in 2010
or alternatively
$936 million in 2009 to $4,270 billion in 2010