The Good, The Bad, and The Ugly: My Initial Comments on the New Dessler 2011 Study

NOTE: This post is important, so I’m going to sticky it at the top for quite a while. I’ve created a page for all Spencer and Braswell/Dessler related posts, since they are becoming numerous and popular to free up the top post sections of WUWT.

UPDATE: Dr. Spencer writes: I have been contacted by Andy Dessler, who is now examining my calculations, and we are working to resolve a remaining difference there. Also, apparently his paper has not been officially published, and so he says he will change the galley proofs as a result of my blog post; here is his message:

“I’m happy to change the introductory paragraph of my paper when I get the galley proofs to better represent your views. My apologies for any misunderstanding. Also, I’ll be changing the sentence “over the decades or centuries relevant for long-term climate change, on the other hand, clouds can indeed cause significant warming” to make it clear that I’m talking about cloud feedbacks doing the action here, not cloud forcing.”

[Dessler may need to make other changes, it appears Steve McIntyre has found some flaws related to how the CERES data was combined: http://climateaudit.org/2011/09/08/more-on-dessler-2010/

As I said before in my first post on Dessler’s paper, it remains to be seen if “haste makes waste”. It appears it does. -Anthony]

Update #2 (Sept. 8, 2011): Spencer adds: I have made several updates as a result of correspondence with Dessler, which will appear underlined, below. I will leave it to the reader to decide whether it was our Remote Sensing paper that should not have passed peer review (as Trenberth has alleged), or Dessler’s paper meant to refute our paper.

by Roy W. Spencer, Ph. D.

While we have had only one day to examine Andy Dessler’s new paper in GRL, I do have some initial reaction and calculations to share. At this point, it looks quite likely we will be responding to it with our own journal submission… although I doubt we will get the fast-track, red carpet treatment he got.

There are a few positive things in this new paper which make me feel like we are at least beginning to talk the same language in this debate (part of The Good). But, I believe I can already demonstrate some of The Bad, for example, showing Dessler is off by about a factor of 10 in one of his central calculations.

Finally, Dessler must be called out on The Ugly things he put in the paper.

(which he has now agreed to change).

1. THE GOOD

Estimating the Errors in Climate Feedback Diagnosis from Satellite Data

We are pleased that Dessler now accepts that there is at least the *potential* of a problem in diagnosing radiative feedbacks in the climate system *if* non-feedback cloud variations were to cause temperature variations. It looks like he understands the simple-forcing-feedback equation we used to address the issue (some quibbles over the equation terms aside), as well as the ratio we introduced to estimate the level of contamination of feedback estimates. This is indeed progress.

He adds a new way to estimate that ratio, and gets a number which — if accurate — would indeed suggest little contamination of feedback estimates from satellite data. This is very useful, because we can now talk about numbers and how good various estimates are, rather than responding to hand waving arguments over whether “clouds cause El Nino” or other red herrings. I have what I believe to be good evidence that his calculation, though, is off by a factor of 10 or so. More on that under THE BAD, below.

Comparisons of Satellite Measurements to Climate Models

Figure 2 in his paper, we believe, helps make our point for us: there is a substantial difference between the satellite measurements and the climate models. He tries to minimize the discrepancy by putting 2-sigma error bounds on the plots and claiming the satellite data are not necessarily inconsistent with the models.

But this is NOT the same as saying the satellite data SUPPORT the models. After all, the IPCC’s best estimate projections of future warming from a doubling of CO2 (3 deg. C) is almost exactly the average of all of the models sensitivities! So, when the satellite observations do depart substantially from the average behavior of the models, this raises an obvious red flag.

Massive changes in the global economy based upon energy policy are not going to happen, if the best the modelers can do is claim that our observations of the climate system are not necessarily inconsistent with the models.

(BTW, a plot of all of the models, which so many people have been clamoring for, will be provided in The Ugly, below.)

2. THE BAD

The Energy Budget Estimate of How Much Clouds Cause Temperature Change

While I believe he gets a “bad” number, this is the most interesting and most useful part of Dessler’s paper. He basically uses the terms in the forcing-feedback equation we use (which is based upon basic energy budget considerations) to claim that the energy required to cause changes in the global-average ocean mixed layer temperature are far too large to be caused by variations in the radiative input into the ocean brought about by cloud variations (my wording).

He gets a ratio of about 20:1 for non-radiatively forced (i.e. non-cloud) temperature changes versus radiatively (mostly cloud) forced variations. If that 20:1 number is indeed good, then we would have to agree this is strong evidence against our view that a significant part of temperature variations are radiatively forced. (It looks like Andy will be revising this downward, although it’s not clear by how much because his paper is ambiguous about how he computed and then combined the radiative terms in the equation, below.)

But the numbers he uses to do this, however, are quite suspect. Dessler uses NONE of the 3 most direct estimates that most researchers would use for the various terms. (A clarification on this appears below) Why? I know we won’t be so crass as to claim in our next peer-reviewed publication (as he did in his, see The Ugly, below) that he picked certain datasets because they best supported his hypothesis.

The following graphic shows the relevant equation, and the numbers he should have used since they are the best and most direct observational estimates we have of the pertinent quantities. I invite the more technically inclined to examine this. For those geeks with calculators following along at home, you can run the numbers yourself:

Here I went ahead and used Dessler’s assumed 100 meter depth for the ocean mixed layer, rather than the 25 meter depth we used in our last paper. (It now appears that Dessler will be using a 700 m depth, a number which was not mentioned in his preprint. I invite you to read his preprint and decide whether he is now changing from 100 m to 700 m as a result of issues I have raised here. It really is not obvious from his paper what he used).

Using the above equation, if I assumed a feedback parameter λ=3 Watts per sq. meter per degree, that 20:1 ratio Dessler gets becomes 2.2:1. If I use a feedback parameter of λ=6, then the ratio becomes 1.7:1. This is basically an order of magnitude difference from his calculation.

Again I ask: why did Dessler choose to NOT use the 3 most obvious and best sources of data to evaluate the terms in the above equation?:

(1) Levitus for observed changes in the ocean mixed layer temperature; (it now appears he will be using a number consistent with the Levitus 0-700 m layer).

(2) CERES Net radiative flux for the total of the 2 radiative terms in the above equation, and (this looks like it could be a minor source of difference, except it appears he put all of his Rcld variability in the radiative forcing term, which he claims helps our position, but running the numbers will reveal the opposite is true since his Rcld actually contains both forcing and feedback components which partially offset each other.)

(3): HadSST for sea surface temperature variations. (this will likely be the smallest source of difference)

The Use of AMIP Models to Claim our Lag Correlations Were Spurious

I will admit, this was pretty clever…but at this early stage I believe it is a red herring.

Dessler’s Fig. 1 shows lag correlation coefficients that, I admit, do look kind of like the ones we got from satellite (and CMIP climate model) data. The claim is that since the AMIP model runs do not allow clouds to cause surface temperature changes, this means the lag correlation structures we published are not evidence of clouds causing temperature change.

Following are the first two objections which immediately come to my mind:

1) Imagine (I’m again talking mostly to you geeks out there) a time series of temperature represented by a sine wave, and then a lagged feedback response represented by another sine wave. If you then calculate regression coefficients between those 2 time series at different time leads and lags (try this in Excel if you want), you will indeed get a lag correlation structure we see in the satellite data.

But look at what Dessler has done: he has used models which DO NOT ALLOW cloud changes to affect temperature, in order to support his case that cloud changes do not affect temperature! While I will have to think about this some more, it smacks of circular reasoning. He could have more easily demonstrated it with my 2 sine waves example.

Assuming there is causation in only one direction to produce evidence there is causation in only one direction seems, at best, a little weak.

2) In the process, though, what does his Fig. 1 show that is significant to feedback diagnosis, if we accept that all of the radiative variations are, as Dessler claims, feedback-induced? Exactly what the new paper by Lindzen and Choi (2011) explores: that there is some evidence of a lagged response of radiative feedback to a temperature change.

And, if this is the case, then why isn’t Dr. Dessler doing his regression-based estimates of feedback at the time lag or maximum response? Steve McIntyre, who I have provided the data to for him to explore, is also examining this as one of several statistical issues. So, Dessler’s Fig. 1 actually raises a critical issue in feedback diagnosis he has yet to address.

3. THE UGLY

(MOST, IF NOT ALL, OF THESE OBJECTIONS WILL BE ADDRESSED IN DESSLER’S UPDATE OF HIS PAPER BEFORE PUBLICATION)

The new paper contains a few statements which the reviewers should not have allowed to be published because they either completely misrepresent our position, or accuse us of cherry picking (which is easy to disprove).

Misrepresentation of Our Position

Quoting Dessler’s paper, from the Introduction:

“Introduction

The usual way to think about clouds in the climate system is that they are a feedback… …In recent papers, Lindzen and Choi [2011] and Spencer and Braswell [2011] have argued that reality is reversed: clouds are the cause of, and not a feedback on, changes in surface temperature. If this claim is correct, then significant revisions to climate science may be required.”

But we have never claimed anything like “clouds are the cause of, and not a feedback on, changes in surface temperature”! We claim causation works in BOTH directions, not just one direction (feedback) as he claims. Dr. Dessler knows this very well, and I would like to know

1) what he was trying to accomplish by such a blatant misrepresentation of our position, and

2) how did all of the peer reviewers of the paper, who (if they are competent) should be familiar with our work, allow such a statement to stand?

Cherry picking of the Climate Models We Used for Comparison

This claim has been floating around the blogosphere ever since our paper was published. To quote Dessler:

“SB11 analyzed 14 models, but they plotted only six models and the particular observational data set that provided maximum support for their hypothesis. “

How is picking the 3 most sensitive models AND the 3 least sensitive models going to “provide maximum support for (our) hypothesis”? If I had picked ONLY the 3 most sensitive, or ONLY the 3 least sensitive, that might be cherry picking…depending upon what was being demonstrated. And where is the evidence those 6 models produce the best support for our hypothesis?

I would have had to run hundreds of combinations of the 14 models to accomplish that. Is that what Dr. Dessler is accusing us of?

Instead, the point was to show that the full range of climate sensitivities represented by the least and most sensitive of the 14 models show average behavior that is inconsistent with the observations. Remember, the IPCC’s best estimate of 3 deg. C warming is almost exactly the warming produced by averaging the full range of its models’ sensitivities together. The satellite data depart substantially from that. I think inspection of Dessler’s Fig. 2 supports my point.

But, since so many people are wondering about the 8 models I left out, here are all 14 of the models’ separate results, in their full, individual glory:

I STILL claim there is a large discrepancy between the satellite observations and the behavior of the models.

CONCLUSION

These are my comments and views after having only 1 day since we received the new paper. It will take weeks, at a minimum, to further explore all of the issues raised by Dessler (2011).

Based upon the evidence above, I would say we are indeed going to respond with a journal submission to answer Dessler’s claims. I hope that GRL will offer us as rapid a turnaround as Dessler got in the peer review process. Feel free to take bets on that. :)

And, to end on a little lighter note, we were quite surprised to see this statement in Dessler’s paper in the Conclusions (italics are mine):

These calculations show that clouds did not cause significant climate change over the last decade (over the decades or centuries relevant for long-term climate change, on the other hand, clouds can indeed cause significant warming).”

Long term climate change can be caused by clouds??! Well, maybe Andy is finally seeing the light! ;) (Nope. It turns out he meant ” *RADIATIVE FEEDBACK DUE TO* clouds can indeed cause significant warming”. An obvious, minor typo. My bad.)

The climate data they don't want you to find — free, to your inbox.
Join readers who get 5–8 new articles daily — no algorithms, no shadow bans.
0 0 votes
Article Rating
514 Comments
Inline Feedbacks
View all comments
Frank K.
September 7, 2011 1:48 pm

Dr. Spencer says:
“I hope that GRL will offer us as rapid a turnaround as Dessler got in the peer review process. Feel free to take bets on that.”
Like G. Karst, I too look forward to the GRL’s new and improved “RAPID REVIEW” ™ process. I’m sure it’ll only take 8 weeks (10 tops)!

September 7, 2011 1:49 pm

Ron Cram:
See the comments under by blog post for the likely source of the factor of ten difference.
I did not include any mention of Dessler’s video because (1) I haven’t seen it, and (2) I was just addressing his peer-reviewed and published paper in this post.
-Roy Spencer

Martin Lewitt
September 7, 2011 1:53 pm

KR,
“To be quite blunt, without such a physical mechanism overriding the water vapor cycle, your assertions of clouds as the forcing driving the ENSO are “Just So Stories”.
The paper was about diagnosis of radiative feedback. The IPCC projections are just plain invalid if there isn’t strong positive cloud feedback, because all of the models have significant positive feedback. As someone at climateaudit pointed out, AR4 states in Section 8.6.2.3:
…in the absence of cloud feedbacks, current GCMs would predict a climate sensitivity (±1 standard deviation) of roughly 1.9°C ± 0.15°C (ignoring spread from radiative forcing differences). The mean and standard deviation of climate sensitivity estimates derived from current GCMs are larger (3.2°C ± 0.7°C) essentially because the GCMs all predict a positive cloud feedback (Figure 8.14) but strongly disagree on its magnitude.
If the AR4 models don’t get their significant positive cloud feedback, then there is plenty to override “the water vapor cycle”, for starters try the speedup of the cycle. Wentz in the journal Science showed that all the AR4 models reproduced less than half in the increase in precipitation seen in the observations. Under representing a negative feedback (the water cycle) combined with over representing a positive feedback pretty much leave the models running wild.
Do you really think this statement from SB11:
“We hypothesize that changes in the coupled oceanatmosphere circulation during the El Niño and La Niña phases of ENSO cause differing changes in cloud cover, which then modulate the radiative balance of the climate system”
was the main result from the paper or can be fairly characterized as: “assertions of clouds as the forcing driving the ENSO”?
How Much More Rain Will Global Warming Bring?
Frank J. Wentz, Lucrezia Ricciardulli, Kyle Hilburn, and Carl Mears
Science 13 July 2007: 317 (5835), 233-235.Published online 31 May 2007 [DOI:10.1126/science.1140746]

Ged
September 7, 2011 1:53 pm

@KR,
I’m sorry, but which models show a -statistically significant- fit that agrees with satellite observations? I sure don’t see any, nor does Dessler.
Taking most and least sensitive was instructive, and looked fine. Why? Because the point was that basing a model on sensitivity alone was not accurately matching the satellite observations. Or are you going to contest this?

September 7, 2011 1:54 pm

DayDay wins the humor contest! 🙂

glacierman
September 7, 2011 1:56 pm

Any possiblity we will ever know who the reviewers for Dessler2011 were?
Great job they did. No puffball review there……very thorough.
I am sure they are watching….probably even commenting. Hang your heads in shame.

September 7, 2011 1:59 pm

BTW, to falsify a hypothesis it is NOT necessary to propose a better alternative one. Either the H0 or acknowledgment of ignorance will do just fine.

KR
September 7, 2011 2:12 pm

GedI’m sorry, but which models show a -statistically significant- fit that agrees with satellite observations? I sure don’t see any, nor does Dessler.
If you read the Dessler paper, the three models that are, incidentally, best known to reproduce the ENSO variations, fall almost entirely within the uncertainty range of the satellite temperature values. And yes, Dessler clearly points those out. Those were _not_ included in Spencer’s graphs, nor were the uncertainty ranges, despite those particular models having a fairly close agreement to the temperature records over the last 10 years.
10 years is a fairly short time in terms of climate – 30 is statistically (based on year to year variations) a better minimum time to estimate climate changes, as opposed to ENSO variations, weather, or simply noise. One of my concerns with the Spencer et al 2011 paper is that it’s really just too short a time frame (10 years?) to look at equilibrium climate sensitivity – it’s a time frame much more appropriate to considering whether climate models match the ENSO. And Spencer did not show the models known to best track the ENSO. Like it or not, that (in my opinion) was a poor selection of data, and gives the impression that counter-evidence was not shown. I really wish he had shown all 14 models that he ran, and discussed his conclusions in that light.

KR
September 7, 2011 2:14 pm

DayDay – Excellent! Almost laughed myself out of my chair…

September 7, 2011 2:16 pm

AR5 is nigh.

Dr A Burns
September 7, 2011 2:29 pm

I would have thought that the fact that monthly Hadcrut3 temperatures vary around 3.5 degrees annually and that they are lowest when the earth is closest to the sun, would have been clear evidence of the strong negative feedback of clouds ?

Bloke down the pub
September 7, 2011 2:31 pm

Perhaps Roy this all just goes to show that you can have good science, and you can have quick science. Asking for both may be a step too far.

HAS
September 7, 2011 2:32 pm

TallDave September 7, 2011 at 12:16 pm
KR September 7, 2011 at 12:43 pm
Jeremy September 7, 2011 at 1:19 pm
I had a quick look at Dessler’s claims that GFDL CM 2.1, MPI ECHAM5, and MRI CGCM 2.3.2A are the best match to ENSO at
http://judithcurry.com/2011/09/06/spencer-braswell-part-iii/#comment-109890
It does seem that based on at least one recent system of categorisation they aren’t.

September 7, 2011 2:32 pm

Roy,
One issue is that the models which show the best correlation are actually form the middle of the pack with ECRs in the range of 3.4 and TCRs in the range of 1.6-2.2
By only showing the high and low sensitivity I think most people assume that the moderate sensitivity models would be in the middle. But they are not. It’s clear that the data and models are not happy campers. What’s not clear is whether you can diagnose ECR by looking at these data, maybe TCR.. anyways, still sorting through the math

John Whitman
September 7, 2011 2:48 pm

Roy Spencer,
Thank you for the post at WUWT.
I think this is science as it should be. I sincerely hope Dessler sees this discourse in a light that reflects well on the openness in science; openness that I think your approach does reflect.
To the DISCOURSE!!!
John

ZT
September 7, 2011 2:49 pm

Dr Spencer: Thank you and Steve McIntyre!
I have thoroughly enjoyed reading Dessler’s ‘Science’ 2010 paper.
All those interested, please see Dessler 2010: ftp://ftp.ingv.it/pub/pietropaolo.bertagnolio/climate/dessler10-cloudFeedbacks.pdf
Here Dessler says:
“the slope using the MERRA is 0.46 T +/- 0.75 W/m2/K” (i.e. positive or negative we don’t know statistically) and later includes the rather witty line “Obviously, the correlation … is weak (r2= 2%)”
As one would typically want an r2 greater than 50% before drawing any form of conclusion using regression, this is an ugly example of the flagrant disregard for science and logic typical of the climatological community. Yet Dessler concludes:
“My analysis suggests that the short-term cloud feedback is likely positive and that climate
models as a group are doing a reasonable job of simulating this feedback, providing some indication that models successfully simulate the response of clouds to climate variations.”
A clear, ugly, example of the corrupt ‘science’ of climatology.
Perhaps Dessler would care to put together a short video explaining his statistical analysis?

Brandon Caswell
September 7, 2011 2:52 pm

This is a chess game.
-Dr Spencer took an offensive move in releasing this paper. If it stands it might have to be considered by the new IPCC report.
-They countered with personal attacks, but they actually left them in a worse position on the board than before.
-it was countered by making fun of the fact they didn’t respond to anything in SB 11 science, and it hit a chord. Now this is getting serious.
-They responded with a sacrifice of the editior of the magazine, which they timed to come out with this new paper. Good move because, the editor move on its own was getting laughed off the board.
-Spencer or another will respond with a rebuttal paper……that will get held up for as long as it takes so they won’t have to consider the initial SB11 paper for the IPCC.
-When reviewer comments at IPCC suggest the SB11 paper, they will point to the dessler 2011 paper as the reason to not consider it because the rebuttal will be too late to consider.
Checkmate. They have lost alot of respect, but they win the battle for the IPCC report. They think the IPCC report will trump all the other small battles like this and give them a clean slate again after it comes out. This is no shock, the same people did the same thing with past papers and IPCC reports. Business as usual. No shame.

Brandon Caswell
September 7, 2011 3:06 pm

You have to remember, this is not about science or even global warming.
This is about careers.
This is about reputations.
This about ideologies.
This is about money.
People have been killed over these things in the past!
Do you really think Someone like Trenberth can even consider admitting he might have been wrong about something? Especially since they made such spectacles of themselves claiming everyone but themselves are morons. His career will now live and die with AGW, like many others. They are now between the proverbial rock and a hard place. They look silly if they continue, they look silly if they don’t.

D. J. Hawkins
September 7, 2011 3:20 pm

@KR
Re your comments on “The Ugly”, you do realize that the point Spencer et al are making is not that the clouds force temperature, but that because the effects are bidirectional it’s currently not possible to determine the true sensitivity?

DocMartyn
September 7, 2011 3:25 pm

Roy, on the Climate4you website they have the plot of the change in the ground and the air temperature, vs time, during the course of a solar eclipse:-
http://www.climate4you.com/images/SolarEclipse%20LYR%20SurfaceAndAirTemp%2020080801.gif
http://www.climate4you.com/Longyearbyen%20SolarEclipse%2020080801.htm
You can see the effect of a cloud and the Longyearbyen Solar Eclipse, August 1, 2008.
Now, one could calculate the light flux, based on time and location, prior to and following the eclipse. What I note is that the ground is a very good radiator and that the air is a very poor radiator. This is a little odd if one believes that the air ‘traps’ the heat.
Using the data provided in the links, one could calculate the non-radiatively forced (i.e. non-cloud) temperature changes versus radiatively (mostly cloud) forced variations. directly.
For myself i would prefer an experimental approach,
Have a large, football field sized, sheet of aluminumized Mylar suspended to tethered weather balloons. Tether each balloon to fours trucks and at different T’s, drive the trucks so that the sheet blocks the sunlight above your spectrophotometer/temperature monitors.

September 7, 2011 3:48 pm

HankH says:
“This causes me to question if they’re truly modeling reality or happen to be two models of the 14 that agree better for the wrong reasons. Statistically speaking two models of the 14 are going to agree better no matter how bad the entire ensemble gets it.”
This is a good point, and we can extend it further by pointing out that when taking the data and comparing it to multiple models, the necessary statistical r^2 value required for a given confidence level must be much higher than if the comparison was to just one model. In a Student T test of r^2 it goes up as the power of the number of comparisons.
Similarily, the Steve McIntyre regression of Dessler shows about 115 – 120 points on the plot. A Student T test would then require an r^2 of 0.032 to be significant at the 95% CI. I think both the zero lag (Dessler) and the 4 month lag (Spencer) correlations would fail a Student T test.

Avondlander
September 7, 2011 3:59 pm

To KR:
Read this excellent article on the possibility of long-term external (cosmic) influences on cloud formation:
http://online.wsj.com/article/SB10001424053111904900904576554063768827104.html

BA
September 7, 2011 4:07 pm

So who were the 3 reviewers who passed SB11, despite its cherry-picked model/data comparison and assumptions chosen to support a particular conclusion? Remote Sensing invites authors themselves to nominate up to 5 reviewers. Did Spencer do so, and were the 3 reviewers actually from his list? If so, SB11 got published in the first place by exploiting a weakness in their peer review system.
Spencer at different times has written that he only knows who one of the reviewers was, but also that all 3 of them were well published in climate modeling. Those statements don’t seem to agree.

KR
September 7, 2011 4:22 pm

D. J. Hawkins“@KR: Re your comments on “The Ugly”, you do realize that the point Spencer et al are making is not that the clouds force temperature, but that because the effects are bidirectional it’s currently not possible to determine the true sensitivity?”
From Spencer and Braswell 2011, Conclusions, first line: “We have shown clear evidence from the CERES instrument that global temperature variations during 2000–2010 were largely radiatively forced.”
Also, “Finally, since much of the temperature variability during 2000–2010 was due to ENSO [9], we conclude that ENSO-related temperature variations are partly radiatively forced.”
Radiatively forced temperatures? By clouds, then, rather than ocean heat distribution (i.e, the ENSO)? This appears to be a clear claim on Spencer and Braswell’s part to me – I really don’t see how to read that in any other fashion. It’s a claim that cloud changes are the forcing component on the temperature changes of the last decade, rather than the ENSO. How would you interpret that???

KR
September 7, 2011 4:24 pm

Dr. Spencer
I would still be interested in your reasons for not including 8 of the 14 models you studied, the ones that don’t support your hypothesis as strongly. I believe your paper would have been much better with those included, along with error bars so that we could evaluate the strength of your hypotheses.