I discovered this climate model failure a while ago, but haven’t published a post about it because, if I were to compare the modeled and observed sea ice area for each hemisphere, I would need to make too many approximations and assumptions. The reasons: The NSIDC sea ice area data through the KNMI Climate Explorer is presented in millions of square kilometers, while the CMIP5-archived model outputs there are presented in the fraction of sea ice area—assumedly a fraction of the ocean area for the input coordinates.
I decided to take a simpler approach with this post—to show whether the models simulate a gain or loss in each hemisphere.
That is, we know the oceans have been losing sea ice in the Arctic since November 1978, but gaining it around Antarctica. See Figure 1.
Figure 1
Then there are the oodles of climate models stored in the CMIP5 archive. They’re the models being used by the IPCC for the upcoming 5th Assessment Report. Would you like to guess whether they show the Northern and Southern Hemispheres should have gained or lost sea ice area over the same time period?
The multi-model ensemble mean of their outputs indicate, if sea ice area were dependent on the increased emissions of manmade greenhouse gases, the Southern Ocean surrounding Antarctica should have lost sea ice from November 1978 to May 2013. See Figure 2.
Figure 2
Well at least the models were right about the sea ice loss in the Northern Hemisphere. Too bad for the modelers that our planet also has a Southern Hemsiphere.
We could have guessed the models simulated a loss of sea ice around Antarctica based on their simulation of the sea surface temperatures in the Southern Ocean. As illustrated in the most recent model-data comparison of sea surface temperatures, here, sea surface temperatures in the Southern Ocean have cooled, Figure 3, while the models say they should have warmed.
Figure 3
STANDARD BLURB ABOUT THE USE OF THE MODEL MEAN
We’ve published numerous posts that include model-data comparisons. If history repeats itself, proponents of manmade global warming will complain in comments that I’ve only presented the model mean in the above graphs and not the full ensemble. In an effort to suppress their need to complain once again, I’ve borrowed parts of the discussion from the post Blog Memo to John Hockenberry Regarding PBS Report “Climate of Doubt”.
The model mean provides the best representation of the manmade greenhouse gas-driven scenario—not the individual model runs, which contain noise created by the models. For this, I’ll provide two references:
The first is a comment made by Gavin Schmidt (climatologist and climate modeler at the NASA Goddard Institute for Space Studies—GISS). He is one of the contributors to the website RealClimate. The following quotes are from the thread of the RealClimate post Decadal predictions. At comment 49, dated 30 Sep 2009 at 6:18 AM, a blogger posed this question:
If a single simulation is not a good predictor of reality how can the average of many simulations, each of which is a poor predictor of reality, be a better predictor, or indeed claim to have any residual of reality?
Gavin Schmidt replied with a general discussion of models:
Any single realisation can be thought of as being made up of two components – a forced signal and a random realisation of the internal variability (‘noise’). By definition the random component will uncorrelated across different realisations and when you average together many examples you get the forced component (i.e. the ensemble mean).
To paraphrase Gavin Schmidt, we’re not interested in the random component (noise) inherent in the individual simulations; we’re interested in the forced component, which represents the modeler’s best guess of the effects of manmade greenhouse gases on the variable being simulated.
The quote by Gavin Schmidt is supported by a similar statement from the National Center for Atmospheric Research (NCAR). I’ve quoted the following in numerous blog posts and in my recently published ebook. Sometime over the past few months, NCAR elected to remove that educational webpage from its website. Luckily the Wayback Machine has a copy. NCAR wrote on that FAQ webpage that had been part of an introductory discussion about climate models (my boldface):
Averaging over a multi-member ensemble of model climate runs gives a measure of the average model response to the forcings imposed on the model. Unless you are interested in a particular ensemble member where the initial conditions make a difference in your work, averaging of several ensemble members will give you best representation of a scenario.
In summary, we are definitely not interested in the models’ internally created noise, and we are not interested in the results of individual responses of ensemble members to initial conditions. So, in the graphs, we exclude the visual noise of the individual ensemble members and present only the model mean, because the model mean is the best representation of how the models are programmed and tuned to respond to manmade greenhouse gases.
CLOSING
Just add sea ice onto the growing list of variables that are simulated poorly by the IPCC’s climate models. Over the past few months, we’ve illustrated and discussed that the climate models stored in the CMIP5 archive for the upcoming 5th Assessment Report (AR5) cannot simulate observed:
Satellite-Era Sea Surface Temperatures
Global Surface Temperatures (Land+Ocean) Since 1880
And in an upcoming post, we’ll illustrate how poorly the models simulate daily maximum and minimum temperatures and the difference between them, the diurnal temperature range. I should be publishing that post within the next week.
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.



‘As illustrated in the most recent model-data comparison of sea surface temperatures, here, sea surface temperatures in the Southern Ocean have cooled, Figure 3, while the models say they should have warmed.’
I thought the Antarctic was meant to be melting due to warmer seas(cf warmer air temps).
http://wattsupwiththat.com/2013/06/13/game-changer-antarctic-melt-due-to-warm-water-not-air-temperature/#more-88085
Presumably that’s why Antarctic sea ice is above normal.
Warm AMO/NAO in the NH. Cold Southern Ocean in the SH. The amount of ice is based on the temperature of the ocean water. What an amazing revelation. Can we get a refund for all that wasted money on models?
As per previous “Scientific Publications”, they just realized that Ice Melt in the Antarctic is caused by Ocean Currents, not Atmospheric warming. How can you make a “model” when you don’t even know the basics of the physical processes???
Of course, one can always make a “statistical model” based on past information. BUT, as they say in Stock Prospective s, “past performance does not guarantee future results”.
When will these “scientists” wake up to the fact that CO2 is not a problem, but a solution to growing more crops, and it has very little to do with “Global Warming”. As per this blog, over half of the perceived temperature increases are due to the “Urban Heat Island Effect”. We will find out if the Sun is the remainder, since we are in a grand observation of a “Quiet Sun”.
Seems to be a coupled system North down, South up and visa-versa. Could be due to multidecadal ocean current changes which we just don’t fully understand and seems to be a cycle of about 60-80years.. I can certainly live with that.
Thanks, Bob. Well said.
The NSIDC Sea Ice data can be seen at http://nsidc.org/data/seaice_index/
Are we still talking about water-world models, or have they actually put continents, ocean currents, and mountains in yet?
Bloke down the pub: I haven’t read the paper associated with that post. But the press release doesn’t mention warming ocean temperatures, as far as I could see.
Regards
Good stuff, Bob.
The Arctic sea ice is the only variable which the models have got close to right. Obviously, it was just a fluke. When you only get 1 out of 13 key climate indicators right, it is by accident, not from good modelling.
Bob:
Why are these two graphs different, although they supposedly represent the same thing?:
No. 3 of this present article:
http://bobtisdale.files.wordpress.com/2013/06/figure-32.png?w=640&h=422
And No. 7 from this other article of yours (http://bobtisdale.wordpress.com/2013/02/28/cmip5-model-data-comparison-satellite-era-sea-surface-temperature-anomalies/)
http://bobtisdale.files.wordpress.com/2013/02/07-so-hem.png
Shouldn’t they be alike?
Perfect! I was just thinking about finding charts of both poles’ ice since 1979, and merging them together, just like figure 1. Got the idea from P. Gosselin’s site.
Heber Rizzo says: “Why are these two graphs different, although they supposedly represent the same thing?”
Nope. They’re not the same thing. One is the Southern Ocean (90S-60S), and the other is the Southern Hemisphere (90S-0).
Regards
Bob said:
Nope. They’re not the same thing. One is the Southern Ocean (90S-60S), and the other is the Southern Hemisphere (90S-0).
Fool of me.
But I was told that melting the melting under the ice shelves caused Antarctica’s expanding sea ice!
http://dx.doi.org/10.1038/ngeo1767
Or are sea surface temps getting warmer in the SH summers? Ahhhhh. Grrrrrr.
I am well familiar with EKG’s and MRI’s. The premis is the same. Brains at rest are noisy with random synaptic activity. Brains that are listening to a signal or doing some kind of functional thinking will have a brainwave component rise out of the noise. I can build a model of this observation or I can simply measure it in-situ and let the software perform calculations on the data (basically adding and subtracting run that is picked up by the electrode “listeners” over and over again) till the noise is cancelled out (which would mean 100’s of runs and even 1000’s of runs). If I performed this test on thousands of different human beings using the same signal and I see a similar pattern, I could say that brain waves can be measured and are not random when a brain is doing some kind of specific task (listening, looking, reading, doing math, thinking about a certain subject, etc). At that point I no longer need the model. All I have to do is measure real brains. Unless the one I am measuring is mine. Apparently I have a non-random, very busy brain. A brainstem response to a click noise could not be located in the noise. My baseline brain at rest is too noisy and non-random.
Models must contain dialed in assumptions about observations that have been purposely biased by the modeler. In other words, in this case, the modeler believes (or is testing his/her belief) that intrinsic and extrinsic drivers of weather pattern variations are random, having no step functions or echoed affects. So they are built to demonstrate this assumption. The runs are completed to demonstrate this noise and its average anomaly. Which should eventually cancel to 0 if enough runs are completed. When they do, the modeler believes he/she has natural climate just right. The anthropogenic dial is a constant factor built to be void of noise and tuned to observations of a previous observed trend thought to be caused by some kind of anthropogenic activity. The dial is added to the model. The runs are done again and abracadabra, a rising anomaly. The result should be no surprise to anyone.
The problem here is that we have only one subject. Earth. And a short number of runs (if we are doing yearly runs, 17 runs is very limited). The result so far is that our modeled termperature response is not matching the current observation. In medical circles this would neither prove or disprove a model of brain activity. Maybe Earth is like Pam’s brain. Very busy and non-random.
TYPO: (basically adding and subtracting runs that are picked up by the electrode “listeners” over and over again)
IPCC AR5 is going to look like a joke without a funny ending. I’m 99% certain it will say we’re melting when everybody in the world will want cheaper heating and lots of it too.
Bob Tisdale says:
June 15, 2013 at 7:36 am
“Bloke down the pub: I haven’t read the paper associated with that post. But the press release doesn’t mention warming ocean temperatures, as far as I could see.”
While the full paper may add something about warmer water, I don’t see that in the press release either. There is major MSM hype each time a big chunk of ice shelf gets a crack and then floats away. This research says a greater proportion of ice melts from the under side in contact with water. That is not surprising and it doesn’t make for good TV images of Manhattan sized bergs.
~~~~~~~
@ur momisugly johnmarshall
“Seems to be a coupled system . . .”
Looking at Bob’s Figure 1, it is easy to leap on to this wagon. At best, I’ve only got another 15 to 20 years so I hope someone figures this out before 2073.
~~~~~~~
Thanks, Bob.
It is worth reading Robert G Brown’s treatise over at Judith Curry’s blog. He sometimes writes here as rgbatduke. He is of the opinion that taking the average of the CMIP ensemble is silly, that it has no meaning. The average of garbage is just average garbage.
Idea for a future article:
How well do the models do at simulating equator-pole temperature gradients?
Suggestion: Break the analysis down geographically — e.g. western ocean boundaries (where gradients are steep) vs. eastern (where gradients are diffuse), etc.
A very interesting article from Bob, precise and to the point as always, shows just how bad the models are at forecasting,
According to my reckoning, there are over 1 million Sq. Km more sea ice in the Arctic than this time last year, and more than the last few years.
http://arctic.atmos.uiuc.edu/cryosphere/timeseries.anom.1979-2008
This may or may not continue, but the air temperature North of the 80th parallel is below average also.
http://ocean.dmi.dk/arctic/meant80n.uk.php
I know Joe Bastadi was forecasting a recovery, is this what we are now seeing, I wonder if this model forecast, the one they got right, is going to suffer the same fate as all the other “fails”.
Regards
Keith Gordon
Thanks Bob
Bill Illis says: “When you only get 1 out of 13 key climate indicators right, it is by accident, not from good modelling.”
What are the 12 that have turned out wrong. It would be nice to have a summary list of all the model predictions and note which are still in play and which have already failed miserably.
Tom Jones says:
June 15, 2013 at 8:54 am
“It is worth reading Robert G Brown’s treatise over at Judith Curry’s blog. He sometimes writes here as rgbatduke. He is of the opinion that taking the average of the CMIP ensemble is silly, that it has no meaning. The average of garbage is just average garbage.”
Assuming that a model run’s temperature time series is the realized temperature time series plus a noise component, and that the noise is normally distributed and each model run has an indipendent noise component, we could reduce the noise component’s amplitude by the root of the number of model runs; so for 100 model runs we could reduce the noise component by an amplitude factor of 10.
But, as I have repeatedly tried to explain, the deviation between the real system and a simulation of the real chaotic system with an iterative model of finite resolution leads to a deviation between real system and simulation that grows beyond all bounds over time. (I said exponentially in an earlier comment on another thread but that is imprecise; the correct definition is it grows beyond all bounds).
When the error grows over time beyond all bounds, it is clearly not ordinary normally distributed noise; and even reducing it by an amplitude factor of 10 only delays the growth beyond a given bound by a CONSTANT time.
Gavin’s argument is therefore bunkum.
There are two processes here.
If one takes a single model with random events built in, one can run it 1000’s of times to get an average where the random noise cancels out and the forcings are left, as Gavin suggested (or one could just turn off the random events and run it once). The spread between runs gives a reasonable estimate of the range random events play in modifying the forcings.
However, I agree with RGB that averaging 50 or so different models is meaningless. Each model should be individually compared to reality, and any that do a poor job should be discarded. When the IPCC plots a bunch of models and then implies that they are good because the actual measurements fall within the spread, they are spouting nonsense.
Tom Jones says:
June 15, 2013 at 8:54 am
It is worth reading Robert G Brown’s treatise over at Judith Curry’s blog. He sometimes writes here as rgbatduke. He is of the opinion that taking the average of the CMIP ensemble is silly, that it has no meaning. The average of garbage is just average garbage.
Tom, he has posted several posts here on this subject, you may want to read through those:
http://wattsupwiththat.com/2013/06/13/no-significant-warming-for-17-years-4-months/#comment-1334821
“Saying that we need to wait for a certain interval in order to conclude that “the models are wrong” is dangerous and incorrect for two reasons. First — and this is a point that is stunningly ignored — there are a lot of different models out there, all supposedly built on top of physics, and yet no two of them give anywhere near the same results!”
I think the post at Judith Curry’s blog is a continuation of the discussion as one reader asked Judith to publish it (see comment posted by RayG)
Bob, at OT this but you may find it relevant.
You may have seen my extension of Willis’ volcano stack idea:
http://climategrog.wordpress.com/?attachment_id=278
I then did a cumulative integral to estimate the volcanic effect once the natural cycles have been removed. (Much of what is usually attributed to volcanic cooling is false attribution of natural cycles).
http://climategrog.wordpress.com/?attachment_id=310
Then I realised that the pattern that comes out after the eruptions is the same as before I flattened it with the integral. That means that after the volcanoes the natural cycle is exaggerated. The climate response is to increase the natural variability.
It then hit me that this is proof of your idea that increased swings (ie the non ENSO neutral years) can inject energy into the climate system.
In the case of volcanoes we see tropical regions are not linear but controlled “like the body temperature of a mammal” as SteveF said recently 😉
I’ve posted a fuller account of all this at the Blackboard but moderation is slow over there. http://rankexploits.com/musings/2013/estimating-the-underlying-trend-in-recent-warming/
regards.