The Climate Model Muddle
Guest post by Ed Zuiderwijk
This is a posting about the epistemology of climate models, about what we can learn from them about the future. The answer will disappoint: not much. In order to convince you of the veracity of that proposition I will first tell you a little story, an allegory if you want, regarding a thought experiment, a completely fictitious account of what a research project might look like, and then apply whatever insight we gained (if any) to the climate modelling scene.
A thought experiment
Here’s the thought experiment: We want to make a compound that produces colour somehow (the mechanism how it does that is not really relevant). However, we specifically want a well-defined colour, prescribed by whatever application it is going to be used for. Say a shade of turquoise.
Now, our geologist and chemistry colleagues have proposed some minerals and compounds that could be candidate materials for our colourful enterprise. Unfortunately there is no information whatsoever what colours these substances produce. This circumstance is compounded by the fact that the minerals are exceedingly rare and therefore extremely expensive while synthetic ones are really difficult to make and therefore even more pricy. So, how do we proceed, how do we find the best compounds to try? Getting a sample of each of the many compounds and testing each of them for the colour it produces is out of the question. Therefore, what we do is to use modelling of the physics of the colour-producing process for each of the proposed compounds in order to find those which render turquoise, if there are any. Sounds straightforward enough but it isn’t because there are several different codes available, in fact 5 in total, that purport to do such a simulation, each with their own underlying assumptions and idiosyncrasies. We run these codes for the proposed compounds and find that, unfortunately, the colours they predict are inconsistent for individual compounds and generally all over the place.
For instance, take the compound Novelium1. The predicted colours range from yellow-green to deep violet with a few in between like green, blue or ultramarine, a factor 1.3 range in frequency; similar for the other candidates. In this situation the only way forward is doing an experiment. So we dig deep into the budget and get a sample of Novelium1, and see what colour it actually produces. It turns out to be orange-red which is pretty disappointing. We are back where we started. And because of our budgetary limitations we are at the point of giving up.
May we here introduce a member of our team. Let’s call him Mike. Mike is a bit pushy because he fully realises that were we to succeed in our aim it would get us some prestigious Prize or another, something he is rather keen on. He proposes to do the following: we take the model that predicted the colour closest to the actual one, that’s the model that gave us yellow-green, and tweak its parameters such that it predicts orange-red instead. This is not too difficult to do and after a few days jockeying on a keyboard he comes up with a tweaked model that produced the observed colour. Alacrity all around except for one or two more skeptical team members who insist that the new model must be validated by having it correctly predict the colour of compound Novelium2. With that Prize riding on it this clearly is a must, so we scrape the bottom of the budget barrel and repeat the exercise for Novelium2. The tweaked model predicts yellow. The experiment gives orange.
We gave up.
What does it mean?
Can we learn something useful from this story? In order to find out we have to answer three questions:
First, what do we know after the first phase of the project, the modelling exercise, before doing the experiment? Lamentably the answer is: nothing useful. With 5 different outcomes we only know for certain that at least 4 of the models are wrong but not which ones. In fact, even if the colour we want (turquoise) shows up we still know nothing. Because how can one be certain that the code producing it is the ‘correct result’ given the outcomes of the a priori equally valid other models? You can’t. If a model gave us turquoise it could just be a happy coincidence when the model itself is still flawed. The very fact that the models produce widely different outcomes tells us therefore that most probably all models are wrong. In fact, it is even worse: we can’t even be sure that the true colour produced by Novelium1 is inside the range yellow-green to violet, even if there were a model that produces the colour we want. In the addendum I give a simple probability based analysis to support this and subsequent points.
Second, what do we know after the unexpected outcome of the actual experiment? We only know for certain that all models are wrong (and that it is not the compound we are looking for).
Third, why did Mike’s little trick fail so miserably? What has happened there? The parameter setting of the original un-tweaked model encapsulates the best understanding – by its makers, albeit incomplete but that’s not really relevant – of the physics underpinning it. By modifying those parameters that understanding is diluted and if the ‘tweaking’ goes far enough it disappears completely, like the Cheshire Cat disappears the more you look at it. Tweaking such a model in hindsight to fit observations is therefore tantamountto giving up the claim that you understand the relevant physics underlaying the model. Any pretence of truly understanding the subject goes out of the window. And with it goes any predictive power the original model might have had. Your model has just become another very complex function fitted to a data set. As the mathematician and physicist John von Neumann once famously said of such practice: ‘with four parameters I can fit an elephant, and with five I can make him wiggle his trunk’. The tweaked model likely is a new incorrect model that coincidently produced a match with the data.
An application to climate models
Armed with the insights gleaned from the foregoing cautionary tale we are now in a position to make some fundamental statements about IPCC climate models, for instance the group of 31 models that form the CIMP6 ensemble (Eyring et al, 2019; Zelinka et al, 2020). The quantity of interest is the Equilibrium Climate Sensitivity (ECS) value, the expected long-term warming after a doubling of atmospheric CO2 concentrations. The predicted ECS values in the ensemble span a range from 1.8C at the low end to 5.6C at the high end, a whopping factor 3 in range, more or less uniformly occupied by the 31 models. Nature, however, may be cunning, even devious, but it is not malicious. There is only one ‘true’ ECS value that corresponds to the doubling of CO2 concentration in the real world.
Can we make any statement about this ensemble? Only these two observations:
First, most probably all those models are incorrect. This conclusion follows logically from the fact that there are many a priori equally valid models which can not be simultaneously correct. At most only one of these models can be correct, but given the remaining 30 incorrect models the odds are against any model at all being correct. In fact it can be shown that the probability that none of the models is correct can be as high as 0.6.
Second, we even cannot be sure that the true ECS is in the range of ECS values covered by the models. The probability of that being the case is 1.0-0.6=0.4, which means that the odds that the true ECS is in the range covered by the models are roughly 2 to 3 (and thus odds on that the true ECS is outside the range). The often made assumption that the ‘true’ ECS value must be somewhere in the range of outcomes from the models in the ensemble is based on a logical fallacy. We have absolutely no idea where the ‘true’ model – number 32, the ‘experiment’ – would land, inside or outside the range.
There are some qualifications to be made. What, for instance, does it mean: the model is ‘incorrect’? It means that it could be incomplete — there are concepts or principles missing in it that should be there — or, conversely, over-complete — with things that are but should not be there — or that there are aspects of it which are just wrong or wrongly coded, or all of those. Further, because many models of the ensemble have similar or even identical elements one might argue that the results of the ensemble models are not independent, that they are correlated. That means that one should consider the ‘effective number’ N of independent models. If N = 1 it would mean all models are essentially identical, with the range 1.8C to 5.6C an indication of the intrinsic error (which would be a pretty poor show). More likely N is somewhere in the range from 3 to 7 – with an intrinsic spread of, say, 0.5C for an individual model – and we are back at the hypothetical example above.
The odds of about 3 to 2 that none of the models is correct ought to be interesting politically speaking. Would you gamble a lot of your hard-earned cash on a horse with those odds? Is it wise to bet your country’s energy provision and therefore its whole economy on such odds?
Hindcasting
An anonymous reviewer of one of my earlier writings provided this candid comment, and I quote:
‘The track record of the GCM’s has been disappointing in that they were unable to predict the observed temperature hiatus after 2000 and also have failed to predict that tropopause temperatures have not increased over the past 30 years. The failure of the GCM’s is not due to malfeasance but modelling the Earth’s climate is very challenging.’
The true scientist knows that climate models are very much a work in progress. The pseudo scientist, under pressure to make the ‘predictions’ stick, has to come up with a way to ‘reconcile’ the models and the real world temperature data.
One way of doing so is to massage the temperature data in a process called ‘homogenisation’ (e.g. Karl et al, 2015). Miraculously the ‘hiatus’ disappears. A curious aspect of such ‘homogenisation’ is that whenever it is applied the ‘adjusted’ past temperatures are always lower, thus making the purportedly ‘man-made warming’ larger. Never the other way around. Obviously, you can do this slight of hand only once, perhaps twice if nobody is watching. But after that even the village idiot will understand that he has been had and puts the ‘homogenisation’ in the same dustbin of history as Lysenko’s ‘vernalisation’.
The other way is to tweak the model parameters to fit the observations (e.g. Hausfather et al., 2019). Not surprisingly, given the many adjustable parameters and keeping in mind von Neuman’s quip, such hind-casting can make the models match the data quite well. Alacrity all around in the sycophantic main-stream press, with sometimes hilarious results. For instance, a correspondent for a Dutch national newspaper enthusiastically proclaimed that the models had predicted correctly the temperatures of the last 50 years. This truly would be a remarkable feat because the earliest software that can be considered a ‘climate model’ dates from the early 1980s. However, a more interesting question is: can we expect such a tweaked model to have predictive power, in particular regarding the future? The answer is a resounding ‘no’.
Are climate models useless?
Of course not. They can be very useful as tools for exploring those aspects of atmospheric physics and the climate system that are not understood, or even of which the existence is not yet known. What you can’t use them for is making predictions.
References:
Eyring V. et al. Nature Climate Change, 9, 727 (2019)
Zelinka M. et al. Geophysical Research Letters, 47 (2020)
Karl T.R., Arguez A. et al. Science 348, 1469 (2015)
Hausfather Z., Drake H.F. et al. Geophysical Letters, 46 (2019)
Addendum: an analysis of probabilities
First the case of 5 models of which at most 1 can possibly be right. What is the probability that none of the models are correct? All models are a priori equally valid. We know that 4 of the models are not correct, so we know at once that the probability of any model being incorrect is at least 0.8. The remaining model may or may not be correct and in the absence of any further information both possibilities could be equally likely. Thus, the expectation is that, as a matter of speaking, half a model (of 5) is correct, which means the a priori probability of any model being incorrect is 0.9. For N models it is 1.0-0.5/N. The probability that all models fail then becomes: F=(1-0.5/N)^N which is about 0.6 (for N > 3). This gives us odds of 3 to 2 that none of the models are correct and it is more likely that none of the models are correct than that one of them is. (If we had taken F=(1-1/N)^N the numbers are about 0.34 with odds of 1 to 2)
Now an altogether different question. Suppose one of the models does give us the correct experimental result, what is the a posteriori probability that this model is indeed correct, given the results of the other models? Or, alternatively, that the model is incorrect even when it gives the ‘right’ result (by coincidence)? This posterior probability can be calculated using Bayes’ theorem,
P(X|Y) = P(Y|X)*P(X)/P(Y),
where P(X|Y) stands for the probability of X given Y and P(X) and P(Y) are prior probabilities for X and Y. In this case, X stands for ‘the model is incorrect’ and Y for ‘the result is correct’, in abbreviated form M=false, R=true. So the theorem tells us:
P(M=false|R=true) = P(R=true|M=false) * P(M=false) / P(R=true)
On the right-hand side the first term denotes the false-positive rate of the models, the second term is the probability that the model is incorrect and the third is the average probability that the result predicted is accurate. Of these we already know P(M=false)=0.9 (for 5 models). In order to get a handle on the other two, the ‘priors’, consider this results table:

The ‘rate’ columns represent a number of possible ensembles of models differing in the badness of the incorrect models. The first lot still give relatively accurate results (incorrect models that often return the about correct result, but not always; pretty unrealistic). The last with seriously poor models which on occasion give correct results (by happy coincidence) and a number of cases in between. Obviously, if a model is correct there is no false-negative (TF) rate. The false-positive rate is given by P(R=true|M=false) = FT. The average true result expected is given by 0.1*TT + 0.9*FT = 0.82 for the first group, 0.55 for the second and so on.
With these priors Bayes’ Theorem gives these posterior probabilities that the model is incorrect even if the result is right: 0.87, 0.82 etc. Even for seriously poor models with only a 5% false positive rate (the 5th set) the odds that a correct result was made by an incorrect model are still 1 to 2. Only if the false positive rate (of the incorrect models) drops dramatically (last column) can we conclude that a model that produces the experimental result is likely to be correct. This circumstance is purely due to the presence of the incorrect models in the ensemble. Such examples shows that in an ensemble with many invalid models the posterior likelihood of the correctness of a possibly correct model can be substantially diluted.
——–
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.
How to drive those with geological and/or chemistry backgrounds.
The chemical formula, very roughly, for turquoise is CuAl6(PO4)4(OH)8 · 4H2O.
Taking another example, an extremely desired coloring material by artists over millennia is lapis lazuli.
Lazurite, the central coloring component of lapis lazuli has a chemical formula of Na6Ca2(Al6Si6O24)(SO4,S,S2,S3,Cl,OH)2.
When Lapis is crushed and mixed with hardening oils/resins the color is reasonably stable. Turquoise, not really.
Aluminum Oxide is technically corundum, better known as sapphire and ruby.
The sapphire/ruby’s formula is Al2O3. The perceived colors are contaminants.
Padparadscha sapphire is apricot, salmon or pink sapphire; colors cause by trace amounts of Fe and Cr3+.
There are other gemstones that are primarily Al, aluminum compounds.
Topaz’s chemical formula is Al2(SiO4)(F,OH)2
Rubellite tourmaline, i.e. pink to red colored tourmaline has a chemical formula A(D3)G6(T6O18)(BO3)3X3Z.
Where:
Amethyst and smoky quartz share a simple chemical formula with clear quartz of SiO2.
Radiation causes the amethyst and the smoky colors.
Now, what was that you are saying about elephants wiggling their trunks?
Using colors and their formulation to invent a simple construct is an oxymoron. Colors and their formulae are anything but simple.
ATheoK
All very interesting, but not really germane to the analogy being presented.
While off topic, you might find this to be of interest:
https://www.msn.com/en-us/news/technology/scientists-finally-solve-the-mystery-behind-a-100-year-old-chemistry-experiment/ar-BB16cNE0
From the essay:
“ There is only one ‘true’ ECS value that corresponds to the doubling of CO2 concentration in the real world.”
Another climate assumption without evidence – that CO2 is the climate control knob. If the doubling of CO2 has any effect on temperature, that effect could depend on any of a multitude of other variable factors, perhaps even allowing for a possible range of outcomes much greater than 3. Perhaps, by the logic of the thought experiment in the essay, the probability is that ECS approaches 0.
There’s no such thing as ECS, in the real world it is swamped by emergent phenomena.
ECS, the Charney sensitivity, is defined as the long-term temperature effect of a doubling of CO2 concentration after the system has settled down in its new state, that is after all feedbacks and other changes have worked themselves through. It is difficult to see why the process, if it were repeated ab initio, should not give the same result, unless the result depends on the history of how the final state was reached. If that were the case then all models per definition are off, because any modeling would be impossible.
The notion has no bearing on whether the CO2 drives the climate or not. That depends on the value. If ECS is only about 0.5, which is what I consider a realistic estimate, then its role in the energy balance of the planet is marginal compared with other factors.
Why did you pick 0 to 2? There are actually more numbers below 2 (-∞ +∞).
Look at the chaos, globally, computer models have foisted on everyone with with the COVID-19 sc@mdemic!
Going back to first principles of science, or ‘natural philosophy’ – it becomes clear that all science is models, and all are incomplete – all are simplified representations of reality that are intrinsically and inevitably* less than the reality, itself.
Incompleteness therefore cannot be the problem, and problem there most certainly is. The problem in the end is simply one of utility. Some models like Newtonian and Einsteinian gravity give useful answers. Climate change models do not, in the sense that they predict the future climate. Of course they are most useful in achieving political control of energy, which is why they are promoted.
I think it is important to avoid trashing climate models for the wrong reasons. That leads to an anti-science mentality that can use the same justification to trash useful theories that do work, on the grounds that they cannot be complete. Indeed the whole thrust of the Left’s anti-science is based on the Post Hegelian and Post Marxist faux proposition, that the truth is relative to culture and culture can be changed, and therefore new truths can emerge. The facts of the world, according to them, are subject to the consciousness that perceives them. This ‘magic’ thinking is behind all the BLM and cultural diversity and gender politics and the attack on religion – these are all attempts to change the reality of the world, by changing how people think about it.
The fallacy, is that whilst they may come to dominate people’s perceptions of the world, they do not change what the world that is being perceived, actually is.
You may utterly believe that gender is a matter of choice. But it won’t – as we coarsely say here – put tits on a bull.
I think this is a very key and very important point that distinguishes between those who are broadly Left-thinking and those who are broadly Right. The Right is, in the end, more humble and does not believe the world can be fundamentally changed by changing people’s beliefs, whereas the Left in its arrogance believes that all it takes is blind faith in emotionally satisfying principles.
Politicians and those in the business of making profits of course realise that for their purposes, what is, in fact, the case, is really of little interest, since their concern is solely the manipulation of people’s perceptions in order to get them to do, or let the manipulators do, what they want [them] to do.
The utility of climate change models (and indeed all politically correct ‘woke’ models of today) as with any religious system, lies not on their ability to accurately predict the future, but in the construction of a public world-view that controls a moral framework that dictates the actions of the masses and sanctifies the actions of a few.
They are all very very bad science, but they are very very clever and very very good examples of ‘headology’ = getting people to believe in stuff that affects their behaviour to the extent you can enslave their minds.
And for those of a Christian persuasion, whilst I would say that at the core, Christianity is no different to any other ‘headology’ – it is a far far wiser and more benign belief system, than the ‘woke’ politics of the Left.
*the reasoning behind that is off topic and long.
Too little attention is paid to both the Old and New Testaments, not necessarily from a spiritual perspective but from a societal basis. We appreciate the Founders of the U.S. for the humanitarian aspects of the Declaration of Independence and the Constitution. We should also revere the authors of the Bible for the foundational principles of how to live together in a just and respectful society.
Didn’t some say with 6 parameters you can make an elephant fly? Where is John Galt?
The problem I have with the precautionary principle (I think someone implies above) is it assumes there is nothing to be lost. But what if we have it wrong, doesn’t that mean we don’t know what’s really going on, we have failed to understand long range weather putting our forecasting of dangerous droughts, floods, storms etc back decades and our climate forecasts way out all meaning we are potentially in an even more dangerous situation. And unnecessary actions have weakened us too much to cope and we didn’t prevent an extinction; other actions would have been more effective for the environment. As a hypothetical situation what if it’s about many indirect solar factors influencing stratosphere, jetstreams, oceans, influencing indirectly weather hence eventually climate and temperature and it’s nothing to do with CO2 or temperatures. So by looking at CO2 and temperatures we completely miss what’s going on, and CO2 has little impact and indirect solar changes weather patterns with disastrous impacts and eventually, indirectly temperature changes.
I too don’t understand the confidence in models if not only did they not forecast the pause (well, I can accept they may have failed to forecast it if something changed they could not be expected to forecast) but to then take many years to explain it, when the relevant measurements from all the sensors around the world, in space etc were presumably coming in every day and could be put straight into the model doesn’t make sense to me how we can be so confident in the models. I must be missing something.
Regards models failing due to it being a very difficult task: I confess I don’t know nearly enough, but what about the option of models failing because the wrong approach is used?
For example; doesn’t focusing on temperatures and averages miss far too much of the thermodynamics. I presume we put sensitivities in (e.g. to CO2) which I do not understand why we do that or think in terms of drivers or sensitivities; i.e. it is what it is but it isn’t; sensitivities irrelevant; isn’t it too complex to simplify that way? It seems to me that kind of approach leaves you dangerously open to assumptions.
Aren’t temperatures heavily influenced by weather? Isn’t weather heavily influenced by jet streams and stratospheric factors? Or have I got that wrong? How well do we understand what influences stratosphere and jetstreams and incorporate that in models?
“The tweaked model likely is a new incorrect model that coincidentally produced a match with the data.”
No, not really a coincidence; the model was repeatedly tweaked to produce the desired result.
A “coincidence” suggests a random concurrence of events or outcomes.
The tweaking of the model was a purposeful and intentional to produce a desired outcome.
That is not coincidence.
Perhaps the thought experiment modelers should have averaged all the results !!!
Current climate models have a large number of ‘adjustables’, in many cases of order 10 or even higher. The point von Neuman made is that you likely can fit the data with any of those models. Just take a randomly selected model and go through the motions. The coincidence is in the fact that there is a model that can be made to fit.
But if you do not believe in the climate models and the deleterious effects that a recent slight increase in atmospheric CO2 would bring, then you risk and international and national CO2 mediation targets never being met — all that fabulous technology of wind and solar gone to waste. All those carbon credit schemes wound-up! The modeled projections MUST be seen as worthy and to some degree accurate (even if they aren’t).
For without a robust belief in the models you risk the loss of levies on CO2, the money drying-up resulting in the UN being left with very little money — not enough money to support it’s continual social & bureaucratic expansion.
So think on, and realise how devastating that would be for the world!
[I need a sarc tag?]
“The real world is muddy and messy and full of things that we do not yet understand. It is much easier for a scientist to sit in an air-conditioned building and run computer models, than to put on winter clothes and measure what is really happening outside in the swamps and the clouds. That is why the climate model experts end up believing their own models.”
Freeman Dyson (RIP)
Huh, this writing is hilariously bad. Even the addendum, the “Analysis of probabilities” is complete garbage.
“We know that 4 of the models are not correct, so we know at once that the probability of any model being incorrect is at least 0.8.”
This is an elementary error. Even if we disregard the fact that a model outcome is not a simple value but multiple with ranges, error bands, and time dependency etc. so even differing predictions don’t necessarily mean any of the models is wrong (signifying a complete lack of understanding of modelling by the author), if we have a one out of n situation, the probabilities aren’t simply 1/n (or (n-1)/n), apart from the extremely simple cases. This is such an elementary error that it makes it obvious that the author has no idea what he’s talking about.
nyolci:
So you would argue that a GCM is not an “extremely simple case,” is that correct? In my thinking, I’m looking at the results of a GCM being a single output value (predicted temperature) based upon variable inputs. Thus in the case of GCM ensembles, don’t we have a “one out of n situation” where one or all of the ensemble predictions are either true or false? And isn’t that an extremely simple case? How am I wrong in this regard?
Thanks!
> And isn’t that an extremely simple case? How am I wrong in this regard?
Completely wrong. An extremely simple case is coin toss. But here’s this one: I’m either levitating right now or not. This is a 1 out of 2 situation. It’s obvious that the probabilities are very different from 50-50%. Assigning probabilities in this manner is an elementary error. And pls note, that this is just a marginal side issue with the article, the whole thing is problematic.
I realise that my conclusions hurt. But given that you so ademantly know that I don’t know what I am talking about I may assume that you yourself do know. So, please, enlighten us and tell us how we should choose the best model from the five and what the likelihood is that we made right choice.
Ed,
If all five models are wrong then how to you choose the best one?
Besides which, if all five models are wrong they the probability is that their average will be wrong as well. They can either be all high, all low, or that they might span the true value. So there is two out of three chances, 67%, that you will get the wrong answer from them or their average.
With no way to validate the models there is no way to know which possibility applies.. Since, in fact, they all run too hot compared to the satellite data and balloon data, it’s a pretty good guess that the first possibility is the most probable.
> But given that you so ademantly know that I don’t know what I am talking about I may assume that you yourself do know. So, please, enlighten us and tell us how we should choose the best model from the five and what the likelihood is that we made right choice.
It’s extremely ridiculous that you got even this one wrong. ‘Cos this is obvious that I can point out your amateurism without having to tell you about which one is the best (or even knowing that). Your sentence “With 5 different outcomes we only know for certain that at least 4 of the models are wrong but not which ones” gave away the fact that you’re clueless about modelling. Okay, I understand that your answer is part of some kinda debating tactic like “turn the tables and make him the one who has to explain”, but of course I’m not the one with the obligation.
Anyway, just to enlighten you, I give you some thoughts about modelling. I can give you these thoughts without knowing the specifics of the models. So models have initial conditions and external variables. All these models above may have these differently, even to the point that they represent different scenarios. I don’t know and I don’t care, for I only know that the modellers are scientists who may get their stuff wrong but these models are the result of a very long vetting and reviewing process so we may safely assume there are no serious errors in them.
Also modellers have numberless runs with slight changes to the external variables. The actual results are coming from the statistical analysis of these runs. It means that a single model doesn’t have an answer like “4” but a lot of variables with time dependence and error bands that may be comparable in magnitude to the result. In this sense even a single model can have a result like 1.8-5 to a single question. This is entirely expected, climate is a chaotic and extremely nonlinear system with tipping points etc. Models are very good at predicting the tendencies and the magnitudes.
All in all it’s very likely that all the models are good even if they predict different things. I know it’s hard for an outsider to swallow this but this is true regardless. (Actually it very likely that error bands overlap and this is what really counts). Equivalently we can say that very probably all models are wrong and this is not a contradiction. Especially in the presence of poorly known tipping points models may have wildly different short term predictions. These tend to even out in the long term, furthermore, models catch tendencies and magnitudes very well.
Hi Griff
“All models are wrong, but some models are useful.” Unfortunately, the utility of climate models is all too often assessed by how much funding they generate for their creators. Or by how well they “support” your biases.
The analogy confused me more than anything.
I think the most important things common to establishment climate models (AKA IPCC) are: uncertainty and flexibility. Uncertainty gives them a huge get-out-of-jail clause. They can make as many (models, projections) as they want but because it’s uncertain we’re not sure (for certain). That allows flexibility to add lots of kludges to the models (which help predict catastrophe), colloquially known as parameterizations; especially for positive feedbacks. But that’s OK, because they’re all based on the “known physics” of “settled science”; as every modeller (who wants a grant) uses the same basic greenhouse gas models descending from papers of Manabe and Wetherald (1967), and Held and Soden (2000). Anyone disputing whether Manabe and Wetherald / Held and Soden, even make sense is apparently disputing “settled science”. So is in “science denial” when they model radiative gas behaviour any other way. That’s why alternative models by: David Evans, saturated GHGE (Miskolczi), mean free path models, … are paid for by Big Oil, and are therefore evil.
Summary: IPCC are angels. People who dispute them are evil. But good and evil are terms associated with sin and religion; so instead the evil people are called deniers, and the good people called “consensus”. Better – sounds far more scientific. The consensus have ‘uncertainty’ on their side allowing them vast scope to exaggerate via parameterizations. No matter what – any model not supporting catastrophic interpretations (directly or indirectly) is suspect; which is all models not cribbed from Manabe and Wetherald (1967), and Held and Soden (2000).
PS: Uncertainty also means it does not matter when they’re wrong; because they never said they were right. Despite hundreds of trillions of climate policy money riding on the projections.
How wrong we all were thinking uncertainty was detrimental to the climate cause!