Pat Frank
LiG Metrology, Correlated Error, and the Integrity of the Global Surface Air Temperature Record has passed peer-review and is now published in the MDPI journal, Sensors (pdf).
The paper complements Anthony’s revolutionary Surface Stations project, in that the forensic analysis focuses on ideally located and maintained meteorological sensors.
The experience at Sensors was wonderfully normal. Submission was matter-of-fact. The manuscript editor did not flee the submission. The reviewers offered constructive criticisms. There was no defense of a favored narrative. There was no dismissive language.
MDPI also has an admirable approach to controversy. The editors, “ignore the blogosphere.” The contest of ideas occurs in the journal, in full public view, and critical comment must pass peer-review. Three Huzzahs for MDPI.
LiG Metrology… (hereinafter LiG Met.) returns instrumental methods to the global air temperature record. A start-at-rock-bottom 40 years overdue forensic examination of the liquid-in-glass (LiG) thermometer.
The essay is a bit long and involved. But the take-home message is simple:
- The people compiling the global air temperature record do not understand thermometers.
- The rate or magnitude of climate warming since 1900 is unknowable.
Global-scale surface air temperature came into focus with the 1973 Science paper of Starr and Oort, Five-Year Climatic Trend for the Northern Hemisphere. By 1983, the Charney Report, Carbon Dioxide and Climate was 4 years past, Stephen Schneider had already weighed in on CO2 and climate danger, Jim Hanson was publishing on his climate models, ice core CO2 was being assessed, and the trend in surface air temperature came to focused attention.
Air temperature had become central. What was its message?
To find out, the reliability of the surface air temperature record should have been brought to the forefront. But it wasn’t. Air temperature measurements were accepted at face value.
Errors and uncertainty were viewed as external to the instrument; a view that persists today.
LiG Met. makes up the shortfall, 40 years late, starting with the detection limits of meteorological LiG thermometers.
The paper is long and covers much ground. This short summary starts with an absolutely critical concept in measurement science and engineering, namely:
I. Instrumental detection limits: The detection limit registers the magnitude of physical change (e.g., a change in temperature, DT) to which a given instrument (e.g., a thermometer) is able to reliably respond.
Any read-out below the detection limit has no evident physical meaning because the instrument is not reliably sensitive to that scale of perturbation. (The subject is complicated; see here and here.)
The following Table provides the lower limit of resolution — the detection limits — of mercury LiG 1C/division thermometers, as determined at the National Institute of Standards and Technology (NIST).
NIST 1C/division Mercury LiG Thermometer Calibration Resolution Limits (2σ, ±C)

These are the laboratory ideal lower limits of uncertainty one should expect in measurements taken by a careful researcher using a good-quality LiG 1C/division thermometer. Measurement uncertainty cannot be less than the lower limit of instrumental response.
The NASA/GISS air temperature anomaly record begins at 1880. However, the largest uncertainties in the modern global air temperature anomaly record are found in the decades 1850-1879 published by HadCRU/UKMet and Berkeley BEST. The 2s root-mean-square (RMS) uncertainty of their global anomalies over 1850-1880 is: HadCRU/UKMet = ±0.16 C and Berkeley BEST = ±0.13 C. Graphically:
Figure 1: The LiG detection limit and the mean of the uncertainty in the 1850-1880 global air temperature anomalies published by the Hadley Climate Research Unit of the University of East Anglia in collaboration with the UK Meteorological Office (HadCRU/UKMet) and by the Berkeley Earth Surface Temperature project (Berkeley BEST).
That is, the published uncertainties are about half the instrumental lower limit of detection — a physical impossibility.
The impossibility only increases with the decrease of later uncertainties (Figure 6, below). This strangeness shows the problem that ramifies through the entire field: neglect of basics.
Summarizing (full details and graphical demonstrations in LiG Met.):
Non-linearity: Both mercury and especially ethanol (spirit) expand non-linearly with temperature. The resulting error is small for mercury LiG thermometers, but significant for the alcohol variety. In the standard surface station prior to 1980, an alcohol thermometer provided Tmin, which puts 2s = ±0.37 C of uncertainty into every daily land-surface Tmean. Temperature error due to non-linearity of response is uncorrected in the historical record.
Joule-drift: Significant bulb contraction occurs with aging of thermometers manufactured before 1885, and is most rapid in those made with lead-glass. Joule-drift puts a spurious 0.3-0.7 C/century warming trend into a temperature record. Figure 4 in LiG Met. presents the Pb X-ray fluorescence spectrum of a 1900-vintage spirit meteorological thermometer purchased by the US Weather Bureau. Impossible-to-correct error from Joule drift makes the entire air temperature record prior to 1900 unreliable.
The Resolution message: All of these sources of error and uncertainty — detection limits, non-linearity, and Joule drift — are inherent to the LiG thermometer and should have been evaluated right at the start. Well before making any serious attempt to construct a record of historical global surface air temperature. However, they were not. They were roundly neglected. Perhaps most shocking is professional neglect of instrumental detection limit.
Figure 2 shows the impact of the detection limit alone on the 1900-1980 global air temperature anomaly record.
Land surface temperature means include the uncorrected error from non-linearity of spirit thermometers. Sea surface temperatures (SSTs) were measured with mercury LiG thermometers only (no spirit LiG error). The resolution uncertainty for the global air temperature record prior to 1981 was calculated as,
2sT = 1.96 ´ sqrt[0.7 ´ (SST resolution)2 + 0.3 ´ (LST resolution)2]
= 1.96 ´ sqrt[0.7 ´ (0.136)2 + 0.3 ´ (0.195)2] = ±0.306 C, where LS is Land-Surface.
But global air temperature change is reported as an anomaly series relative to a 30-year normal. Differencing two values requires adding their uncertainties in quadrature. The resolution of a LiG-based 30-year global temperature normal is also 2s = ±0.306 C. The resolution uncertainty in a LiG-based global temperature anomaly series is then
2s = 1.96 ´ sqrt[(0.156)2 + (0.156)2] = ±0.432 C

Figure 2: (Points), 1900 – 1980 global air temperature anomalies for: panel a, HadCRUT5.0.1.0 (published through 2022); Panel b, GISSTEMP v4 (published through 2018), and; Panel c, Berkeley Earth (published through 2022). Red whiskers: the published 2s uncertainties. Grey whiskers: the uniform 2 s = ±0.432 C uncertainty representing the laboratory lower limit of instrumental resolution for a global average annual anomaly series prior to 1981.
In Figure 2, the mean of the published anomaly uncertainties ranges from 3.9´ smaller than the LiG resolution limit at 1900, to 5´ smaller at 1950, and nearly 12´ smaller at 1980.
II. Systematic error enters into global uncertainty. Is temperature measurement error random?
Much of the paper tests the assumption of random measurement error; an assumption absolutely universal in global warming studies.
LiG Met. Section 3.4.3.2 shows that differencing two normally distributed data sets produces another normal distribution. This is an important realization. If measurement error is random, then differing two sets of simultaneous measurements should produce a normally distributed error difference set.
II.1 Land surface systematic air temperature measurement error is correlated: Systematic temperature sensor measurement calibration error of proximately located sensors turns out to be pair-wise correlated.
Matthias Mauder and his colleagues published a study of the errors produced within 25 naturally ventilated HOBO sensors (gill-type shield, thermistor sensor), relative to an aspirated Met-One precision thermistor standard. Figure 3 shows one pair-wise correlation of the 25 in that experimental set, with correlation r = 0.98.
Figure 3: Histogram of error in HOBO number 14 (of 25). The StatsKingdom online Shapiro-Wilk normality test (2160 error data points) yielded: 0.979, p < 0.001, statistically non-normal. Inset: correlation plot of measurement error — HOBO #14 versus HOBO #15; correlation r = 0.98.
High pair-wise correlations were found between all 25 HOBO sensor measurement error data sets. The Shapiro-Wilk test, has the greatest statistical power to indicate or reject the normality of a data distribution and showed that every single measurement error set was non-normal.
LiG Met. and the Supporting Information provide multiple examples of independent field calibration experiments, that produced pair-wise correlated systematic sensor measurement errors. Shapiro-Wilk statistical tests of calibration error data sets invariably indicated non-normality.
Inter-sensor correlation in land-surface systematic measurement field calibration error, along with non-normal distributions of difference error data sets, together falsify the general assumption of strictly random error. No basis in evidence remains to permit diminishing uncertainty as 1/ÖN.
II.2.1 Sea-Surface Temperature measurement error is not random: Differencing simultaneous bucket-bucket and bucket-engine-intake measurements again yields the measurement error difference, De2,1. If measurement error is random, a large SST difference data set,De2,1, should have a normal distribution.
Figure 4 shows the result of a World Meteorological Organization project, published in 1972, which reported differences of 13511 simultaneously acquired bucket and engine-intake SSTs from all manner of ships, at low and high N,S latitudes and under a wide range of wind and weather conditions. The required normal distribution is nowhere in evidence.
Figure 4: Histogram of differences of 13511 simultaneous engine-intake and bucket SST measurements during a large-scale experiment carried out under the auspices of the World Meteorological Organization. The red line is a fit using two Lorentzians and a Gaussian. The dashed line marks the measurement mean.
LiG Met. presents multiple independent large-scale bucket/engine-intake difference data sets of simultaneously measured SSTs. The distributions were invariably non-normal, demonstrating that SST measurement errors are not random.
II.2.2 The SST measurement error mean is unknown: The semivariogram method, taken from Geostatistics, has been used to derive the shipboard SST error mean, ±emean. The assumption again is that SST measurement error is strictly random, but with a mean offset.
Subtract emean, and get a normal distribution of with a mean of zero, and an uncertainty diminishing as 1/ÖN.
However, LiG Met. Section 3.4.2 shows that the semivariogram analysis doesn’t produce ±emean, but instead ±0.5Demean, half themean of the error difference. Subtraction does not leave a mean of zero.
Conclusion about SST: II.2.1 shows measurement error is not strictly random. II.2.2 shows ignorance of the error mean. No grounds remain to diminish SST uncertainty as 1/ÖN.
II.2.3 The SST is unknown: In 1964 (LiG Met. Section 3.4.4) Robert Stevenson carried out an extended SST calibration experiment aboard the VELERO IV oceanographic research vessel. Simultaneous high-accuracy SST measurements were taken from the VELERO IV and from a small launch put out from the ship.
Stevenson found that the ship so disturbed the surrounding waters that the SSTs measured from the ship were not representative of the physically true water temperature (or air temperature). No matter how accurate, the bucket, engine-intake, or hull-mounted probe temperature measurement did not reveal the true SST.
The only exception was an SST obtained using a prow-mounted probe, but iff the measurement was made when the ship was heading into the wind “or cruising downwind at a speed greater than the wind velocity.”
Stevenson concluded, “One may then question the value of temperatures taken aboard a ship, or from any large structure at sea. Because the measurements vary with the wind velocity and the orientation of the ship with respect to the wind direction no factor can be applied to correct the data. It is likely that the temperatures are, therefore, useless for any but gross analyses of climatic factors, excepting, perhaps, those taken with a carefully-oriented probe.”
Stevenson’s experiment may be the most important investigation ever carried out of the veracity of ship-derived SSTs. However, the experiment generated scant notice. It was never repeated or extended, and the reliability question of SSTs the VELERO IV experiment revealed has generally been by-passed. The journal shows only 5 citations since 1964.
Nevertheless, ship SSTs have been used to calibrate satellite SSTs probably through 2006. Which means that earlier satellite SSTS are not independent of the large uncertainty in ship SSTs.
III. Uncertainty in the global air temperature anomaly trend: We now know that the assumption of strictly random measurement error in LSTs or SSTs is unjustified. Uncertainty cannot be presumed to diminish as 1/ÖN.
III.1 For land-surface temperature, uncertainty was calculated from:
- LiG resolution (detection limits, visual repeatability, and non-linearity).
- systematic error from unventilated CRS screens (pre-1981).
- interpolation from CRS to MMTS (1981-1989).
- unventilated Min-Max Temperature System (MMTS) sensors (1990-2004).
- Climate Research Network (CRN) sensors self-heating error (2005-2010).
Over 1900-1980, resolution uncertainty was combined in quadrature with the uncertainty from systematic field measurement error, yielding a total RMS uncertainty 2s = ±0.57 C in LST.
III.2 For sea-surface temperature uncertainty was calculated from Hg LiG resolution combined with the systematic uncertainty means of bucket, engine intake and bathythermograph measurements scaled by their annual fractional contribution since 1900.
SST uncertainty varied due to the annual change in fractions of bucket, engine-intake and bathythermograph measurements. Engine intake errors dominated.
Over 1900-2010 uncertainty in SST was RMS 2s = ±1.38 C.
III.3 Global: Annual uncertainties in land surface and sea surface again were combined as:
2sT = 1.96 ´ sqrt[0.7 ´ (SST uncertainty)2 + 0.3 ´ (LST uncertainty)2]
Over 1900-2010 the RMS uncertainty in global air temperature was found to be, 2s = ±1.22 C.
The uncertainty in an anomaly series is the uncertainty in the air temperature annual (or monthly) mean combined in quadrature with the uncertainty in the selected 30-year normal period.
The RMS 2s uncertainty in the NASA/GISS 1951-1980 normal is ±1.48 C, and is ±1.49 C in the HadCRU/UEA and Berkeley BEST 1961-1990 normal.
The 1900-2010 mean global air temperature anomaly is 0.94 C. Using the NASA/GISS normal, the overall uncertainty in the 1900-2010 anomaly is,
2s = 1.96 ´ sqrt[(0.755)2 + (0.622)2] = ±1.92 C
The complete change in air temperature between 1900-2010 is then 0.94±1.92 C.
Figure 5 shows the result applied to the annual anomaly series. The red whiskers are the 2s quadratic annual combined RMS of the three major published uncertainties (HadCRU/UEA, NASA/GISS and Berkeley Earth). The grey whiskers include the combined LST and SST systematic measurement uncertainties. LiG resolution is included only through 1980.
The lengthened growing season, the revegetation of the far North, and the poleward migration of the northern tree line provide evidence of a warming climate. However, the rate or magnitude of warming since 1850 is not knowable.
Figure 5: (Points), mean of the three sets of air temperature anomalies published by the UK Met Office Hadley Centre/Climatic Research Unit, the Goddard Institute for Space Studies, or Berkeley Earth. Each anomaly series was adjusted to a uniform 1951-1980 normal prior to averaging. (Red whiskers), the 2s RMS of the published uncertainties of the three anomaly records. (Grey whiskers), the 2s uncertainty calculated as the lower limit of LiG resolution (through 1980) and the mean systematic error, combined in quadrature. In the anomaly series, the annual uncertainty in air temperature was combined in quadrature with the uncertainty in the 1951-1980 normal. The increased uncertainty after 1945 marks the wholesale incorporation of ship engine-intake thermometer SST measurements (2s = ±2 C). The air temperature anomaly series is completely obscured by the uncovered uncertainty bounds.
IV. The 60-fold Delusion: Figure 6 displays the ratio of uncovered and published uncertainties, illustrating the extreme of false precision in the official global air temperature anomaly record.
Panel a is (LiG ideal laboratory resolution) ¸ Published. Panel b is total (resolution plus systematic) ¸ Published.
Panel a covers 1850-1980, when the record is dominated by LiG thermometers alone. The LiG lower limit of detection is a hard physical bound.
Nevertheless, the published uncertainty is immediately (1850) about half the lower limit of detection. As the published uncertainties get ever smaller traversing the 20th century, they get ever more unphysical; ending in 1980 at nearly 12´ smaller than the LiG physical lower limit of detection.
Panel b covers the 1900-2010 modern period. Joule-drift is mostly absent, and the record transitions into MMTS thermistors (1981) and CRN aspirated PRTs (post-2004). The comparison for this period includes contributions from both instrumental resolution and systematic error.
The uncertainty ratio now maxes out in 1990, with the published version about 60´ smaller than the combined instrumental resolution plus field measurement error. By 2010, the ratio declines to about 40´ because ship engine-intake measurements make an increasingly small contribution after 1990 (Kent, et al., (2010)).
Figure 6: Panel a. (points), the ratio of the annual LiG resolution uncertainties divided by the RMS mean of the published uncertainties (2s, 1850-1980). Panel b. (points), the ratio of the annual total measurement uncertainties divided by the RMS mean of the published uncertainties (2s, 1900-2010). Inset: the fraction of SSTs obtained from engine-intake thermometers and hull-mounted probes (a minority). The drop-off of E-I temperatures in the historical record after 1990 accounts for the declining uncertainty ratio.
V. The verdict of instrumental methods:
Inventory of error and uncertainty in the published air temperature record:
NASA/GISS: incomplete spatial coverage, urban heat islands, station moves.
Hadley Centre/UEA Climate Research Unit: random measurement error, instrumental or station moves, changes in instrument type or time-of-reading, sparse station data, urban heat island, bias due to changes in sensor exposure (screen type), bias due to changes in methodology of SST measurements.
Berkeley Earth: non-climate related noise, incomplete spatial coverage, and limited efficacy of their statistical model.
No mention by anyone of anything concerning instrumental methods of analysis, in a field completely dominated by instruments and measurement.
Instead, one encounters an analysis conveying no attention to instrumental limits of accuracy or to the consequences attending their technical development, or of their operational behavior. This, in a field where knowledge of such things is a pre-requisite.
Those composing the air temperature record display no knowledge of thermometers. Perhaps the ultimate irony.
No appraisals of LiG thermometers as instruments of measurement despite their preponderance in the historical temperature record. Nothing of the very relevant history of their technical evolution, of their reliability or their resolution or detection limits.
Nothing of the known systematic field measurement errors that affect both LiG thermometers and their successor temperature sensors.
One might expect those lacunae from mathematically adept science dilettantes, who cruise shallow numerical surface waters while blithely unaware of the instrumental depths below; never coming to grips with the fundamentals of study. But not from professionals.
We already knew that climate models cannot support any notion of a torrid future. Also, here and here. We also know that climate modelers do not understand physical error analysis. Predictive reliability: a mere bagatelle of modern modeling?
Now we know that the air temperature record cannot support any message of unprecedented warming. Indeed, almost no message of warming at all.
And we also now know that compilers of the air temperature record evidence no understanding of thermometers, incredible as that may seem. Instrumental methods: a mere bagatelle of modern temperature measurement?
The climate effects of our CO2 emissions, if any, are invisible. The rate or magnitude of the 20th century change in air temperature is unknowable.
With this study, nothing remains of the IPCC paradigm. Nothing. It is empty of content. It always was so, but this truth was hidden under the collaborative efforts of administrative embrace, partisan shouters, character assassins, media propagandists, and professional abeyance.
All those psychologists and sociologists who published their profoundly learned insights into the delusional minds, psychological barriers, and inadequate personalities plaguing their notion of climate/science deniers are left with egg on their faces or in their academic beards. In their professional acuity, they inverted both the order and the perceivers of delusion and reality.
We’re faced once again with the enormity of contemplating a science that has collapsed into a partisan narrative; partisans hostile to ethical practice.
And the professional societies charged with embodying physical science, with upholding ethics and method — the National Academies, the American Physical Society, the American Institute of Physics, the American Chemical Society — collude in the offense. Their negligence is beyond shame.
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.





Repeated measurements of different things, as with the temperature at different times and places, cannot be more accurate than any one individual measurement.
So claims otherwise are tendentious.
Those measurements are then “adjusted” in the process of assembling the anomaly records. It is ludicrous to assume or assert that those adjustments are random, or that they correct the inaccuracies of the measurements.
removing structural Bias is Standard in all Science.
“Correcting” temperature records in standard in Climate “Science”.
Pat Frank has done it again! His ground-breaking 2019 paper on the effect of propagation of error in a single initial condition in the wretched models showed that any temperature prediction that fell between -12 and +12 K was statistically meaningless. I had the honor to be present at the annual planetary-emergencies meeting of the World Federation of Scientists at which Pat first presented his findings. He was shouted down by the rabble of profiteers of doom, but bravely stood his ground. His paper on the failure of clahmatawlagiests to understand such elementary statistical techniques as error propagation was submitted to 13 successive journals, 12 of which rejected it with spectacularly half-witted “reviews” that amounted to little more than “This paper does not accord with the Communist Party Line and, since the Communist Party Line cannot be wrong, the paper must be wrong and must not be printed.” Fortunately, 13th time lucky, the paper was reviewed by (among others) Professor Karl Wunsch, who is a believer in the official narrative but, unlike nearly all other believers, retains enough intellectual honesty to know when the Party Line must be amended, at least so as to prevent the Party from looking more ridiculous than it already does. He could find no fault with the paper, and recommended publication. Though there was some strikingly inept whining from the usual paid trolls for climate Communism, both in comments here and in the wider blogosphere, Pat’s paper of 2019 stands unrefuted to this day. It was, until now, the most important paper published in global warming studies.
Now the second shoe drops with Pat’s latest paper. First of all, warmest congratulations to him for getting it past the gatekeepers and into print in a leading journal. He has very fairly stated that there are physical indications that the world is warming somewhat, but has concluded that those who measure global land and sea surface temperature know as little about the physical characteristics of thermometers as they do about statistical precision. As Pat rightly points out, if data are ill-resolved as well as incorrectly sampled (as at sea) with inconstantly-performing instruments (as on land), and then have erroneous statistical analyses applied to the meaningless results, all the predictions fall well within the correctly-evaluated uncertainty envelope and, therefore, tell us nothing – absolutely nothing whatsoever – about the amplitude of global warming in recent times. He is correct to conclude that one cannot use terrestrial temperature records as the basis for attempting to derive global temperature anomalies, still less a trend therein. Already the paid trolls are piling in with their pseudo-statistical gibberish, as usual divorced from the reality that if each temperature measurement is itself meaningless it is not possible to derive any realistic trend in such meaningless pseudo-measurements (except by inadvertence) whatever technique of statistical prestidigitation is adopted. All that can legitimately be pleaded is that the weather worldwide seems by direct observation of potentially-temperature-correlated physical indicators to have become warmer to an unknown and thus far net-beneficial degree.
Pat’s latest paper, which he kindly sent me a couple of days ago, is a most delightful read. And, like his head posting here, it is expressed with a powerful eloquence rare among the scientific community. I look forward to seeing how the trolls try to attack his fascinating real-world finding that the glass used in thermometer bulbs changes its characteristics over time. The trolls would be well advised to stay away from Pat on this, for materials science of this kind is one of the specialisms of the Stanford Linear Accelerator Laboratory, where Pat is now a most distinguished Emeritus Professor. As usual, the Marxstream media will follow their usual tactic: since the paper is damaging to the Communist Party Line but cannot legitimately be attacked scientifically they will simply refuse to publish it, so that almost no one ever gets to hear of it. However, I shall see to it that this second ground-breaking paper is put into the hands of the governments that I quietly advise on the climate question.
With this paper, what little scientific credibility the official climate-change narrative once possessed has collapsed. The world will in due course have very good cause to thank Pat Frank for saving us from the darkness and abject poverty to which the profiteers of doom would callously subject the countries of the West and – revealingly – only the countries of the West.
As I see it, this new paper also implicates – as meaningless – any of the observation-based estimates of a climate system response (whether ECS or TCR) to CO2 “forcing” using such time series records of temperature anomalies. This is not meant as a criticism of any of those investigators.
The earth’s energy imbalance is the most important number, since it tells us directly how far away from equilibrium we are. There is a belief by many that we can control the EEI by controlling CO2 emissions. What’s interesting is the EEI appears to be quite insensitive to changing incoming radiation, according the the CERES satellite data. In fact, it would not be unreasonable to conclude that the climate sensitivity is essentially zero. The earth is deciding what EEI it wants and not even a variation in the sun’s luminosity can change that. By extension, neither can CO2.
Thanks for your reply. Sorry for the late response.
“The earth’s energy imbalance is the most important number, since it tells us directly how far away from equilibrium we are.”
What is that number, and what is the uncertainty associated with it? I know what NASA says. Not too sure about that 10 milliwatts per square meter precision or the +/- 100 milliwatts per square meter uncertainty. [ https://ceres.larc.nasa.gov/science/ ]
Consider that the planet as a whole experiences a surface warming and cooling response of about 3.8C on an annual cycle – because absorption of solar radiation in the NH and SH is not symmetric. How far away from equilibrium are we? That cyclic surface temperature response is displayed here from the output of a reanalysis model.
https://climatereanalyzer.org/clim/t2_daily/
I agree that “In fact, it would not be unreasonable to conclude that the climate sensitivity is essentially zero.” Or perhaps put it this way – The surface temperature response to incremental increases in non-condensing GHGs cannot be reliably distinguished from zero by any means we have available to us.
I have yet to see an analysis of the ~ 4.3c peak variation in equilibrium temperature which must occur from peri- to ap- helion around our orbit . It’s confounded with hemispherical differences , but can be calculated . Your 3.8c variation is the first I’ve seen citing the annual variation , but I’d like to see an analysis factoring the orbital contribution .
I use the “3.8C” value because of Figure 10.2 in Javier Vinos’ recent book. I haven’t progressed that far actually reading his book yet to see if there is a deeper analysis.
https://judithcurry.com/wp-content/uploads/2022/09/Vinos-CPPF2022.pdf
By eyeball, the reanalysis web page I linked gives about the same value.
Thanks for your kind words, Christopher, and the accurate summary.
You’ve been a good friend and a comrade-in-arms for years, in the struggle for science and sanity. So you deserved first notice.
Still, the promotion to Professor wasn’t necessary. 🙂 Scientific Staff Emeritus is all I am.
Here, here! Well stated, CMoB!
Dear karlomonte,
You probably mean: hear, hear!
(I wonder if there is an ISO standard for that?)
All the best,
Bill
Wrong.
And your dig on ISO standards is lame, even by your standards (pun intended).
“However, I shall see to it that this second ground-breaking paper is put into the hands of the governments that I quietly advise on the climate question.”
I don’t have such a mailing list but I’m thinking an abstract of this paper & Pat’s previous one should be sent to every government official even remotely connected to acting on the ‘climate change’ scam. But also to every ‘climate scientist’ and anyone who might review their papers. After having knowledge of these 2 papers if any supposed ‘climate scientist’ tries to publish a paper that doesn’t address the propagation of errors in any model they use and the measurement error inaccuracies in data is committing outright fraud. Not just a bit of ‘professional misconduct’…fraud.
You grossly overestimate the intellectual prowess and scientific understanding of Government bureaucrats.
Not to mention their unwillingness to hear or see anything outside of their ideaology-driven mindset. It’s difficult to work out who might actually be receptive to this with a voice that might be heard.
Thunderous applause for this message Monckton, at least from this quarter. The temperature record is not fit for purpose and neither are the climate models!
“With this paper, what little scientific credibility the official climate-change narrative once possessed has collapsed.”
Does this mean no more “pause” articles or realitymeters? If it’s impossible to know what the trend is how can you compare it to the models.
The pathetic Bellman, who lacks the courage to publish in its own name, is perhaps unfamiliar with the difference between the surface and space, where the UAH satellite measurements are taken. The microwave sounding units on the satellites do not rely on glass-bulb mercury thermometers but on platinum- resistance thermometers. We use the satellite record, not the defective terrestrial record, for the Pause analyses.
The Pause may well end with the new El Niño.
Yet the satellite and surface data agree much more than they disagree. Some here insist the uncertainty in the UAH data is much bigger than that claimed here for the surface data. Maybe Pat Frank could say what he thinks the uncertainty is for UAH.
A distortion bordering on a lie. But what else should be expected from a trendologist?
So what do you think the monthly uncertainty of UAH is now?
In the past I think you’ve quoted figures of over 1°C, and you’ve called me a Spencer groupie when I thought it was probably a bit more certain.
Another lie.
You don’t really care what a real number is because it might upset your pseudoscience notions of tiny “uncertainties”,.
Sorry, you are correct it was actually ±3.4°C.
“A lower estimate of the UAH LT uncertainty would be u(T) = ±1.2°C; combined with a baseline number this becomes: u(T) = sqrt[ 1.2^2 + 1.2^2 ] = 1.7°C, and U(T) = 2u(T) = ±3.4°C”
https://wattsupwiththat.com/2021/12/02/uah-global-temperature-update-for-november-2021-0-08-deg-c/#comment-3401727
Oh look bellcurvewhinerman has little ol’ me stashed in his enemies files!
*preens*
And you still can’t read with comprehension.
Oh look, the troll calls me a lier, then whines when I quote his own words back at him.
Have you heard of this thing called the internet? It allows you to do searches through other people’s comments. Maybe you should remember that next time you deny saying something.
The truth is you don’t care what a real number might be; your task is to denigrate anything that puts the precious GAT trend lines in doubt. This is the realm of pseudoscience. Thus you have to nitpick what Pat Frank writes regardless of relevancy.
So very Stokesian.
And you still can’t read with comprehension.
“And you still can’t read with comprehension.”
As he quoted:
“A lower estimate of the UAH LT uncertainty would be u(T) = ±1.2°C; combined with a baseline number this becomes: u(T) = sqrt[ 1.2^2 + 1.2^2 ] = 1.7°C, and U(T) = 2u(T) = ±3.4°C”
It’s hard to see how that means anything other than that the uncertainty is ±3.4°C.
Oh look! Nitpick can’t read either.
No surprise here.
What interests me is the fact that they ADDED THE INDIVIDUAL UNCERTAINTIES of temp and baseline (in quadrature). Why is the same thing not done in climate science when computing anomalies??
Oh, and
“So he can argue with all the other Spencer groupies like yourself?”
https://wattsupwiththat.com/2022/06/01/uah-global-temperature-update-for-may-2022-0-17-deg-c/#comment-3527275
Heh more enemies files.
Got any more? This is amusing.
T. Mo (2017) Postlaunch Calibration of the NOAA-18 Advanced Microwave Sounding Unit-A
Figure 1 shows the temperature specification error is ±0.3 C.
Post-launch error was about ±0.2 C for 250 days during 2005. From Table II, MSU avg. NedT: specification, 0.43±0.27; measured, 0.25±0.17.
Does that mean the UAH monthly and annual average should be that large?
Stokes: “ERRORS ALL CANCEL!”
But Frank believes uncertainties do not cancel if they are caused by resolution. That’s why I’m asking if he applies the same logic to UAH. I’m hoping so as it means the end of all the pause nonsense.
If the instruments are the same, LIG in this case, then how can the resolution uncertainty cancel?
Once again, you are showing, for everyone to see, that you have absolutely no understanding of the concept of uncertainty!
I’m not going to keep explaining this to you. You claim to have done all the exercises in Taylor, bit seem to have forgotten the ones demonstration how this can happen.
But in this case you are ignoring the point. I’m not asking about surface temperature, but about UAH. If your logic is correct then it isn’t possible to reduce the uncertainty in the satellite data. If you can’t do that then Monckton is wrong to claim that UAH data does not have the same issues is claimed by Frank.
“Pipe down, class! Da expert ix tellin’ y’all like it is!”
Still waiting for your expert opinion on the uncertainty of UAH.
[Free clue, feeling generous this AM]
And what you continually to fail to grasp is that whatever I think or believe about this number is totally irrelevant.
Finally something we can both agree on.
The WHOOOSH sound was the clue going over your head at Mach 5.
[no doubt the clue will go directly into your stupid enemies files]
“[but] seem to have forgotten the ones demonstration how this can happen.”
Not with the lower limit of instrumental detection.
The UAH uncertainties I posted are from calibrations. They’re not instrumental detection limits, which are categorically different. Pay attention.
He can’t.
Bellman July 1, 2023 5:07 am
The main surface datasets disagree with both each other and the satellite data …
w.
But as I said, they agree more than they disagree. I’m talking particularly of the monthly and annual variations, which your very smoothed graph hides.
I’m also a bit puzzled why HadCRUT5 is showing less warming than UAH, assuming the red and black line is UAH.
Here are the three main ones set to the same base period. HadCRUT and GISS are nearly identical, but UAH is noticeably slower.
Here’s a scatter plot of UAH6 and GISS monthly anomalies. The correlation coefficient is 0.83.
And there’s the same but showing the claimed ±1.6°C 2σ uncertainty range.
Finally, here’s the same showing annual values.
Correlation coefficient is 0.91.
In his paper Frank claims the 2σ uncertainty of the global average temperature anomalies is around ±2 C (see figure 19). You’re graph shows no where near that level of disagreement.
BTW…I’m confused about UAH in your graph. The absolute values of UAH TLT are around -10 C or about 25 C lower than the surface temperature so why is the red-black line assuming it is UAH TLT so high up in the graph?
All the surface air temperature compilations use the same data, bdgwx. Resemblance doesn’t mean physically correct.
The only valid UAH-surface temperature comparisons would employ raw data.
But UAH is not surface data. The satellite data is the closes thing I can think of to an independent comparison with surface data. The uncertainty in the difference between the two should be greater than the uncertainty of the surface data alone.
They’re both temperature records. Whether they’re compared depends on what one desires to know.
In any case, uncertainty is not error..
The magnitude of error in each record, the surface and UAH, is not known. Subtracting the records won’t tell you about error.
But if you *do* subtract them, the uncertainty (not the error) in each records adds in quadrature to condition the difference.
He’s been told all of these many, many times. Gave up long ago that there is any hope he might ever understand.
“In any case, uncertainty is not error.”
As I’ve been trying to tell you. That’s why the uncertainty of the mean is not the same as the mean uncertainty. If everything has the same uncertainty, it does not mean that every measurement will have the same error.
The problem is that many here take the mantra “error is not uncertainty” as an excuse for saying the uncertainty can be anything you wish. They’ll say the uncertainty could be ±500°C, and it’s impossible to prove them wrong because “error is not uncertainty”.
But in my view, it doesn’t matter how you define uncertainty – either as the deviation of errors around the true vale, or as an interval that characterizes the dispersion of the values that
could reasonably be attributed to the measurand – you are still making a claim about how much difference there is likely to be between any measurement and the measurand. If you say there is a 95% confidence interval of ±2, then you are claiming that around 5% of the time you might get a measurement that differs from the measurand by more than 2.
If you make 500 measurements and it turns out that none of them where more than 2 from the correct value then you might wonder if your uncertainty interval is a little too large. If it turns out that all your values only differed from the true value by a a few tenths of a degree, it seems pretty clear that your uncertainty was far too big.
You do realize that every one of those 500 readings have their own uncertainty. With LIG, was the meniscus properly read, resolution uncertainty, systematic uncertainty, what the GUM calls “influence effects”, etc. Electronic thermometers have their own set of uncertainties. Ultimately, each reading has its own distribution surrounding it. Variances do add, thereby expanding the deviation.
With temperatures, many months have skewed distributions. NIST in TN 1900 assumes a Student t distribution that has longer tails, will more faithfully represent the uncertainty range after expansion. Knowing that there are skewed distributions, using simple arithmetic averaging is a joke unto itself. The GUM requires that a mean is proper only with symmetric distributions. It never details how to handle skewed distribution in computing uncertainty.
Here is a good question for you. How do you calculate uncertainty from a skewed distribution of measurements?
Dr. Frank has mentioned that climate science from the beginning has never done a scientific analysis of uncertainty in measurements nor of how to combine various locations together properly.
From the beginning someone said let’s just take a simple average and assume it is correct. Thn somewhere along the line, someone decided what they were seeing wasn’t making sense, so they concluded, without evidence, that pas temperature data needed correction. What’s worse, they made changes so that measurements that were changing due to different “influence effects” all matched so LONG RECORDS could be manufactured. No science whatsoever.
He doesn’t care, at all. His task is a political one.
As usual with you I’ve no idea what any of this has to do with the point I was making, but there a few interesting points.
“With temperatures, many months have skewed distributions. NIST in TN 1900 assumes a Student t distribution that has longer tails, will more faithfully represent the uncertainty range after expansion.”
I’m sure I’ve told you this before, but you are misunderstanding this. The point of using a Student t distribution has nothing to do with the data being skewed. On the contrary it assumes the data is Gaussian. TN1900 states that they are assuming the daily maximums are a normal distribution, though with such a small sample it’s impossible to say either way. They even point out you might get a different uncertainty if you don’t assume a normal distribution and talk about the procedure developed by Frank Wilcoxon to estimate the different uncertainty.
The point of a Student t distribution is to compensate for a small sample size – specifically caused by having to use the sample standard deviation, which will be an uncertain estimate of the population deviation.
“Here is a good question for you. How do you calculate uncertainty from a skewed distribution of measurements?”
A number of ways – but I’m not an expert on any of this.
For a start, the CLT suggests that the sampling distribution of a skewed distribution will still tend towards normality as sample size increases. If your sample size is large enough the skewed nature of the data doesn’t matter. (Obliviously you need to decide on what “large enough” means.)
Then there are specific special case statistical methods that can be applied to special cases (but I couldn’t tell you what any are.)
You might also be able to apply a transform to the data to make it normal.
And I suspect this is more usual nowadays, there are various Monte Carlo, and resampling techniques. This is what I think NT 1900 recommends, and avoids a lot of the statistical issues. Most of the statistics we are talking about where developed in the 19th and early 20th century, and are based on simplifying equations in order to make it possible to do the math. With computers a lot of this is less useful and it’s possible to just simulate taking multiple samples and seeing what the standard deviation of the means is.
“For a start, the CLT suggests that the sampling distribution of a skewed distribution will still tend towards normality as sample size increases.”
The CLT REQUIRES multiple samples in order to build a distribution of sample means that tends to Gaussian. *NOT* your “single sample” can define a distribution!
Another face-plant. Fortunately real metrologists care nothing for your view.
“you are still making a claim about how much difference there is likely to be between any measurement and the measurand.”
No.
Your statement shows a complete lack of understanding of uncertainty.
Ironically, after you agreed that uncertainty is not error, right there you went on to portray uncertainty as error.
Uncertainty is an ignorance width. You’ve heard that many times, Bellman. Clearly it’s never penetrated.
“No.”
Then say how you define uncertainty. In particular what would you mean by a 95% level of uncertainty?
“Ironically, after you agreed that uncertainty is not error, right there you went on to portray uncertainty as error.”
No. I said that uncertainty can be defined in terms of error, not that it is error.
“Uncertainty is an ignorance width. You’ve heard that many times, Bellman. Clearly it’s never penetrated.”
Because it’s hand waving. What are you ignorant about? What does a 95% ignorance width mean? It just feels like a phrase used to avoid understanding that probabilities are involved, and as a way of avoiding having to justify any conclusions.
“Because it’s hand waving.”
Roy, C.J. and W.L. Oberkampf, A comprehensive framework for verification, validation, and uncertainty quantification in scientific computing. Comput. Methods Appl. Mech. Engineer., 2011. 200(25-28): p. 2131-2144.
“Aleatory (random) uncertainties in model inputs are treated as random variables, while epistemic (lack of knowledge) uncertainties are treated as intervals with no assumed probability distributions.” (my emphasis)
Helton, J.C., et al., Representation of analysis results involving aleatory and epistemic uncertainty. Int’l J. Gen. Sys., 2010. 39(6): p. 605-646
“epistemic uncertainty deriving from a lack of knowledge about the appropriate values” (my emphasis)
[1] Vasquez, V.R. and W.B. Whiting, Accounting for Both Random Errors and Systematic Errors in Uncertainty Propagation Analysis of Computer Models Involving Experimental Measurements with Monte Carlo Methods. Risk Analysis, 2006. 25(6): p. 1669-1681
“[S]ystematic errors are associated with calibration bias in the methods and equipment…
“[E]ven though the concept of systematic error is clear, there is a surprising paucity of methodologies to deal with the propagation analysis of systematic errors. The effect of the latter can be more significant than usually expected. Important pioneering work describing the role of uncertainty on complex processes has been reported by Helton (1994, 1997) and references therein, Paté–Cornell (1996), and Hoffman and Hammonds (1994). Additionally, as pointed out by Shlyakhter (1994), the presence of this type of error violates the assumptions necessary for the use of the central limit theorem, making the use of normal distributions for characterizing errors inappropriate.” (my emphasis)
You’re ignorant of the subject, Bellman, you don’t know you’re ignorant, you don’t want to know you’re ignorant, you reject the discussions of professionals, you’re foolishly aggressive about promoting your ignorance, and you evidently think that everyone else is ignorant and aggressive in the same ways you are.
All of that makes you a nearly hopeless case. Perhaps to be ameliorated after a long bout of honest and very critical introspection.
Thanks. Though the first paper is behind a paywall and the other two links are dead.
I don’t disagree with anything said in your quotes, but they don’t answer my question – what do you mean by a 95% ignorance interval? What does the 95% mean if it’s not referring to a 95% of a measure being within the interval of the true value?
Saying we will never know the true value so the question is irrelevant, just leaves it as your word against anyone else’s; an unfalsifiable hypothesis.
Look at the passage from Bevington that Tim Gorman was talking about. There it’s just a simple estimate of the standard uncertainty interval, but he points out that the interval should mean that 7 out of 10 times your measureemnt should fall within the interval, and that exaggerate the interval to be on the safe side.
You’ve been given the answer to this question in the past, but of course you pooh-poohed and rejected it.
“You’ve been given the answer to this question in the past”
Which was?
Standard (GUM 4) versus expanded (GUM 6) uncertainty:
6.2 Expanded uncertainty
6.3 Choosing a coverage factor
Which is exactly what I would say. I’ve no idea why you keep claiming I disagree with this at all. You have the standard uncertainty, and multiply it by a coverage factor to get the expanded uncertainty.
The links to Helton and Vasquez terminate in periods. Apologies for the oversight. Remove the periods and the links work.
You asserted my statement that, “Uncertainty is an ignorance width” is handwaving.
It isn’t. it is analytically exact and has considerable literature precedent.
A 95% uncertainty (ignorance width) means statistically there’s ~95% chance that the physically correct value is somewhere between the two bounds. But there is no way to know where.
Uncertainty bounds wider than any possible physically reasonable value means that the ignorance is total. No quantitative information is available about correct value, at all.
For example, the diagnosis of total ignorance follows from the ±15 C 1σ uncertainty bound around the centennial air temperature projections made using climate models of the CMIP5 generation.
The ±15 C ignorance width is far larger than any physically reasonable air temperature expected for any year of the 22nd century (a concept beyond the grasp of every single climate modeler I have encountered). CMIP5 air temperature projections are physically meaningless.
“leaves it as your word” Leaves it as demonstrated by the quantitative analysis within LiG Met.
Unless you know a way to correct every single air temperature measurement taken between 1850 and 2023 for the impact of local environmental variables, a way to correct for unknown Joule-drift, and a way to remove the lower limit of detection from all LiG thermometers in use prior to 1981.
Know any way to do that? Neither do I. That leaves us with unknowable.
Bevington is about random error.
Your responses make so much sense. I wish I had studied more about the subject.
My earliest experience was from working on engines and other equipment. Here is a nice video about overhauling an old tractor engine. Notice the exact measurement for valve guide insertion. Then the exact measurements and angles on the valve seats.
My lab training was in chemistry, physics, and EE labs in college.
My real world experience came from designing and building high performance HF transmitters and receivers. A real eye opener about high quality test equipment to make sensitive measurements. Resolution limit is very apparent when dealing in microvolts.
I forgot to add the link for the video.
https://www.instagram.com/reel/CuNmL6Ru2Jv/?igshid=MzRlODBiNWFlZA==
“A 95% uncertainty (ignorance width) means statistically there’s ~95% chance that the physically correct value is somewhere between the two bounds. But there is no way to know where.”
Which is what I would have expected it to mean. This still leaves the question of why you won’t consider that a test that suggests the physically correct value is not outside the bounds ~5% of the time doesn’t suggest a flaw in the analysis.
“what do you mean by a 95% ignorance interval?”
You’ve been given this multiple times and you can’t seem to remember it.
You even spit out what Bevington said but you refuse to understand it!
He did *NOT* say that it exaggerates anything. You just make crap up and expect people to believe it!
Once again, here is *exactly* what Bevington said so that others won’t assume you are correct.
“A study of the distribution of the results of repeated measurements of the same quantity can lead to an understanding of these errors so that the quoted error is a measure of the spread of the distribution. However, for some experiments it may not be feasible to repeat the measurements and experimenters must therefore attempt to estimate the errors based on an understanding of the apparatus and their own skill in using it. For example, if the student of Example 1.1 could make only a single measurement of the length of the table, he should examine his meter stick and the table and try to estimate how well he could determine the length. His estimate should be consistent with the result expected from a study of repeated measurements; that is, to quote an estimate for the standard error, he should try to estimate a range into which he would expect repeated measurements to fall about seven times out of ten. Thus he might conclude that with a fine steel meter stick and a well-defined table edge, he could measure to about +/- 1mm or +/- .001m. He should resist the temptation to increase this error estimate, “just to be sure.”
Several points should be gleaned by anyone that can read (that lets out bellman).
Once again, bellman shows that he simply cannot read and understand simple written English. He doesn’t accept that statistical analysis of measurements only works when you have totally random error generated from multiple measurements OF THE SAME THING. If there is any systematic bias then statistical analysis will not be able to identify it and the accuracy of the mean suffers because of it.
We all know that bellman, bdgwx, Stokes, and most of the climate science cadre simply do not believe this. They have their own truth – “all error is random, Gaussian, and cancels, even systematic bias associated with multiple measurements of different things.
We all also know that bellman is going to make a fool of himself trying to refute this simple truths of metrology. I can’t wait.
“He did *NOT* say that it exaggerates anything. You just make crap up and expect people to believe it!”
What are you on about now? Who said anything about exaggerating anything. I merely pointed out that Bevington says you should not exaggerate the interval.
“Once again, here is *exactly* what Bevington said so that others won’t assume you are correct.”
Which is the same section I quoted. You seem to be reading much more into this than I meant.
“Several points should be gleaned by anyone that can read (that lets out bellman).”
This might be ironic if it turns out you are misunderstanding anything.
“This only works for repeated measurements of the same thing, as Bevington takes pains to state.”
In this case the says, in the passage you just quoted
“The uncertainty interval should *NOT* be exaggerated as bellman accuses Bevington of saying.”
You’ve accused me of saying this many times in the last few posts and it’s such a bizarre misreading. All I said was to take note of the sentence where Bevington explicitly says you SHOULD NOT exaggerate. Why you think that means I’m suggesting Bevington is saying the opposite is incomprehensible.
If you can’t figure out why I pointed to that sentence, then you really aren’t paying attention to this article.
So you didn’t say:
bellman: “Look at the passage from Bevington that Tim Gorman was talking about. There it’s just a simple estimate of the standard uncertainty interval, but he points out that the interval should mean that 7 out of 10 times your measureemnt should fall within the interval, and that exaggerate the interval to be on the safe side.” (
Sorry. I hadn’t spotted that embarrassing mistake. It should have said “and the you should not exaggerate the interval…”
No excuses for my lax typing. But it would have been easier if you’d asked me about that comment there, rather than shouting about it in a different thread.
““Aleatory (random) uncertainties in model inputs are treated as random variables, while epistemic (lack of knowledge) uncertainties are treated as intervals with no assumed probability distributions.” (my emphasis)”
Thanks for confirming what I have been saying for two years. I never found a confirming reference. Uncertainty intervals have no probability distribution. If they did you could guess at the true value – like bellman always wants to do.
He doesn’t understand that uncertainty means you don’t know. Why he can’t accept that is beyond me.
“Uncertainty intervals have no probability distribution.”
I’m not sure that is what the paper is saying.
Aleatory uncertainties definitely have a probability distribution. What the do in the paper is to not assume a probability distribution for the epistemic uncertainties. Not assuming doesn’t mean it doesn’t have one it’s just you are not assuming you no what it is.
But it still effectively means you are assuming a uniform distribution.
If you don’t know what it is then how do you know it even exists? How do you analyze it if you don’t know the distribution?
It doesn’t appear that you have any better understanding of aleatory uncertainty any more than you do of measurement uncertainty.
Assume I am rolling a die you can’t see. I give you the value of 4 rolls – 2, 3,4, 5. What size die am I rolling? You can eliminate d2, d3, and d4 dies because you would never get a 5 from them. But how do you know if it is a d6, d8, d10, d20, or even a d100? Each of those is a possible generator of those values. That is aleatory uncertainty and there is *NO* probability function or distribution that you can use to allow you to determine which size die its. Thus you have an unknown that you can’t resolve, an uncertainty.
Your uncertainty interval would range from d6 to infinity (assuming you could build a die that large). It is not a uniform distribution because there is only ONE DIE being rolled. It is the true value with 100% probability of being the true value. The probability of all the other dies being the true value is zero. The problem is that you don’t know which die is the true value. The probability distribution telling you the probability for each size die is unknown – which means saying it is a uniform distribution is nothing more than a subjective way for you to make sense of it! But it has no meaning in the real world. And it *still* doesn’t allow you to resolve the uncertainty – it all just remains an unknown! The numbers on a d6 *do* have a uniform probability distribution – each of them is equally likely to be the “true value” on the next roll. A far different thing.
Instrumental uncertainty you *can* quantify, at least within limits, based on the construction of the apparatus. And you can certainly find an average value for the instrumental uncertainty related to the physical constraints of the apparatus.
That he tries to glean information about distributions or true values from uncertainty intervals indicates he still doesn’t grasp the basics.
“How do you analyze it if you don’t know the distribution?”
Don’t ask me. That’s the subject of the paper.
That is *NOT* the subject of the paper! The paper is about the uncertainty INTERVAL, not the probability distribution of the possible values.
You need to admit to yourself at some point that you simply do not understand the concept of uncertainty and abandon your cult dogma. You’ll never fit your cult dogma into the reality of metrology. Neither will bdgwx. Neither will Stokes, Nor will climate science.
You need to learn a few basic truths of metrology. They are simple and easy to understand if you approach with an open mind instead of cherry-picking stuff to try and disprove the fundamental truths of uncertainty.
Learn these and you’ll stop posting assertions that don’t work in the real world. Leave your universe of “all uncertainty is random, Gaussian, and cancels”.
“That is *NOT* the subject of the paper!”
To be clear I meant the paper the quote was from:
A comprehensive framework for verification, validation, and
uncertainty quantification in scientific computing
So what difference does it make? You *still* don’t know anything about the subject but feel that you can make unsubstantiated claims that something about it is wrong.
“Aleatory uncertainties definitely have a probability distribution.”
The aleatory uncertainties are treated as a random variable in the input to a model.
from the paper: “ aleatory – the inherent variation in a quantity that, given sufficient samples of the stochastic process, can be characterized via a probability density distribution,”
Note carefully the words “sufficient samples”. How do you get sufficient samples of temperature when you get ONE measurement of one thing?
Is this part of your magic crystal ball that can determine a distribution from one value?
“But how do you know if it is a d6, d8, d10, d20, or even a d100?”
Are those the only options? Don;t you have a d12?
“That is aleatory uncertainty and there is *NO* probability function or distribution that you can use to allow you to determine which size die its.”
Firstly, do you mean “aleatory” here. Aleatory is uncertainty caused by randomness, it’s epistemic uncertainty that’s caused by a lack of knowledge.
Secondly, of course you can use probability to determine the likelihood or probability of any given die (depending on your definition of probability). It’s what statisticians have been doing for centuries.
Assuming these were your only 8 dice then in frequentist terms the likelihood of any die is the probability that you would roll those 4 numbers. It’s inevitably going to mean the d6 is most likely, because you are more likely throw any given number on a d6 than a d8 etc.
In Bayesian terms, you need a prior, but as you are already pleading ignorance, you can just assume all of the 8 dice are equally likely. The the probability of any specific die is the probability of getting those 4 numbers with that die divided by the overall probability of getting those 4.
My quick calculations (be sure to check them as I’ could easily have made a mistake) gives the following probabilities
2: 0.00000
3: 0.00000
4: 0.00000
6: 0.68770
8: 0.21759
10: 0.08913
20: 0.00557
100: 0.00001
Quite likely it is a 6 sided die, and whilst it’s possible it might have been a d100, the odds are very small.
Since the rolls are random you cannot predict sequences. The probabilities of each individual roll do not sum to predict a sequence.
You are perfectly demonstrating the difference between aleatory and epistemic.
You can’t use probabilities to predict the next roll, it is random, no matter what size of die is involved. Each number on the die has an equal probability of being rolled. If you could predict the next roll every casino in Las Vegas would be out of business. If you can’t predict the next roll then you can’t predict any specific sequence either. Aleatory uncertainty.
You do *not* know which size of die it is. It could be any of them. Ignorance is epistemic.
Measurement uncertainty of temperature certainly has a major component of epistemic uncertainty. You simply don’t know the systematic bias let alone the random uncertainties. Ignorance.
“Epistemic uncertainty is also known as systematic uncertainty, and is due to things one could in principle know but does not in practice. This may be because a measurement is not accurate, because the model neglects certain effects, or because particular data have been deliberately hidden”
What Pat has done is to pull the random curtain away from some of the components of measurement uncertainty, with respect to the measure of temperature, by defining the minimum ignorance interval. It is still ignorance.
Such a pity. You finally come up with an interesting question, and one that doesn’t involve measuring bits of wood you’ve picked out of the sewer – and when I give my answer you mover the goal posts again.
“Since the rolls are random you cannot predict sequences. The probabilities of each individual roll do not sum to predict a sequence”
You were not asking about predicting the next number in a sequence. You were asking about the the likelihood of the particular die you were rolling.
“You can’t use probabilities to predict the next roll, it is random, no matter what size of die is involved.”
The point of probability is not to predict the next roll, it’s to tell you the probability.
“You do *not* know which size of die it is. It could be any of them. Ignorance is epistemic.”
But as we see, we can generate a probability distribution based on the evidence. We are not completely ignorant. And we can then use that probability to work out the probability of the next throw. It’s very unlikely to be 100, becasue it’s very unlikely you are using a d100. It’s more likely to be between 1 and 6, becasue a 6 sided die is most likely.
“Measurement uncertainty of temperature certainly has a major component of epistemic uncertainty.”
And if you want to do an uncertainty quantification you need to estimate what they are. But just claiming they must be big, seems like wishful thinking if as you say you are ignorant of them. Any modelling then needs to see how these uncertainties will affect the outcome. This would I expect depend on whether you think all instruments have the same epistemic uncertainty or if they are all different. And if they had the same values during the base period as they do now.
“It is still ignorance.”
That’s how it seems to me.
“You were not asking about predicting the next number in a sequence. “
ROFL!! I didn’t give you a sequence of numbers? Wow, just WOW!
“The point of probability is not to predict the next roll, it’s to tell you the probability.”
Cop-out! Completely expected. You thought you could tell what die was being used and when shown you couldn’t you punt!
“But as we see, we can generate a probability distribution based on the evidence”
You said you can’t predict the next roll! That makes it what kind of uncertainty? You can run. You can pettifog. You can equivocate. But you can’t hide!
“It’s very unlikely to be 100, becasue it’s very unlikely you are using a d100.”
We use a d100 in D&D role play lots of time to choose items from a table. You can’t even get this one right!
“But just claiming they must be big, seems like wishful thinking if as you say you are ignorant of them.”
Like Pat pointed out, you just don’t get it. It’s an ignorance interval. That can be huge, especially when made up of components whose ignorance interval is huge!
“Any modelling then needs to see how these uncertainties will affect the outcome”
You don’t need to model instrumental uncertainty that can be quantified as being resolution limited. You are *still* trying to get out from under having to consider uncertainty as a real thing.
Face it, instrumental uncertainty doesn’t cancel. It just adds and adds and adds – all the way through!
“ROFL!! I didn’t give you a sequence of numbers? Wow, just WOW!”
Stop rolling around and try to remember. You said
“Cop-out! Completely expected. You thought you could tell what die was being used and when shown you couldn’t you punt!”
I said you could tell which die was most likely to be used. Not that I knew which one was being used. Once again your ignorance is showing.
“You said you can’t predict the next roll!”
No idea what your problem is. You cannot know what the next roll is. You can make statements about the probability of the next roll.
“We use a d100 in D&D role play lots of time to choose items from a table. You can’t even get this one right!”
I’d have thought Tunnels and Trolls was more appropriate. Sorry, but you can just not be this gormless. You know full well what I said, it’s all described in the previous comment. It’s unlikely to be a d100 because you rolled 4 numbers between 2 and 5. It’s therefore much more likely you were using a smaller die. The odds of getting 4 numbers <= 20 on a d100 is about i in 500. The chances of getting 4 <= 6 is more like 1 in 80000. It’s far more likely you where using a smaller die.
“Like Pat pointed out, you just don’t get it. It’s an ignorance interval. That can be huge…”
Only if you don;t care that it’s meaningless. There seems to be this assumption that becasue youve started calling it an ignorance interval, reality doesn’t matter.
“Face it, instrumental uncertainty doesn’t cancel. It just adds and adds and adds – all the way through!”
It doesn’t bother you that Pat Frank is actually dividing by N all the way through. IKf he followed your advice the ignorance would be beyond believe. Intervals of millions of degrees.
“And it *still* doesn’t allow you to resolve the uncertainty – it all just remains an unknown!”
At which point you are giving up on any type of uncertainty analysis. Just becasue you don’t know whnat something is doesn;t mean you can’t say how likley it is. The point of the paper you are citing is trying to calculate probabilities that something will succeed or not. They don’t just say there’s no way to know becasue it could be anything?
You just don’t get it. How do you resolve something you can’t possible know? That doesn’t mean you are giving up on uncertainty analysis – it is just the opposite in fact! You estimate your ignorance interval!
Do *YOU* know the systematic bias in the temperature measuring station at Kansas State University? Of course *YOU* do I’m sure. But how do the climate scientists know? How do they assign a probability to *any* value in the ignorance interval?
If you don’t know the systematic bias then how do you calculate the random uncertainty?
“But it still effectively means you are assuming a uniform distribution.”
Wrong.
Assignment of a box interval designates a width of ignorance. The position of the correct value within the width is utterly unknown, including even if it is within the width.
It’s best guess.
Great statement. I am going to save your entire response for future use.
Ditto.
It’s never going to penetrate. It would mean having to admit he’s been wrong about uncertainty since forever. So he’ll keep on making stupid assertions that a six year old could refute.
“The problem is that many here take the mantra “error is not uncertainty” as an excuse for saying the uncertainty can be anything you wish”
Again, you simply can’t read, or you refuse to. I gave you what Bevington said about the uncertainty interval. As usual, you didn’t bother to read it. He says the estimate of the uncertainty interval should be such that seven out of ten measured values fall within the interval!
That is *NOT* “anything you wish”!
Jeesh, man, give it a break. Nothing, absolutely NOTHING, you have said in the past two days makes any sense at all!
“you are still making a claim about how much difference there is likely to be between any measurement and the measurand.”
Since you do *NOT* know the true value then how do you come up with a difference? The uncertainty interval gives you a clue as to what the maximum and minimum differences are likely to be but they claim NOTHING about what it actually *is*.
You *STILL* don’t have the slightest clue as to the concept of uncertainty. You keep trying to force it into your basket of “error”. Uncertainty is *NOT* error. Go write that down on a piece of paper 1000 times. Maybe it will sink in but based on past experience with you IT WON’T.
“If you make 500 measurements and it turns out that none of them where more than 2 from the correct value then you might wonder if your uncertainty interval is a little too large”
OMG! Then you refine your uncertainty interval estimate! Not everyone is as perfect as you are.
From the GUM: “4.3.1 For an estimate xi of an input quantity Xi that has not been obtained from repeated observations, the
associated estimated variance u2(xi) or the standard uncertainty u(xi) is evaluated by scientific judgement
based on all of the available information on the possible variability of Xi .” (bolding mine, tpg)
You are the ultimate cherry-picker. You have never actually studied anything associated with uncertainty. You cherry-pick Taylor. You cherry-pick Bevington. You cherry-pick Possolo. You cherry-pick the GUM. Not once have you *ever* actually sat down a read the tomes in detail, worked out the examples, and figured out why what you come up with for answers never match what the experts say the answers are.
Type B is a perfectly legitimate value for uncertainty. And it is based on SCIENTIFIC JUDGEMENT. Not on wacky assertions that one data point defines an unknown distribution!
“I gave you what Bevington said about the uncertainty interval. As usual, you didn’t bother to read it.”
There are hundreds of angry comments directed at me every day. I don’t notice all of them or follow the demands of each.
In the mean time here’s his definition of uncertainty.
“He says the estimate of the uncertainty interval should be such that seven out of ten measured values fall within the interval!”
Seems like a good rule of thumb, presumably talking about a standard deviation.
“That is *NOT* “anything you wish”!”
Exactly, that’s good probabilistic definition of uncertainty based on the error. Bevington is not the one shouting “uncertainty is not error”. Quite the reverse.
Here’s the passage in question. It’s talking about estimating an uncertainty when you only have a single measurement.
As I said the 7 out of 10 is talking about the standard error. And the last sentence may be something to remember.
Why is it that 70% was chosen? Before you answer, you might consider what the percentage of values that 1 sigma covers!
Because, as I said, and as Bevington said he’s talking about a standard error. For a normal distribution a standard error covers about 68%.
I really which I knew where all these questions are supposed to be leading. You never seem to realize how much you are contradicting yourselves. All this was meant to be demonstrating why uncertainty has nothing to do with error, yet you are quoting someone who talks about uncertainty entirely in terms of error.
The *ONLY* one contradicting themselves appears to be you.
From the GUM: “NOTE 3 “Experimental standard deviation of the mean” is sometimes incorrectly called standard error of the mean.”
This is *exactly* what Bevington is talking about – standard deviation of the mean based on experimental results. Terminology you simply won’t accept.
“From the GUM: “NOTE 3 “Experimental standard deviation of the mean” is sometimes incorrectly called standard error of the mean.””
And they are wrong. In the larger world standard error is the correct term and standard deviation at best confusing.
“This is *exactly* what Bevington is talking about – standard deviation of the mean based on experimental results.”
Except in this case he specifically isn’t because he’s talking about a case where you are only taking one measuement and have to estimate the standard error.
And you are mixing up to different things again. The GUM quote is talking about the standard error of the mean. Bevingotn is talking about the standard deviation of the results.
We are not talking at this point about taking a mean to reduce uncertainty.
“And they are wrong. In the larger world standard error is the correct term and standard deviation at best confusing.”
I see. So *YOU* are the final arbiter of what is wrong and right in the GUM. Have you written to the NIST to tell them that this statement is wrong? It’s your duty to correct them if they really are wrong. Are you going to shirk your duty?
“Except in this case he specifically isn’t because he’s talking about a case where you are only taking one measuement and have to estimate the standard error.”
What’s that got to do with anything? *You* can’t even tell that what Bevington is saying is that from one measurement there is no way to find the standard deviation of the sample means!
“And you are mixing up to different things again. The GUM quote is talking about the standard error of the mean. Bevingotn is talking about the standard deviation of the results.”
I *know* that. Again, you are throwing crap against the wall hoping something will stick! Without a large observation population there is no way to do sampling in order to determine the standard deviation of the sample means let alone the standard deviation of the population. So you have to estimate it.
*You* said you could determine the distribution from one sample meaning you could also derive the standard deviation of that distribution. Now you are trying to say that you can’t?
Another faceplant!
“I see. So *YOU* are the final arbiter of what is wrong and right in the GUM. ”
No. I merely expressed my opinion that they are wrong on that point. Why do you think I should not be allowed to say they are wrong, but it’s fine for them to say every statistician and text book is wrong to call it the Standard Error?
Rest of hysterical rant ignored.
You opinion doesn’t count. Can you give a rational explanation as to why it is wrong?
No one is saying that it is wrong to call it Standard Error. They are saying that it is misleading as all git out! It *is* the standard deviation of the sample means, pure and plain. Call it the Marx Index if you want.” A rose by any other name would smell just as sweet”. The “Marx Index” would still be the standard deviation of the sample means.
The fact that the name “Standard Error” has certainly confused *YOU* as to what it actually stands as mute proof that you don’t know the basics.
“You opinion doesn’t count.”
So you keep telling me.
“Can you give a rational explanation as to why it is wrong?”
Well, for a start what they describe as a standard deviation isn’t.
https://www.scribbr.com/frequently-asked-questions/standard-error-vs-standard-deviation/
https://statisticsbyjim.com/basics/difference-standard-deviation-vs-standard-error/
““No one is saying that it is wrong to call it Standard Error.””
The exact words were
“incorrect” implies “wrong” in my book, and saying “sometimes” is a slight when it’s almost invariably called “standard error”.
“They are saying that it is misleading as all git out!”
Where do they explain why they think it’s confusing. Given how many times people here have mixed up “standard deviation of the mean” with the “standard deviation”, or insist that everyone actually calls it “standard deviation of the means”, I’m not sure how this is less confusing.
“The “Marx Index” would still be the standard deviation of the sample means.”
Why not call it the thing everyone else calls it? And why say calling it what everyone has called it for 100 years is incorrect?
“The fact that the name “Standard Error” has certainly confused *YOU* as to what it actually stands as mute proof that you don’t know the basics.”
And back to the in-substantive personal insults.
And this coming from someone who still thinks the only way to know what the SEM actually is, is to take hundreds of different samples and find their standard deviation.
Why if you think the GUM is wrong do you even use the “maths”?
Because you falsely believe it gives the answers you want.
“Why if you think the GUM is wrong do you even use the “maths”?”
I know in your world view everything is either totally good or totally evil. And which is which depends on whether they agree with you or not.
But in my world there is nuance. The GUM is mostly correct, and there’s nothing I can see wrong with their equations – they are pretty standard after all. But I still find parts badly written and confusing, and in this case they throw in a snide aside about not using the term “standard error” with no explanation.
The GUM equations *are* correct – for random uncertainty associated with multiple measurements of the same thing using the same device under the same conditions.
What they write is *NOT* badly written or confusing. It only seems that way to you because you don’t (or won’t) understand the basics of metrology.
And can’t.
You are correct, he worships at the altar of the God of Averages, and his guru is Nitpick Nick Stokes.
I can’t get past the fact, al least for me, that the average is not a measurement. It may have the same value as a member of the dataset but that is coincidence. If it *was* a measurement it would have the same uncertainty as any other member of the dataset inherently and could only be quoted based on the measurement device’s resolution ability. You couldn’t increase the resolution of the average past that of its measurement device.
So the cultists want it both ways, they want it to be a measurement and they want it to not be a measurement, depending on what it needs to be for them in the moment.
Typical cognitive dissonance.
Absolutely.
The JCGM had the temerity to gore his sacred SEM cow, so they have to be “wrong”.
Yep!
And Pat Frank thinks they are wrong about resolution, so maybe he’s a heretic in your eyes as well.
If you find the phrase “standard error” incorrect explain why.
RTFM
“I can’t get past the fact, al least for me, that the average is not a measurement”
I’ll ask again. If, for you, an average is not a measurement, why do you keep talking about the measurement uncertainty of an average?
“It may have the same value as a member of the dataset but that is coincidence.”
And irrelevant. You still don’t understand that an average is not the same as one physical thing. It’s the mean. A figure that characterizes a collection of things or a distribution. It is not the same as saying an average thing.
“If it *was* a measurement it would have the same uncertainty as any other member of the dataset inherently and could only be quoted based on the measurement device’s resolution ability.”
Which is what Pat wants to do. But as I keep trying to tell you the uncertainty of the average is not the same as the average uncertainty. You keep saying you accept this, but every time you circle back to saying it is the same thing.
“I’ll ask again. If, for you, an average is not a measurement, why do you keep talking about the measurement uncertainty of an average?”
*I* don’t, at least not when considering single measurements of different things using different devices, i.e. multiple measureands.
*I* talk about needing to know overall uncertainty so I can judge if a beam made up of multiple components will span a foundation and support the design load. I could care less about the “average” length, I want to know the possible minimum and maximums based on the overall uncertainty.
I may want to know the mean instrumental systematic uncertainty If I’m using different devices to measure different things, that’s not the same thing as wanting to know an “average” value for a measurement.
As I’ve said multiple times, the average uncertainty is not the uncertainty of the average.
“You still don’t understand that an average is not the same as one physical thing.”
If it’s not a physical thing then it doesn’t live in the *real* world of measurement. If that is true then it is worthless. You may as well argue how many angels can fit on the head of a pin.
“A figure that characterizes a collection of things or a distribution. It is not the same as saying an average thing.”
So what? It’s not a measurement! It’s a statistical descriptor. And the statistical descriptor know as the “mean” is really only useful if you have a symmetrical distribution – which is highly unlikely when the data set is made up of different things. A random collection of boards collected from construction sites around the city are highly likely to have a very skewed distribution – making the mean truly useless.
Now, come back and tell us how all distributions are random, Gaussian, and their uncertainty all cancels out!
“The GUM equations *are* correct – for random uncertainty associated with multiple measurements of the same thing using the same device under the same conditions.”
But not just for that. I’m talking about equation 10 – could be used for repeated measurements of the same thing, but is mainly used when combining different things to make a new measurand. To use one example that you accept is correct, finding the volume of a cylinder from the height and radius. height and radius are not the same thing.
Or there example,
Lots of different things there, not even measured with the same instrument.
“What they write is *NOT* badly written or confusing.”
Yet it seems to confuse you a lot of the time, such as when you think equation 10 is about combining the same thing using the same measurement.
“But not just for that. I’m talking about equation 10 – could be used for repeated measurements of the same thing, but is mainly used when combining different things to make a new measurand.”
You *still* don’t get it, do you? You are talking of a MEASUREMENT MODEL OF ONE THING! If you have to measure the length and width of a rectangular table then its area will have the uncertainty in the length and width propagated on to it! But you are still measuring for the same thing! “A” table.
Measuring the length and width of 100 different tables is going to tell you what? Will the average area tell you which tables will fit in your dining room? Will it tall you how big and how small tables can be?
You can’t even understand what you write yourself! “make a new measurand” PLEASE NOTE THAT “MEASURAND” IS SINGLULAR. You are finding the characteristics of ONE MEASURAND. You may have to find the length and width of that one measurand and propagate the uncertainties associated with those measurements but you still have only ONE MEASURAND.
You may have to measure the mass, pressure, and volume of a gas sample in order to know its temperature but that is still FOR JUST ONE MEASURAND.
Those measurements of the component parts simply won’t tell you anything about a different parcel of air thousands of miles away – i.e. multiple measurands.
Yet that is what climate science tries to do with temperature. It’s what *YOU* try to do with temperature.
You have no idea what you are talking about. Quit until you have some knowledge from someone familiar with metrology.
Equation 10 is used when a measurand consists of multiple meausements put together in a defined mathematical equation that defines a quantity of an object. (A=πr²h) (I=εAσT⁴) (v=d/t) (P=nRT/V) These functional relationships all have variables with different effects on the measuand so you need to do a partial differential to determine a sensitivity value for each. Do you need to do this when the measures all have the same units and simple addition or subtraction is done? Nope. Plain old equations 4 and 5 will suffice.
Yep, and Tbar = sum(T_i) / N is NOT a measurand.
the thing is – you are the ones who insist that measurement uncertainty of individual temperature reading must be propagated in some way into the average temperature. Yet you keep insisting the equations for propagating uncertainty don’t work if the things you are propagating are different, or they don’t work for averaging, or whatever. Sol how do you want anyone to propagate these uncertainties?
You don’t have a problem propagating the uncertainties to the sum of temperature readings, even though they are different things, and the sum of temperatures is not a real thing. You don’t object to any of the propagation Pat Frank does even though it involves combing different things to make an average you insist doesn’t exist. So which equations do you want to use?
OMG!
No one says you can’t propagate uncertainty in measurements of different things. In fact we are saying you MUST propagate those uncertainties.
What you CAN’T do is assume they are random and Gaussian and therefore cancel!
You can’t even get this one correct after its been pointed out multiple times!
“You don’t have a problem propagating the uncertainties to the sum of temperature readings, even though they are different things, and the sum of temperatures is not a real thing.”
you can’t even get this one right! The uncertainties can certainly be propagated. But the uncertainty of the average is the uncertainty of the sum! It is not the average uncertainty! The average uncertainty just takes the total uncertainty and spreads it evenly across all of the elements. What sense does that make? Especially when you have measurements of different things with different uncertainties? The average uncertainty isn’t goin to tell you how long a beam will be when made up of random selections of different boards!
“You don’t object to any of the propagation Pat Frank does even though it involves combing different things to make an average you insist doesn’t exist. So which equations do you want to use?”
Again, the problem isn’t propagating uncertainty from different things. The problems are 1. assuming they are all random, Gaussian, and cancel, and 2. assuming the average uncertainty is somehow the uncertainty of the average!
There are treatments for hysteria.
Could you suggest some to your brother.
“you are the ones who insist that measurement uncertainty of individual temperature reading must be propagated in some way into the average temperature.”
Not correct. The methodology in LiG Met. is not as you suggest.
The uncertainty stemming from instrumental resolution applies to every one of its measurements and to every mean of those measurements.
The uncertainty arising from unavoidable systematic errors revealed by sensor calibration experiments provides an estimate of the reliability of instrumental measurements and necessarily propagates into a measurement mean.
They don’t get and they will *never* get it.
By “you are the ones” I wasn’t talking about your methodology. Just the three or so on here who have been insisting that the rules for propagation require the uncertainty of the average to be the same as the uncertainty of the sum, and never except that this is not what the methods of propagation they use actually leads to that conclusion.
“The uncertainty stemming from instrumental resolution applies to every one of its measurements and to every mean of those measurements.”
No disagreement there, just that unless you can explain how all the uncertainties from resolution are the same systematic uncertainty, the uncertainty of the mean will not be be the mean uncertainty.
bdgwx disagrees with you.
When apprising the uncertainty due to instrumental resolution only (calibration uncertainty due to systematic measurement error not included), the mean of the uncertainty will be the uncertainty of the mean (and also of every measurement).
Maybe he does, maybe he doesn’t. You don’t provide a reference. I don’t have to agree with him and he doesn’t have to agree with me. But if he says what you claim, I’d definitely disagree.
What would be true is that when measuring the same thing repeatedly, especially with a precise instrument, the resolution would be a systematic error. But my argument is that when you are measure different things, which vary well beyond the resolution, it will mostly be random.
There is a nice little summary of uncertainty of the mean and repeatability at http://bulldog2.redlands.edu/fac/eric_hill/Phys233/Lab/LabRefCh6%20Uncertainty%20of%20Mean.pdf
Very good treatment of uncertainty. The interval within which the population mean might lie is really only applicable when you have what the document calls “repeatable” measurements. Otherwise the average itself makes little sense as far as metrology is concerned.
I especially like the treatment of “range” as an index into uncertainty estimates. The range is directly related to the variance in a data set. As the variance goes up so does the uncertainty associated with the mean. That’s why it is important to know either the standard deviation or variance of a data set and not just the average value. Something sadly lacking in climate science.
I also like this part: “Here is a general rule for distinguishing a repeatable from unrepeatable measurements:
A measurement is repeatable if (1) you are sure that you are really measuring the same quantity each time you repeat the measurement, (2) your memory of the earlier measurement will not affect your later measurement, and (3) your measurement device is sufficiently precise that you will see variations in its readout when you repeat a measurement.”
Pat’s paper is all about (3). It’s unavoidable with measuring devices. Resolution is not the only factor involving uncertainty but also how small of an increment can be recognized. It doesn’t to any good to have resolution marks down to the thousandths digit if the instrument can only respond to increments in the tenths or hundredths digit. This is why trying to identify temperature anomalies in the hundredths digit is a fools errand. Averaging can’t increase either resolution or the smallest increment that is recognizable. This, more than anything else, should tell those trying to use the GAT that it is a garbage calculation. The uncertainty will be in at least the unit digit for such a large data set consisting of measurements of different things. Meaning you simply *DO NOT KNOW AND CANNOT DISCERN” changes in the hundredths digit.
This is true even for the UAH data set. The inherent uncertainties in knowing the exact conditions for each measurement, even in the same grid, makes the uncertainty interval for each data point quite wide. All kinds of things can affect the irradiance of the atmosphere, from clouds to smoke to aerosols to etc. I’m sure the UAH crew thinks those can be “averaged” out but its a false hope. If nothing else it will increase the range of values seen (i.e. the variance), even at the same location, which means the uncertainty of the average value goes up, not down.
The argument that “it’s always been done it this way by very smart people” is just hokum. It’s an argumentative fallacy, Argument to Tradition.
This is a good resource and I have used several times. I have it stored in my notes file.
Here is a summary of some of my problems.
1) Daily Average – At a minimum, the Type B estimate of LIG uncertainty is 0.5° F. They add in quadrature. √(0.5² + 0.5²) = 0.7. This is the measurement uncertainty in the daily average. It should be propagated into any further calculations.
2) NIST TN 1900 declares the monthly average of Tmax as the measurand. They use the same equation as this document to find the same method to “estimate the uncertainty Uₘ of the mean of a normally distributed set of measurement using just one set of N measurements as follows:”
Uₘₑₐₙ = (ts) / √N
While NIST assumed measurement uncertainty was neglible this isn’t really the case. Two points,
a) “s” was used since it is the deviation of the distribution surrounding the mean and dividing by √N calculares the SEM;
b) √N is not applicable for dividing measurement uncertainty because it is not part of the deviation surrounding the mean, it is the deviation of the distribution surrounding each measurement;
c) the total measurement uncertainty is something like,
Uₘₑₐₛ = √(30 • 0.7²) = 3.8;
d) Assume the interval of the mean is 1.8 from NIST TN 1900;
e) the combined uncertainty would be
Uᴄᴏᴍʙɪɴᴇᴅ = √(3.8² + 1.8²) = ±4.2
Not an unreasonable overall uncertainty.
“the total measurement uncertainty is something like,Uₘₑₐₛ = √(30 • 0.7²) = 3.8;”
You really ought to explain to Pat Frank where he went wrong.
Do you understand combined uncertainty? These are two pieces. There are other components that need to be evaluated to understand the total. Look at the GUM. Components add.
The point is that your calculation for that one thing, combing 30 days of the uncertainty in the daily mean, is not what Pat Frank does. I think he’s wrong but you are wronger. You are just falling back on the same mistake of thinking the uncertainty in an average is the same as the uncertainty of a sum.
You do this twice. First in the uncertainty of the daily average which you think can be bigger than either of the max or min uncertainties. Then when you take the uncertainty of the mean of 30 days, you are just adding this uncertainty 30 times in quadrature.
Compare this to PF’s equation 4 – the uncertainty of T_mean from the resolution uncertainty of the two measurements. He adds the two uncertainties in quadrature and the divides by 2. (Ignoring the typo in the paper.) This gives a standard uncertainty of 0.194°C, or a 95% uncertainty of ±0.382°C.
Then in equation 5 he does what you are trying to do, and workout the uncertainty of the monthly mean, though uses the value of 30.417 days in a month. But here he adds the squares and divides by N, before taking the square root. the result is the monthly uncertainty is the same as the daily one. What he does not do, is what you are doing and multiply the uncertainty by √30.
A highly technical term…
√30 times wronger than Pat to be more exact. And considering how wrong Pat is, that’s a whole lot of wrongness.
And still you remain — unskilled and unaware.
So nows the time to prove you are skilled and aware. Just say who you agree with more.
What should the measurementuncertainty of a monthly average, when the daily uncertainty is ±0.7 and the month has 30 days?
Is it √(30 × 0.7²) as Jim claims?
Or is it, √(30 × 0.7² / 30) as Pat claims?
You can’t even tell the difference between total uncertainty and mean uncertainty. Pat calculated the MEAN uncertainty due to instrument physical characteristics, not the total uncertainty of a set of measurements of different things.
“You can’t even tell the difference between total uncertainty and mean uncertainty.”
OK, I’ll bite. When you say “total” uncertainty, do you mean the uncertainty of the sum, or do you mean all uncertainties combined. Because what Jim said was
I assumed this was meant to be the uncertainty in a mean monthly values. I assume that U_meas is a typo for U_mean, and that the 30 refers to the number of days in a month.
If it’s only meant to be the uncertainty in the sum of 30 measurements, then fair enough. I’ve just no idea why you would want such a thing. The sum is meaningless, except as a step towards the mean.
But then he says at the end
Where 3.8 is the figure he got from c), and 1.8 is an assumed uncertainty in the monthly mean. So what would be the point in combining the uncertainty of a sum of 30 days, with the uncertainty for a mean of the same 30 days? And why does he say 4.2 is not an unreasonable uncertainty? What is it an uncertainty of?
“OK, I’ll bite. When you say “total” uncertainty, do you mean the uncertainty of the sum, or do you mean all uncertainties combined. Because what Jim said was”
This has been answered for you multiple times. Yet you’ve never bothered to understand it. As Pat said, it’s useless to try and explain it to you. You just do not want to understand!
“If it’s only meant to be the uncertainty in the sum of 30 measurements, then fair enough. I’ve just no idea why you would want such a thing. The sum is meaningless, except as a step towards the mean.”
It’s meaningless to you because you’ve never actually worked out *anything* in Taylor’s book. You can’t even understand his Eq. 3.18 and 3.19.
If you have q = x/w then the relative uncertainty in q is the sum of the relative uncertainty in x PLUS the relative uncertainty in w. It is *NOT* the relative uncertainty in x *divided* by w!
It simply doesn’t matter if q is an average or something else. It’s where you and bdgwx keep going wrong with q = x/n. You do *not* divide the uncertainty of x by n (n is the number of elements in the average). You *add* the relative uncertainties of x and of n. It’s what Taylor does, Bevington does, and what Possolo does. You’ve never once worked out any of their examples, NOT ONCE!
You haven’t eve figured out what Jim was doing. The terms “combined uncertainty” and “expanded uncertainty” are truly foreign concepts to you because you’ve never actually worked any examples out from anywhere. You just keep applying the meme that all uncertainty is random, Gaussian, and cancels. You don’t even realize that you apply that meme *all the time*!
And yet more personal abuse rather than answer a simple question.
“If you have q = x/w then the relative uncertainty in q is the sum of the relative uncertainty in x PLUS the relative uncertainty in w. It is *NOT* the relative uncertainty in x *divided* by w!”
And you still don’t understand the implication of relative uncertainty.
You don’t divide the relative uncertainties, but the absolute uncertainties will be affected by w. Specifically if w has no uncertainty than u(q) / q = u(x) / x, which (for some reason you are incapable of seeing), means u(q) = u(x) / w. It has to because q = x / w.
Rest of hysterical rant ignored.
Relative uncertainties ARE NOT AVERAGES!
Period, Exclamation point!
You *still* have read Taylor, done any of the exercises, or understand “weighting” when calculating uncertainty.
You quoted Possolo as having multiplied the radius uncertainty by 2. It’s apparent that you didn’t even understand what he did!
R^2 is R*R. When you propagate uncertainties you add the uncertainties of each factor. R appears twice as a factor. The uncertainty of R^2 is u(R) + u(R). TWO R uncertainty values! There is a reason why he didn’t include pi in his uncertainty calculation and my guess is that you have exactly ZERO understanding of why.
“Relative uncertainties ARE NOT AVERAGES!”
Why on earth would they think they were?
“You quoted Possolo as having multiplied the radius uncertainty by 2.”
Do I have to keep explaining to you how this works? Using Gauss’s formula – equation 10 from the GUM, the square of the uncertainty of a combination of values, is equal to the sum of the squares of each value, multiplied by the square of the partial differential for each term. In this case for the Radius the volume depends on R², hence the partial derivative is 2.
More completely, you have the formula V = πR²H.
Partial derivative for H is πR²
Partial derivative for R is 2πH
Hence
u(V)² = (πR²)² × u(H)² + (2πH)² × u(R)²
This can be simplified by dividing through by V², which nicely turns everything into relative uncertainties
[u(V) / V]² = [u(H) / H]² + [2u(R) / R]²
Any more questions?
“There is a reason why he didn’t include pi in his uncertainty calculation and my guess is that you have exactly ZERO understanding of why.”
There was no point including π as a term as it has zero uncertainty. But as you never seem to understand it is there in the equation. It’s in both of the coefficients in the first form, and then when we divide by V² it disappears from the right hand side, becasue it gets cancelled from both terms. But now it appears on the left hand side, in the form of V, which remember is πR²H.
But it would appear in the partial differential in each term, right? So he didn’t drop it because it has zero uncertainty. Using your logic it should appear in each term, just like you want (1/n) to appear in each term.
So why did Possolo drop it? Try again.
“But it would appear in the partial differential in each term, right?”
Yes. As I explained in my previous comments.
u(V)² = (πR²)² × u(H)² + (2πH)² × u(R)²
“So he didn’t drop it because it has zero uncertainty.”
The term π as a term is not included as a separate uncertainty as it will just be zero.
“Using your logic it should appear in each term, just like you want (1/n) to appear in each term.”
It does appear in each term, as I wrote. And yes, finally you are getting it, when the function is an average there is a (1 / n)² in each term.
“So why did Possolo drop it? Try again.”
Do you think you’ll understand it better if I explain again.
If pi is dropped because it appears in each term and is zero why is n not dropped because it appears in each term and has zero uncertainty?
You seem to want to have your cake and eat it too.
Answered below, but probably not to your satisfaction.
You wind up with 2 * u(R). u(R) + u(R). It’s a WEIGHTING factor. You didn’t answer why he dropped pi out of the uncertainty calculation. Kind of inconvenient for you maybe?
I’ll ask again. Why did he drop pi?
How many times are you going to repeat this nonsensical question. As a term in the equation for a volume it’s a constant and has no uncertainty. If you wanted to you could include it in the equation as if it were a parameter to the function, but as it’s uncertainty is zero it will just disappear. But it isn’t really a parameter, it’s just a constant so there isn’t a need to have it as an additional term.
I doubt I can help you much if you still don’t understand the concept.
“n” is a constant with zero uncertainty. So why isn’t it dropped when figuring the uncertainty of the average?
“But it isn’t really a parameter, it’s just a constant”
So is “n” in the average!
““n” is a constant with zero uncertainty. So why isn’t it dropped when figuring the uncertainty of the average?”
Good grief, we’ve been going over this for 2 years and you still haven’t learnt a thing.
Lets take this real slow.
In the formula for the volume π is a constant with no uncertainty. It does not need a separate term because it has no uncertainty. But that does not mean it does not appear in the uncertainty equation. It appears as a modifier to each term. Remember, I told you what the equation is for u(V)
u(V)² = (πR²)² × u(H)² + (2πH)² × u(R)²
See π appears twice. Now if you remember I explained that you then can simplify the equation by dividing through by V². Now this works because V is actually πR²H. So when we divide each of the terms by V² we are actually dividing by (πR²H)². This means that for each term all the values cancel, except for the particular term, which is left as a divisor. Thus creating a relative uncertainty.
E.g. for height we had
(πR²)² × u(H)²
Divide this by (πR²H)²
(πR²)² × u(H)² / (πR²H)²
= u(H)² × [(πR²)² / (πR²H)²]
= u(H)² × [π² / (πH)²]
= u(H)² × [1 / H²]
= u(H)² / H²
And so as if by magic the π symbol disappears. But you can’t actually just get rid of multiplying everything by π. If π was integral to the original equation what happened to it. Well remember we divided by V² = (πR²H)². We couldn’t just divide the right hand side of the equation, because in algebra everything has to balance so we also divided the left hand side as well.
so
u(V)²
becomes
u(V)² / V²
and balance is restored. But V² = (πR²H)²
so we are really saying the left hand side became
u(V)² / (πR²H)²
and so that sneaky little π is still there, and partly determines the value of u(V).
So now what about the average? For simplicity I’ll just talk about an average of two values, X and Y, and leave expanding it to n values as an exercise for the reader.
Let’s call the mean of X and Y, M, where
M = (X + Y) / 2
Our equation for the uncertainty then just becomes
u(M)² = A² × u(X)² + B² × u(Y)²
where A and B are the partial derivatives for X and Y respectively. What are the values. I say with some trepidation, as I know this is the part you claimed was wrong before, but as
(X + Y) / 2 = X / 2 + Y / 2
then we have A = B = 1 / 2.
So substituting back we have
u(M)² = (1 / 2)² × u(X)² + (1 / 2)² × u(Y)²
= (1 / 2)² × (u(X)² + u(Y)²)
and this is a lot simpler than the Volume uncertainty as it’s just addition with a constant modifier. We can juts take the square root and get
u(M) = √[u(X)² + u(Y)²] / 2
In the volume example, you divided the RHS by volume to get the relative uncertainty of the volume.
Yes. As I said, and the LHS as well. That’s why this equation leads to the standard rule that when adding and subtracting numbers you add the absolute uncertainties, but when multiplying or dividing you add the relative uncertainties.
“In the formula for the volume π is a constant with no uncertainty. It does not need a separate term because it has no uncertainty. “
Again, how is this any different than “n” in the formula for the average.
You didn’t answer. My guess is that you can’t.
I did answer, but as always I over-estimated your ability to understand simple equations.
Let my try again. There is no difference between having a function which includes a constant π and one with a constant 1 / n. They both result in a constant in the partial derivative for the term, that partial derivative appears in the uncertainty formula as a weight for each term.
Where the difference occurs is in how you can simplify the equation. In the volume example (becasue it’s an equation involving multiplying) you have a complicated set of modifiers on each term, and it simplifies the equation to divide through by V². This has the effect of removing all the factors modifying each term, which includes π. The effect is to tern the equation from one about absolute uncertainties into one about relative uncertainties.
In the equation for the uncertainty of the mean, there is little point in changing it becasue it is already simple enough. You could if you wanted multiply through by n², but all that would give you is an equation for n times the uncertainty of the mean. You’d still end up having to divide by n if you wanted the uncertainty of the mean.
If you have any further questions please try taking course in algebra first.
“Where the difference occurs is in how you can simplify the equation. In the volume example (becasue it’s an equation involving multiplying) you have a complicated set of modifiers on each term, and it simplifies the equation to divide through by V². This has the effect of removing all the factors modifying each term, which includes π. The effect is to tern the equation from one about absolute uncertainties into one about relative uncertainties.”
q_avg also has “n” as a factor. When you multiply by q_avg why don’t all the “n” divisors cancel out?
Those *modifiers* you are talking about are weighting factors. If you have a common weighting factor in all terms and its uncertainty is zero then you remove it from the uncertainty equation, just like Possolo did with pi. (1/n) becomes a constant weighting factor with an uncertainty of zero. Removing doesn’t change the overall weighting of the other factors.
“q_avg also has “n” as a factor. When you multiply by q_avg why don’t all the “n” divisors cancel out?”
I’m sorry. I can’t help you any further with your homework. If you still don’t understand my explanations, you’ll need to pay for a private tutor or accept that maybe this is beyond you’re abilities.
You tell another yoke-ah here?
“u(M)² = A² × u(X)² + B² × u(Y)²”
And how is this any different than pi in the volume formula?
Here’s a quiz for you: why is the term “standard error of the mean” deprecated in the GUM?
It isn’t deprecated. They say it’s incorrect. I don’t know they say it, that’s why I keep asking. It’s simply stated it’s incorrect with no explanation.
It makes no odds what they call it though. Experimental standard deviation of the mean, is just a more clumsy, less correct name for the same thing.
No, they ae not same. One is the standard deviation of the sample means and the other is the standard deviation of the measurements ASSUMING NO UNCERTAINTY IN THE MEASUREMENTS THEMSELVES.
One is how precisely you have calculated the population mean. The other is the variation of the data elements in the distribution!
Its very simple, go read the definition of the term “error” and see if you can sort it out.
So you don’t know either.
Idiot, all you have to do open, and read.
Ahem.
B.2.19
error (of measurement)
result of a measurement minus a true value of the measurand
There is no true value associated with s/root(N), the SEM does not and cannot produce one. The term is incompatible with uncertainty.
Duh!
B.1 Source of definitions
The definitions of the general metrological terms relevant to this Guide that are given here have been taken from the International vocabulary of basic and general terms in metrology (abbreviated VIM), second edition, 1993* [6], published by the International Organization for Standardization (ISO), in the name of the seven organizations that supported its development and nominated the experts who prepared it: the Bureau International des Poids et Mesures (BIPM), the International Electrotechnical Commission (IEC), the International Federation of Clinical Chemistry (IFCC), ISO, the International Union of Pure and Applied Chemistry (IUPAC), the International Union of Pure and Applied Physics (IUPAP), and the International Organization of Legal Metrology (OIML). The VIM should be the first source consulted for the definitions of terms not included either here or in the text.
Do you think they just pulled these terms out of a hat?
2 Definitions
2.1 General metrological terms
The definition of a number of general metrological terms relevant to this Guide, such as “measurable quantity”, “measurand”, and “error of measurement”, are given in Annex B. These definitions are taken from the International vocabulary of basic and general terms in metrology (abbreviated VIM)* [6]. In addition, Annex C gives the definitions of a number of basic statistical terms taken mainly from International Standard ISO 3534-1 [7].
Annex D
“True” value, error, and uncertainty
The term true value (B.2.3) has traditionally been used in publications on uncertainty but not in this Guide for the reasons presented in this annex. Because the terms “measurand”, “error”, and “uncertainty” are frequently misunderstood, this annex also provides additional discussion of the ideas underlying them to supplement the discussion given in Clause 3. Two figures are presented to illustrate why the concept of uncertainty adopted in this Guide is based on the measurement result and its evaluated uncertainty rather than on the unknowable quantities “true” value and error.
Oh yeah, you are a prime example of the second sentence.
Not once has bellman ever completely read anything for meaning. He is a champion cherry-picker and simply doesn’t care what the context of anything is as long as he thinks he can use it to throw up against the wall in the faint hope something will stick someday.
Nope!
Correct
“There is no true value associated with s/root(N), the SEM does not and cannot produce one. The term is incompatible with uncertainty.”
Meaningless gibberish.
The true value is the population mean. The error is the sample mean minus the true mean. Nobody says the standard error of the mean produces a true mean. It’s a measure of the estimated “dispersal” of the error expected from any sample means.
“Do you think they just pulled these terms out of a hat?”
N, but as far as I can see the VIM does not say anything about the correctness or not of the term “standard” error. I don’t think they even mention means or sampling. The only time they refer to standard deviation is to say that standard measurement uncertainty is defined as a standard deviation.
Granted, I’m not in the habit of keeping up with all international standards reports, so there maybe something I’ve missed. But so far you haven’t quoted anything form the VIM saying standard error is incorrect terminology.
“In addition, Annex C gives the definitions of a number of basic statistical terms taken mainly from International Standard ISO 3534-1“
The same document I’ve taken the definition of standard error from. At no point does it say standard error is an incorrect term.
No, no, and no.
Defend the GAT hoax at any cost!
The Standard Error of the Mean is not the uncertainty in the mean. If you understand what the sample means distribution truly is, then you will recognize that the SEM is the standard deviation of the sample means distribution. So what you have is the mean of the sample means distribution and standard deviation of the sample means distribution.
The standard deviation of the sample means distribution is what you need to define to your self.
It is defined as an interval within which the estimated mean lays. The value of the sample mean is determined by the values that are in the samples and the uncertainty in that calculation is determined by the values in the samples.
The SEM therefore doesn’t determine the “error” in the value of the estimated mean.
The SEM defines the interval where the sample mean may lay. That interval will get tighter and tighter with an increase in the size of each sample. In other words, the ESTIMATE of the population mean gets closer and closer to the actual population.
The reason the SEM has been depricated is because a way too many people, like you, thought it defined the accuracy, precision, and resolution of the value of the mean. It does not. It has nothing to do with the VALUE and UNCERTAINTY of the mean.
Think about it. To make it easy let’s look at data in integers. This is where Significant Digit rules come into play.
1) Data – integer
2) Mean – integer
3) Samples – integer
4) Sample Means – integer
5) Sample Mean – integer
Note1: 2 thru 4 may carry one extra digit to insure rounding accuracy.
Note 2: The standard deviation of the sample means may have any number of significant digits since it defines an interval.
Experimental uncertainty is something different. The GUM has changed from using Standard Error, another descriptor of the SEM, to STANDARD DEVIATION OF THE MEAN. This should make sense with the above.
With experiments, you are not sampling a popuation. Your are instead creating a population using several experiments.
From these you can calculate the population mean and the population standard deviation.
Using the equation:
SEM = σ/ √n
you can now calculate an SEM, which is what the GUM calls the experimental standard deviation of the mean. NIST then explains that this value should be expanded to obtain an experental uncertainty at a 95% confidence level.
If you understand statistics, this should be something you can undrstand if you try.
None of this has anything to do with what Pat Frank has done with instrument resolution uncertainty.
“None of this has anything to do with what Pat Frank has done with instrument resolution uncertainty.”
Indeed. It’s been a massive distraction.
“And considering how wrong Pat is,…”
Something you continually declare and never demonstrate.
“ou are just falling back on the same mistake of thinking the uncertainty in an average is the same as the uncertainty of a sum.”
And you just refuse to understand that the SEM is *NOT* the uncertainty of the average or of the population.
The uncertainty of the average is DRIVEN by the uncertainty of the elements making up that average, not by how accurately you can calculate that average. You simply cannot decrease the uncertainty associated with the average by calculating it more and more accurately.
The average uncertainty, when totaled, will give you the total uncertainty. It is useful in that it can be *used* as an average uncertainty across a population of different things when trying to find the total uncertainty. That is what Pat has done with his analysis of the physical characteristics of the LIG.
If you had *ever* bothered to actually read Taylor, Chapter 3, for meaning instead of cherry-picking it for support of your CAGW dogma, you would understand this. If you had actually ever built a stud wall for a house or built a beam to span a foundation or built a race engine this would all become clear. The total uncertainty in the horsepower output of a race engine is not the average uncertainty of each cylinder but the *sum* of the uncertainties in each cylinder’s output (and the intake manifold, and fuel injectors, etc). It is *exactly* the same thing with the Earth’s temperature. The uncertainty of that value is the *total* uncertainty in the elements making up that average, it is *not* the average uncertainty of the elements.
“You simply cannot decrease the uncertainty associated with the average by calculating it more and more accurately.”
Do you ever listen to yourself?
“The average uncertainty, when totaled, will give you the total uncertainty.”
What do you mean by totaled? Do you mean add up, or combine all the different uncertainties?
“It is useful in that it can be *used* as an average uncertainty across a population of different things when trying to find the total uncertainty. ”
You must stop using AI to write your comments.
“That is what Pat has done with his analysis of the physical characteristics of the LIG.”
Whatever you think he’s done, I’m pretty sure it’s not what you think.
“The total uncertainty in the horsepower output of a race engine is not the average uncertainty of each cylinder but the *sum* of the uncertainties in each cylinder’s output (and the intake manifold, and fuel injectors, etc).”
Thus making it a poor model for talking about the uncertainty of an average.
“It is *exactly* the same thing with the Earth’s temperature.”
You think the earths temperature is the sum of all the thermometer readings?
“What do you mean by totaled? Do you mean add up, or combine all the different uncertainties?”
You’ve never once actually built anything, it is truly obvious in the questions you ask and the assertions you make. NOT ONCE!
Have you ever put anything together as a kit even? Ever had to ask “where did the hole go for this bolt”?
“Whatever you think he’s done, I’m pretty sure it’s not what you think.”
I know *exactly* what he did. He found a mean uncertainty in the design of LIG’s that can be used as a factor for the uncertainty total of temperature measurements. If I take an LIG out in the backyard to measure the temperature then part of the total uncertainty for that measurement includes the instrumental uncertainty associated with the design of that instrument. If I try to combine multiple single measurements made by multiple LIG’s then the individual uncertainties propagate onto the data set, including its average value.
It just becomes more and more obvious with every assertion you make about Pat’s study that you don’t have even a basic concept of metrology. You are stuck in your box of “all uncertainty is random, Gaussian, and cancels”. You don’t even realize you are stuck in a box.
“Thus making it a poor model for talking about the uncertainty of an average.”
You didn’t even understand the simple example I gave you. It’s hopeless trying to explain it to you.
“You think the earths temperature is the sum of all the thermometer readings?”
Again, you just don’t understand. You can’t even distinguish between stated values and the uncertainty interval that goes with the stated values, let alone how that uncertainty propagates. It comes from you always assuming that all uncertainty is random, Gaussian, and cancels. You can’t help yourself. You don’t even know that you do it.
“You’ve never once actually built anything, it is truly obvious in the questions you ask and the assertions you make. NOT ONCE!”
Rather than going through your usual pathetic ad hominems you could have actually answered the question. I’m simply trying to figure out what you meant by “totaled” in that context. It wasn’t a trick question, I just didn’t want to go down the wrong path in the argument.
Which right after you wrote this:
Did you really expect Tim to take your question seriously?
“You are just falling back on the same mistake of thinking the uncertainty in an average is the same as the uncertainty of a sum.”
No one thinks that RMS is the same as RSS. You’re just hyping a strawman.
“Then in equation 5 he does what you are trying to do,…”
You’ve never understood the resolution argument, and still don’t.
He doesn’t know enough to understand that he doesn’t understand. But he thinks he does.
“No one thinks that RMS is the same as RSS. You’re just hyping a strawman.”
Tim Gorman:
https://wattsupwiththat.com/2023/06/29/the-verdict-of-instrumental-methods/#comment-3745807
I just said that RMS is not the same as RSS in my statement. And you can’t even figure that out. What *is* your primary language?
I *told* what the average uncertainty is good for. It is good for THE SAME MEASURING DEVICE. You just can’t separate “same” from “different”. Why is that?
Do you or do you not think that the uncertainty of the mean is the uncertainty of the sum? Your evasion on this point is getting tedious.
In case you can’t figure out why I quoted your comment look at the comment I was replying to. I said
“You are just falling back on the same mistake of thinking the uncertainty in an average is the same as the uncertainty of a sum.”
and Pat replied
“No one thinks that RMS is the same as RSS. You’re just hyping a strawman.”
Obviously, he’s using RMS and RSS as shorthand for the uncertainty of the mean and the uncertainty of the sum – and if he isn’t his reply was a complete non sequitur.
He doesn’t understand that when evaluating the uncertainty of the volume of a barrel that the uncertainty in both the radius and height add.
What are you fantasizing about now? We went over this topic months ago. I agreed entirely with the Possolo example, and with how to use GUM 10, to derive the uncertainty. It is not just the sum of the uncertainty of the height and radius. You need to add in quadrature the relative uncertainties. And the radius uncertainty needs to be doubled.
(u(V) / V)² ≈ (2 × u(R) / R)² + (u(H) / H)²
You never did admit that Possolo assumed no measurement uncertainty, either random error or systematic bias. You wouldn’t admit that this is *exactly* what you and climate science do.
Possolo was trying to find the variation in one thing at one location – Tmax at one measuring station and made the assumption that measuring Tmax on different days was the same as multiple measurements of the same thing.
It works as an example for how to handle experimental measurements but the assumptions necessary in the example make it almost impossible to do in the arena of widespread temperature measurements where multiple locations and multiple measuring devices are involved – yet climate science does it regularly.
And you just don’t get it. It’s too far outside your religions dogma.
“You never did admit that Possolo assumed no measurement uncertainty, either random error or systematic bias.”
Time for a nap. Of course Possolo assumes measurement uncertainty, that’s the whole point of the exercise. The calculation of the uncertainty of the volume is based on the measurement uncertainty of the radius and height. Of course this is assuming random uncertainty, that’s the purpose of the exercise.
How can you think this is assuming no measurement uncertainty?
“””””Of course Possolo assumes measurement uncertainty, “””””
From TN 1900:
“””””Assuming that the calibration uncertainty is negligible by comparison with the other uncertainty components, and that no other significant sources of uncertainty are in play,”””””
The purpose of this example is to determine experimental uncertainty in the data, not measurement uncertainty.
Quality people do this all the time to assess the tolerance specs for products. They must assess all kinds of uncertainty. Calibration, measurement/resolution, experimental, temperature, humidity, etc. TN 1900 develops one type, experimental.
Read the entire damn document, slowly, carefully, while taking notes. There is a plethora of information in it.
We were not talking about TN1900. We were talking about the book about uncertainty, where he uses the example of the volume of a cylinder. This constant moving of goal posts is getting increasingly tiresome.
“Of course Possolo assumes measurement uncertainty, that’s the whole point of the exercise.”
A measurement consists of two parts, the stated value and the uncertainty interval. Possolo only uses the stated values, he ignores the uncertainty intervals.
You simply can’t understand why a trend line developed solely from stated values can’t be assumed to be the “true” trend line. You can’t understand using stated value variation either. Any uncertainty in the stated values *will* also affect the calculation of the variation in the stated values thus adding to the total uncertainty of the data set. The *only* way to make Possolo’s example work is to assume zero measurement uncertainty – and that violates real world constraints. It’s an EXAMPLE, meant for people to learn from. About analyzing experimental measurement results. Again, it doesn’t work in the real world.
” Of course this is assuming random uncertainty, that’s the purpose of the exercise.”
And that is exactly what you do when trying to support GAT calculations!
“This is often done by regarding the measured value, 8.40 m, as the observed value of a random variable whose mean is the true value of R, and whose standard deviation is 0.03 m. “
You can’t even understand that this only applies for multiple measurements of the SAME THING! And even then it ignores systematic uncertainty in the measuring tape!
You simply don’t live in the real world. You can’t distinguish between examples with simplifying assumptions meant as a learning tool and real world applications. You want to live in an alternate reality where measurement uncertainty, both random and systematic, don’t exist!
“Possolo only uses the stated values, he ignores the uncertainty intervals. ”
Now you are lying about Possolo. He wrote a whole book on measurement uncertainties. That’s the book we are talking about – the one that shows how to calculate the uncertainty in the volume of a cylinder, when your measurements for the radius and the height are uncertain. How can you say he ignores uncertainty?
“You simply can’t understand why a trend line developed solely from stated values can’t be assumed to be the “true” trend line.”
Apart from all the times I’ve tried to explain to you how to calculate the uncertainty in a trend line based on stated values?
But, another nice deflection. Why can’t you ever stick to a subject?
“The *only* way to make Possolo’s example work is to assume zero measurement uncertainty”
Again, the whole point of bringing up Possolo was to talk about the example of the cylinder. You were the one who used it as an example in the first place. This has nothing to do with TN1900 example 2.
“You can’t even understand that this only applies for multiple measurements of the SAME THING! ”
Where does it say you have multiple measurements of the same thing?
“And even then it ignores systematic uncertainty in the measuring tape!”
Take it up with Possolo. It’s an example to explain how you can propagate uncertainty – not something that need you to account for every possible issue.
“You simply don’t live in the real world. ”
Yet you were the one who brought up this non-real world example. You were the one who said
Was that meant to be about living in the real world? Do you think in the real world the uncertainty in the volume is just the adding of the uncertainty in both the radius and height?
You are so lost in the forest you can’t even keep track of what is going on in the thread. EXPERIMENTAL uncertainty was the subject of Ex 2 in TN1900 and *it* was what we were discussing. I only brought up the volume of a barrel in order to give you another example of where Possolo didn’t consider systematic bias in the measuring device.
“Apart from all the times I’ve tried to explain to you how to calculate the uncertainty in a trend line based on stated values?”
You calculate the best-fit residuals — ASSUMING THE STATED VALUES USED FOR THE TREND LINE ARE 100% ACCURATE!
It’s your same old meme over and over again. You assume all uncertainty is random, Gaussian, and cancels so all you have to do is look at the stated values to get a “true” trend line that you can data match to. You don’t even realize you do it. Even after example after example of showing that if the stated values have uncertainty then so does the trend line!
“Take it up with Possolo. It’s an example to explain how you can propagate uncertainty – not something that need you to account for every possible issue.”
EXACTLY! Which means it is a “learning example” not meant to describe the real world of measurement uncertainty – AND YET YOU TAKE IT AS BEING A REAL WORLD EXAMPLE!
“Was that meant to be about living in the real world? Do you think in the real world the uncertainty in the volume is just the adding of the uncertainty in both the radius and height?”
The uncertainty in the volume of the barrel *is* the propagated uncertainty from the measurements! But it is ALL of the uncertainty – not just the random error introduced from eyeball reading of a measuring tape!
You make a statement like thsi, never actually understanding that what you are saying is the the uncertainty of the individual measurements must be propagated onto the average in order to determine the uncertainty of the average.
You’ll turn around in the next post and say that the uncertainty of the average is the average uncertainty which is *NOT* the propagated uncertainty of the measurement elements! You just say what works for you at the time, regardless of whether it is consistent or not. And you wonder why everyone thinks you are a troll?
“You are so lost in the forest you can’t even keep track of what is going on in the thread.”
The thread, or at least this part of it, was you claiming about me
Everything since then has been you trying to deflect from it.
And moths ago you couldn’t accept that GUM 10 does *NOT* apply when systematic uncertainty is present. And you STILL can’t.
You can’t even distinguish when quadrature addition is justified and when it isn’t.
Just making stuff up now. If I ever said that, quote the exact words and provide a link.
GUM 10 is for random independent uncertainties. Systematic uncertainty, by definition is not independent.
When you quote Gum 10 you are quoting something that ignores systematic uncertainty. It is part of your meme that “all uncertainty is random, Gaussian, and cancels”. Why do you never want to consider that field measurements ALWAYS have systematic uncertainty? Pat’s study is about one factor that contributes to systematic uncertainty in an LIG device. Yet you continue to hold that his study is wrong for some reason. Systematic uncertainty is different for every different measuring device. You *never know* where a device is exactly with calibration drift. Calibration drift is a time dependent factor as well as an equipment factor. NOT KNOWING is what the uncertainty interval is supposed to be used for. Device design shortcomings, however, are not calibration drift. You *can* analyze them and determine a mean value for the uncertainty introduced by the design shortcomings and then apply that across the universe of those devices as systematic bias. Something you just can’t seem to get your head around!
GUM 10 is for repeated measurements of the same thing using the same device under identical conditions. Yet you want to apply it to *everything* – including temperature measurements from thousands of different locations using thousands of different devices with wildly different variances and wildly different systematic uncertainties.
Every thread you get involved in here on WUWT is focused on real temperature measurements in the real world and yet you want to try and cram everything into the same box with you – where all uncertainty is random, Gaussian, and cancels. You can’t even admit that Possolo’s Ex 2 in TN1900 is a LEARNING example and not something that is useful in the real world of “global average temperature”.
If all you can talk about is life in your alternate universe then you are wasting everyone’s time and bandwidth on here. No one cares what happens in your alternate universe.
TN 1900 is an example of one component of of a combined uncertainty. The GUM goes to great length to raise the awareness of including ALL uncertainty. TN 1900 does not address measurement uncertainty or any systematic uncertainty. Hubbard and Lin showed the systematic uncertainty must be addressed at each station individually. There are too many microclimate variables to assume a standard value. Anthony has proved that with his station sitng studies.
“When you quote Gum 10 you are quoting something that ignores systematic uncertainty.”
You were claiming I said that it did. I’m asking you to back up that claim or admit it was a lie.
“Yet you continue to hold that his study is wrong for some reason.”
For a number of reasons.
“Systematic uncertainty is different for every different measuring device.”
Which is one of them. If every device has a different systematic, then you can’t count those errors as being the same across all measurements. For the purpose of one instrument it’s systematic, but across all instruments it’s more random.
“GUM 10 is for repeated measurements of the same thing using the same device under identical conditions.”
It is not. There is no requirements for any of the inputs to be based on repeated measurements.
“Yet you want to apply it to *everything* – including temperature measurements from thousands of different locations using thousands of different devices with wildly different variances and wildly different systematic uncertainties.”
Nobody, as far as I know is applying GUM 10 to any uncertainty analysis of a global mean anomaly. It’s much more complicated than that.
“where all uncertainty is random, Gaussian, and cancels.”
For the benefit of any new readers here. This is a lie that Tim repeats over and over again in the hope that someone will be gullible enough to believe it.
“You can’t even admit that Possolo’s Ex 2 in TN1900 is a LEARNING example and not something that is useful in the real world of “global average temperature”.”
That’s exactly what I keep trying to tell Jim.
“If all you can talk about is life in your alternate universe then you are wasting everyone’s time and bandwidth on here.”
Nobodies forcing you to waste it. Normally all that happens in these threads is I’ll ask a simple question or make a simple observation, and then I’ll be bombarded with countless attacks, which I like a fool have to respond to. And then we get stuck in an infinite loop.
Stop posting nonsense that you don’t understand.
I think that’s the total, so the variance needs to be scaled for the average.
It might be worth looking at relative uncertainties to see how they compare.
“The interval within which the population mean might lie is really only applicable when you have what the document calls “repeatable” measurements.”
You keep wanting to have this both ways. You take a document that talks about repeatability in a laboratory setting, but then use it to make claims about statistical field experiments. The document says nothing about the population mean, it’s only talking about how to get a better estimate of the measurement of a single thing by repeated measurements. Yet somehow draw the conclusion it’s saying the idea of a confidence interval isn’t even applicable without repeated measurements.
“I especially like the treatment of “range” as an index into uncertainty estimates.”
Which particular range are you talking about? They use it somewhat loosely in two different contexts. As a confidence interval, and as the range of the entire data set. Two very different ranges.
“The range is directly related to the variance in a data set.”
You realize this is exactly what I’ve been trying to explain to you. The SEM is directly related to the SD of a data set. It’s just you keep insisting it’s impossible to derive the uncertainty of a mean from just one sample.
“Pat’s paper is all about (3).”
Really? You think the resolution in the thermometers means there is no difference in the measurements. That all temperatures recorded anywhere and anytime are always the same, becasue the resolution of the LiG can’t detect the difference between 10°C and 20°C?
What (3) is saying is what I keep saying – that repeated measurements of the same thing, will nopt help you if you get the same reading each time. Because the random errors in the measurement is less than the resolution (what they call precision, incorrectly) of your device.
This is not a problem when you are measuring lots of different things of different sizes, becasue then the variation is going to be bigger than the resolution.
I’m fairly sure this is just spurious precision arising from the calculation.
The variation of different things is of dubious value at best. The average of SH and NH temps, which have different variances and ranges, tells you exactly as much about the temperature of the Earth as would combining the heights of Shetland ponies with the heights of Arabians would tell you about the heights of horses. The average of a bi-modal or multi-modal distribution tells you little about the distribution and the total variance of the combined data just gets huge.
“I’m fairly sure this is just spurious precision arising from the calculation.”
Lots pf people seem to believe this, and tell me repeatedly. What I never see is an explanation or proof or demonstration as to why all the precision is spurious.
To me iot seems to come down tot he simple fact that people here don’t understand what an average is.
You mean that people that have had to LIVE with real world measurements and results don’t understand. Only you know what the average of different things really is in the real world.
“You mean that people that have had to LIVE with real world measurements and results don’t understand.”
Arguments from spurious authority carry little weight with me. Unless you can demonstrate in some way why it’s impossible for the precision of an average to be less than the individual measurements, I’m not interested. As things stand you are just arguing from tradition.
I’ve given you demonstrations showing that resolution of an average is not destroyed when you reduce the resolution of the data. The fact you can’t even accept that, or any over piece of evidence, makes it clear this is just an article of faith for you. Man can never fly you say, therefore that plane cannot really be flying.
“Only you know what the average of different things really is in the real world.”
Only me, and any one who’s studied statistics.
“Arguments from spurious authority carry little weight with me.”
Taylor, Bevington, and Possolo are “spurious authority”? Only for you, only for you!
“Unless you can demonstrate in some way why it’s impossible for the precision of an average to be less than the individual measurements, I’m not interested. As things stand you are just arguing from tradition.”
Don’t try using argumentative fallacies, you don’t understand them any better than you understand uncertainty.
Even a six year old can understand that the average can’t have any more precision than the measuring device provides. This has been explained to you over and over ad infinitum. And you simply refuse to learn. If this were true you could get infinitely precise measurements from a yardstick. You’d just have to make enough measurements. Karlomonte’s ask of you to patent such a micrometer-yardstick *should* have gotten through to you bit it didn’t. It never will. You absolutely, stubbornly refuse to learn.
“Taylor, Bevington, and Possolo are “spurious authority”? Only for you, only for you!”
Truly pathetic. I’d say a new low even for you, but I expect you’ve done worse.
Anyone can follow this thread up to see what I was talking about – and it was not any of those authors. I’m specifically talking about any of these labs who it is claimed say that “Thou must never quote an average to more decimal places than the individual measurements”.
None of those sources say anything of the sort – and if they did I’m sure you could quote their exact words. They all talk about the sensible rules, based on actually calculating the uncertainty, and which can easily lead to you having more digits in the average.
“Even a six year old can understand that the average can’t have any more precision than the measuring device provides. ”
I don’t take advise from a six year old anymore than I would from some random physics lab. Arguments from incredulity are not very convincing. Just provide the mathematical prove, or a clear explanantion as to why it’s impossible – not just claim that it seems obvious.
“This has been explained to you over and over ad infinitum.”
So that’s an now argumentum ad nauseam. Any more fallacies you would like to demonstrate.
“If this were true you could get infinitely precise measurements from a yardstick.”
No you could not for multiple reasons I keep explaining to you.
When are you going to start?
After you find more “errors” in the ISO terminology?
“None of those sources say anything of the sort – and if they did I’m sure you could quote their exact words. “
I’ve given you their exact quotes on this at least three times. I’m not going to repeat myself. Go look at Taylor, section 2.5 and Bevington, Chapter 1, Pages 2 and 3.
For the umpteenth time, you need to READ BOTH AUTHORS COMPLETELY, FOR MEANING, AND DO THE EXAMPLES!
STOP CHERRY-PICKING AND LEARN THE SUBJECT!
Nor have you given any reasons. You just keep saying the average can be more precise than the actual measurements – meaning a yardstick can be used as a micrometer if only you take enough readings.
Which “average” would you like? There are plenty to select from?
Assuming you mean the arithmetic mean, it’s the sum of the observations divided by the number of observations. It doesn’t necessarily lie in the possible sample space.
I did this before, but there’s been a lot of to-ing and fro-ing between lots of people so would have been easy to miss.
Let’s try the faces of our familiar 6-sided die. The observations are {1, 2, 3, 4, 5, 6} The sum is 21, and the mean is 21/6. Taking out common factors, it’s 7/2.
Converting to decimal, you get 3.5, which has added a significant digit.
On the other hand, if the set was {2, 4, 6, 8, 10, 12}, you would have a sum of 42 (an auspicious number), and a mean of 42/6, or 7. No additional significant digits.
Similarly, {1.0, 2.0, 3.0} is 6.0 / 3, or 2.0, whereas {1.0, 2.0, 4.0} is 7.0 / 3, or 2.333333333333333333 to however many digits you care to express it.
Has the presence of 4.0 rather than 3.0 in the set of observations improved the precision of the mean?
“Converting to decimal, you get 3.5, which has added a significant digit.”
Indeed it has. But that doesn’t mean that extra digit is spurious. 3.5 very much is the correct mean of a fair 6-sided die, and insiting on rounding it to 4 gives a very misleading result.
The fact that the .5 is not spurious cvan easily be seen by throwing very large numbers of dice. You can be reasonably certain that the answer will always be close to 3.5, and not 4.
“Similarly, {1.0, 2.0, 3.0} is 6.0 / 3, or 2.0, whereas {1.0, 2.0, 4.0} is 7.0 / 3, or 2.333333333333333333 to however many digits you care to express it. Has the presence of 4.0 rather than 3.0 in the set of observations improved the precision of the mean?“”
Which is just a problem of trying to push fractions into a decimal system. The extra digits aren’t spurious, it’s just the correct answer is 2 1/3. It’s just saying the correct answer in decimal can be as accurate as you want or need.
But in most real world situations this isn’t a problem. There’s enough uncertainty in the sampling to make too many digits not useful or meaningful.
What none of this means is that an average cannot by more precise, have more digits than were used in the individual measurement.
First, dice are not measurements. Measurements of temperature are samples of a continuous function. Totally different from a discreet function.
As an example, what is the probability function of multiple fair die rolls. IOW, triangular, normal, Poisson, binomial, rectangular, etc.
Now let’s look at Tavg. What is the probability function of Tavg? How about a monthly distribution of Tavg.
Second, the mean may be 3.5, but what is the probability of rolling 3.5? You need to discuss 3.5 in terms of Expected Value. Then discuss a single measurement in terms of an Expected Value, i.e., the probability of a certain temperature.
“First, dice are not measurements.”
It wasn’t my example.
“Totally different from a discreet function.”
Not totally different. For many circumstances they work the same, it’s just the maths becomes more complicated when dealing with continuous functions.
“As an example, what is the probability function of multiple fair die rolls”
How many do you want. For large enough throws the CLT tells you it can be approximated by a normal distribution.
You can work out an exact solution for different numbers of dice, but I’m not sure if the distribution has a name.
“What is the probability function of Tavg?”
You’ll need to give some context. For an individual station on a specific day, or for all places on the globe over all time? Are you talking about the actual temperature or the measurements. If just the measurements rounded ot a specific number of digits, you have a discrete distribution in any case.
“Second, the mean may be 3.5, but what is the probability of rolling 3.5?”
Zero. Or as close to it as makes no difference.
“You need to discuss 3.5 in terms of Expected Value.”
You can, or you can discuss it as the average of a number of rolls.
“Then discuss a single measurement in terms of an Expected Value, i.e., the probability of a certain temperature.”
Why. You’re the only ones who seem to worry that any given value is not always the expected value. Expected does not mean the number we expect for all or any value, it’s the expectation of the average, in an absolute sense, the average of all possibilities.
And to get back to the point – the 0.5 is not spurious. It’s exactly what you expect on average. If it were spurious, any value between 3 and 4 would be equally valid. But roll the dice enough times and you can be sure the average will converge to 3.5, not 3.7 or 3.1.
It’s a fair cop with the halves, but once you have remainders from dividing by larger primes you introduce a lot of spurious decimal precision.
Try {1. 1. 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 2} (14/13)
Yes, it’s a problem with base conversion, but what people see is the decimal notation.
Not talking about sampling; purely the perceived precision of the mean.
You just agreed about the decimal conversion of thirds.
“Try {1. 1. 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 2} (14/13)”
The trouble here is you are trying to talk about spurious precision in exact averages. There is nothing spurious about 14/13. It’s exactly the average of those numbers. As a mathematician that’s the only acceptable way of writing it 😉 But if you insit in converting it to a decimal, then any number of places will be an approximation (ignoring the ways you can indicate repeating digits). The more decimal places the less spurious the answer.
However this is only tangential to the problem being posed, which is what happens when you have a messy real world average of multiple different values all measured with a degree of uncertainty and to a specific number of digits.
The claim made here, and elsewhere, is that if you measure everything, say, to the nearest unit, it’s impossible to know the average to less than the nearest unit. And that’s where I disagree. With a large enough sample size the uncertainty in the mean, whether coming from sampling or from the measurements, can be less than a unit. If you calculate the mean as 3.5123, and the uncertainty as ±0.12, you should write this as 3.51 ± 0.12, rather than 4.
The idea that the 0.5 is spurious is wrong in my opinion – it carries a lot of useful information which is lost if you round it to the nearest unit, and just quoting 4 as an answer not only adds uncertainty it distorts the result. The implied uncertainty of just stating 4 is that the answer could be anything from 3.5 to 4.5, when we know it’s almost unlikely to be larger than 3.63 but could be less than 3.5.
And this why you need to patent your yardstick-micrometer…
I’ll tell you what. I’ll patent for you, once you explain how you can assure there are no systematic errors above a micrometer, a practical way of taking trillions of measurements, and how we ensure that the errors in each measurement are greater than several yards.
Until then you should concentrate on your straw man gun.
At this point I see you have two choices:
1) write to the chairman of the ISO terminology committee and tell him how they gored your sacred cow.
2) go talk to the Forestry Dept. and see if they can fit you in, before it is too late.
Or 3) let engineers use whatever terminology they like, whilst the rest of the world continues to use the correct term.
So the ISO committee consists of nothing but engineers?
You know this … how exactly?
Where did I say that? What I said was I didn’t care if engineers wanted to use their own terminology, as long as they didn’t try to enforce it on others.
Here’s what the ISO says about the term standard error.
https://www.iso.org/obp/ui/#iso:std:iso:3534:-1:ed-2:v2:en
And you STILL haven’t read the GUM, no surprise.
Like you said, bellman believes only engineers use the GUM.
I’m sorry if you find the word “engineer” derogatory. It was not meant to be. I merely used it to differentiate the people who are using the terminology in the GUM to those who use the more common definitions. And this only applies to the question of whether you think standard error is an incorrect term.
So just who do you think wrote the GUM?
Or is this another yoke-ah?
“””””Commonly, in practice, standard error is implicitly referring to the standard deviation of the sample mean. “””””
You need to work out in your brain what the “sample mean” truly is. I don’t think ISO defines it the same way you do. Read the preceding items and their references in the document and carefully work out the examples.
You will find that ISO is pretty much like the GUM. Their documents just haven’t been updated recently.
Up above he claimed its a “true” value!
“You need to work out in your brain what the “sample mean” truly is”
In my brain it’s the mean of a sample. But lets see what the ISO says:
Seems about right. Why, what did you think it meant?
OK, let’s work through the examples
97 / 10 = 9.7. Sound’s like a mean to me.
Look, it’s a long day, and this thread is just more distraction. If you have something to say about the ISO definition, just say it rather than expect me to find it.
“You will find that ISO is pretty much like the GUM. Their documents just haven’t been updated recently.”
To paraphrase an old joke, the great thing about standard is there are so many to choose from.
“The claim made here, and elsewhere, is that if you measure everything, say, to the nearest unit, it’s impossible to know the average to less than the nearest unit. And that’s where I disagree. With a large enough sample size the uncertainty in the mean, whether coming from sampling or from the measurements, can be less than a unit. If you calculate the mean as 3.5123, and the uncertainty as ±0.12, you should write this as 3.51 ± 0.12, rather than 4.” (bolding mine, tpg)
The interval in which the average might lie, the standard deviation of the sample means, is *NOT* the precision of the mean. It has nothing to do with the precision of the mean, only of your sampling and its prediction of the mean.
The precision of the mean is determined by the precision of the measurements in the data set. The precision of the mean is determined by the data element with the least precision.
This is such a simple concept that a child can figure it out.
Plus 1 digit for rounding. bellman caught me fair and square on that.
That extra digit is only for intermediate calculations. It is supposed to be truncated (by rounding or whatever) in the final result so that the final result has no more precision than the constituent measurments.
Supposed to by whom. Again none of the sacred texts tell you to have a final result that has no more digits than the constituent measurements.
“Again none of the sacred texts tell you to have a final result that has no more digits than the constituent measurements.”
No more digits? Huh?
Again none of the sacred texts tell you to have a final result that has no more digits than the constituent measurements.””
Do you suffer from short-term memory loss? You’ve been given this at least once in this thread and multiple times over multiple threads!
Taylor: Section 2.2
“Rule 2.5: Rule for Stating Uncertainties – Experimental uncertainty should almost always be rounded to one significant figure.”
“Rule 2.9: Rule for stating answers – The last significant figure in any stated answer should usually be of the same order of magnitude (in the same decimal position) as the uncertainty.”
His example is stating a measured speed as 6051.78 +/- 30 m/s is obviously ridiculous. The uncertainty of 30m/s means the digit 5 really might be as small as 2 or as large as 8. Meaning the trailing digits of 1, 7,8 have no significance at all! Calculating out an average past where the uncertainty applies is just ludicrous. It’s what a statistician does when they have no relationship to the real world.
Bevington, Chapter 1.1, Page 4:
“In other words, we let the uncertainty define the precision to which we quote our result.”
Exactly the same as Taylor.
Since the uncertainty is driven, at least in part, by the resolution of the measuring device, which, in turn, determines the precision with which a result can be accurately stated, the average value falls under the same limitation – it should not be quoted to a precision level greater than the uncertainty limit. Since the uncertainty magnitude is based on the least precise data provided by a measurement, it also determines the precision with which the result can be stated.
Why do you just keep on making these idiotic assertions? None of my five statistics books even cover this. In fact, of the two I have at my desk right now, neither have the words “resolution”, “precision”, or “measurement” in the index! That should be an indication of how much emphasis introductory statistics textbooks put on the science of metrology – NONE! Is it any wonder that so many in science and math today don’t have a clue about the real world of measurements and uncertainty? ?That includes *YOU*!
He has a mental block to anything that is out-of-phase with his GAT hoax predilections.
So now the GUM is “wrong”, without him putting eyeball to page.
he doesn’t *read* anything, he just cherry-picks.
“Experimental uncertainty should almost always be rounded to one significant figure.”
Exactly. That’s the uncertainty not the result. If the uncertainty of a mean is less than the uncertainty of the measurements you can quote the result to more digits than the measurements. He tells you that explicitly. You should no that as you’ve “done all the exercises”.
“In other words, we let the uncertainty define the precision to which we quote our result.”
Again, exactly my point. The uncertainty defines the “precision”. Not some silly rule about the number of digits in the original measurements – the uncertainty.
“Exactly. That’s the uncertainty not the result.”
Why didn’t you read the rest?
“The last significant figure in any stated answer should usually be of the same order of magnitude (in the same decimal position) as the uncertainty.””
You can’t even read and understand two paragraphs. You have to cherry-pick one thing!
” If the uncertainty of a mean is less than the uncertainty of the measurements y”
The “uncertainty of the mean” the way you use it is the interval in which the population average might lie. As we keep telling you that has nothing to do, ABSOLUTELY NOTHING, with the accuracy of the mean. You can calculate the population mean down to the millionth digit and it can still be wrong by a million units.
The SEM is *NOT* the uncertainty of the average. You simply don’t understand precision vs accuracy. You can put five shots on target into the same hole (high precision) but they may be far from the center of the target. You can calculate the population mean down to “one hole” and it can still be right at the edge of the target (highly inaccurate).
Do you understand *anything* about the real world? It’s been explained to you often enough.
“Why didn’t you read the rest? ”
Just because I didn’t quote something doesn’t mean I haven’t read it. The second half of the quote continues to agree with me, and is just saying what I said in my previous comment – the uncertainty determines the number of digits to report.
“The “uncertainty of the mean” the way you use it is the interval in which the population average might lie.”
Not really, but lets not quibble.
“As we keep telling you that has nothing to do, ABSOLUTELY NOTHING, with the accuracy of the mean.”
Have you ever read Taylor. He makes it clear that he regards the SEM as the uncertainty of the mean. As always this doesn’t include the possibility of systematic errors, but that doesn’t mean it tells you ABSOLUTELYNOTHING about the accuracy, let alone the uncertainty.
“You can calculate the population mean down to the millionth digit and it can still be wrong by a million units.”
I’m sure you could. But you would have to have a really bad experiment to do so.
“The SEM is *NOT* the uncertainty of the average. You simply don’t understand precision vs accuracy.”
Yes, we keep on going round in circles over this. Precision is not accuracy. Accuracy is not trueness (as defined in the VIM). You can be very precise and still wrong. You can be very true and still not accurate. None of this is an argument for ignoring the precision of your measurement or average.
“Do you understand *anything* about the real world? It’s been explained to you often enough.”
And it doesn’t matter how many times I say I agree you will still keep claiming I don’t.
The problem is you keep using this argument as a distraction. This argument has nothing to do with how many significant figures you report or whether it’s better to have a large or small sample size, or whether the SEM is an indication of the uncertainty a sample mean.
“ He makes it clear that he regards the SEM as the uncertainty of the mean.”
Give me a section and page number. My guess is that you will quote something out of Chapter 4 or later where he is talking about random and Gaussian uncertainty — WHICH MEANS MEASUREING THE SAME THING MULTIPLE TIMES.
“As always this doesn’t include the possibility of systematic errors, but that doesn’t mean it tells you ABSOLUTELYNOTHING about the accuracy, let alone the uncertainty.”
It tells you *NOTHING* about the real world. Again, if you would stop cherry-picking and actually read and understand Taylor and Bevington you *might* get a glimpse of this. You haven’t even bothered to read Taylor’s Chapter 1!
It is so tiresome typing in here from Taylor’s book that I can’t believe it. YOU COULD JUST READ HIM FROM COVER TO COVER!
Chapter 1 Intro:
“Experience has shown that no measurement, however carefully made, can be made completely free of uncertainties.”
“This chapter then describes how (in some simple cases, at least) the magnitude of the experimental uncertainties can be estimated realistically, often by means little more than plain common sense.”
Section 1.4: “In the applied sciences, for example, the engineers designing a power plant must know the characteristics of the materials and fuels they plan to use. The manufacturer of a pocket calculator must know the the properties of its various electronic components. In each case someone mus measure the required parameters, and having measured them, must establish their reliability, which requires error analysis. Engineers concerned with the safety of airplanes, trains, or cars must understand the uncertainties in driver’s reaction times, in braking distances, and in a host of other variables; failure to carry out error analysis can lead to accidents such as that shown on the cover of this book.”
This is what we are trying to explain to you and you refuse to accept. Quoting the average with more precision than the measurements themselves support can lead to every single on of the instances above. The average simply can’t be made more reliable by calculating it out to more digits than the measurements themselves justify.
You only confirm that you have never, EVER, been in a situation where health and welfare depend on your measurements. And you simply won’t believe that significant digit rules apply ALWAYS, not just when *you* choose to follow them.
“I’m sure you could. But you would have to have a really bad experiment to do so.”
You just whiffed again. You simply can’t bring yourself to admit the truth!
” None of this is an argument for ignoring the precision of your measurement or average.”
You just don’t get the concept of uncertainty and the role it plays in the real world. You never will until a significant emotional event brings you face-to-face with reality. And since you will never be involved in a health and welfare situation you will never have that significant emotional event.
“The problem is you keep using this argument as a distraction”
It’s *NOT* a distraction – it has real world consequences. And you refuse to accept that. It’s because it goes against your CAGW religious dogma of “all uncertainty is random, Gaussian, and cancels”.
“Give me a section and page number.”
What. I thought you had read the entire book, done all the exercises and memorized it for meaning.
“where he is talking about random and Gaussian uncertainty — WHICH MEANS MEASUREING THE SAME THING MULTIPLE TIMES. ”
and there are the goal posts moving. You claim that when reporting an average the number of digits must be based on the number in the measurements. Now you are going to claim that this rule doesn’t apply when you are actually just measuring the same thing multiple times. How would this make any sense?
Of course he’s talking about measuring the same thing becasue that’s the point of the book. It’s a book about measuring things, not statistics. But here goes. Page 102, section 4.4 (missing a few symbols to save time).
Note, the mean is more reliable than any one measurement, and σ divided by √N is the uncertainty.
“Quoting the average with more precision than the measurements themselves support can lead to every single on of the instances above.”
Yet Taylor never bothers to warn about this. And actually says the mean may be more precise than the measurements.
“And since you will never be involved in a health and welfare situation you will never have that significant emotional event.”
Glad you can predict my future. But a few dead relatives and a few health scares myself, suggests your prediction may not be that accurate.
He still doesn’t understand uncertainty versus error.
There seem to be various opinions on this, but one rule of thumb appears to be that the number of significant figures is of the same order as the sample size. That makes sense, because the mean is the sum divided by the count.
That is totally separate from the resolution, accuracy or precision of the data points.
For determining the mean value of a series of measurements of a single item, then truncation does seem appropriate. It’s probably context sensitive.
In physical science and engineering it is extremely important to not give the impression that you have measured things to a higher resolution that you actually did. Adding significant figures to an average value over and above what you actually measured gives others the impression you have used higher resolution measurements than you actually did.
It’s why the resolution of the average should be no greater than the resolution of the measurement with the least resolution in your data set. The resolution is always limited by that one value, unless, of course, you decide to delete it from the set. But then you have to justify doing that.
“In physical science and engineering it is extremely important to not give the impression that you have measured things to a higher resolution that you actually did.”
But as you keep saying, an average is not a measurement. It’s a statistical calculation based on, possibly, hundreds or thousands of individual measurements.
“Adding significant figures to an average value over and above what you actually measured gives others the impression you have used higher resolution measurements than you actually did.”
Only people who don’t understand what an average is. (Which if comments here are to be believed might include most physics and chemistry students.)
“It’s why the resolution of the average should be no greater than the resolution of the measurement with the least resolution in your data set.”
The main point of averaging is to be able to detect differences in populations, with the well understood point being the larger your sample size the easier it is to detect small differences. The size of the difference you can detect is the definition of resolution.
Take the die example. You have a die, and want to test if it’s fair or biased. A simple test is to throw the die a number of times and see what the average is. If it’s significantly different from the expected 3.5 you can say with some confidence that it is not fair. The more throws you make in your test the easier it is to detect small significant differences.
Clown. You are obviously not part of this set.
The wheels on the bus go round and round!
“But as you keep saying, an average is not a measurement. It’s a statistical calculation based on, possibly, hundreds or thousands of individual measurements.”
*YOU* are the one that wants to present it as a measurement! In the case where you have multiple measurements of the same thing under the same conditions and systematic uncertainty is reduced to insignificance, the average is a representation of the “true” value”. You simply don’t want to mislead others as to how precisely you made your measurements.
In situations where this is not the case it actually becomes even more critical to not overstate the precision. Suppose you are travelling in a capsule to Mars. Your engine burn times are based on the angular velocity you have with the planet. Do you want to mislead everyone that you measured that angular velocity to a precision you can’t justify with the measurements you made of it? Doing so could lead you into plowing into the planet or missing your insertion window by a wide margin because of the distance travelled at the wrong direction and velocity.
“Only people who don’t understand what an average is. (Which if comments here are to be believed might include most physics and chemistry students.)”
NO, *including* people that understand what an average is. People who assume that you know not to misrepresent your findings. People who depend on the precision with which you made your measurements to make safe products.
“The main point of averaging is to be able to detect differences in populations, with the well understood point being the larger your sample size the easier it is to detect small differences. The size of the difference you can detect is the definition of resolution.”
Huh? “differences in populations”? You mean like the height differences between Shetland ponies and Arabians? Or do you mean differences in results from similar populations? Like a control group vs a test group where both groups are similar populations?
Once again, you betray your alternate universe. The subject at hand is PHYSICAL MEASUREMENTS such as temperatures. Combining different temperatures from around the world *IS* like finding the average height of the animals in a corral that are a mix of Shetland ponies and Arabians. They are not common populations. Neither are temperatures taken at Manitoba and Rio de Janeiro common populations.
He won’t understand what angular velocity is.
“Your engine burn times are based on the angular velocity you have with the planet. Do you want to mislead everyone that you measured that angular velocity to a precision you can’t justify with the measurements you made of it?”
If I was traveling to Mars and my life depended on knowing the angular velocity, and for some reason it could only be measured by taking the average of a large number of measurements with an imprecise instrument – then I would definitely not want them to round the best estimate of the mean to the nearest integer, or whatever. Why would I possibly want to change a value that was a best estimate, for one you knew was probably wrong. It;s not much consolation as I crash into Mars to know that at least they didn’t overstate the precision.
So you would rather have them depend on a value you give them that is more precise than your measurements justify. And you can’t even comprehend the ramifications of that!
If they *know* the uncertainty in your average then they can allow for it – something they can’t properly do if you misrepresent the uncertainty by giving a more precise average than the measurements justify.
I feel like Michael Palin 🙁
and what is that if not spurious precision?
It all looks very sciency, but 14/13 (or 15/13 or 25/13 or …) is purely a ratio of integers. The decimal representation leads one to think that the observations are also continuous values which have been measured to the same resolution.
Rather than the discrete values, repeat the exercise with measurements to 2 decimal places {1.00. 1.00. 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 1.00, 2.00}. Each of those measurements has an implicit resolution of +/- 0.005, so it’s now 14.00 +/- .065 (feel free to add in quadrature if you prefer), so are you still going to go to 20 decimal places?
You are going to have to define “spurious” in this context. To me it’s implying any digits after a certain value can be discarded because they are essentially random.
“Each of those measurements has an implicit resolution of +/- 0.005, so it’s now 14.00 +/- .065 (feel free to add in quadrature if you prefer), so are you still going to go to 20 decimal places?”
No. As I keep saying you need to estimate the uncertainty and use that. In your case if the uncertainty of the sum is ±0.065, then the uncertainty of the mean is ±0.005. You can report the average as 1.077 ± 0.005.
This, though, supposes the uncertainties really are ±0.005 based only on the number of digits, they could be much larger. Secondly, you are only talking about the measurement uncertainty here. Assuming this is a sample, the standard error of the mean is a better guide. In this case that would be 0.074, and using a 2σ uncertainty interval, you could say 1.08 ± 0.15. This would be true whether you measure to the nearest integer, or 1/100th.
Thank you.
That’s over-complicating it. Make it as simple as possible, but no simpler.
“Make it as simple as possible, but no simpler.”
But that’s why I don’t like the SF rules. They are far too simple and result in either too much or too little uncertainty.
Learn what the word means prior to using it in sentences, sounds just like the IPCC.
“They are far too simple and result in either too much or too little uncertainty.”
Wrong! They help PREVENT either too much or too little uncertainty! They help get just the right amount of uncertainty!
I don’t know why people waste their time writing all these books on uncertainty. If you can get just the right amount of uncertainty by just counting the digits in your measurements, why care about all these propagation rules?
Because people’s lives and fortunes depend on doing measurements right and presenting your results correctly.
The truth comes out. You don’t LIKE Signifiant Digit rules. Too bad.
You are displaying your lack of knowledge and appreciation of what physical measurements consist of.
SF rules were developed for a reason.
One, so that people attempting to replicate an experiment would get similar results. The measurements could be refined by using better methods and devices.
Two, people replicating experiments would not think they needed to purchase better instruments with better resolution in orfer to achieve similar results.
It is very much why trained engineers and scientists in other fields who live and die with the measurements they make simply cant believe anomaly values from the early 1900’s that were measured to the nearest units digit yet quoted to the 1/100ths resolution and uncertainty in the 1/1000ths. These are mathematician values and not measurement values.
Yep!
And this clown thinks the GUM is “wrong” when they gored his scared ox.
I wish you would make your minds up. Do I worship at the later of the GUM, as Pat Frank was accusing me of, or do I despise it and all it’s works because I disagree with one comment.
Poor whinerman, so oppressed.
Get a brain, PDQ.
Irrelevant to the issue.
Did I say it was relevant to any particular issue? I’m just pointing out the lack of consistency in all these ad homs.
Saying I worship at the alter of GUM is irrelevant to any issue, saying I think the GUM is wrong becasue they killed my sacred ox, is irrelevant to the issue. Unfortunately it’s the standard of argument I’ve come to expect here.
bgwxyz certainly worships the GUM.
“The truth comes out.”
You mean that truth I’ve been telling you about since day 1, however many decades ago that was. I don’t like significant figure rules.
To be clear, for the most part I’ve nothing against them. When used properly they are quite useful. What I don’t like is the very simplistic rules that seem to be taught at an early stage, and are then treated as gospel, by students who don’t learn better.
But what I particularly don’t, is this completely arbitrary rule about averaging. The rule is introduced uniquely to deal with averages, even though it goes against all the other rules, which is a big clue that there is no mathematical basis for these rules.
“One, so that people attempting to replicate an experiment would get similar results.”
That seems like a bad reason. If you get different results by using a more refined instrument, isn’t that an important thing to know?
“Two, people replicating experiments would not think they needed to purchase better instruments with better resolution in orfer to achieve similar results.”
See my point above. The point of replicating experiments isn’t to confirm them. It’s to see if you can falsify them. Being able to get a significantly different result just by using better methods and devices is good as it falsifies the original experiment. The result was down to poor measuring devises.
Significant digits put your GAT hoax numbers to the lie, so you have to “dislike” them.
You NAILED it!
“””””But what I particularly don’t, is this completely arbitrary rule about averaging. The rule is introduced uniquely to deal with averages, even though it goes against all the other rules, which is a big clue that there is no mathematical basis for these rules.””””””
The “rule” is not about averaging. They are about maintaining the amount of information conveyed in a measurement, neither adding or subtracting to the original quantity of information.
How would you develop a rule that prevents someone measuring a measurand to the nearest inch but by averaging multiple measurements, reporting the average to 1/100th of an inch and the uncertainty to 1/1000th of an inch?
What rule would you recommend when adding a measurement with units resolution to one with a 1/10th resolution? Would you allow adding resolution to the units measurement?
“The “rule” is not about averaging.”
The one I’m talking about is. The one that needs to be stated as the usual rules would result in too many digits in an average.
“They are about maintaining the amount of information conveyed in a measurement, neither adding or subtracting to the original quantity of information.”
And as I keep saying they don’t do that.
“How would you develop a rule that prevents someone measuring a measurand to the nearest inch but by averaging multiple measurements, reporting the average to 1/100th of an inch and the uncertainty to 1/1000th of an inch?”
Why is the question, not how. How’s very easy, you just say “though must round any average to the same digits as thine measurements, regardless of the uncertainty of the average”.
“What rule would you recommend when adding a measurement with units resolution to one with a 1/10th resolution?”
The ones described in all the books about propagating uncertainties. If you don’t know the uncertainty of your two measurements, but have to estimate it from the number of digits, you could follow Taylor’s suggestion and say one uncertainty is ±1, and the other is ±0.1. Then the uncertainty of the two added to is √(1² + 0.1²) ~= 1. This is the general rule of adding uncertainties in quadrature the smaller one tends to vanish, so you only really need to worry about the bigger ones.
Now, what is the uncertainty if you add 100 things, each measured with a “resolution” of 0.1?
Your rules would say that as all have a 1/10th digit the anser can also have a 1/10th digit, implying the uncertainty of the sum is the same as the individual values. Whereas the better rules would say that the uncertainty increases with the square root of the number of things being added, so you only really know the units figure.
As I see it, the rules or guidelines are fine for what they are, just simple rules to be used in simple circumstances, but they don’t scale. In the class room you are only adding up or taking the average of a few things. The rules don’t help with knowing what the uncertainty of the average of 10000 things are. They don’t even consider issues like sampling, or non independent errors.
Nice word salad, blob should be proud.
Once again you show that you live in an alternate universe from the rest of us.
“The point of replicating experiments isn’t to confirm them.”
That’s why we now know the speed of light to a much higher resolution than we did a century ago? Someone falsified the speed of light?
My virology PhD son would disagree with you. They do experiments all the time to confirm the results that other’s experiments produce. Not to falsify anything but to better understand the processes involved in the result! if this drug does this in this strain of mouse does it to the same in a different strain? That isn’t falsifying *anything* having to do with the first experiment.
If it’s a repeating decimal then you have infinite precision according to bellman’s theory.
Of course he’s going to say you have to pick some point at which to truncate it. But he doesn’t want to have to use the significant digit rules to determine where to truncate. Those “rules” are too limiting for a man of his talents with statistics.
“If it’s a repeating decimal then you have infinite precision according to bellman’s theory.”
If you are talking about a constant, yes, all constants rational or irrational have infinite precision. That’s got nothing to do with the uncertainty of a sample mean. The precision of that is determined by the SEM.
“But he doesn’t want to have to use the significant digit rules to determine where to truncate.”
If by SF rules, you mean the ones presented in the GUM or Bevington or Taylor, I do want to use them. If you mean the simplistic ones that used as an approximation for actual uncertainty rules – as I may have said, I’m not a fan. But they are probably adequate for most purposes they are designed for. Just become senseless when people start insisting they mean that global annual average temperatures have to be rounded to the nearest degree.
“If you are talking about a constant, yes, all constants rational or irrational have infinite precision. That’s got nothing to do with the uncertainty of a sample mean. The precision of that is determined by the SEM.”
An average of a population can be a repeating decimal. Is it also infinitely precise? The SEM is meaningless if you know the population so it’s of no help at all in such a case.
“If by SF rules, you mean the ones presented in the GUM or Bevington or Taylor, I do want to use them. If you mean the simplistic ones that used as an approximation for actual uncertainty rules”
Both Taylor and Bevington reproduce the “simple” rules in their books! Word for word. So you can either use them or you can say they are wrong and you do *NOT* want to use them. But you can’t have it both ways!
You’ve just proven once again that you are a cherry-picker. You’ve never actually read either Taylor or Bevington for meaning on anything. You just cherry-pick things, take them out of context, and then complain when caught doing so.
He can’t even be bothered to read the terminology sections in the GUM!
“An average of a population can be a repeating decimal. Is it also infinitely precise? ”
Obviously. That’s the point of the population mean. It’s the thing it is. It’s infinity precise because if you had a god like ability to actually know it, you would always get the same answer.
“The SEM is meaningless if you know the population so it’s of no help at all in such a case.”
Well yes. If you know the exact value of something there’s little point in estimating it.
“Both Taylor and Bevington reproduce the “simple” rules in their books! Word for word. ”
Then quote the parts where they say word for word that an average cannot be reported to more digits than the measurements. All you ever do is point out that they say what I agree with, that the uncertainty should define how many digits you report. Not the digits of your measurements, the digits of the uncertainty,
I am *NOT* your slave! I don’t exist to type in excerpts from the books for you to ignore on here and never read. I’ve done it often enough to know that you will *NOT* read them at all but will just continue with your asinine assertions.
I’ve given you the references from both Taylor and Bevington.
Taylor: Section 2.2
Bevington: Pages 1-6 (the very START of his book)
It is up to you to go read them and understand them We all know you won’t do it. We all know it would put the lie to your assertion so you will simply ignore what they say in their books.
It’s up to *YOU* to prove they don’t discuss the reason for significant digits in metrology.
“It’s up to *YOU* to prove they don’t discuss the reason for significant digits in metrology.”
This is idiotic.
You claim that there is a rule that must be followed in all cases when taking an average, that the number of significant digits of the mean must be the same as that of the individual measurements.
I point out that if this is such an important rule it’s odd that none of the books on error analysis actually state this rule, nor does the very document dedicated to explaining how to express uncertainty.
You fail to show that any of them do, claiming it’s too much effort, then say it’s up to me to prove they don’t. How? By publishing the entire book?
“Bevington: Pages 1-6 (the very START of his book)”
He mentions significant figures on pages 4 and 5. He says what I keep saying, you have to let the uncertainty define the number of digits.
On page 54, Example 4.1, about the standard error of the mean, he has an experiment with a ball and a timer. The uncertainty of the timer is said to be ±0.020 s. From Example 1.2 and Figure 1.2, we can see that the measurements are reported to 2 decimal places. He makes 50 timings and takes the average at 0.635s with a standard deviation of 0.020s. The standard error of the mean is 0.020 / √50 = 0.0028. He quotes the result as 0.635 ± 0.003 s.
The uncertainty of the mean is less than the uncertainty of the individual measurements, and can be quoted to an extra decimal place, along with the uncertainty.
“Taylor: Section 2.2”
From that section:
Again, zero mentions of this special rule for averages.
And as I keep pointing out and you ignore, or do your usual special pleading, at least two exercises explicitly point out that they demonstrate that the result of an average can have more digits than the individual measurements.
You’ve already replied to this comment, but you just accused me of living in an alternative universe.
Exercise 4.15.
My emphasis.
4.17
Answer to 4.17
This is stupid. We have providd you with web page after web page from unversity labs detailing the fact that an average can not exceed the resolution of the measured quantities. This is formalized in Significant Digit rules that everone who takes physical lab classes must follow.
If you can find a lab instruction, book, or text that lets you do as you claim, POST IT!
Listen carefully. I know you have never been responsible for quality measusurements by your silly posts. If you would have taken this type of averaging to a boss, the first thing they would have asked is when did we buy new high precision measuring equipment. When you told him/her that we didn’t, we just started averaging to more decimal points, you would have been fired on the spot.
He ran away from the terminology quiz, I wonder why.
“I wonder why”
Maybe it’s becasue I’m too busy today to respond to every 10000 word essay.
Normally you whine about how short my posts are, now you whine about how long they are.
Hypocrite.
I point out your posts are short and devoid of content. I’m pointing out that the long posts, not yours, are very long and take a very long time to reply to.
And it’s going to take even longer now, as I’ve just accidentally deleted a too long reply to Jim’s magnum opus.
And so whinerman … whines.
And so the troll, trolls, even as he demands I answer every question posed.
Also, the troll says
You call me a troll — hilarious!
Go talk with the Forest Dept. this technical stuff ain’t working out for you.
N.B.: in this context, “troll” refers to anywho who dares to not give whinerman the respect he thinks he deserves.
“N.B.: in this context, “troll” refers to anywho who dares to not give whinerman the respect he thinks he deserves.”
No, troll in this context means someone who posts 50 odd messages a day, consisting of nothing but one line sneering insults. Someone who even as he objects to being called a troll will call me “whinerman”.
Poor whinerman, so abused.
“We have providd you with web page after web page from unversity labs detailing the fact that an average can not exceed the resolution of the measured quantities.”
Stating something as a fact is not detailing it. None of the pages you’ve shown actually deliver any kind of evidence for why you can’t do it. It’s just “it doesn’t feel right” or “becasue we say so”.
And don’t you think it’s odd that you keep relying on web pages for university labs as authorities, rather than any of the ones actually describing uncertainty and measurement, including the GUM, VIM or NIST. If these rules are so important, why do none of the documents tasked with explaining how to report uncertainty mention them?
“If you can find a lab instruction, book, or text that lets you do as you claim, POST IT!”
Taylor (2.9)
(Note Taylor want the uncertainty to usually be stated to 1 significant figure unless the first is a 1 in which case he says to use 2)
Section 3.4 and the infamous 200 sheets of paper example. Thickness of 200 sheets measured with an uncertainty of ±0.1″. Thickness of 1 sheet calculated by dividing by 200 is reported as 0.0065 ± 0.0005″
You can scream, correctly that that is not strictly an average. But you still have to explain why it’s possible for one sheet of paper to be reported to 2 digits more than the original measurement, and then why that isn’t acceptable when reporting an actual average.
Exercise 4.15.
My emphasis.
4.17
Answer to 4.17
“Stating something as a fact is not detailing it. None of the pages you’ve shown actually deliver any kind of evidence for why you can’t do it. It’s just “it doesn’t feel right” or “becasue we say so”.”
You simply don’t get it. It’s so that people reading your result won’t believe you are making measurements at a higher resolution than you actually have!
How hard of a concept is that to grasp? It’s not a feeling or “because we say so”. It’s because it has real world consequences if people think, for instance, that the manufacturing tolerances in a product is greater than it actually is! It could actually cause death and destruction in the real world, like a bridge collapse because someone thought you measured shear strength to a precision that was past your ability to actually discern!
But we all know you live in an alternate universe. I guess death and destruction don’t mean anything there.
His latest line about the GUM being the product of “engineers”-only is beyond incredible.
Could someone get a “sense of humor” implant to karlo with due haste.
Literally all I said was that “engineers” could use what ever term they wanted. This minor joke is apparently regarded now as some blasphemy against the sacred GUM, because apparently it implies that that the GUM was written by engineers and not something that fell from the heavens on a golden cloud.
Oyvey! He tell yok-ah, Ho ho ho!
Saturday Night Live needs scab joke writers.
And again, just who do you think wrote the GUM?
Getting back to your original example, why is the mean of {1, 2, 3, 4, 5, 6} more precise than the mean of {2, 4, 6, 8, 10, 12}?
It isn’t. Both are equally precise, at least in the measurement definition of the word. If you just mean in the sense if how many digits you are giving, again it’s just a matter of how you write it or which base you use. In this case both are exact numbers so have an infinite number of significant figures. 3.5 or 7.0, or 3.500 and 7.000. No difference.
So 3.5 and 7 are to the same precision?
There’s no measurement involved in discrete values, just a count (cue muppet with cape and pointy teeth). A mean isn’t a measurement, either. It’s a calculated statistical measure of centrality
“So 3.5 and 7 are to the same precision?”
In this case yes. To be clear I’m talking about the standard VIM definition of precision. The variability in repeated measurments, or some such. But I expect this is getting confused with the traditional meaning of the number of decimal places in an argument, which is more to do with resolution.
BZZZZZT.
“standard VIM definition of precision” — which you also think contains mistakes and errors!
Which mistakes and errors do you imagine I’m claiming the VIM makes?
Get some clues PDQ, you need help.
It was YOU who claimed the terminology in the GUM is wrong.
Backpedaling now?
The VIM is not the one claiming the term Standard Error is incorrect.
Duh!
Go read what I spoon fed you, dolt.
The VIM 2021 doesn’t eveñ mention “Standard Error”! No wonder it doesn’t call it incorrect!
Yes, that’s the point. I was accused of disagreeing with the VIM.
No you were not; if you had eyes to read you might have realized thus, prior to taking another leap into the oily blackness.
You’ll never convince bellman of this. To him the average *is* a measurment.
I keep saying I don’t care one way or another if you define it as a measurement or not. All I’ve said is that if you insist it isn’t a measurement, then trying to apply the concept of measurement uncertainty to it is a waste of time.
Why do you insist on never reading what people say? The mean uncertainty in an instrument type because of design (an average) *is* useful. But it is a *factor* in the overall uncertainty estimate, not a measurement of a measurand.
You simply refuse to understand the difference between a measurement and a statistical treatment of a set of measurements.
I can go out and measure the amount of wear left on 50 rear tractor tires. I have a set of 50 masurements. I can treate those statistically and find a mean and standard deviation. I can even calculate an SEM for that set of measurements.
Yet the mean is NOT a measurement. The mean is simply nothing more than a description of the central tendency of the set of measurements. Can this mean have an uncertainty? Certainly. Each of my measurements will have an uncertainty consisting of several different things.
The mean of a set of temperature measurements from different devices is nothing more than a set of tire wear measurements. It is not a measurement in and of itself. The mean is simply the value of the central tendency.
Yet the mean is calculated from values that do have uncertainty. That uncertainty must be propagated into the calculated mean. You can’t just say there is no uncertainty because it is not a measurment in and of itself.
Succinct and to the point. Of course it will go right over bellman’s head.
“You simply refuse to understand the difference between a measurement and a statistical treatment of a set of measurements.”
Quite the reverse. You are the ones claiming that a mean has to be treated like a measurement. You insist that the precision of a mean can only be the same as any individual measurements,. I’m saying the mean can have a higher precision becasue it isn’t a measurement as such.
“That uncertainty must be propagated into the calculated mean.”
Nobody is saying that isn’t the case. What the argument is about what that propagation is. With some saying it’s the same as the sum and therefore grows with sample size. Others saying it is the same as the average uncertainty and therefore stays the same with increasing sample size, and the rest (including me) saying that in general it reduces with sample size.
But, what we also say is that in general, the uncertainty caused by the random sample from a population is usually much bigger than the measurement uncertainty and makes it largely irrelevant.
Count Count!
It’s Count von Count, you heathen.
You just asserted that every physical measurement lab using Significant Figure Rules don’t know what they are doing! Not physics lab. Not chemistry labs. Not electronic labs. Not commercial labs. And, not even NIST.
Congratulations.
I’ve been telling you for years I don’t agree with the “rules” as a set of arbitrary commandments. They may well have their place in those sorts of environments, but if they are using the rules as a means of figuring out the uncertainty in situations involving hundreds or thousands of measurements – then yes I think they are wrong. They need to listen to statisticians or metrologists, as appropriate, especially if talking about averaging a large sample.
“And, not even NIST.”
Could you point me to the NIST list of rules for significant digits? All I could find was
https://www.nist.gov/system/files/documents/2019/05/14/glp-9-rounding-20190506.pdf
which makes no mention of the rules you keep going on about, just the usual round the uncertainty to 2 significant digits and report the value to the same order.
“You take a document that talks about repeatability in a laboratory setting, but then use it to make claims about statistical field experiments.”
Measurements are measurements be they in a lab or in the field. The difference is that in the lab you can many times, not always but many times, make multiple measurements of the same thing. In the field, you get one chance at measuring temperature – no multiple measurements of the same thing.
“it’s only talking about how to get a better estimate of the measurement of a single thing by repeated measurements.”
Correct. Something you seem to ignore when talking about the global average temperature calcultion.
“Yet somehow draw the conclusion it’s saying the idea of a confidence interval isn’t even applicable without repeated measurements.”
It isn’t applicable when you are measuring different things! The average of different things is nothing but a mathematical calculation leading to no actual knowledge of the set of different things.
The average height calculated from a set consisting of the heights of Shetland ponies and Arabian horses is useless. It doesn’t exist. You have a bi-modal distribution. It’s no different with temperatures in the SH combined with temperatures in the NH – a multi-modal distribution whose average value is meaningless physically. Each have different variances as well as values making the use of anomalies a fool’s errand.
“Which particular range are you talking about?”
It’s pretty obvious what they are talking about. It’s your reading ability that is lacking. It is the range of values that determines the uncertainty. Did you not look at the table at all? Probably not, you never read for meaning, only for cherry-picking possibilities.
“You realize this is exactly what I’ve been trying to explain to you. The SEM is directly related to the SD of a data set.”
Malarky! The SEM is the SD of the SAMPLE MEANS. That is *NOT* the variance or SD of the population. You keep on making the mistake of equating the two no matter how often it is explained to you!
” It’s just you keep insisting it’s impossible to derive the uncertainty of a mean from just one sample.”
That’s because the SEM is the SD of sample MEANS. If you only have one sample then the CLT does not guarantee that it is Gaussian at all. *YOU* just assume that one sample can perfectly describe the population. The only way to calculate the SEM is to have multiple samples – then the CLT comes into play.
Really? You think the resolution in the thermometers means there is no difference in the measurements”
“You honestly do not understand Pat’s paper at all.” He has tried to explain it to you multiple times and you just stubbornly refuse to listen. Resolution *is* a factor in the repeatability of measurements. The problem is that you are never measuring the same thing when it comes to temperature measuring stations in the field.
“That all temperatures recorded anywhere and anytime are always the same, becasue the resolution of the LiG can’t detect the difference between 10°C and 20°C?”
All you are doing is highlighting your lack of understanding concerning Pat’s paper. The fact that a measuring device has a variation detection limit does *NOT* mean they all similar devices read the same. I have no idea where you got this idea.
“What (3) is saying is what I keep saying – that repeated measurements of the same thing, will nopt help you if you get the same reading each time.”
MALARKY! Field temperature measurements are *NOT* measurements of the same thing. And each single measurement of different things will be biased (systematic bias) by its detection limit.
Why do you *INSIST* on looking at things as always being multiple measurements of the same thing? It’s obvious that your belief that all uncertainty is random, Gaussian, and cancels drives such a worldview and you seem to be incapable of abandoning it.
Detection limits *are* a factor in multiple measurements of the same thing, even in the lab. Detection limits are derived from the physical limits of the measuring device. They have to be included in uncertainty calculations. You DO NOT KNOW what you do not know. You seem to be unable to grasp the concept of DO NOT KNOW. It biases your entire view of uncertainty.
“The average height calculated from a set consisting of the heights of Shetland ponies and Arabian horses is useless.”
Then stop doing it. Honestly, you can come up with useless calculations all day. It says nothing about the useful calculations.
“It’s pretty obvious what they are talking about.”
I asked what you were talking about, not them. As I said, they use range in two different senses.
“It is the range of values that determines the uncertainty.”
So the standard deviation then.
“Did you not look at the table at all?”
My mistake. You mean the range of the data set. Not sure why you find that impressive though. It’s just making a less exact estimate of the SEM without bothering to calculate the standard deviation.
They say calculating the SD can be tedious, which may be true if you are doing it by hand. But it’s trivial with a computer. I’m just surprised you want to use this quick and dirty approach, when before you were insisting the only way to get the SEM was to take many separate samples.
“The SEM is the SD of the SAMPLE MEANS. That is *NOT* the variance or SD of the population”
You missed the bit where they take the SD and divide by root N.
“My mistake. You mean the range of the data set. Not sure why you find that impressive though. It’s just making a less exact estimate of the SEM without bothering to calculate the standard deviation.”
You *still* don’t get it. The SEM is *NOT* the uncertainty of a measurement. It is how precisely you have identified the mean of a set of data. You can can make the SEM vanishingly small and *still* not know the accuracy of the average! The accuracy of the measurements is *NOT* the SEM. Write that down 1000 times, Maybe you’ll remember it but I doubt it.
If you have the standard deviation of the population then it means you already know the average of the population so the SEM becomes meaningless.
It’s obvious you didn’t bother to read the paper. It just shows with every post you make. When you can only make one measurement you can estimate the uncertainty associated with it using the table. That has nothing to do with the SEM.
STOP CHERRY-PICKING and actually *study* something for once!
“They say calculating the SD can be tedious, which may be true if you are doing it by hand.”
This is a STUDENT’S handout. You *want* students to learn how to do it manually so they understand what the computer is doing. You are a prime example where that has not happened for you. You’ve not worked out one single exercise in Taylor or Bevington. All you do is cherry-pick stuff you think you can fool people with on your baseless assertions. You can’t really understand the difference between SEM and population accuracy till you’ve done the work! And you’ve never done the work!
He can’t even make it past the terminology, yet now he’s lecturing Pat Frank about being “wrong”.
Another climastrology psychic that “knows” how I “view everything”.
This is lame, even for you, whinerman.
“””””But in my world there is nuance.”””””
A plaintiff lawyer would love for you to use the word “nuance”!
“Well just how did you NUANCE the measurement?”
“Why did you feel the need to NUANCE a measurement?
“Did you insure your NUANCED measurement would not influence the end result such your design would or could be considered negligent?”
“Did you insure your NUANCED measurement would not influence the end result such your that your design would cause the device not to meet specified requirements.”
“Why did you not follow internationally accepted design methods?”
You may think these are made up hypotheticals, but I assure you they happen often. You have obviously never taken any engineering ethics courses, if you just make nuanced decisions based on your gut.
Fortunately I’m not on trial at the moment, and it isn’t a crime to disagree with things said in the GUM.
“You have obviously never taken any engineering ethics courses, if you just make nuanced decisions based on your gut.”
Seriously? You think it’s unethical to use the term “standard error”?
You just made my point! You obviously support making poor people poorer through higher costs of energy.
You have no skin in the game. If you are wrong, so what, right?
“””””You think it’s unethical to use the term “standard error”?”””””
Like Willis, quote my words. If you read my post, I never used the words, “standard error”.
As usual you never understand simple ideas. Nuance is not a term in use in engineering when discussing measurements. Conservatism all the way in values and their uncertainty!
“You obviously support making poor people poorer through higher costs of energy.”
Pathetic. If you can’t argue your case without this level of ad hominem, then I doubt you have much confidence in it.
“Like Willis, quote my words. If you read my post, I never used the words, “standard error”.”
My point was that I’m being attacked for daring to say I disagreed with something in the GUM, at the same time being attacked for worshiping the GUM, when the only thing I’ve disagreed with is them saying it’s incorrect to use the term “standard error”. You now suggest it’s unethical to have a nuanced opinion on the GUM.
“Nuance is not a term in use in engineering when discussing measurements.”
Maybe it should be. It doesn’t make sense to be too certain about uncertainty.
“Conservatism all the way in values and their uncertainty!”
You should resist the temptation to increase the error estimate “just to be sure”.
Quote the words I said, not what you think they mean. I never said “increase the error estimate”.
I said, “Conservatism all the way in values and their uncertainty!”. I don’t know how you misinterpreted that. I can tell you do not have physical lab training.
For your information!
Conservativism in values means properly using Significant Figure rules to insure all people know the resolution of what was measured. It means not “adjusting” data to suit your preconceived ideas and instead calling the data not fit for purpose..
It means not using averages to increase reported resolution. This the most egregious thing I have ever heard. It requires the creation of information that simply does not exist. It is what Pat Frank tried to show. If you have ten numbers whose measurements have one decimal digit, that is all a sum, division, multpication, or subtraction can support. Why? Because any additional decimal digits exist within the uncertainty interval. There is no mathematical operation that can act like a crystal ball to determine what the value should be.
Sometime in the past he claimed to have “spent years in the software industry”, which could mean many different things.
He’s either as clueless and obtuse as the posts indicate, or he’s an agent provocateur for the IPCC clowns.
I have done my share of programming. Fortran, Cobol, Basic, Unix tools, and C.
Programmers are number manipulators. Everything is numbers, even characters. No concept of measurements, they are just another number to be manipulated.
You still believe “the error can’t be that big!”
You look at the words in the GUM but don’t understand what they mean. You can’t even get past the terminology!
Do you just post these randomly. This has nothing to do with my comment. But if you insist
“You still believe “the error can’t be that big!””
If you mean the errors in this paper, I can easily believe it, especially having spent some time conversing with the man.
If you mean the uncertainty intervals in his paper, then yes I find it difficult to believe. They just have no relation to the actual data.
But I wouldn’t want to let you think that is the only reason. I think the assumptions are wrong as well.
“You look at the words in the GUM but don’t understand what they mean. You can’t even get past the terminology!”
What bit of
Do you think I misunderstood?
I think you’re an agent for the IPCC communists.
“They just have no relation to the actual data.”
Really? You have not shown one iota of proof that LIG thermometers do *NOT* suffer from the uncertainties inflicted on the measurements by the apparatus design.
Yet you can sit there and say that they have no relationship to the actual data? That’s not fact based, it is subjective opinion based on a complete lack of knowledge of the subject and biased by cult dogma from the CAGW religion.
All he has is the magic of averages which allows him to see things that aren’t there.
You’re almost getting it. Statistics can be a magical tool, much like a microscope, which allows you to detect things that are there, just invisible to the naked eye.
Your yardstick-micrometer remains a fantasy that exists only inside your skull.
You realize this is a strawman that exists only wherever you keep your brain?
Hardly, you are the pseudoscientist who believes it is possible to increase resolution via the magic of averaging.
Are you backtracking now?
Embrace your identity!
Except a microscope sees things that exist in reality!
Calibration is done in a controlled environment. Constant air temperature, humidity, air pressure, air movement, immersion times, etc. As soon as they are placed in a screen in the environment all those careful conditions go out the window.
That simply must cause some difference in systematic uncertainty.
“You have not shown one iota of proof that LIG thermometers do *NOT* suffer from the uncertainties inflicted on the measurements by the apparatus design.”
I’m not saying anything about that. I’m assuming he’s got that bit correct – it doesn’t seem an unreasonable level of uncertainty to me. The data I’m talking about is the global average anomalies, either for a month or year. That’s where the claimed uncertainties are just reflected in the data. I don’t care how ignorant you are, there is no way it makes sense to claim the interval is 3 – 4°C for an annual global average.
“cult dogma from the CAGW religion.”
You do realise how much of a tell it is when you say things like that?
“The data I’m talking about is the global average anomalies, either for a month or year. That’s where the claimed uncertainties are just reflected in the data. I don’t care how ignorant you are, there is no way it makes sense to claim the interval is 3 – 4°C for an annual global average.”
In other words you believe that all uncertainty cancels and you don’t need to worry about it affecting a calculated average!
When you subtract two anomalies their uncertainties cancel, right?
Yep!
You’re a total idiot if you really believe this (and it isn’t a front for something).
Yep. Within sample (or entire population) vs between samples.
Ambiguity strikes again…
According to https://stats.libretexts.org/Bookshelves/Introductory_Statistics/Introductory_Statistics_(Shafer_and_Zhang)/06%3A_Sampling_Distributions/6.01%3A_The_Mean_and_Standard_Deviation_of_the_Sample_Mean the SEM is the standard deviation of the sample means, hence the intra-sample standard deviation of the value of interest (in this case the mean)
What do you think “standard error” *IS? Even the GUM says the standard deviation of the mean is sometimes incorrectly called the standard error of the mean. And you refuse to admit to this.
And no one has said *anything* about exaggerating the uncertainty interval “just to be sure”. You are delusional! Instrumental uncertainty is *NOT* exaggerating the uncertainty interval!
“What do you think “standard error” *IS?”
Well in Bevington, it’s being used to mean what the GUM would call standard uncertainty.
“Even the GUM says the standard deviation of the mean is sometimes incorrectly called the standard error of the mean. And you refuse to admit to this.”
I admit that’s what the GUM says. I just think they are wrong. They never justify or cite reasons why it’s incorrect to call it the standard error of the mean. They just don’t seem to like it.
But, as I said in another comment, Bevington is not talking about the standard error of the mean here.
“And no one has said *anything* about exaggerating the uncertainty interval “just to be sure”.”
Bevington was the one who said it. “He should resist the temptation to increase this error estimate, “just to be sure.””
I just think it’s an important point. I can’t think why you are so upset about it.
You left out the fact that he was talking about repeated measurements OF THE SAME THING. Which global temperatures are *NOT*.
And Bevington *is* saying that uncertainty is not error. How you get that he is doing so is just beyond belief!
He states in his book: “Error is defined by Webster as “the difference between an observed or calculated value and the true value”. Usually we do not know the “true” value; otherwise there would be no reason for performing the experiment.”
If you don’t know the true value then there is no way to determine *error*. What you *can* specify is the interval within which the true value may lie. That is *NOT* error it is UNCERTAINTY. They are NOT the same thing. One is based on knowing, the other is based on *not* knowing. Universes apart.
You’ve spent over two years trying to refute this simple and basic fact of metrology. “Not knowing” is different from “knowing”. And you simply refuse to understand that.
There is overlap for sure, but they do not use the same data. And in the case of reanalysis they use wildly different observations and methodologies.
UAH does not measure the surface temperature; a fact evidenced by it’s ~263 K baseline.
So, are you implying that the UAH trend is incorrect? I’ve not seen any papers that refute the accuracy of the trend, only that the absolute temperatures don’t agree with surface temps. Do you have a paper that refutes the trend? You should know that NOAA Star now has essentially the same results as UAH.
You are quite simply wrong about this 263K number (probably another meaningless average).
For the dweeb who mindlessly downvoted my comment, here is the UAH baseline temperature distribution for August:
There are only so many temperature measurement stations out there. EVERYONE uses the same data – the data those stations collect.
People may use widely different methodologies but they only have available what the temperature measurement stations collect, at least for the global average temperature. Harvard, MIT, etc don’t each have a network of measurement stations in Brazil, or Vietnam, or even in the US.
Great response about a stunning finding by Dr. Frank. Climate science has been enchanted by statistics without the ability to deal with the underlying issues.
Professor Karl Wunsch “could find no fault with the paper.”
You sure?
“Carl Wunsch commented September 2019
I am listed as a reviewer, but that should not be interpreted as an endorsement of the paper. In the version that I finally agreed to, there were some interesting and useful descriptions of the behavior of climate models run in predictive mode. That is not a justification for concluding the climate signals cannot be detected! In particular, I do not recall the sentence “The unavoidable conclusion is that a temperature signal from anthropogenic CO2 emissions (if any) cannot have been, nor presently can be, evidenced in climate observables.” which I regard as a complete non sequitur and with which I disagree totally.
The published version had numerous additions that did not appear in the last version I saw.
I thought the version I did see raised important questions, rarely discussed, of the presence of both systematic and random walk errors in models run in predictive mode and that some discussion of these issues might be worthwhile.
CW”
Italics mine.
https://pubpeer.com/publications/391B1C150212A84C6051D7A2A7F119
Truth be told, this paper was the peer reviewed equivalent of Michael Jackson baby dangling. Harmless eccentrics might be interesting, even informative. But as David Letterman said of Michael, post child endangerment “Gosh guys. I think Michael might be getting a reputation as an oddball”. Hence the avoidance of both Michael and Dr. Frank above ground, since then….
You always create new ways to demonstrate your ignorance, bob. With a dash of dishonesty on the side.
Guess what this means: “there were some interesting and useful descriptions of the behavior of climate models run in predictive mode.”
That’s Prof. Wunsch agreeing with the analysis: propagating error through air temperature projections.
Doing so provided, “some interesting and useful descriptions of the behavior of climate models run in predictive mode.” Namely that climate models have no predictive value.
“The published version had numerous additions that did not appear in the last version I saw.”
The review continued through a third round, one more after Carl Wunsch had signed off on the manuscript and recommended publication.
Further changes were requested by reviewer #2 in the third round. Of course Prof. Wunsch did not see them, as he had already signed off.
“which I regard as a complete non sequitur and with which I disagree totally.”
In PubPeer comment #7, I pointed out that last sentence was prominent in every single version Prof. Wunsch saw.
But in your commitment to honesty, you missed comment #7, didn’t you bob. Given your level of discernment, one supposes you didn’t look past #5.
v.1 final sentences: “Even advanced climate models exhibit poor energy resolution and very large projection uncertainties. The unavoidable conclusion is that an anthropogenic temperature signal cannot have been, nor presently can be, evidenced in climate observables.”
v.2 final sentences: “Any impact from GHGs will always be lost within the uncertainty interval. Even advanced climate models exhibit poor energy resolution and very large projection uncertainties. The unavoidable conclusion is that an anthropogenic temperature signal cannot have been, nor presently can be, evidenced in climate observables.”
v.3 final sentences: “Any impact from GHGs will always be lost within the uncertainty interval. Even advanced climate models exhibit poor energy resolution and very large projection uncertainties. The unavoidable conclusion is that a temperature signal from anthropogenic CO2 emissions (if any) cannot have been, nor presently can be, evidenced in climate observables.”
Every version. And as the final sentence of the Abstract. Every version of it.
And if Prof. Wunsch, or anyone else, thinks the last sentence is a non-sequitur, given a demonstration of the utter predictive impotence of climate models, then I’d like to know how he, or anyone else, plans to show a causal effect of CO₂ on climate observables.
Your blind oppositional opportunism drops you into a slime pit every time, bob. Where you seem quite at home.
I read all of the comments. Yes, Dr. Wunsch might very well be covering up for his carelessness.
https://www.amazon.com/PROBLEM-SOLVING-FLOW-CHART-Notebook/dp/B09K1HRHZW
But the fact remains that as soon as he realized how badly he effed up he ran away from you.
But I must admit, you’re pretty much hermetically sealed against criticism beyond the inside of your eyelids. Even within this fawning fora you deflect. And you also don’t meet head on (pretty much universal) criticism from outside, save for your self assertions that you “won the argument”. The fact that this paper will be functionally ignored in superterranea will also be poo pooed by you, with you comparing yourself to Einstein, Maxwell. In fact, you deserve your own version of the linked flowchart. It’s supposed to rain here this afternoon and I can work on it for you…
Clown.
“””””which I regard as a complete non sequitur and with which I disagree totally.”””””
You are making an argumentative fallacy in the way you are using this statement.
You fail to include in your response that the statement you reference was not also included in statements from other reviewers, therefore the truth of the statement is in doubt.
You would fail a high school debate by not including that other reviewers accepted the conclusion.
From an accepted source of argument judging.
https://www.grammarly.com/blog/appeal-to-authority-fallacy/
“””””The appeal to authority fallacy is the logical fallacy of saying a claim is true simply because an authority figure made it.”””””
Logical fallacies are pretty much the only ammunition the GAT hoaxers have in the tank.
“But the fact remains that as soon as he realized how badly he effed up he ran away from you.”
Willful dyslexia.
I reiterate Prof. Wunsch: “there were some interesting and useful descriptions of the behavior of climate models run in predictive mode.”
That is Prof. Wunsch agreeing with the analysis: propagating error through air temperature projections. Which demonstrates they’re predictively useless.
He made no mistake he did not run away from his review or from my paper.
Prof. Wunsch merely dissociated himself from, “The unavoidable conclusion is that a temperature signal from anthropogenic CO2 emissions (if any) cannot have been, nor presently can be, evidenced in climate observables.” which appeared in all versions of the manuscript; the last line of the Abstract and the final sentence.
The rest of your comment is a rant of lies — (don’t meet … criticism … self assertions … comparisons with blah, blah) all refutable by mere inspection of the record — and the bitter insults of someone who has lost the debate and is in consequence losing the bit of his mind that might have been rational.
It was already a stretch — you bringing up the old discussion of Propagation… to discredit LiG Met. Truly a fool’s errand.
Profound thanks again to Anthony and Charles for publishing my work and providing a forum for free speech about science.
I am sure that many of us are glad to have had your work drawn to our attention. It appears to be thorough and rigorous and reaches the parts that others fail to reach. Many congratulations.
Ever since I encountered photographic thermometers, designed to measure darkroom chemical temperatures with gradations marked at 0.2C intervals I have wondered whether weather thermometers shouldn’t have been constructed to similar designs. Of course the photographic thermometer covers a much more limited temperature range, so there would have to be several to cover the span of real world temperatures, and the consequences of a thermometer spending time outside its natural recording range might be as bad for accuracy as the defects you have identified.
Thanks for the kind words, Idau.
Back when meteorological stations were designed and installed. no one cared about knowing air temperature to ±0.1 C.
The thermometers are of a special design, that preserve the liquid column at the T_max or T_min.
Your idea is good in principle, but having many small-range accurate thermometers in a Stevenson screen likely would not have worked in practice.
“The lengthened growing season, the revegetation of the far North, and the poleward migration of the northern tree line provide evidence of a warming climate. However, the rate or magnitude of warming since 1850 is not knowable.”
__________________________________________________________________
Also provides evidence of increased atmospheric carbon dioxide. It’s not just temperature and it’s not a problem:
1. More rain is not a problem.
2. Warmer weather is not a problem.
3. More arable land is not a problem.
4. Longer growing seasons is not a problem.
5. CO2 greening of the earth is not a problem.
6. There isn’t any Climate Crisis.
Or
Has it been warmer in the past?
Has it been colder in the past?
Has it been clouder in the past?
Has there been less cloud in the past?
Has there been more rain in the past?
Has there been less rain in the past?
Have there been more hurricanes in the past?
Have there been fewer hurricanes in the past?
Have sealevels been higher in the past?
Have sea levels been lower in the past?
Has the been more atmospheric CO2 in the past?
Has there been less atmospheric CO2 in the past?
With the past starting yesterday the answer to all those questions is yes. The answer to “Is any of this a problem?” is no
Apart from
Has there been less atmospheric CO2 in the past?
Which is problematic
there is a climate emergency cult crisis
All of the climate fanatics like that the temperature data is awful because it allows them to do adjustments to it, adjustments that have the effect of lowering past temps and raising current ones. It’s why we never hear about the USCRN from the climate fanatics because it shows no warming over its history. It’s my belief that at some point NOAA will stop publishing the USCRN data because it doesn’t show what they want it to show.
Thanks for reminding me of USCRN
The paper starts with the usual Pat Frank nonsense:
“The published 95% uncertainty of the global surface air-temperature anomaly (GSATA) record through 1980 is impossibly less than the 2σ = ±0.25 °C lower limit of laboratory resolution of 1 °C/division liquid-in-glass (LiG) thermometers.”
GASTA is not a measured temperature. It is an average. Much information goes into it; many errors cancel. The average is known more accurately than each reading that makes it up.
Years ago I set out here, empirically, how averaging improves precision for the simple task of averaging over a month. You can downgrade instrumental precision by rounding, and still get basically the same result. Here is an extract, covering Melbourne daily max:
The reason is clear if you look at the individual month. The spread of temperatures is far larger than could be explained by instrumental error. It is caused instead by weather. And that is what determines the uncertainty of the monthly average for one place, and even more for the global average. And as more data goes into the average, the precision improves by the (almost) universally accepted accumulation of information. The error is basically sampling error, and it does diminish roughly with √ N. There may be issues of independence, but the cancellation process dominates.
Again, the limit to accuracy of the average is determined by weather sampling issues, not instrumental accuracy.
Nitpick Nick is in panic mode…
Nick, each temperature measurement is of a different thing, not multiple measurements of the same thing. So how do they “average out”?
Because some deviate high and some deviate low, for whatever reason. And when you add them, the deviations cancel.
Liar. The usual 1/root-N mantra of the trendologists.
No, nooo, noooooo, the average makes more sense than each reading making up the average, which is noise. While single observations have uncertainity intervals, averages have error bars for example.
Here is maximum temperature measured at Albany, Western Australia on 4 January 1907: 17.9 +/- 0.3 degC.
So what use is a single number? And why do you use inflammatory, aggressive and ungentlemanly terms such as liar?
Are you so angry that you are bereft of etiquette.
Why not take an ISO 17025 nap and hop-out spreading joy and goodwill on the sunny-side?
Yours sincerely,
Dr Bill Johnston
http://www.bomwatch.com.au
Because Stokes is a liar, by his deceptive quoting of people (among other things).
Seeing you defend Nitpick is no surprise, at this point.
Your obsession with me is noted, get professional help.
Bill thinks all errors are random “noise” and cancel with the magic of averaging, just like all the rest of the trendologist ilk.
Accurate to the nearest half degree means accurate to the nearest half degree. All the measurements are independent, so canceling error is impossible.
You are reifying “temperature” to somehow get accuracy greater than that of the instruments used to make each individual reading.
So how do you explain that example? I took the temperatures accurate to 0.1°C, got monthly means. I then rounded to 1°C, and the means did not vary on that scale. They varied on the scale expected by cancellation, which was σ/sqrt(31), where σ is the sd of the uniform distribution over 1°C.
Nice trick. Temperature varies over the period of a day, a week, a year. What, pray tell, are you measuring more accurately than any instrument used to make any given measurement?
This is just the maximum temperature for each day, and working out the average max for the month.
Then why do we go to the great expense of making instruments with high precision and calibrate them, when, by your claim, all we have to do is make multiple gross measurements from a cheap instrument?
Thanks to Nick’s explanation my tape measure has become a micrometer.
If you want an accurate measure of the place where you are, high precision is great. If you want to know the global average, the obstacle is coverage and sampling error, not local precision.
Because Clyde Spencer, ‘we’ don’t.
How much have you actually contributed to “making instruments with high precision and calibrate them”. Are they “high precision” in the first place?
Most people commenting here have never seen a meteorological thermometer let alone taken regular weather observations using one. So aside from drawing a tribal repose, who of the royal “we”, know anything about measuring weather parameters.
How about a straw-poll of those who have undertaken regular standard weather observations, verses the ‘experts’ who have sat on the sidelines sipping cool-aid and urging the referee, but have never played the game?
Cheers,
Dr Bill Johnston
http://www.bomwatch.com.au
You forgot to whine about ISO 17025 in your latest Bill-rant, Bill.
It sounds like you are saying you believe that technicians reading a thermometer, that don’t have a clue about how it was constructed, would be expected to understand the issues of accuracy related to bore diameter consistency, bulb dimensional stability over time, and the role of surface tension and wettability of the glass bore are better prepared to assess the function than an engineer. I suppose most anything is possible.
How many thermometers have you constructed in your career?
Dear Clyde,
Perhaps you did not hear what I wrote.
You previously said “we go to the great expense of making instruments with high precision and calibrate them” ….
I don’t have to have made any thermometers to highlight that met-thermometers with 1 degC (or 0.5 degC) indices are not “high precision instruments”. If you want a high precision instrument, by all means buy one of those, but don’t buy a met-thermometer. One reason you might want to buy one is to cross-calibrate a met-thermometer in a water-bath, for instance.
I believe, parallax error (reading a thermometer at angle from horizontal), reading the wetted-perimeter (not the meniscus) and unconscious rounding, affect the precision of an estimate to a greater extent than the precision of the thermometer itself.
All the best,
Bill Johnston
“How about a straw-poll of those who have undertaken regular standard weather observations,”
Well, I used to work for a railroad where we recorded the temperatures at every railroad station at 6am, 12 noon, 6pm and midnight. This is done by all railroads. I actually had to observe a thermometer to get my particular station’s readings. Does that count?
Thanks Tom,
This becomes a problem because standard weather observations include Max for the previous day, and overnight Min (observed at 9am). Dry bulb and wet bulb at 9am local time and 3pm (for relative humidity ranges and dew point) rainfall, visibility, wind etc….. This all cobbles together to make predictions and inform the locals how hot it is and to use sun-block and get vaccinated …
However, an airport may be very interested in ambient-T, RH, and wind strength and direction through the day in order to calculate parameters affecting load, speed, takeoff direction and ‘lift’ in relation to runway length … upper-air, fuel consumption .. etc
The railroad may want to predict expansion in lengths of rail as that may affect buckling etc etc. The local wind farm or solar array may want to know something else – wind droughts and hailstorms come to mind …. The local sewerage treatment woks may want to know about evaporation, fermentation …
You get the picture.
In any event, the conditions under which observations are made and the diligence with which they are made are vital to the outcome. More importantly from a climate perspective the consistency of repeated observations through time …
Cheers,
Bill
Thanks, Bill.
How do you know the measured values are a correct indication of the physically true air temperatures, Nick?
He doesn’t know. He doesn’t even mention how he propagated the measurement uncertainties in his comparison! He can’t actually know that the means came out the same.
It is the standard calculation of mean for any numbers. Do you have any example of how “physically true” numbers would behave differently?
If your numbers are meaningless, your mean is meaningless.
Pat, you’ve thumped the table about how uncertainty propagates which is at variance with many years of established scientific practice. But you have never given a verified example of how it works with “true” numbers (or how the calculation knows they are “true”).
Nick you are talking about a mathematically derived probability not error ranges. What you describe works for pure abstract numbers but not for real-world instrument readings. Gain a little understanding for once.
Then tell us. If the methods, equations, and tools that NIST, JCGM, all of the other internationally recognized standards bodies have produced fail in “real-world” scenarios then why does everybody trust them for calibration certification and what methods, equations, and tools are we supposed to use in their place?
You forgot to spam the NIST uncertainty machine link in this post.
HTH
Most of these bodies *specifically* work with multiple measurements of the same thing and assume that generates a random and Gaussian error distribution – just like Bevington does. And then they assume that all that random, Gaussian error cancels out. From there they assume one of two things: 1. the SEM is the uncertainty of the mean or 2. The variation in the stated values is the uncertainty in the mean.
When you have multiple measurements of the different things you simply cannot assume the uncertainty will give you a random, Gaussian distribution. Therefore you cannot justify either of the two assumptions. And, in fact, this is what Pat has proven – the uncertainty is *NOT* random and Gaussian for temperature measurements.
It is only climate alarm cultists that have to fall back on the assumption that all uncertainty is random and Gaussian FOR EVERYTHING. In the real world that just isn’t the case. As Pat said, it is an assumption of convenience to make things *easier*. The *easy* way is seldom the right way.
Because the calibration protocols result in a competitive environment with various calibration labs auditing themselves and others.
The protocols use multiple measurements of the same thing under the same environment.
They don’t go out and take single measurements of different things and apply the average to a single instrument being calibrated.
You *have* lived in your parents basement for too long. Get out into the real world. Go get a job as an apprentice machinist. You’ll find out quickly how measuring different things tells you nothing about the thing you are creating.
Because the bodies are staffed by statisticians and computer modellers who understand the manipulation of abstract numbers but have never taken measurements in the field and do not understand that there is a big difference between abstract numbers and temperature readings.
NIST has never taken measurements in the field? Are you being serious here?
And now we come full circle. NIST and JCGM were the institutions that were going to take down climate science once and for all. Yet here we are with WUWT commenters taking down NIST and JCGM because their methods, equations, and tools are inconsistent with a contrarian publication in a predatory journal.
bgwxyz sure puts his true feathers on display here with this inane, irrational rant.
The accusation that Frank’s work was published in a “predatory journal” is really an implied ad hominem. You are attempting to demean the work not by addressing the substance, but by attacking Frank’s character by suggesting where it was published is of more importance than what he presented. That is like the trolls who refuse to even read anything published here at WUWT.
You have destroyed what little credibility you may have had by demonstrating a lack of objectivity and engaging in what is universally understood to be an act of desperation to win an argument.
Just because MDPI is listed as a predatory journal does not mean Pat Frank’s character is bad. In fact, I’d be willing to bet the vast majority of authors including Frank who publish in MDPI do so in good faith. In fact, there is precedent for describing authors who publish in a predatory journal as being “wronged”, “misled”, and needing “protection”. Don’t hear what I’m not saying. I’m not saying that USC 15 § 45 will be applied by the FTC in this case or not. I don’t know. I’m just saying in the past when they have they have not attacked the character of the authors. In fact, they’ve defended them.
And if pointing out that 1) equation (4) has an arithmetic mistake 2) equations (5) and (6) are not consistent with NIST and JCGM formulas and 3) that no attempt was made to propagate the LiG resolution uncertainty in the spatial domain does not qualify as “addressing the substance” then what would qualify?
Yet YOU tried to beat Pat Frank over the head for using a journal you don’t approve of.
Hypocrite.
Clyde has you pegged.
And if pointing out that 1) equation (4) has an arithmetic mistake
It has a misprint. The result of the calculation is correct.
2) equations (5) and (6) are not consistent with NIST and JCGM formulas
Eqns. 5 and 6 are correct. Your repeated complaints display a fully refractory ignorance.
and 3) that no attempt was made to propagate the LiG resolution uncertainty in the spatial domain
Irrelevant. The paper concerns instrumental performance.
does not qualify as “addressing the substance”
Your 1-3 are insubstantial. Your capacities for willful ignorance and an insistent foolishness are exceptional.
then what would qualify?
An argument from knowledge, of which you are very evidently incapable, bdgwx.
Which is fine. I misprint all of the time. In fact, I’d venture to guess I misprint more than you do. That’s not what I’m concerned about. What I’m concerned about is the calculation in (4) is performed differently than in (5) and (6) which is very odd since all 3 of them are proport to be the uncertainty of a mean just over different time periods.
To be pedantic no. Even with the division typo corrected there is still a mistake. You either needed to have 1.96σ on the LHS or multiply by 2 on the RHS. Again, though, I’m not really concerned with that detail per se. But it does make me seriously question the peer-review process here.
Says who?
Of course it is relevant. You claim to compute the global average temperature uncertainty. That can only be done correctly with the consideration of the grid mesh used to compute the global average temperature or at the very least a recognition of the spatial domain.
“since all 3 of them are proport to be the uncertainty of a mean just over different time periods.”
No, they do not. As already noted to you ad nauseam.
Says me, bdgwx. And for good reason: eqns. 5 & 6 compute the relevant quantities, namely the RMS of uncertainty over the time ranges.
Your continued misrepresentation is, charitably, a persistence in ignorance.
“the global average temperature uncertainty.”
The lower limit of instrumental resolution and accuracy in an air temperature anomaly. Your view is entirely wrong.
If you’re going to boil it down to either believing you or NIST, JCGM, ISO, etc. then can you really blame me if I choose the later?
You’re headed into the cornfield regardless.
What does spatial spread have to do with instrument capability? You aren’t making any sense. If an LIG has a resolution limit in Bangor, Maine it will have the same resolution limit in Phoenix, AZ. What in Pete’s name does geography have on resolution limit?
bg thinks because he can stuff the average formula into GUM 10 and the vaunted NIST uncertainty machine, he is now the world’s foremost expert on the subject.
Let it me known that I think NIST calculates the uncertainty of the output of a measurement model that computes the mean correctly.
Ask me if I care what you believe…
You don’t even know the difference between the SEM and the accuracy of the population mean. So how can you judge whether the NIST is doing what you think it is doing?
Yup, yup, and yup.
They don’t fail, they are just not used correctly.
Quit while you are behind, Stokes, by yapping about true values you are only highlighting your abject ignorance.
karlomonte you need another nap.
b.
Give it up, Bill-clown.
Oh karlomonte, you missed your afternoon nap, yap or whatever
So hard to keep up with yet so far behind.
Hypocrite. Is this what you call “gentlemanly etiquette”?
I’ve propagated errors in my experimental work and put the results in my published papers. I first learned error propagation in Analytical Chemistry and later in Instrumental Methods labs.
…at variance with… You don’t know what you’re talking about, Nick.
You’re right about calculations. They don’t know about true. But true is a non-issue.
Accuracy is the crux. Accuracy statements are qualified. Experimental scientists know the reliability of their data. They propagate the known uncertainties into their final result to convey its reliability to others.
LiG Met. is about accuracy statements. The behavior of numbers reveals nothing about accuracy. Calibration reveals accuracy.
If inaccurate numbers sum to zero, the zero is inaccurate.
A self-obvious truth that somehow seems to always elude the climate alarm cultists.
If, if, if, but how do you personally know numbers are meaningless?
This is about your paper, not putting-down Nick Stokes.
I want to know about your paper and your justifications, not your putting other people down.
Cheers,
Bill Johnston
So go read it, duh.
Where have I put anyone down, Bill.
> Where have I put anyone down
Is that a trick question, Pat?
Pay me 10 bucks per put down and I might find them back for you.
If you have evidence, present it Willard.
What’s in it for me, Pat.
A demonstration that your pejorative inference is honest.
I already have that demonstration ready, Pat, and inference might not mean what you make it mean.
If you want to know about my paper and my justifications, and you don’t find an answer in the paper, then ask a specific question, Bill.
Wrong again.
Nick has obviously no clue when the law of large samples can be used and when it can’t.
You could have ended your sentence after “no clue”.
Figuring out how close you are to the population mean doesn’t tell you how accurate that population mean is.
Why is the difference so hard to understand?
Because they never earned their rifle merit badge as a Boy Scout. When scoring targets, we just assume that the rifle sights can be off by two or three inches at the target (the accuracy of the population mean). But if the scout has a good sight picture, good sight alignment, good hold control, good breath control, can gently squeeze the trigger, and follows through the shot, they will always be able to place 5 shots inside a one inch circle (a very small population mean). As long as a quarter covers all five shots, they pass, even if the quarter is totally outside the black. The scout controls the population mean. The armorer maintaining the rifles controls the accuracy of the population mean when adjusting the sights.
You got it! And the SEM is the same thing as the quarter!
But Pat, this becomes an inflammatory argument – how do you know they are not?
Argue the point not the person making the point.
All the best,
b.
Bill, if someone is speaking from ignorance, it is not a personal insult to point that out.
You might want to follow your own advice and refrain from saying things like, “Oh karlomonte, you missed your afternoon nap, yap or whatever …”
I agree Clyde. Tedious as it is, this thing with karlomonte has a history and seems to have developed a life of its own.
b.
I called you out on your irrational obsession with Jennifer M., and now you have transferred this obsession onto me,
You say “irrational obsession with Jennifer M.”. She was wrong in her assertions, but frustratingly kept on repeating the same things. Somehow that turns into an “irrational obsession”.
Furthermore I prepared a report detailing why and how she was wrong. I followed that up the other day with a second report that showed the protocols I suggested she follow, were robust. Both reports are available here: https://www.bomwatch.com.au/statistical-tests/
I have no obsession with you, dear karlomonte. In fact you say some things that are interesting and detailed and in context quite handy to know. You let yourself down with your extreme language and accusational tones tho.
All the best,
Dr. Bill Johnston
http://www.bomwatch.com.au
The problem is that you ignored the uncertainty in BOTH cases! You simply can’t know if they came out the same or not! What was the variance of each data set? Did you round up, down, or at random. Rounding doesn’t improve uncertainty, it just carries it along!
Again you are dealing with a distribution. Did you check to see if the distribution of that data was normal? Was it the same distribution after rounding.
Let me point out that these are measurements with a certain resolution. Significant digits apply. If you found the mean using data with 1 decimal place was the same as using data with only units, you were lucky or you fiddled the data.
A mean of integer data should be an integer. Then to be equal, the average of data in the tenths would have to come out 20.0.
In your dreams, Nick.
Relative errors and uncertancies always add in a sum or a mean. They never cancel each other except in “cancel culture”.
Then explain why when I enter (x0+x1)/2 into the NIST uncertainty machine the result is an uncertainty that is less than the individual uncertainties of x0 and x1? Is NIST wrong? Is JCGM 100:2008 wrong? Is Bevington section 4 wrong?
Ever hear of garbage-in-garbage-out?
Apparently not.
He *still* believes that the average uncertainty is the uncertainty of the average!
Because when dividing the sum by 2, the quadrature-combined uncertainty is divided by 2 as well, to keep the fractional uncertainty constant.
Suppose each uncertainty is ±0.2. Then the uncertainty in the sum of x_0 + x_1 is sqrt(2x(0.2)^2] = ±0.28.
But when dividing the sum by 2, the uncertainty is divided by 2 as well.
Uncertainty in the mean is then ±0.14.
All that really tells you is how accurately you have calculated the mean. It doesn’t tell you what the accuracy of the mean is.
In addition, why is the uncertainty divided by 2 when dividing the sum by 2? The number “2” has no uncertainty. If you follow the rules for uncertainty, the uncertainty of the average would be the relative uncertainty of the two factors plus the relative uncertainty of the constant (i.e. Zero). You can’t just arbitrarily reduce the uncertainty just because you have found an average. In
Tim, division by 2 keeps the proportionate uncertainty constant,
Ahhh! But that is still the average uncertainty and not the uncertainty of the average.
One thing we can agree on. But it will have effect on Tim. It’s a article of faith for him. You must never reduce an uncertainty, so the uncertainty of an average is equal to the unsertainty if the sum, no matter how illogical this becomes.
Says the man that has never once built a beam to span a foundation, designed a staircase, built a stud wall for a room addition, built a racing engine, used a milling machine or lathe to build even something simple like a bolt, measured crankshaft journals or cylinder bores in an engine, built a self-guiding robot, built an electric signal amplifier, or anything else in the world requiring actual measurements.
Uncertainties add, they always add. You simply cannot reduce them by averaging, not in the real world. You simply cannot assume that all uncertainty is random, Gaussian, and cancels. Without that assumption your whole world falls apart.
Don’t me tell Pat Frank. He’s the one with paper doing what you say is wrong.
Why can’t I get my phone to correct my bad typing. That should have been
Don’t tell me, tell Pat Frank. He’s the one with the paper doing what you say is wrong.
Exactly!
And when you follow that procedure you get σ_T_month_mean = 0.035 C and σ_T_annual_mean = 0.010 C in place of (5) and (6) respectively. You can multiple those by 2 to get 2σ [1].
See the difference. Above you divided by 2 after doing the square root. But in your publication you divide by 30.417 and 12 before doing the square root.
[1] In your (5) and (6) you multiple by 1.96 which is the coverage factor for 95% CI; not 2σ. I’m ignoring that for now since it does not significantly effect the outcome.
The x_0 and x_i are different values, bdgwx. Then taken in an arithmetic average.
Eqns. 5 and 6 are RMS of a constant uncertainty. They’re not arithmetic averages.
It doesn’t matter what x0 and x1 are. What matters is what u(x0) and u(x1) are. And you said “Suppose each uncertainty is ±0.2” so that’s what I entered into the NIST calculator.
I know. They are quadratic averages. Not that it matters because when you take an average of the same value repeatedly you still get that same value as result regardless of whether it is an arithmetic average or a quadratic average. Regardless the uncertainty u((x0+x1)/2) is not either the arithmetic nor quadratic average of u(x0) and u(x1) as is plainly obviously from the result of the NIST calculator.
and is 0.14 not less than 0.2?
Did I dispute that?
Let u(x0) = 0.195, u(x1) = 0.195, …, u(x11) = 0.195.
Let y = (x0 + x1 + … + x11) / 12
Your equation (6) computes 2u(y) = 0.382 C.
NIST computes 2u(y) = 0.113 C.
See the problem?
The problem is, there’s no ‘y’ in eqn. 6.
See your problem?
y is the measurement model that computes the annual average. Equation (6) is said to compute u(y). Refer to JCGM 6:2020 for details on developing measurement models.
No. What specifically about the NIST calculator is a problem?
bdgwx’s problem is that he doesn’t understand using equations.
Let’s take his (x0+x1)/2 as the functional description of a measurement. So let’s make sample 1, x0 = 70, x1 = 71,
Then, q₁ = 71 and q₂ = 72
What is the formula for the mean? How about GUM equation 3.
q̅ = (1/n)Σqₖ
Now, we will put in values.
According to the functional description of (x0 + x1)/2, we get:
q̅ = [(70 + 71) / 2] / 1 = 70.5
Hmmmm? What is the problem here. We only have one item (sample). So we divide by 1, and voila, a simple mean occurs.
Let’s try again with several measurements
q₁ = (1+2+3+4)/4 = 2.5
q₂ = (6+8+10+11)/4 = 8.75
q₃ = (2+6+11+7)/4 = 6.5
q̅₁ = (2.5 + 8.75 + 6.5) / 3 = 5.92
What is?
q̅₂ = 1+2+3+4+6+8+10+11+2+6+11+7) / 12 = 5.92
Why is this?
Because “4” is a common denominator. when you divide by 3, you multiply 4*3 = 12. Consequently you get the mean either way.
Since these are experimental values. The GUM Section 4.2.2 (equation 4) applies. We assume uncertainty is ±0.5.
s²(qₖ) = (1/(n-1))Σ(qⱼ – q̅)²
s²(qₖ) = (1/2)Σ(2,5 – 5.92)² + (8.75 – 5.92)² + (6.5 – 5.92)²
s²(qₖ) = (1/2)(11.7 + 8.0 + 0.34) = 10.02
So now, equation 5 gives:
s²( q̅) = s²(qₖ) / n
s²( q̅) = 10.2 / 3 = 3.34
s( q̅) = √3.34 = ±1.83
You’ve been told why this is multiple times. But it violates your religious beliefs that you can increase resolution and decrease uncertainty by averaging.
Average uncertainty is *NOT* the uncertainty of the average!
Yeah. I know. That’s what I’m trying to tell Pat. (5) and (6) in his publication compute the average of the uncertainty (specifically the quadratic average) and not the uncertainty of the average.
Nobody cares about the average of the uncertainty. It is of little use because it does not tell you how certain you about the average itself.
To be precise here:
average of the uncertainty = Σ[u(x_i), 1, N] / N
uncertainty of the average = u(Σ[x_i, 1, N] / N)
Now can you explain this to Pat?
Nobody cares about the average of the uncertainty.
That’s exactly what I cared about.
The uncertainty is the lower limit of detection and is therefore of constant weight in a daily mean, in a monthly mean, and in a yearly mean. Eqns 5 and 6 are there to show that.
Years ago I collaborated in a project with a guy who didn’t understand the X-ray absorption spectroscopy I used. When we wrote our paper, he kept editing what I wrote into what he thought it should mean. He was wrong and his changes made the paper wrong. I kept having to change it back, til we solved the meaning problem.
You’re doing the same thing, bdgwx. You don’t understand the work. You change it around in your mind to what you think it should mean. Your concept is wrong. And then you address your mistaken understanding.
And since the paper doesn’t match what you think it should be, you think I made a mistake. But I didn’t. You did. But you don’t understand the work, and so you don’t grasp your mistake.
Rinse and repeat. Over and over and over.
Then you need to tell people that your publication is only determining the quadratic average of the individual temperature measurements over a period period of time and not the uncertainty of the average itself.
You even use the phrase “The uncertainty in Tmean for an average month (30.417 days) is the RMS of the daily means“. In other words, you are labeling (5) as being u(Σ[Tavg_i, 1, N] / N) but calculating Σ[u(Tavg_i)^2, 1, N] / N instead.
“…you need to tell people that your publication…”
That part of the paper is about the lower limit of resolution, applicable to all 1C/division meteorological LiG thermometers.
This is exactly what I tell people. But you obviously didn’t grasp it.
You called it “The uncertainty in Tmean”, but all you did is calculate the quadratic average of the individual measurement uncertainties. That is not the uncertainty in Tmean.
It is when the uncertainty entering the mean is restricted to the lower limit resolution of LiG thermometers.
“…field-conditions lower limit of visually-read resolution limited…“
Who says that? Do you have a reference?
And why does that change the uncertainty of the mean formula?
Can you derive the RMS formula you used in equations (5) and (6) via the law of propagation of uncertainty?
There is nothing to explain. The two are not the same.
u(Σ[x_i, 1, N] / N)
break that down: Σ[x_i, 1, N] / N is the AVERAGE VALUE.
Adding the u in front of it doesn’t tell you anything. You don’t KNOW “u”.
Σ[u(x_i), 1, N] / N does nothing but spread the uncertainty factors evenly across all members in the data set. That doesn’t tell you the uncertainty of the average.
2,3,4 has an average of 3. When you find the average uncertainty you convert that to 3,3,3. So what? The uncertainty of the average is not the average uncertainty, it is the uncertainty propagated from each member onto the average. The propagated uncertainty of 2,3,4 using root-sum-square is 5.4. The propagated uncertainty of 3,3,3, is 5.2. You don’t even get the same answer for the propagated uncertainty! So how can the uncertainty of the average be the same as the average uncertainty?
An example measurement:
A K-type thermocouple is immersed in a constant-temperature oil bath and connected to an oscilloscope capable of 2 Gsamples/s. The oscilloscope is programmed to digitize the TC voltage for one second, at the max sampling rate. Then the voltages are downloaded and the mean is calculated.
Blindly stuff the mean formula into the GUM uncertainty propagation we get:
u(Tbar) = u(T_i) / sqrt(2×10^9) = u(T_i) / 45000
Vanishingly small, and this u(Tbar) goes to zero in the limit as the number of samples goes to infinity. This is a non-physical result.
Alternatively, the standard devision can be used to calculate the uncertainty (completely in line with the GUM). But this leads to:
u(Tbar) = sigma / 45000
With good instrumentation practices the standard deviation from the oscilloscope is likely pretty small, so this again leads to tiny numbers that make no sense.
Anyone experienced with temperature measurements using thermocouples should immediately see the problem here — no reference junction. Without subtracting the reference junction voltage, the TC voltages are not proportional to absolute temperature.
In other words, this is a relative temperature measurement with an allegedly tiny measurement uncertainty. This makes no sense.
The point is that averaging these billions of data points has not reduced uncertainty, at all. The uncertainty of the sampling must be combined separately with the results from the huge uncertainties associated with the thermocouple itself.
These are NOT reduced by endless averaging.
Using the techniques of the GUM without understanding only leads to nonsense.
Except they are always +/-, never just ++.
I wood have fought 1 -1 = zero or plus three or somfing
What are you talking about?
BUT YOU DON’T KNOW THE MAGNITUDES OF *ANY* OF THE ERRORS THAT YOU THINK CANCEL!
So how can you possible know they cancel? Pat Frank has shown that the temperature distributions uncertainties are NOT NORMAL. Non-normal distribution cannot just be assumed to be random and cancel out!
All you are doing is repeating the unproven meme of the the climate science cult – “ALL ERROR IS RANDOM, GAUSSIAN, AND CANCELS”.
Pat has shown how this meme simply cannot be true. But you, as a true cultist, believe it in the face of actual, physical proof otherwise.
You simply can’t be embarrassed, can you?
No, but he can tell you to go play with the NIST uncertainty machine.
“… when you add them, the deviations cancel.” Wrong Nick – you cannot possibly predict whether the deviations will cancel or multiply – either is entirely possible and completely random. As you add more readings to your average, the instrument and reading errors will make the final result completely worthless. A complete lack of precision and understanding, despite being able to parrot the religious mantra’s word-perfect.
WRONG again.! The deviations do not cancel.
You are applying maths with is basically just WRONG.
The “spot the agenda”- Nick From the “hottest eva” school of “weather=climate” bollox….Meteo France & UK Met office school.
The deniers or eliminators of the MWP or the Roman Periods, which didn’t have thermometers at all…..just a few vineyards close to Scotland.
I guess I prefer to believe funny old historical evidence about green Greenlands and Roman vineyards or non heated villas in England 2000yrs ago eh?
When averaging things, the “reason” really does matter. Errors cannot cancel if the reasons are different (except coincidentally and so not in the long term).
Say one area is unusually cold because it is shaded by a steep hill and local winds run down the hill and through it. That could balance with a sunny place surrounded by warm water.
Now say a cold front comes over.
Clouds block the sun cooling the warm anomaly. And the winds increase, increasing the cold anomaly. Averaging is not balancing anymore.
You cannot average measurements of different things just because the measurements are in the same units. ‘T at place A’ is not the same thing as ‘T at place B’.
Your argument is wrong.
“Errors cannot cancel if the reasons are different”
This is arithmetic. 1-1=0. It doesn’t matter how it got to be 1.
This is reality, not pure mathematics.
If all 1s become 2s and all -1s stay the same you are arguing that:
2-1=0
And you can’t say your snapshot measurements respond in the same way if you don’t know what’s going on. It does matter how it got to be 1 if you are going to average it with other measurements. If you don’t know, you have no reason to say it’s the same sort of measurement.
.
Simply put, you can’t average apples and oranges.
“This is reality, not pure mathematics.”
It’s just the way numbers work. If you add them, and they vary in sign, there will be cancellation.
You may have some argument for saying you shouldn’t add them. But if added, that is what will happen.
Nope! You can’t just consider sign, you also have to consider magnitude. There may be SOME cancellation but you simply cannot just assume TOTAL cancellation unless you actually know the distribution of the errors AND their magnitude. Both of which are impossible to know when you have a multiplicity of measurements of different things made by different instruments.
It’s what root-sum-square addition is used for. PARTIAL cancellation.
Nope. The population growth of swans in Denmark is highly correlated the the annual human birthrate in Denmark. They each have their own uncertainty interval. You can’t simply subtract the two and say there is cancellation of the uncertainty intervals. They are not physically related and so the uncertainty intervals don’t actually cancel. Statisticians would tell that they do but they only look at the numbers, not at the causal relationship.
Temperatures taken at different places and times *are* just like the swans and babies. They are DIFFERENT things. No causal relationship. The uncertainty in Station A doesn’t cancel the uncertainty at Station B. When you combine the two the total uncertainty goes UP, not down.
You have no reference to back this up. This is an unsupported assertion and is not true. You are trying, probably unknowingly, to conflate experimental data variance with measurement uncertainty. You can’t do that. Dr. Frank has shown that isn’t true. Did you even read his summary above?
You should know that “adding” measurements requires adding uncertainties. You have no measurement experience or you wouldn’t assert this.
You still didn’t get what Tom Halla was saying!
And, it isn’t just the temperature that varies with every unique parcel of air that passes a weather station. The humidity, pressure, and density varies with each parcel.
During the Friday before the 3-day 4th of July holiday in 1968, I, my wife, and my younger brother were camping along the North Fork of the American River, not far from Auburn. It was one of the hottest days I have ever experienced. At one point in the afternoon I went up to my pickup truck and turned on the radio to see if I could get a weather report. I got a station in Sonora (CA) that reported 120 deg F. Sonora is basically along the same longitude, and during such heat spells, the isotherms tend to be parallel to the Sierra Nevada, which runs north-south. All three of us spent the entire day immersed in the water to keep cool. Occasionally, a gust of wind would come up the canyon, and all three of us would start coughing. It was like being in a steam sauna. Apparently, the hot air gust was flash evaporating water from the surface of the river. I doubt that such circumstances of rapid change in humidity is recorded by weather stations. However, it is important because it controls the heat index and reflects greater enthalpy of the air parcels. The air parcels are all part of the global atmosphere, but they have physical properties that vary over short time scales, and are thus different enough that they shouldn’t be considered to be the same thing, just variants of the same class.
Exactly. Temperature isn’t a measure of the heat content of air. So, an average global surface temperature can tell us nothing about the total heat content of the atmosphere. At best it’s a weak proxy for heat content. Increased heat content is predicted with increased CO2. Even if we had a hundred-year record of temperatures measured every second at a billion equally spaced locations on the surface of the earth, that would not tell us if the heat content of the atmosphere had increase. Heat content is enthalpy (heat units per mass unit of air and its associate water vapor). A 110 °F day in Phoenix and have the same heat content as an 80 °F day in Houston, just 1000 miles away. The temperature difference is 30 °F the heat contents is the same.
More formally, temperature is a proxy for the more important property of enthalpy; however, temperature has a low correlation with enthalpy in humid climate zones. Therefore, the difference between enthalpy and the estimate of the enthalpy of air parcels varies with the climate zone and season.
Pat Frank also shows that the error distribution is not normal. This means, even if errors (on many measurements of the same thing, not many different measurements) would cancel out in the case of a randomly distributed error, for temperature measurements, this is not the case.
Errors cancel even if not normally distributed (which is not the same as random).
One can’t assume errors cancel when one doesn’t know the errors. To do so is self-serving.
You took the words right out of my mouth. In a skewed distribution there is no guarantee that the left side will cancel out the right side. The probability distribution is basically a histogram telling how often a value occurs. Unless matching values for appearance occur on both sides of the mean there is no guarantee of cancellation. It’s why the median in a skewed distribution isn’t the same as the mean.
You would think Stokes would know this but he’s blinded by his cultish beliefs.
No. Histograms have fixed bins, it is a probability destiny function.
Skew is also measurable, so too its effect on the mean/median. While rainfall is not, daily temperature is mostly robust to skew. (Compare annual mean-T with median-T for instance).
b.
If the distribution is unsymmetrical, they cannot completely cancel. Therefore, there is a residual bias. That could be small or large. It is impossible to say without knowing the distribution. You are engaging in wishful thinking.
They won’t completely cancel, whatever the distribution. But the mean of the errors will tend to zero, whatever the distribution. That is what the Law of Large Numbers says. OK, exceptions for distributions with tails too fat to define a second moment.
When measurements are inaccurate, a zero mean is inaccurate.
Provide examples.
b.
For the several hundred sites in Australia that I have examined, daily data distributions are bimodal. However, removing the day-of-year cycle may result in long-tailed distributions but with no change in the mean or median.
Multiple analyses show that the change from 230-litre Stevenson screens to 60-litre screens increased the likelihood of upper-range extremes.
Something for another day at this point.
Cheers,
Dr Bill Johnston
http://www.bomwatch.com.au
Visualize a skewed probability distribution function with a median of zero. Add all the numbers constituting the left-hand side to the numbers on the right-hand side. The sum will not be zero.
Yes Clyde,
I routinely use a variety of normality/distribution tests as well as histograms and Q-Q plots.
Cheers,
Bill
It doesn’t matter. The errors cancel all the same as long as they are random. You can actually prove this out for yourself with the NIST uncertainty machine.
An error that is for example not randomly distributed is a systematic error. These don’t cancel out.
It depends on the measurement model. If the measurement model is y = a – b then the systematic error cancels. If the measurement model is y = a + b then the systematic error doubles.
The general rule is that when the product of all partial derivatives of the measurement model is less than zero then systematic error is reduced and vice versa when positive. Refer to JCGM 100:2008 equation 16 for why this happens.
You can prove this out for yourself with the NIST uncertainty machine using the correlation feature.
Being “reduced” is not the same as “cancelled.” Unless one can specify how much the systematic error is reduced, then one is dealing with an specified uncertainty and cannot claim to know what the uncertainty is. The best one can do is to bracket the uncertainty with inequalities.
It is when the measurement model is y = a – b. This is because the product of the partial derivatives ∂y/∂a * ∂y/∂b = -1.
An example where you only get partial cancellation (or only a reduction) is y = a – b/2. In that case the product of the partial derivatives ∂y/∂a * ∂y/∂b = -0.5.
That’s what the law of propagation of uncertainty does. When the product of the partial derivatives is -1 all of the systematic uncertainty is cancelled. When it is between -1 and 0 then only some of it is cancelled.
More spam. YOU go play with your machine.
No, the systematic error does *NOT* cancel out. This has been pointed out to you over and over ad infinitum.
1+1+1+ 1 = 4. Add a systematic bias to each and you can get 2+2+2+2=8. Where is the cancellation? The average of the first is 1. The average of the second is 2. Where is the cancellation of the systematic error?
Stop repeating cult dogma and think for a minute!
No it does not. Systematic error is not additive in time, it adds to each value and therefore to a measurable step-change in the mean.
As interesting and uninformative as this discussion has been, it is snooze time for me.
Good night (OZ time),
Bill
Just seen this. It needs to be corrected in case a schoolchold sees it uncritically and fails their 11+.
Wrong. Whether the model is {y = a – b} or {y = a + b} the errors add. They don’t cancel.
Let us demonstrate this.
Consider the effect of the units you are working in. If we create a new unit, exactly the same scale except zero is more than b lower, then -b becoems a +b. How can arbitrarily changing the units (not the scale,or the measurement, just the zero) change whether we know more or less about the accuracy of the measurement?
It can’t.
If we don’t know one thing, we cannot take away what we don’t know from another thing we don’t know and get no things we dont know left.
We just have two bits we don’t know. The errors add.
This has been pointed out to them multiple times – but it never sinks in. Just like they believe the SEM somehow describes the accuracy of a population mean.
It’s correct. Systematic error cancels when y = a – b. JCGM 100:2008 equation 16. Assume u(x) = u(a) = u(b) for simplicity. Notice that the uncorrelated term reduces to 2u(x)^2 for both y = a – b and y = a + b. Then notice that the correlated term reduces to -2u(x)^2 for y = a – b and +2u(x)^2 for y = a + b when r(a, b) = 1. You can also verify this with the NIST uncertainty machine.
I don’t know what “zero is more than b lower” means. Can you present a concrete example?
It’s the same situation with C or K. If a = 283.15 ± S K and b = 273.15 ± S K where is S is entirely systematic thus r(a, b) = 1 then y = 10 ± 0 K. Or if we change the units but keep the scaling the same like is the case with C we have a = 10 ± S C and b = 0 ± S C. Then y = 10 ± 0 C. In either case the systematic uncertainty S cancels since r(a, b) = 1. Nothing has changed with combined uncertainty u(y). It’s still 0 as long as S is entirely systematic meaning r(a, b) = 1.
Ah, but that is the power of algebra. We don’t have know what the value of a variable is to know that it can cancel in a subtraction. If there is a systematic effect S then both a and b will have the S effect. Thus y = (a + S) – (b + S) = a – b. Notice that S cancels out and we don’t even need to know what’s value is.
Not when r(a, b) > 0 they don’t. Here is a concrete example using JCGM 100:2008 equation 16.
y = a – b
a = 5
b = 3
u(a) = u(b) = 1
When r(a, b) = 0.00 then y = 2 ± 1.41
When r(a, b) = 0.25 then y = 2 ± 1.22
When r(a, b) = 0.50 then y = 2 ± 1.00
When r(a, b) = 0.75 then y = 2 ± 0.71
When r(a, b) = 1.00 then y = 2 ± 0.00
Note that r is the degree to which u(a) and u(b) correlate. r = 0 means there is no correlation. r = 1 means there is complete correlation. If all of u(a) and u(b) is systematic then r = 1. If none of u(a) and u(b) is systematic then r = 0.
This can be verified with the NIST uncertainty machine.
Liar. The GUM does NOT say this.
Uncertainty INTERVALS simply do not cancel. They do not have a scalar value that can be defined.
The uncertainty interval goes from minimum to maximum value. Those values *add* no matter what. See Taylor, Section 2.5.
for u = p – v
For variable “p” with +/- ẟp where p_max = p_best + ẟp and p_min = p_best – ẟp
for v you get v__max = v_best + ẟv and v_min = v_best – ẟv
When you subtract them the maximum possible uncertainty interval is p_max – v_min (or alternatively v_max – p_min.
THERE IS NO CANCELLATION.
If you think there might be some cancellation then you add the uncertainties in quadrature – BUT THEY STILL ADD!
They add whether you have p + v or p – v.
You *still* have absolutely *NO* understanding of uncertainty. Thinking it has a scalar magnitude that can be subtracted in order to cancel it is just one more piece of religious dogma you refuse to give up on.
Again with the uncertainty machine spam from bee’s wax.
NIST and GUM used to be the authority that would put climate science right. Now they are classed as know-nothings too. The only authority left is Pat Frank.
Idiot.
The LiG detection limits in Table 1 are from NIST.
Who says that the uncertainty in measurements of different things represents a random distribution that is normal? I can guarantee you that the wear in eight pistons in a racing engine is *NOT* normal even though each one represents a random value.
I suggest you read, “Bernoulli’s Fallacy: Statistical Illogic and the Crisis of Modern Science.”
Thanks. I’ll check it out. Does Aubrey Clayton challenge the law of propagation of uncertainty, NIST, JCGM, etc.?
Nick’s analysis requires perfect rounding, which in turn requires perfect instrumental accuracy and infinite precision. Neither is available in the physically real world.
Nick is wrong and should know he’s wrong, because we already had that discussion here on WUWT seven years ago. You cleverly deleted my reply posted to mohyu, remember Nick?
Limits of instrumental resolution mean limits on information. There’s no magicking knowledge out of thin air.
There’s an excellent discussion of accuracy, precision and resolution at Phidgets, which I recommend to all. Quoting, “resolution determines the upper limit of precision. The precision in your data cannot exceed your resolution.”
Instrumental detection limits and systematic measurement error will ensure refractory errors in measured data, especially with respect to the physically true magnitude (air temperature).
There’s no getting around it, except in numerical Never-Never-Land.
Just yesterday Stokes deceptively misquoted me when I tried to explain that a formal uncertainty analysis of the UAH temperature measurement has not be performed:
https://wattsupwiththat.com/2023/06/26/uncertain-uncertainties/#comment-3740021
Fundamentally, Stokes fails to understand the distinction between accuracy and precision. For him, they are the same thing.
If he did do so, his entire act would collapse.
As Maxwell Smart would say –
“Aaaahhh. It’s the old numbers game again, 99”
If you measure the length of a board with a tape measure that is one inch off, you can average all you want but you will still be wrong by one inch. Averaging reduces random uncertainty but not calibration issues, especially when both liquid in glass thermometers and platinum resistance thermometers tend to drift in one direction. For mercury thermometers, if the bulb shrinks, it reads high, not low. As a PRT ages, stress builds up in the platinum wire and increases its resistance and the measured temperature. Non-random errors cannot be corrected by mathematical approaches meant for Gaussian distributions.
The problem comes when you know your tape measure is off by up to 1 inch, but have no idea whether it is longer or shorter or by exactly how much – then consider it may change each and every time you take a measurement!
“and it does diminish roughly with √ N
Just flat out WRONG.
These are individual readings in time, NOT a series of repetitive readings of the same thing.
Go back and learn about basics about errors of measurement and when to apply “√ N”, and when NOT TO.
He won’t, 1/root-N is his religion.
NIST has here a set of examples of application of the GUM. Their E2 is exactly the calculation of the average temperature for a month, with data on 22 days. So how do they find the uncertainty of their average? They divide the sd by sqrt(22).
Which is NOT what you and the rest of the trendologists rant about!
NIST is taking the variance of the sample set as the uncertainty, which is the standard deviation of the mean over the square root of N minus the degrees of freedom.
bg-whatever is blindly stuffing the mean formula into the uncertainty propagation equation and declaring that the uncertainty of the average is the uncertainty of a single temperature measurement divided by root-N.
sigma / sqrt(N-df) =/= u(T_i) / sqrt(N) !!
You are lying, again.
And when do climastrologers EVER report standard deviations of the myriad averages they do? NEVER
“the usual Pat Frank nonsense” Win the argument by contemptuous dismissal. Gavin Schmidt does similar. So does Noam Chomsky. You’re in good company, Nick.
Pigs only fly backwards from east to west, so it is possible that the temperature is the same from day-to-day but the instrument keeps changing or something …. sinusoidal calibration or lack of, perhaps daily isobaric correlation … ISO thingy, GUM, ask karlomonte. Oh wait, karlomonte has never observed the weather.
More seriously Nick, it is a refreshing change that the paper was published don’t you think? As I see it, you and karlomonte could join forces and submit a rebuttal.
Just an idea … pigs do fly after-all.
Dr Bill Johnston
http://www.bomwatch.com.au
And how do you know that I have “never observed the weather”?
Are you psychic?
No, you are just a clown (one who now seems to have transferred his obsession with Jennifer M. onto me).
Have you or are you just a BS-artist? Observing is different to recording.
b.
Yes, you are obsessed with me, ankle-biter.
“Yes, you are obsessed with me, ankle-biter.” wins the argument by contemptuous dismissal. Pat does similar. So does Rush Limbaugh. You’re in good company, Mr. Three Cards.
“Pat does similar.”
Another accusation sans evidence. Put up or shut up Willard.
This comment starts with the usual Nick Stokes ignorance.
and continues from there.
Oh, the trendologists won’t like this one!
I have said from the beginning … the data is not fit for purpose … any attempt to run it thru a computer model or statistical exercise is just an exercise in “Look how smart I am” but does nothing to reveal any truths about the weather or climate … its one big intellectual circle j*rk … sure, people have devoted their entire careers to it … its just cold fusion on a huge scale because of the money involved …
Finally! Finally! Thank you, Mr. Frank! I’ve suspected as much for decades, but without the wit to explain it!
but this makes empirical curve fitting and arbitrary parameterization seem less informative. that’s annoying.
When it comes to climate, first we find the average station temperature for the day. Then we average up all those daily means to find the annual station average. Then we take all the annual station averages from the equator to the poles winter spring summer and fall and compute the annual global temperature. Then that is compared to a base line and an anomaly is reported. Finally all those anomalies for the last 150 years are plotted out to see if there’s a trend or not, and if there is … But that’s not the end of the story, all that averaging is reviewed every month, and corrections are made all the way back to January 1880
Variance of the data record is a clue to the uncertianty associated with the data set. The higher the variance the less certain one can be of anything statistical analysis provides.
Yet with each average you list out climate science studiously omits any variance calculations for the underlying data, for the the variance of the daily averages, the monthly averages, the annual averages, the global average, or the baseline average.
Yet it is basic statistics that when you combine random variables the variance of those variables add – meaning that the resulting data set is less and less certain. The normal distribution is assumed to have the average value happen far more often than other values as you move away from the mean. As the difference in occurrence gets smaller and smaller with larger variance the certainty of whether or not the mean is *really* the most likely outcome gets less and les.
Just more basic science that the climate cult ignores – every single time. it would just be too damning to have to show the variance in the global average!
Good points, however, even with the “daily average temperature” calculations, information is lost regarding variations in Tmax and Tmin. For example: CO2 (or something else) may cause a change in cloud cover which may both make Tmax cooler and Tmin warmer. Therefore, dealing only with Tavg variations conceals possible greater variations within daily temps. The same can be said in going from daily to annual temps, and from local to global. Each statistical operation loses information from within the data in order to describe the data set as a whole.
There is absolutely no reason to assume that changes in a single variable (CO2) will cause uniform variations in all temperature measurements, daily, annual, and global, as implied by GAT variation
Most of the climate enthusiasts have shown themselves to be statisticians, not professional scientists.
Would an Expert in Applied Mathematics please explain to me how you can use the RMS value of points on a graph to show the TRUE value of the “missing” data.
I ask this as a result of my experience way back in 1974 when I was taking a Statics/Dynamics course in college. HP had placed the HP-35 “Pocket Calculator,” the first, on the market the year before and the professor allowed the use of it in class and for exams. This created a problem. All of the problems and answers up until this time had relied upon the use of a Slide Rule. More than half of the solutions to the problems did not agree with the answers when performed on the HP-35, however, were correct when performed on the Slide Rule. Problem was that the calculator did not have the precision of the SR. The calculator was only accurate to seven (7) places! Thus, when working with small angles precision was lost. IMHO the same “error” effect would be presented in predictive calculations for small changes in a parameter that is predicting an event far in the future. E.g. How many decimals of accuracy does the velocity and trajectory angle of a rocket need to be to land on the moon, intact, when no corrections to either can be made mid-course?
Also, I have been taught that the RMS value of a Sine Wave of sinusoidal Alternating Current represents the Direct current/voltage that dissipates the same amount of power as the average power dissipated by the alternating current/voltage. I have heard that temperatures are taken at ~8 AM and ~8PM. and also The day “High” and day “Low.” However, the RMS value of a Sine Wave = 0.707, Square Wave = “peak,” Triangular Wave = 0.577. None of these are representative of the actual HEATING/Cooling effect of the sun’s radiation or lack thereof. The actual heat collection potential curve is closer to a triangular wave with no lower lobe. Definitely NOT sinusoidal. This seems like another confusion factor originated by the uninformed and brainless, those lacking knowledge of advanced mathematics or those with an agenda.
The “average” of a sine is not the same as the RMS value of the sine. The average is about .63 of the peak. The sun does follow a sine wave path above a point as the earth turns. So the amount of radiation received at a point on the earth is also a sine wave. That is why if you look at daytime temps they are pretty much a sine wave from sunrise to about 3pm when the radiation from the earth overtakes the injection of energy from the sun and the temperature after about 3pm becomes an exponential or polynomial decay. (it’s 3pm because of the thermal inertia of all the components involved in the system)
Nit Pick – the average of a true sine wave is 0.0000000…..
The portions above and below 0 are equal, if not, then it is not a Sinewave.
The temperature profile during the day is not a true sine wave. It is a truncated one, typically ending around 3pm local time. Thus there *will* be an average. In addition, the part of the temperature curve that is sinusoidal always has an offset, a DC value if you will. That also keeps the sinusoidal temperature curve from averaging to zero.
Where did you get the formula used in equation (5) and (6) in your publication?
Is it Bevington 4.22 like what you told me last time?
Standard RMS bdgwx.
Then you used the wrong formula. Refer to the law of propagation of uncertainty equation 16 for the general case and equation 10 for the uncorrelated case in JCGM 100:2008. Notice that it reduces to u(y) = u(x) / sqrt(N) when y = Σ[x_i, 1, N] / N and u(x) = u(x_i) for all x_i and r(x_i, x_j) = 0. This can also be found in Bevington 3.13 for the general case or 3.14 for the uncorrelated case. Alternatively you can use equation 4.14 or 4.23 if the goal is to determine the uncertainty of the average. You can use the NIST uncertainty machine to verify this if there are any doubts.
The formula for RMS is used to calculate the RMS, bdgwx. That’s what I did.
As usual, you don’t know what you’re talking about,. As usual, that doesn’t stop you from holding forth.
He lives for telling people to “play around” with his cherished NIST uncertainty machine, refusing to acknowledge that if garbage is input to the machine, the output will be garbage.
That’s the problem. RMS is not the same thing as the uncertainty of the average. See Bevington section 4. Don’t believe me or Bevington? Use the NIST uncertainty machine and prove this out for yourself.
Clown, is this all you can do when people call you out on your nonsense? Over and over and over and over…
And this reminds me, it is almost for you to spam those lame IPCC graphs, again, for 90th time.
RMS is the root-mean-square of the errors, not the mean error of an average.
I know. That’s the problem. All you’ve done is compute the quadratic mean of the individual measurement uncertainties. And since you’ve assumed they all have the same uncertainty it is hardly shocking that the quadratic mean, like the arithmetic mean, is also the same.
I didn’t say anything about “the mean error of the average”. I’m talking about the uncertainty of the average as described by Bevington via equation 4.14 which is obviously different than RMS. I’ll repeat. RMS is not the same thing as the uncertainty of the average.
He doesn’t understand the difference!
You *still* insist on cherry-picking, don’t you?
In Section 4.1 Bevington states:
“In Chapter 2 we defined the mean u of the parent distribution and noted taht the most probable estimate of the mean u of a random set of observations is the average x_bar of the observations. The justification for that statement is based on the assumption that the measurements are distributed according to Gaussian distribution.”
He goes on to talk about the “Calculation of the Mean” and the “Estimated Error in the Mean”. Both of these are related to the SEM – i.e. how close you can estimate the mean from samples. It is *NOT* the uncertainty of the mean, it is only how close you have calculated the mean.
He goes on to say: “Equation 4.12 might suggest that the error in the mean of a set of measurements x_i can be reduced indefinitely by repeated measurements of x_i. … There are three main limitations to consider: 1. those of available time and resources, those imposed by systematic errors, and those imposed by nonstatistical fluctuations. ”
You *never* want to consider the limitations and always fall back on the assumption that how closely you can calculate the mean is how uncertainty that mean is physically – while ignoring systematic errors and nostatistical flucuations.
You just assume that all errors are random (even systematic ones), Gaussian, and that they cancel.
You simply don’t live in the real world. Neither do most climate scientists.
Bevington’s 3.13 is the standard root-sum-square equation including a covariance term. From Bevington: “The first two terms in the equation are the averages of squares of deviations weighted by the squares of the partial derivatives, and my be considered to be the averages of the squares of the deviations in x produced by the uncertainties in u and in v, respectively.”
He goes on to say that if the measured quantities u and v are uncorrelated, then, on the average, …. we should expect the term to vanish in the limit of a large random selection of observations.
You are back to cherry-picking things you think might prove you right. They don’t. You need to try harder.
BTW, Bevington, in his lead-in to the book STATES SPECIFICALLY that the following analyses assume totally random and Gaussian measurement errors. Something that simply doesn’t apply to the real world of measurement using field instruments whose calibration is questionable at best.
Tim,
Thank you particularly for that last paragraph.
bdgwx repeatedly fails because he deals with theoretical numbers when the numbers from real life have warts and all.
As Pat Frank has noted. As you have noted over and over. As Bill Johnston has noted over and over. As I have noted over and over.
Nick Stokes needs to think about the difference between theoretical numbers and field measurements also.
Geoff S
“Nick Stokes needs to think about the difference between theoretical numbers and field measurements also.”
So what different arithmetic is to be done?
As I complained above, no-one espousing these weird notions ever does an actual worked example, like I did with the Melbourne data. I’m just told that the Melb data, direct from the AWS, is somehow not physical.
is somehow not physical.
No. Inaccurate.
Nick.
The arithmetic that is missing is in the quantitative analysis of all factors that affect the final result. Pat Frank here has worked hard to gather such data for instruments like LIG thermometers. I have not seen such a comprehensive list before. Surely, it contains fundamentals for a useful estimate of uncertainty.
Without that, you cannot discern whether numbers cancel out, because you do not have all of the numbers, so you have to make guesses.
I do not know of ways to estimate the uncertainty of guesses.
Geoff S
So if NIST, JCGM, etc. and the equations and tools they publish produce incorrect results for “real life” scenarios then where I do go for the math that works for “real life” scenarios?
They don’t; rather it is the ridiculous climastrologers who abuse these tools with their religious dogma.
Without 1/root-N, they can no longer ascribe deep meanings to every little squiggle in the UAH graph.
Try Chapter 3 in John Taylor’s book. Chapter 4 and on typically considers situations where the uncertainty is random and Gaussian, just like Bevington does.
The very first sentence in Chapter 4 is:
“We have seen that one of the best ways to assess the reliability of a measurement is to repeat it several times and examine the different values obtained.”
This is specifically talking about multiple measurements OF THE SAME THING. That simply doesn’t apply to single measurements of multiple things!
He goes on to say: “As noted before, not all types of experimental uncertainty can be assessed by statistical analysis based on repeated measurements. For this reason, uncertainties are classified into two groups: the random uncertainties, which can be treated statistically, and the systematic uncertainties which cannot. This distinction is described in Section 4.1. Most of the remainder of this chapter is devoted to random uncertainties. Section 4.2 introduces, without formal justification, two important definitions related to a series of measured values x1, …, xn, all of single quantity x.”
Note carefully that very last sentence. Bevington says the exact same thing in his book.
The climate alarm cult, and that includes you, tries to shoehorn single measurements of different things, typically with systematic biases, into the same shoe as multiple measurements of the same thing. You do so by making the totally unjustified assumption that all uncertainty is random and Gaussian and cancels no matter what.
You, Nick, bellman, etc have never, NOT ONCE, worked through all the examples in Taylor’s book. You just cherry-pick things that look like it supports your unjustified treatment of the temperature data.
In this very thread Nick pulled out the NIST TN and declared: “See? Averaging reduces uncertainty!” without understanding what he was referencing.
The point is, averaging can reduce uncertainty if several criteria are observed, not the least of which is that that same thing must be measured by the same measuring device, under unchanging environmental conditions, at least 100 times to justify one more significant figure. One can do that in a laboratory with a physical object, chemical solution, or electronic device. It is the essence of the idea of a ‘controlled experiment.’
However, in the real world, where none of the environmental parameters are under the control of the experimenter, one is primarily an observer, and the best that can usually be done is to make one high-accuracy/precision measurement with one calibrated instrument, or, simultaneously take several measurements with measuring devices that are very similar, but not identical.
It is the unwarranted assumption that temperatures measured at different times, in different places, with different instruments, can be treated as in the first case above, that is the issue.
Exactly! And these people have been told this a zillion times, yet somehow it can’t penetrate their a priori pseudoscience.
“…not the least of which is that that same thing must be measured by the same measuring device.”
Same tripe, different day. For decades, petroleum reservoir simulators have “averaged” – sometimes with weightings, sometimes not – dozens of geological and rheological parameters for their inputs. They come from different instruments, at different times, most without being gathered at the same point in subspace. These values all have resolutions – often non normal – and are evaluated together. They also often correlate, which is taken into account. And many of these different parameters sometimes cross correlate, sometimes don’t – think porosity and permeability.
But somehow, in spite of the fact that the inputs are certainly worse than those under discussion here. petroleum reservoir simulation, using many of the same techniques as those used in spatial temperature and sea level evaluations, has firmly supplanted the empirical relationships used by cranky doubters like me for most of my career. And in toto, the oil and gas bizzes are trillions better off for it.
And this morning — a blob word salad!
*golf clap*
“…a blob word salad”
Ah, the all purpose, thought free rejoinder you keep on speed paste. Right up there with your middle school “Hey, F*** you, man!“. BTW, sorry about the lunch money. I’m not that guy any more…
Not a single word from blob about Pat’s post or paper, just an inane snipe toward Clyde S.
Correct. I was responding to a fact free claim by Mr. Spencer. Refutation of it pokes holes in the larger claims.
Do the measures of petroleum reservoirs change from minute to minute or are they stable over time?
Big difference!
Depletion. Water influx. Fracs. Steam. Nuff said?
Bigger pic, I was responding to Mr. Spencer’s ridiculous limiting assumptions, which were not only “at the same time”. Just the plethora of techniques to snag a permability estimate number over a dozen….
And none of your factor vary during a day, let alone a month. Tell us the resolution you get on your simulation. Barrel, fractions of a barrel, gallons. Remember we are talking about fractional percents. At 70 degrees and 0.001 uncertainty you have 0.001/70 = 1.4×10^-5. that’s 0.0001%. Are your simulation that accurate?
No, much wider. We commonly run many realizations to accommodate the referenced error distributions. But, as with climate simulations, perfectly fit for purpose. You data is orders of magnitude better than ours w.r.t. what you need to use it for…
Says who? And why is NIST, JCGM, etc. not following this rule?
Says who? The NIST, JCGM, etc.
They specify a MEASURAND. That’s singular. It does *NOT* imply a conglomeration of single measurements of multiple different things.
Taylor, chptr 4: “Most of the rest of this chapter is devoted to random uncertainties. Chapter 4.2 introduces, without formal justification, two important definitions related to a series of measured values x1, …, xn, all of the single quantity x.”
A single measurand “x”. Not multiple measurands x1, …, xn but multiple measurements of x.
Bevington is more wordy but says the same thing: “If we take a measurement x1 of a quantity x, we expect our observation to approximate the quantity, but we do not expect the experimental data point to be exactly equal to the quantity. If we make another mesurement, we expect to observe a discrepancy between the two measurements because of random errors, and we do not expect either determination to be exactly correct, that is,equal to x.”
Again, multiple measurements of the SAME THING. Single measurements of different things do not describe a a distribution around a true value. They just don’t. And if they don’t than an average derived from them is basically meaningless except for mental masturbation by climate scientists. The whole concept of a global average is not fit for purpose from the very get go because it is based an averaging different things.
NIST, JCGM, etc all recognize this. It is *you* that keeps trying to fit what they have developed into a round peg you can jam into a square hole using statistical tools and descriptors that simply aren’t fit for the purpose you are trying to impose on them.
“You, Nick, bellman, etc have never, NOT ONCE, worked through all the examples in Taylor’s book.”
If there were an exercise that showed how averaging increases uncertainty you would point me to it. Expecting me to work through every exercise in a book just in case there may be something that proves your point is a waste of time.
None of the standard authors address averaging of different things because an average of different things is *NOT* a measurement, it is a statistical descriptor. They all *do* address the standard deviation of the sample means obtained from a random, Gaussian distribution – which assumes complete cancellation of uncertainty which leads to the average being the true value of the population. The standard deviation of sample means is only a measure of how close you are to the population mean and only has meaning if the population is random and Gaussian.
You and the rest simply can’t seem to make the leap that measurements from different things is DIFFERENT than measurements of the same thing.
It always comes back to the meme of: “All uncertainty is random, Gaussian and totally cancels.”
You can argue you don’t assume that but it just shines through everything you post. This post is no different.
Its the Way Of The Trendologists — pseudoscience because their conclusions are made prior to analysis, and nothing must be allowed to interfere with the conclusions.
Yes, its true, you can’t read English.
¿Qué?
The battle for truth is raging in every discipline. Thank you to the brave people waging it here against the Climate Communists.
Hmm, however, satellites, radiosondes, and GPS radio occultation also show warming and similar rates. I wonder how they manage to reproduce the same mistakes.
I don’t think this paper is meant to discount the possibility of surface warming.
However, there is a discontinuity between estimated warming of the surface and that of the lower and mid tropospheric layers during the satellite era.
The surface is warming too fast, or maybe the free troposphere is warming too slow.
MIPs seem to be calibrated to the surface estimates, while the satellite tropospheric layers are allowed to deviate.
This is a problem considering the vertical tropospheric profiles are not matching model expectation. It’s usually explained away as systematic bias in MSU tropospheric satellite retrievals.
However, the are at least three distinct possibilities:
1) That the problem truly is satellite bias.
2) That there is a bias created by surface estimates.
3)That there are other physical mechanisms, not yet described, needed to explain the discrepancy.
Such as the discontinuous, spotty nature of the NOAA satellite sampling at latitudes less than about 60°,
Are they measuring the same thing? Or different things? What is actually happening where?
Do satellites, radiosondes, and GPS actually measure surface temperatures? Or only atmospheric temps?
They measure a proxy for temperature in the atmosphere which I forget the details of. It isn’t a direct temperature measurement though – accuracy depends on how closely the proxy correlates with air temperature.
In Pat Frank’s paper it is mentioned that the satellite measurements were “calibrated” based on surface SST measurements from ships which at best were full of non-quantifiable systematic errors. So in any case, we’re back to START. The first thing to do would be to calibrate the satellites with carefully measured actual data where the errors are quantifiable.
There are two series of satellite measurements which don’t match exactly to one another (different instrumentation or whatever). This could also be due to calibration errors.
Javier,
As Bill Johnson shows on Bomwatch blog, usual daily temperatures have a variance that mathematically has a big component from rainfall variation. Rain cools. The satellite microwave measurements might also have a moisture variance component, but I cannot recall any study of it.
It is not likely that surface LIG temperatures have identical mechanisms to the microwave ones, so there should be a difference. Which is more correct remains debatable because of an absence of comparable absolutes, such as can exist in pure number stats, but not in the physical measurement world under discussion. Geoff S
Noice.
Crickets.
Satellites were calibrated against in situ SSTs during the 20th century, Willard. Their surface temperatures are not independent. Mentioned in the paper. Maybe Javier missed that.
The paper is about the historical surface air temperature record to about year 2000. Maybe Javier missed that limit, too. it seems you did.
I presume that the symbol “Ö” was meant to be “sqrt( ).”
You’re right, Clyde, Looks like the translation software didn’t recognize the special ‘sqrt’ character. All the Greek lettering is wrong, too.
Well damn! Good job Pat. I had forgotten about your ‘promise’ to take on the instrument record error after your last seminal paper on model error.
You are effectively the reason I came to WUWT and keep reading it in so much as I stumbled upon your description of your challenges getting anyone to understand, much less publish your paper on climate model error.
Ultimately the proper handling of propagation of errors in a formula and proper handling of measurement error, especially the latter is what makes GOOD science very hard. It’s why I wanted to stick to theoretical physics…no errors other then in my calculations. 🙂
In the undergrad physics course I taught when taking my degree it was the handling of errors that I dinged people for most not whether they demonstrated an understanding of the physical laws they were testing. They get the latter from the Prof & tests they had to take. I wanted them to come away with an understanding that dealing with error would be the single greatest challenge in any career they may take up whether science or not.
In any case I fear that this 2nd seminal paper will go unnoticed & ignored by the ‘powers that be’. When your whole world of action & study is basically thrown asunder in 2 papers it’s hard for people to accept that reality.
It drives me to no end of grief to read postings even by ‘skeptics’ on WUWT that don’t include a representation of the model/measurement error or unrealistically small values.
With that said I still hold hope that your papers will somehow make a difference in the community.
Thanks, ycfs. Everyone I worked with killed themselves to get their data right. They sweated every bit of systematic error, which always showed up, and lived within instrumental resolution.
So it really frosted me to see all of that just waived away with self-serving assumptions in consensus climatology. Making things easy for themselves.
As much as anything, the abuse of method kept me involved in this mess.
Anyway, thanks for your kind words. Your knowledgeable support is very welcome.
First I need to correct 1 minor error in my previous post..it should say ‘undergrad physics LAB course’. I am not a Physics professor, never was one & didn’t play one on TV.
And frankly you are my ‘hero’ of sorts because of your dogged determination to write scuh detailed papers on something very important but for most people so ‘mind numbing’ that they couldn’t get through the first sentence of an introductory text on the subject. 🙂
You remind me of all the experimental physicists at a medium energy accelerator in Canada I hung out with in taking my degree. I couldn’t have done what they did daily. Checking, rechecking, updating calibrations. No detail was too small for these guys especially given the subject of the experiments & trying to detect very small effects.
In any case keep up the great work.
Your experience is why physicists will commonly report findings with 6-sigma uncertainty and self-identified climatologists rarely even provide 1-sigma.
You and your students weren’t the only ones to learn about uncertainty in a LAB course. Most engineers do. Many physics majors don’t because the professors don’t understand it either. It doesn’t help when not a single example in any textbook I own on physics or statistics analyzes experimental results with values given as “stated value +/- uncertainty”.
Thanks, ycfs. I’m humbled.
Pat, thanks for this paper, I really enjoyed it, well written, and very interesting in the historical aspects of thermometry. I have been to Thuringia and know their history of glass making generally and thermometers, specifically. (see table 2)
Some great materials science there, and good 19th century great physical measurements using the ice point by such scientific giants as Joule, whose name is on our basic unit for energy.
As for the uncertainty complainers, I bet these precision Engineers would run circles around these amateurs.
Thanks for the kind words, a_s. I’m glad you enjoyed it. Working through Joule drift and watching results drop out of the analysis was enjoyable.
At the start, I knew lead glass and soft glass existed, but had no idea about any of the details — the mixed alkali effect, etc. It took some research tracking down all those glass compositions.
One indelible impression is the competence and determination of the 19th century scientists. They worked hard, they paid attention to detail, they were candid about their instruments. Their contrast with a certain modern lot is striking.
Focused attention of competent and uncompromising precision engineers would be very welcome.
I often try to relate uncertainty in temperature data to what is found with car engines, especially racing engines. If you could measure the cylinder wall diameter in 10,000 engines with numerous different applications (short track racing, drag racing, short-haul freight carrying, long-haul freight carrying, street racing, endurance racing, and on and on and on) would the average value of the wear actually tell you anything if you don’t know how all of them were measured and recorded? Some using micrometers, some using analog calipers, some using digital calipers, some maybe even using a tape measure or ruler. There would be so many avenues of uncertainty to creep in the average would be of little use. Oil change intervals, oil type, piston ring specs, etc, – all unknowns and hardly able to be considered to be random and Gaussian.
Yet we are expected to believe that the global average temperature, calculated from data with at least as much uncertainty as the cylinder wear in a universe of car engines, is somehow fantastically derived with almost infinite precision by assuming all error is random, Gaussian, and cancels.
Even more basic, the standard deviation of the sample means (the SEM if you will) only tells you how close you are to the population mean. It tells you nothing about the accuracy of the population mean unless you assume all measurement uncertainty is random, Gaussian, and cancels. Even if you had every single member of the total population at hand and could calculate the population average with infinite precision you still wouldn’t know how accurate it is without considering the uncertainty associated with each and every member in the data set and propagating that uncertainty onto the average you calculate. It is that propagated uncertainty that determines the accuracy of the mean.
Based on this analysis of uncertainty, the generally reported measurement changes are less than the uncertainty by a considerable amount. Therefore, the actual warming since 1850, in degrees C, could be as much as 3C+, no?
An interesting conjecture but, I suspect impossible due to lack of data, if not for other reasons in such a complicate system:
If there were enough physical data of changes in amount of ice, changes in tree line, changes in locations of animals and plants northward (northern hemisphere), and other “stuff”, might it be possible to calculate a more accurate value for temperature change?
“Impossible due to lack of data.” is one of the messages of Pat Frank’s paper. The temperatures given are merely a mathematical probability of what they might be but, despite their propaganda, we simply don’t know what the precise temperatures actually are.
Yes.
So we could be at 2.86°C in just 110 years.
That’s about 0.3°C warming every decade!
It’s worse than we thought. Almost a degree of warming by 2050!!!
Thanks Dr. Frank, I agree with Monckton. Between this and your 2019 paper on the propagation of error you have placed the final nails in the coffin for CAGW. Of course the adherents of CAGW refuse to accept that fact until they have developed a replacement, their livelihood depends on it. Besides, there’s no way they can leave the gaggle of true believers they have assembled until they have been converted to a follow-on belief.
Now, where have I heard that before?
No idea but you hum it and I’ll join in!
Pat, what would be your critique of this temperature chart (Hansen 1999)?
Fit for purpose, or not fit for purpose?
Temperature numbers accepted at face value, Tom. No error bars, no uncertainty bounds.
Instrumental detection limit and visual repeatability alone would add about ±0.4 C (95%) up through about 1980 and about ±0.2 C to 2000 (assuming instantaneous switch to MMTS at 1981), darkening much of the face of the plot.
The uncertainty would about halve, were temperatures reported rather than anomalies.
And Joule-drift renders everything prior to 1900 unreliable.
Pat,
I’ll read your paper later today.
From your comments already, it looks like a much-needed “return to the basic senses” paper.
You criticise in closing the professional societies and groups who should have made your paper unnecessary if they had acted properly. By accident of timing, I hope to finish today a 3 part article on the subject of corruption of US scientist by some with money, power and beliefs. I use the Rockefeller Foundation as an example, tracing the history of the Linear No Threshold method to relate toxin dose to harm. Ed Calabrese at Amherst has done decades of diligence on this.
I am hoping that WUWT will accept it.
Pat, thank you for your persistence in injecting reality into major issues like climate models and thermometry that reality history will show were little more than fairy tales. Geoff S
Thanks, Geoff. Let me know what you think. 🙂
cool a Chinese Journal
https://www.linkedin.com/company/mdpi/people/
The academic editor is Japanese, Steve. Do you want to denigrate him, too?
BTW.
China lead the world in the number of patents this year. The quickest advancing nation should not be disparaged for their technical expertise.
https://worldpopulationreview.com/country-rankings/patents-by-country
It’s probably because they have the manufacturing industry and thus real-world problems to solve.
It was reported that most of the technology that was found on the Chicom spy balloon was American.
That is not surprising. The balloon was never going to be recovered by China.
If they had more advanced technology they wouldn’t put it on that.
mosh is a racist, what a surprise.
It is also a predatory journal that publishes over 800 articles per day.
Idiot.
To me, the first table giving the readability of a mercury in glass thermometer is really its reading precision, not its accuracy. The accuracy is a function of the construction of the thermometer and being filled correctly, and then being calibrated properly. In other words, you can read the thermometer to ±0.3°C but if it has drifted or is otherwise out of calibration, you do not know the true temperature to 0.3°C. Your reading is less accurate than that. This error does not disappear by averaging lots of readings. The reading uncertainty may be reduced if it is a random error. The out-of-calibration error is not reduced.
Yep. But the climate alarm cultists just assume, with no justification, that the out-of-calibration error cancels out.
MDPI is a predatory journal. And with a publishing rate of over 800 articles per day I think you’re being awfully cavalier when you say the publication passed peer-review. Who knows…if there is a class action lawsuit against MDPI like what we saw with OMICS then one day you might be able to claw back some of your publishing fees.
I am psychic…
Speaking in ignorance, bdgwx.
I am amused by the posts from bdgwx.
I fail to understand why some peoplw are annoyed by the posts from bdgw when bdgwx is an internet troll who only posts nit-picking snark, snears and smears from behind the cowards’ screan of anonymity.
The future may impose dementia on any of us, so I suggest we try to be kind and laugh at the demented behaviour of bdgwx instead of getting annoyed by it.
He is 100% clue-resistant.
While I agree with your assessment, the problem is that he abuses standard publications and twists them, falsely claiming they support his a priori irrational ideas about “climate”. People reading his nonsense who are not familiar with these subjects could be led into viewing him as some kind of expert, if he is not challenged.
While gadflies can sometimes be useful, they are still invariably annoying. They should work harder at being more informative and less annoying.
MDPI publishes 427 journals in total. Granting your rate, that’s just under two articles per day per journal. Hardly excessive.
Sensors registered a 3.9 impact factor in 2022, which is a good number. It seems you have no case, bdgwx.
It wasn’t my case. It comes from predatoryreports.org.
Just shows you shouldn’t take other’s opinions uncritically, bdgwx.
A 2020 article about MDPI
“I originally studied the MDPI portfolio expecting to find something nefarious behind their success, and I intended to contrast them to other “healthier” OA portfolios. Upon further study, I concluded that my expectations were misplaced, and that MDPI is simply a company that has focused on growth and speed while optimizing business practices around the author-pays APC (article processing charge) business model. As I discovered later, I was not the first skeptic to make a U-turn (see Dan Brockington’s analysis and subsequent follow-up). What follows is an analysis of MDPI’s remarkable growth.”
Sounds a lot like factcheck.com…
Dr. Frank is correct in that, given the business plan of MDPI, the number of daily pubs does not necessarily make it predatory. Rather, the claim is amply backed up by your link.
The best metric for the actual contribution of this paper to our understanding will be it’s citations, Unless you are of the Dr. Evil world wide conspiracy bent, with tens of thousands of scientists ignoring their competitive drive to prevent a Stanford sinecurist with time on his hands from making a ground breaking contribution (that they could exploit for their own benefit), then this article will be widely cited. Maybe we could check back at Dr. Franks similar pubs to see how well he did – without the many self cites….
Another clown in the rings now…
“The best metric for the actual contribution of this paper to our understanding will be it’s citations”
Back in 2001, I published a paper on the iron-molybdenum-sulfur cofactor of Nitrogenase.
It presented a solution to what had been a vexed and unsolved problem of ligand-binding, it reconciled 20 years of data that had lain unexplained in the literature, and it proposed a rational structural model for the solution structure of the cofactor.
The behavior of the cofactor was more complicated than anyone had expected. People left the field for other research endeavors, after that paper came out.
Since then, its gotten only 5 citations.
That work was critically central to the field as it was then, got me an invitation to an exclusive conference, and earned the respect of every first class scientist working on Nitrogenase.
All this is a long-form observation that you don’t know what you’re talking about bob.
As regards LiG Met. why would anyone expect it to be cited by hubristic incompetents when their careers are down in flames consequent to it.
Work stands on its internal merits. Not on its citation index or whether a field shown obviously partisan chooses to notice it.
Amen!
+7000
How many citations do papers from Einstein (relativity) or Planck (heat) or Newton (motion) or Maxwell (EM) receive, yet they are seminal works that have withstood the test of time.
You are doing a specious argument that has nothing to do with the validity of the paper. You would fail a high school forensics/debate class.
A different time. Very few people read Einstein’s original paper. When his work became known, the tenets of that paper have been indirectly cited almost 190,000 times.
So, sorry Mr. Gorman. Citations of peer reviewed papers are one of the main stocks in trade of modern scientific progress. To ignore good work puts you at a disadvantage. Making Dr. Frank’s work – well….
Keep dancing, clown.
Unable to present a single technical objection, you instead throw this shite against the wall, hoping it will stick.
“Citations of peer reviewed papers are one of the main stocks in trade of modern scientific progress”
No, they’re not. Accurate work is the stock in trade of scientific progress.
Citations prove nothing, as fully proved by the 1283 citations garnered by the scientific crock that is MBH98.
“Citations prove nothing, as fully proved by the 1283 citations garnered by the scientific crock that is MBH98.”
Bam!
I would say that MBH98 is as good an example as any, showing that “citations prove nothing”.
https://www.woodfortrees.org/graph/gistemp/from:1978/detrend:0.85/mean:6/plot/uah6/detrend:0.5/offset:0.4/mean:6
The above is a comparison of GISS LOTI and UAH 6, after detrending, offsetting and 6 month moving mean. UAH6 is a temperature calculated from radiation intensity that is dependent on air temperature and GISS is calculated from thermometer measures of maximum and minimum over a 24 hour period. The later is not an intensive property of anything, so the average after Krigging or whatever they do, is meaningless. Still, they are very similar. The overall trends differ significantly but the above plot shows that it’s mainly the estimates at earlier El Nino and and some La Nina periods that GISS has a much lower deviation from the trend.
There is no way in hell that one of them hasn’t done something dodgy.
UAH and GISTEMP measure different things.
Pat Frank
Thanks for a fine and cogent paper. It provides a clear ‘take home message’ that global temperature “measurements” are nonsense. Your finding is not news but is additional confirmation of information which has been known for decades.
The underlying problem is that
(a) there is no agreed definition of “global temperature”
and
(b) if there were an agreed definition of “global temperature” there would be no possibility of an independant calibration standard for any measurement and/or estimate of it.
These facts (itemised here as a and b) enable the teams which provide values which purport to be ‘global temperature’ to alter their data for past years to fit whatever narrative they choose to promote. And each of the teams does change their data series in most months.
I congratulate you on your success in obtaining publication of your paper in the formal literature because – decades ago – I failed to get my paper on these matters published in the formal lierature.
However, these matters were fully explained in my submission to the UK Parliament’s Select Committee Inquiry (i.e. whitewash) of ‘climategate’ that was conducted in early 2010. The submission can be read from Hansard (i.e. the permanent record of the UK’s Parliament) at https://publications.parliament.uk/…/387b/387we02.htm
The linked submission explains the importance of one of the emails from me which were among the amails ‘leaked’ from the Climate Research Unit (CRU) of East Anglia University.
The email is provided as Appendix A of the submission which explains that the email is part of my objections to the improper activities which were used to block publication in the formal literature of my paper. Appendix B is a draft version of that paper and I think it contains much of use to all who are interested in the methods of ‘climate scirntists’ and/or ‘climate science’.
Again, congratulations on your paper and its publication. I am pleased that I have lasted long enough to see the reality of global temperature measurements (i.e. they are all bunkum) being published in the formal literature, and I hope all interested in real science will publicise your paper.
Richard
The link did not work. I will try again
https://publications.parliament.uk/pa/cm200910/cmselect/cmsctech/387b/387we02.htm
Richard
Thank-you Richard. You’re a true veteran of the climate wars, on the science side of the front lines right from the start. I’m happy to acknowledge your precedence in critiquing the mess that is the air temperature record.
I also understand your frustration at having your paper blocked from publication. I faced similar opposition during the six years and 13 submissions it took to get Propagation… competently reviewed and published.
Years ago, when starting to construct myself as an adult, I consciously chose knowledge over delusion. Perhaps naively, I decided if knowledge brought sorrow, I’d rather live knowing and in sorrow than in blissful delusion. Delusion seemed far too dangerous.
In that light, you must have lived for years with considerable sorrow, knowing as you do while seeing people living in delusion determinedly driving themselves and everyone else over a cliff.
So best wishes, Richard. When we’re both gone and the history of this benighted episode is written, your name will be there remembered for your analytical sanity when so many had volunteered into Bedlam.
“So best wishes, Richard. When we’re both gone and the history of this benighted episode is written, your name will be there remembered for your analytical sanity when so many had volunteered into Bedlam.”
Well, said. Thanks for all you have done, Richard.