Guest Post by Willis Eschenbach
Over at Judith Curry’s excellent blog there’s a discussion of Trenberth’s missing heat. A new paper about oceanic temperatures says the heat’s not really missing, we just don’t have accurate enough information to tell where it is. The paper’s called Observed changes in top-of-the-atmosphere radiation and upper-ocean heating consistent within uncertainty.
It’s paywalled, and I was interested in one rough number, so I haven’t read it. The number that I wanted was the error estimate for their oceanic heating rates. This error can be seen in Figures 1a and 3a on the abstract page, and it is on the order of about plus or minus one watt/m2. This is consistent with other estimates of upper ocean heat content measurement errors.
I think I can conclusively demonstrate that their claimed error is way too small. To understand why, let me take a detour through the art, science, and business of blackjack.
In a fit of misguided passion, some years back I decided to learn how to count cards at blackjack. I had money and time at the same moment, an unusual combination in my life, so I took a class from a guy I’ll call Jimmy Chan. Paid good money for the class, and I got good value. I’ve always been good with figures, and I came out good at counting cards. Not as good as Jimmy, though, he was a mad keen player who had made a lot of money counting cards.
At the time they were still playing single deck in Reno. And I was young, single, and stupid. So I took twenty thousand dollars from my savings for my grubstake and went to Reno. It was an education about a curious business.
Here are the economics of the business of counting cards.
First, if you count using one of the usual systems as I did, and you are playing single deck, it gives you about a 1% edge on the house. Not much, to be sure, but it is a solid edge. And you can add to that by using a better counting system or a concurrent betting system, where better means more complex.
Second, if you play head-to-head (just you and the dealer) you can typically play about a hundred hands an hour.
Doesn’t take a math whiz to see that if you don’t blow the count, you will win about one extra hand an hour.
And therein is the catch. It means that in the card counting business, your average hourly wage is the amount of your average bet.
It’s a catch because of the other inexorable rule of counting blackjack. This regards surviving the swings and arrows of outrageous luck. If you don’t want to go home empty-handed, you need to have a grubstake that is a thousand times your average bet. Otherwise, you could go bust just from the natural ups and downs.
Now, twenty thousand dollars was all I could scrape together then. So that meant my average bet couldn’t be more than twenty dollars. I started out at the five dollar level.
I’d never spent any time in a casino up until then. I felt like the rube in every movie I ever saw. I played a while at the five dollar level. You never win or lose much there, so nobody paid any attention to me.
After a day or so making the princely sum of $5 per hour, I started betting larger. First at the ten-dollar level. Then at the twenty-dollar level. That was good money back in those days.
But when you start to make a bit of money, like say you hit a few blackjacks in a row and you’re doubling down, they start paying attention to you, and the trouble begins. First they use the casino holodeck to transport a somewhat malignant looking dwarf armed with a pad and a pencil to your table. He materializes at the shoulder of the dealer, and she starts to sweat. I say she because most dealers were women then and now. She starts to sweat because the casino doesn’t really care about card counters. I was making $20 an hour on average? Big deal, everyone in the casino management made that and more.
What scares casino owners is collusion between dealers and players. With the connivance of the dealer a guy can have a “string of luck” that can clean out a table in fifteen minutes and be out the door, meeting the dealer later to split the money. That’s what casino owners worry about, and that’s why the dealer started sweating, she knew she was being watched too. The dwarf peered through coke-bottle thick glasses, and wrote down the number of chips on each stack in the dealer’s rack, how much money I had, how much other players had. He gave the dealer a new deck. He wore a suit that cost as much as my grubstake. His wingtip shoes were shined to a rich luster. He looked at me as though I were a rich man with a loathsome disease. He watched my eyes, my hands. I started sweating like the dealer.
If I continued to win, the holodeck went into action again. This time what materialized were two large, vaguely anthropoid looking gentlemen, whose suits were specially tailored to conceal a bulge under the off-hand shoulder. They simply appeared, one at each shoulder of the aforementioned vertically challenged gentleman, who looked even dwarfier next to them, but clearly at ease in his natural element. They all three stared at me, and when that bored them, at the dealer. And then at me again.
And if the dealer was sweating, I was melting. I’m not made for that kind of game, I’m not good at that kind of pretence. I found out you can take the cowboy out of the country, but you can’t make him go mano-a-mano with the casinos for twenty bucks an hour.
I lasted a week. I logged my hours and my winnings. During that time, I worked well over forty hours. I only made enough money to pay for the flight and the hotel, and that’s about it. I was glad to put my twenty grand back in the bank.
I couldn’t take the constant strain and pressure of counting and not looking like I was counting and trying to stay invisible and feeling like a million eyes in the sky were watching my every eyeblink and having an inescapable feeling of being that guy in the movies who’s about to be squashed like a bug. But for those who can make it a game and keep it up, what an adventure! I’m glad I did it, wouldn’t do it again.
The part I liked the least, curiously, was something else entirely. It was that my every move was fixed. For every conceivable combination of my cards, the dealer’s card, and the count, there is one and only one right move. Not two. Not “player’s choice”. One move. I definitely didn’t like the feeling that I could be replaced by a vaguely humanoid 100% Turing-tested robot with a poor sense of dress and a really, really simple set of blackjack instructions
But I was still interested in the math of it all. And I had my trusty Macintosh 512. And Jimmy Chan had an idea about how to improve the odds by changing his counting method. And so did some of Jimmy’s friends. And he had a guy who tested their new counting method for them, at some university, for five hundred bucks a run.
So I told Jimmy I’d do the analysis for a hundred bucks a run. He and his friends were interested. I wrote a program for my Mac to play blackjack against itself. I wrote it in Basic, because that was what was easy. But it was sloooow. So I taught myself to program in C, and I rewrote the entire program in C. It was still too slow, so I translated the critical sections into assembly language. Finally, it was fast enough. I would set up a run during the day, programming in the details of however the person wanted to do the count. Then I’d start it when I went to bed, and in the morning the run would be done and I’d have made a hundred bucks. I figured that I’d finally achieved what my computer was really for, which was to make me money while I slept.
The computer had to be fast because of the issue that is at the heart of this post. This is, how many hands of blackjack did the computer have to play against itself to find out if the new system beat the old system?
The answer turns out to be a hundred times more hands per decimal. In practice, this means at least a million hands, and many more is better.
What we are looking at is the error of the average. If I measure something many times, I can average my answers. Is the resulting mean value the true underlying mean of what I am measuring? No, of course not. If we flip a hundred coins, usually it won’t be exactly fifty/fifty.
But it will be close to the true average of the data. How close? Well, the measure of how close it is expected to be to the true underlying average is what is called the “standard error of the mean”. It is calculated as the standard deviation of the data divided by the square root of the number of observations.
It is the last fact that concerns us. It means that if we double the number of observations, we don’t cut the error in half, but only to 0.7 of the original value. One consequence of this is that if we need one more decimal of precision, we need a hundred times the number of observations. That is what I meant by a hundred times per decimal. If our precision is plus or minus a tenth (± 0.1) and we want to know the answer to one more decimal, plus or minus one hundredth (± 0.01), we need one hundred times the data to get that precision.
That is the end of the detour, now let me return to my investigation of their error estimate for the ocean heating rate for the top 1800 metres of the ocean. If you recall, or even if you don’t, that was 1 watt per square metre (W/m2).
Now, that is calculated from temperature readings from Argo floats, about 3,000 of them during the study period.
Let me run through the numbers to convert their error (in w/m2) into a temperature change (in °C/year). I’ve comma-separated them for easy import into a spreadsheet if you wish.
We start with the forcing error and the depth heated as our inputs, and one constant, the energy to heat seawater one degree:
Energy to heat seawater:, 4.00E+06, joules/tonne/°C
Forcing error: plus or minus, 1, watts/m2
Depth heated:, 1800, metres
Then we calculate
Seawater weight:, 1860, tonnes
for a density of about 1.03333.
We multiply watts by seconds per year to give
Joules from forcing:, 3.16E+07, joules/yr
Finally, Joules available / (Tonnes of water times energy to heat a tonne by 1°C) gives us
Temperature error: plus or minus, 0.004, degrees/yr
So, assuming there are no problems with my math, they are claiming that they can measure the temperature rise of the top mile of the global ocean to within 0.004°C per year. That seems way too small an error to me. But is it too small? If we have lots and lots of observations, surely we can get the error down to that small?
Here’s the problem with their claim that the error is that small. I’ve raised this question at Judith’s and elsewhere, and gotten no answer. So I am posing the question again, in the hope that someone can unravel the puzzle.
We know that to get a smaller error by one decimal, we need a hundred times more observations per decimal point. But the same is true in reverse. If we need less precision, we don’t need as many observations. If we need one less decimal point, we can do it with one-hundredth of the observations.
Currently, they claim an error of ± 0.004°C (four thousandths of a degree) for the annual average upper ocean temperature from the observations of the three thousand or so Argo buoys.
But that means that if we are satisfied with an error of ± 0.04°C (four hundredths of a degree), we could do it with a hundredth of the number of observations, or about 30 Argo buoys. And it also indicates that 3 Argo buoys could measure that same huge volume, the entire global ocean from pole to pole, to within a tenth of a degree.
And that is the problem I see. There’s no possible way that thirty buoys could measure the top mile of the whole ocean to that kind of accuracy, four hundredths of a degree C. The ocean is far too large and varied for thirty Argo floats to do that.
What am I missing here? Have I made some major math mistake? Their claimed error seems to be way out of line for the number of observations. I’ve not been able to find a good explanation of how they come up with these claims of extreme precision, but however they’re doing it, my math doesn’t support it.
And that’s the puzzle. Comments welcome.
Regards to everyone,
w.
Discover more from Watts Up With That?
Subscribe to get the latest posts sent to your email.
Gives you an edge of 1%, eh? So in one hour of 100 hands, you’d be winning 50.5 – 49.5.
Now, if you’d combined that with an optimal betting strategy (the 1-2-3-4-5-6, for example), you might have made enough to be taken into the casino parking lot and given a lecture on the expense of major dental treatment….
I have looked through these postings, and the arguments associated with them.
I am not, nor have I ever been, a statistician.
As far as I can see, there are two threads to this issue. One is the esoteric field of statistical estimation of data from individual readings. This looks fun for mathematicians. The other is an attempt to apply this to the real requirement to measure ocean temperatures. I think this is simply not possible.
My concern is that we just do not know how variable the actual water temperatures are in the real ocean. We know that the main currents differ from their surroundings, but we do not know precisely where the edges of the currents are at any one time, probably to within a mile or so. I wonder if the amount of heat stored in a cylinder of water a mile wide on the circumference of each major current is of the same order as the missing heat? Or the columns of cold water which I assume sit under each iceberg?
I suspect that the oceans have many microcells of variable temperatures, and even several thousand ARGOS buoys are unlikely to hit one of these cells during their operational lifetime. And if these cells contain enough heat to satisfy the ‘missing heat’ hypothesis, we will never have a hope of finding it.
I have not done the math but I always thought that the problem with imprecision was not in the ocean heat content (or really change in ocean heat content) but with the radiative numbers at the top of the atmosphere. The Argo numbers for ocean heat (derived from temperature measurements) are much more precise than the radiative flux numbers at the top of the atmosphere. The latter are only good to 1% and you are trying to measure an in going vs. outgoing flux difference of 0.1% which implies you need measurements to 0.01%. My sense of the paper was that it was like a golf ball hit into the rough and may be out of bounds. The ball is the ocean heat content data the grass height is the precision of the radiative flux data. If the grass is short you should be able to find the ball and determine if it’s out of bounds. However, if the grass is very tall, the ball is very difficult to find and if you don’t find it you can’t tell if you are out of bounds. They are using the ball lost in the tall grass to argue they cannot yet determine if they are out of bounds.
We are dealing with climate ‘scientists’ therefore the first thing to do is ensure that they are using the correct terms in the correct way.
I believe this is a simple undergraduate level error of thinking that instrument precision is the same as measurement accuracy. So I get the wrong result (low accuracy) but to 10 places of decimals (high precision). This is then compounded by ‘clever’ statistical massaging of the high precision but low accuracy results.
What is needed is some idea of the accuracy of the measurement of the world ocean average temperature – but then validation is required against a baseline to check accuracy and there is no baseline. All that can really be done is identify the change since the last measurement, but as the floats are not static this metric is meaningless too.
This is then added to the unsupported claim that there is such a thing as an ‘average’ world ocean temperature and that mathematically averaging output from 3000 randomly placed and moving floats actually provides anything meaningful.
This is truly climate ‘science’ at its best.
That we can not find the missing heat is no longer a travesty.
===============================
E M Smith
spot on. Average temperature is scientifically meaningless yet we spend billions on computer models trying to predict this meaningless metric. Has anything ever been as broken as climate ‘science’!
The error is the measurement error of the individual buoy.
David Falkner asks:
Larry Fields observes:
The international quantitative standard for calculating the full uncertainty is the root mean square combination of all the errors.
See NIST TN 1297 , Guidelines for Evaluating and Expressing Uncertainty of NIST Measurement Results.
For Willis and other quantitatively inclined, see the Law of Propagation of Uncertainty etc.
In How well can we derive Global Ocean Indicators from Argo data? K. von Schuckmann and P.-Y. Le Traon observe:
That assumption of ignoring systematic errors needs to be tested! Sensor drift could overwhelm the rest. Bias error is often as large as the statistical error. Thus their total error could well be understated by ~ 41% (the square root of two.)
Furthermore, on earlier measurements they note:
The times series methods also need to be examined.
Statistician William Briggs opines:
Statistics Of Loeb’s “Observed Changes In Top-Of-The-Atmosphere Radiation And Upper-Ocean Heating Consistent Within Uncertainty”
How To Cheat, Or Fool Yourself, With Time Series: Climate Example
I have yet to read the paper – it takes time to get without paying – but as I understand the abstract, they are saying that the ‘missing heat’ from the additional forcing they expect from the CO2 model, COULD BE down there in the ocean depths, because the accuracy of the measurments is not good enough to say that it ISN’T there.
I looked at it from the other direction….from the Top of the Atmosphere downwards. It is at the TOA that they ‘measure’ the missing heat. I am not yet sure whether this additional heating is real measurement or modelled – I guess it is the same problem – a mix – because a global average is produced – but the modelling must be dealing with a much simpler situation. So – if the flux at the TOA they have detected is a near-decadal signal of 0.5 watts per square metre, they can then assume that this amount has gone into the deeper ocean rather than the surface laters of the atmosphere where it would be more readily detected. The TOA average flux is 340 watts/square metre. So the TOA signal that they detect is at the level of 0.01%. Given that the solar output varies by 0.1% over the eleven year cycle and has other irregular phases…..I don’t immediately see how they justify that signal as statistically significant, but hope to look more closely.
So they then go looking for the heat. And because the error margins of the ocean heat content calculations to 1800m are large enough to hide their missing heat….they state that their model is consistent with the TOA data (i.e. the observed flatlining of surface temperature as well as heat content in the upper 700m) which is consistent with the CO2 model. What amazes me is that oceanographers are not more sceptical….there is little heat exchange below 200m let alone 700m…..with about 80% of the late 20th centennial rise in upper ocean heat content held within the first 200m.
With all of this mathematics and statical uncertainty, we keep getting told that the root of the entire problem is…………………………..
CO2 !!!
This is one more excellent example of Willis’ ability to get to the heart of the matter, in this case, the impossibility of determining the ocean’s heat content with severely aliased measurements. I mean spatially aliased, before someone attacks that statement.
I love the way he illuminates the dank corners of Post-Normal Science. The mold in those corners is starting to cringe away from the piercing light.
Keep it up Willis!
Nice Post about the Black Jack, as a banned BJ player myself I can appreciate your story. Most card counters go broke becuase they dont have a big enough bankroll. We played with a ROR of .5% calculated on 50 million simulated hands!
God I hate being banned!! Best money I ever made, lasted 3 years!
BTW I trained blackjack players and had a team of 10 players 🙂
Until I see discussions including convectional heat transfer as a major factor, I see arguments over the radiation budget to be lacking. It has been estimates that 85% of the energy lost to space is first transferred to altitude by convection of warm, humid air. As the oceans appear to be cooling, it might be conjectured that convection is doing a good job.
You cannot achieve more accuracy than the base accuracy of the temperature device you are using. If the ARGO bouys are accruate to +/- 0.1 C then the maximum accuracy of your sea temperature measurements can only be +/- 0.1 C. If I have a thermometer that is acurate to +/- 0.1 C and I want to measure my temperature, I cannot achieve a result of greater accuracy than +/- 0.1 C NO MATTER WHERE, OR HOW MANY TIMES, I TAKE MY TEMPERATURE.
Peter Taylor says:
January 27, 2012 at 6:29 am
“What amazes me is that oceanographers are not more sceptical….there is little heat exchange below 200m let alone 700m…..with about 80% of the late 20th centennial rise in upper ocean heat content held within the first 200m.”
Imo here is the explanation to the so called GHE, the assumed 33K deficit in temp.
I you let the solar that hits the oceans heat the very thin upper slice of water + the atmosphere above it, imo there won’t be much of a deficit, if any.
The thin layer of ocean above the thermocline is the buffer that carries the accumulated daytime heat to the night.
The deep oceans just sit there doing their ocean things without much heat exchange with the upper layer OR the hot core below, BUT they have a temperature of ~275K, not the 0K a blackbody approach assumes.
Willis
I am not trying to lecture you or to say anything about 1800 being a small sample.
Looking back, I guess I was commenting in response to this.
“And it also indicates that 3 Argo buoys could measure that same huge volume, the entire global ocean from pole to pole, to within a tenth of a degree.”
Moving from the stats of 3000 to the stats of 3 is risky business, that is all.
James
The key here is that we need an Analysis of Variance (ANOVA) of the total ocean ‘system’. There are other sources of variability besides just the instrument precision.
Lets say we make several plastic parts that go into our product, and these parts all need to be exactly the same shade of our trademark Ocean Blue color. And we have an analytical technique that measures Ocean Blue color very precisely to determine if we are meeting our company targets. We make batches of plastic and each batch can create 10 or 100 or 1000 parts per batch (depending on the size of the part). Right there we need to sample for Batch-to-Batch variability and by sampling multiple parts from each batch we assess Part-to-Part variability. In addition, lets say we have 10 production lines which can vary a little because perhaps the batch mixing is a bit different Line-to-Line and also production lines make different size and shape parts and may use different types of plastic in the batch formulation. We might even find a seasonal difference if the parts run through the machines faster or cool differently at different times of the year. The ANOVA would determine variances for each of those and statistically combine them to determine the overall Ocean Blue color performance.
What we would probably find is .. the instrument precision was tiny compared to the variability in all of the other parameters!
How does that apply to our Ocean temperature problem? Frequent WUWT readers will be familiar with the detailed work of Bob Tisdale on ocean data, which shows that all oceans are not the same. My quick list of parameters might include:
*Time window (month or season)
*Ocean subdivided into Ocean region (eg N S E W or Gulf Stream – sub sections don’t necessarily need to be the same size and shape)
*Latitude (to account for such things as Gulf Stream cooling as it moves from Bahama to Iceland)
*Ocean ‘phase’ (such as AMO, PDO, El Nino, etc)
*Thermocline (percentage of an Argo profile above vs below the thermocline)
Having established fixed sampling sectors from the above, one might then determine statistics from all Argo floats that happen to supply data within that sector for the given time window. It then becomes possible to look at season-to-season variability or phase variability which are likely larger sources of variance than whether there are 10 or 100 floats in a sector at a given time. One might find that something like the percentage of readings above vs below the thermocline might be the largest source of variability. Or not. But unless you do the ANOVA you really don’t know.
O/t maybe, but what we are talking of here is the shortcomings of metrology and statistical error assessment. Give a thought to that figure that economists haunt us with – GDP. Britain has recently announced that in the last quarter of 2011, GDP fell by 0.2%, National statistic offices and their masters love to believe that GDP (which is not a terribly useful measure in the first place) can be stated with 0.1% precision. That is accuracy of within 1 in a thousand. In reality accuracy of one in one hundred is far beyond attainment and five times that figure would be impressive.
GDP, even in relatively small countries, has many, many times 30.000 recording points – much, much more data. Unfortunately not all of that data makes it to the final count and there are large areas of the economic ocean for which there are no measurements. People forget to send the numbers in, or put it in twice and others put it in the pending basket. Numbers for the ‘no recorded data’ have to be guessed (estimated is the conventional word). There are many guessers, some more able and some less so. Each one of them is subject to some degree of political pressure. No member of any sitting government wants to see a negative answer. (“Young man I think that your estimate for the street value of illegal drug dealing in Chicago is many millions too low. And as for your estimate as to how much activity goes unrecorded in back scratching transactions, I can tell you that my dentist hasn’t paid money for a car repair since he left medical school!)
From all of this we are asked to believe that accuracy of one in a thousand is a purely objective measure. We, the public, wherever we live, are all too gullible.
From my experience, the law of large numbers breaks down at about three sigma because typical error distributions for experimental data have long tails that are definitely not Gaussian. They are usually much higher, e.g., Lorentzian which has an infinite variance.
Why do they expect to to see any “missing” heat at depths of 1800 m? Even UV, which has an absoption length of about 100 m, won’t get anywhere that deep (http://www.lsbu.ac.uk/water/vibrat.html).
There’s no math error, but there’s a huge rules-of-measurement error here.
That calculation shown in your article, Willis, demonstrates how you would try to calculate your theoretical limit of error, not your actual measurement error. The calculation starts from a poorly understood term, that is the presumed error of your forcing. How is the forcing known to such a precision in the first place? What is the relationship between your presumed forcing and your thermocouples taking your readings? I’d wager there is no easy answer to that.
I look at that calculation and I see one possible theoretical limit of error, NOT an actual measurement error. An actual measurement error is always calculated from the measurements themselves or it is meaningless.
Willis,
I see several people have mentioned this but I feel it must be made very explicit:
Accuracy and precision are different from each other. Accuracy is how close a value is to a true physical value. Precision is what resolution in units of measure a value can be read.
Instruments with high accuracy but low resolution are possible but those with low accuracy but high precision are much more likely. (Walmart digital thermometers for example)
Averaging temperature readings of a stable process value over multiple instruments and time can reduce the effects of noise to improve the precision of the measurement. That average thought does not improve the accuracy of the measurement. Accuracy can never be better than instrument error.
A claim that averaging readings from multiple instruments of the same process value reduces instrument error is invalid. This is based on the assumption that instrument errors are random and not systemic. For that assumption to be valid, all instruments must of entirely different design, manufacture, and operating principle.
Also, instrument accuracy must also include the how the instrument is coupled to the process being measured. Is the instrument sensor in a protective well or sleeve? I is the instrument sensor located in a clean area of the process or in a stagnant pocket. Does the instrument sensor itself modify the flow or temperature of the process?
So, in industry, and hopefully science, value accuracy may never be assumed to be better than instrument calibration accuracy. Any claim of precision better than instrument accuracy must include a description of how it was achieved. As an example here we might say +/- .004 degrees plus +/- 0.005 instrument accuracy.
If you look close enough it is incredible the amount of information you can get from a single tree or a single thermometer in a very complex system. Truly incredible.
http://dictionary.reference.com/browse/incredible
All measuring instruments are specified with a given measurement uncertainty. This uncertainty must be included every time a measurement is taken and is a systematic error. It cannot be removed by averaging. Otherwise a cheap meter could replace an expensive one if enough measurements were taken and this is nonsense. Furthermore each measurement taken will also contain some random error. This random error can be reduced by averaging as the random errors will tend to cancel each other out. So, how can I reduce this systematic error? Well, if I used multiple instruments to measure the temperature, each with its own uncertainty then I believe the average of their readings would be closer to the true temperature. However, the ocean is not isothermal and the buoys are not measuring at the same point. Hence averaging their results is meaningless IMO.
Their claimed accuracy is a total fallacy.
The square root of N principal applies to repeated measurements of the same thing and assumes the errors in the measurement are normally distributed. (ie nice gaussian bell , random errors).
There is absolutely no way that measuring a time varying quantity at different times in less than random 3D positions at different depths and different geographical in a medium that has is very significant variation in both depth and latitude/longitude plus seasonal changes can be construed as “measuring the same thing”.
If the temperature sensor in a buoy was placed in a swimming pool and took 10,000 measurements within a short period of time, one may be justified in dividing the uncertainty of one reading by 100 due to repeated measurements .
If you do one reading in every swimming pool in the town you will have 10,000 readings with same uncertainly as you have in one single reading. You may then calculate the mean temperature of all the pools and state that has the same accuracy as one single measurement error.
You have not measured the mean 10,000 times. Only once. You have no reason to divide your single measurement uncertainly by anything except one !
This whole idea of dividing by the root of the number of individual data points is one gigantic fraud.
Any qualified scientist making such a claim is either totally incompetent or a fraudster.
[Fixed the bolding – use angle brackets… -ModE ]
I guess it just goes to show that precision is not the same as accuracy.