Lots of pressure to publish this post early and it is raining this morning here in The Woodlands, so no golf. I checked it over and think it is OK. Here you go!
By Andy May
In Part 1 of this series, we examined the data and analysis that was presented in AR6 to support their conclusion that sea level rise is accelerating. In Part 2 we looked at a serious examination of the observational record for sea level rise over the past 120 years and the modeled components of that rise. We concluded in Part 1 that the statistical evidence presented in AR6 for acceleration was crude and cherry-picked. In Part 2 we saw that the error in both the estimates of sea level rise and in estimating the components of that rise is very large. The error precluded determining acceleration with any confidence, but the data revealed an approximately 60-year oscillation of the rate of sea level rise that matches known natural ocean cycles.
Modern statistical tools allow us to forecast time series, like GMSL (global mean sea level) change, in a more valid and sophisticated way than simply comparing cherry-picked least squares fits as the IPCC does in AR6. Our forecast is based on pure statistics. It is done in the correct way, but not necessarily correct, statistics are like that. We will not know for sure until 2100. That said, let’s do it. If you have a certain kind of nerdy mind, you will enjoy this.
Figure 1 is a plot of the data we will use—the NOAA sea level dataset. Simply looking at it we can tell it is autocorrelated, which means that each quarter’s mean sea level estimate is highly dependent upon the previous quarter’s value. Autocorrelation is important to consider in least squares regression, especially when forecasting time series, but routinely ignored by the IPCC.
Figure 2 plots each sea level estimate versus the previous estimate, this is called a plot of the first lag and the correlation of the two is a measure of autocorrelation. The R2 of the first lag is 0.97, so sea level is very autocorrelated. This is obvious but means that normal least squares linear fit statistics are invalid, the least squares statistics, such as R2, assume that the errors of regression are independent. Least squares, as used in AR6 to show acceleration, is inappropriate with a dataset like this. Most of any given value is heavily dependent upon the previous value. This means the mean-square-error (MSE) will be much too small, causing the error of the fit to be too small. As a result, any least squares line of the data in Figure 1 or any portion of that data is statistically useless, unless the autocorrelation is accounted for.

So how can we forecast GMSL in a statistically valid way? We clearly cannot use least squares and need to apply more advanced techniques. The first step is to remove the autocorrelation from the data, this is normally done by subtracting the previous GMSL value from the current one and progressing in this way throughout the data set. We have done this and show a plot of the result in Figure 3.

The first difference data from GMSL looks pretty good, very much like white noise. This is exactly what we want for valid statistical analysis and forecasting. We will be using an R function called “arima” to create our GMSL forecast, and this function requires three parameters to work, they are called p, d, and q. These parameters tell arima how to condition the input data and build a model that can project valid future values. The plot in Figure 3 shows us that “d” is one. That means taking one difference of adjoining values removes autocorrelation. We also need the data to be stationary, that is the statistical properties do not change with time (left to right). The original dataset (Figure 1) was clearly not stationary, and this is OK, we just do not want the way GMSL changes to be a function of time for this analysis. The R Augmented Dickey-Fuller Test (ADF) function confirms this, as the original dataset has an ADF p value of 0.79, meaning it is non-stationary. The arima p value is not the same as the statistical p test.
The differences plotted in Figure 3 have an ADF p value of 0.01, well below 0.05, the threshold needed to show stationarity. Data are stationary when the distribution over the period being studied is evenly distributed around the mean. That is the distribution, up and down, does not vary significantly with the time axis (x).
Next, we need to derive the arima p and q values. For this we need the ACF (autocorrelation) and PACF (partial autocorrelation) plots shown in Figure 4.

Analyzing the GMSL time series gives us an arima parameter set of (1,1,2) for (p,d,q). We can also run an R function called auto.arima to see what parameters it recommends. We find that it settles on (1,1,2) as well. This is good confirmation that our parameter selection is correct. Figure 5 plots the results.

Figure 5 tells us that the model is successfully capturing the essence of the trends in mean sea level from 1880 through 2020. The model residuals show no trend and they are not autocorrelated. Figure 6 shows the arima forecast from the (1,1,2) model.

Figure 7 is a plot of the forecast from Excel that is easier to read. The forecast we created predicts that GMSL will rise between 148 (6 inches) and 258 mm (10 inches) by 2100. Many researchers call this alarming, but humans have successfully adapted to much higher rates of sea level rise in the past as we can see in Figure 2 of Post 1, and they did so without the technology we have today. When we consider that the average open ocean daily tide range is 1,000 mm or three feet, eight inches of sea level rise over 100 years does not seem like much. In the 20th century sea level rose 5.5 inches, did anyone notice or care, aside from a few researchers?

Conclusions
In the United States we would call the AR6 attempt to convince us that the rate of GMSL rise is accelerating, using adjoining cherry-picked least squares lines “high school,” meaning unsophisticated. Their method is problematic because GMSL is heavily autocorrelated and non-stationary, rendering their cherry-picked least squares fits and least squares statistics invalid.
Our fit, using the R function arima, is at least statistically valid. We specifically corrected for autocorrelation and forced the series to be stationary. We also addressed the minor partial autocorrelation that was left at one quarter and three quarters. The residuals of our model passed both the overall Ljung-Box test and multiple-lag Ljung-Box tests for white noise, meaning the arima model properly captured the 140-year trend in the NOAA sea level data.
Thus, while AR6 cherry-picked periods to support their conclusion that GMSL is accelerating, we reached the opposite conclusion using all the data in a statistically valid way. This does not mean that our forecast is correct, but it does mean that the AR6 speculation that sea level might rise 5 meters by 2150 is extremely unlikely and is best characterized as irresponsible speculation. Our analysis found no statistical evidence of acceleration and produced a linear extrapolation.
While warming of Earth’s surface is clearly the reason land-based glaciers are melting, which does contribute to rising sea level, AR6 provides no evidence the warming is caused by human activities. They use models to infer humans caused it, but unfortunately their models are also not statistically valid as shown in Part 2, here, and by McKitrick and Christy (McKitrick & Christy, 2018). We can all agree that humans probably have some impact on atmospheric warming, but we do not know how much is caused by humans and how much is natural, because we are emerging from the unusually cold Little Ice Age—the “preindustrial” period. Further, as we saw in Part 2, the 30-year rates of sea level rise reveal a distinctly natural-looking oscillation. Glacial ice and ice sheet melting is likely responsible for most of sea level rise, as AR6 states, but the human fraction of that warming might be quite small.
Thus, from a purely statistical point of view, the AR6 claims are childishly invalid. A proper analysis of the data leads to a forecast of roughly 20 cm (~8 inches) of sea level rise by 2100. In the year 2100, our descendants will know who was right.
The data and R code to create the figures in this chapter can be downloaded here. The R code and spreadsheet provide much more detail about the arima forecast, including references not supplied below.
Works Cited
McKitrick, R., & Christy, J. (2018, July 6). A Test of the Tropical 200- to 300-hPa Warming Rate in Climate Models, Earth and Space Science. Earth and Space Science, 5(9), 529-536. Retrieved from https://agupubs.onlinelibrary.wiley.com/doi/full/10.1029/2018EA000401

So why have 80% of coral atoll islands grown in size over the last 40 years?
Of course the young Charles Darwin worked this out on his journey of discovery over 180 years ago.
And Prof Kench has been carrying out these studies and reporting on his findings for decades.
https://www.rnz.co.nz/national/programmes/saturday/audio/2018640643/climate-change-in-the-pacific-what-s-really-going-on
Sorry 40% of islands have grown and 40% are stable,
Thanks
You made a good academic job i guess.
But I still like to look at this graph for acceleration check:
https://tidesandcurrents.noaa.gov/sltrends/sltrends_station.shtml?plot=50yr&id=140-012
Even if it rises an entire 12 inches by 2100, don’t you think we can adjust in 80 years? It’s not a disaster.
This post explains why the claims about rising Global Sea Levels are untrustworthy from a statistical standpoint. But meanwhile, the green PR scare machine uses webpages and desktop publishing tools to frighten people living or invested in coastal settlements. I recently did a tour of several US cities to illustrate how imaginary flooding is promoted far beyond anything appearing in tidal gauge records. For example, Boston:
Example of Media Warnings
Could it be the end of the Blue Line as we know it? The Blue Line, which features a mile-long tunnel that travels underwater, and connects the North Shore with Boston’s downtown, is at risk as sea levels rise along Boston’s coast. To understand the threat sea-level rise poses to the Blue Line, and what that means for the rest of the city, we’re joined by WBUR reporter Simón Ríos and Julie Wormser, Deputy Director at the Mystic River Watershed Association.
As sea levels continue to rise, the Blue Line and the whole MBTA system face an existential threat. The MBTA is also facing a serious financial crunch, still reeling from the pandemic, as we attempt to fully reopen the city and the region. Joining us to discuss is MBTA General Manager Steve Poftak.
The computer simulation of the future:
Imaginary vs. Observed Sea Level Trends (2021 Update)
Already the imaginary rises are diverging greatly from observations, yet the chorus of alarm goes on. In fact, the added rise to 2100 from tidal gauges ranges from 6 to 9.5 inches, except for Galveston projecting 20.6 inches. Meanwhile models imagined rises from 69 to 108 inches. Clearly coastal settlements must adapt to evolving conditions, but also need reasonable rather than fearful forecasts for planning purposes.
Seven US cities presented at
https://rclutz.com/2022/01/27/sea-level-scare-machine-2021-update/
Excellent analysis. Now try it on a tide-gauge-only dataset and a satellite-only dataset. I’m curious to see the result. The NOAA data you used appears to be the infamous hybrid of the two where sea level rise increases suddenly starting in the mid-1990’s when the tide gauge data is discarded and the satellite data is grafted in. There are two distinctly different, essentially linear, slopes in the second plot. If you draw a straight line from 7/22/1926 to 1/1/1995 you can easily see that the trend is linear. And if you draw a line from 1/1/1995 through the end in 2022 the trend is also linear, but at a steeper slope; satellite data (which has consistently shown a higher trend) grafted onto tide gauge data. The tide gauges never stopped measuring but they appear to have discarded the tide gauge in the 1990’s in favor of the higher-trending satellite data. I wonder why.
I’m sorry, but your analysis is not correct (about Andy May’s I can’t say anything because I lack math and stat skill to evaluate it – except that a trial to eliminate autocorrelation through first differentiation looks a bit strange).
” The NOAA data you used appears to be the infamous hybrid of the two where sea level rise increases suddenly starting in the mid-1990’s when the tide gauge data is discarded and the satellite data is grafted in. ”
” The tide gauges never stopped measuring but they appear to have discarded the tide gauge in the 1990’s in favor of the higher-trending satellite data. ”
Where do you have such strange allegations from?
*
Here you see four different evaluations of the PSMSL data, of which mine is the most simple one, because it is based only on tide gauge data, whereas Thomas Frederikse’s and Sönke Dangendorf’s are much more complex:
The red plot is the yearly averaging of the NOAA data made available by Andy May in a zip file.
*
The reason why you believe this 1990 story is that like many others, you confound, in the tide gauge evaluations you read about
In the chart below, you may compare, for all four evaluations above, five-year distant, consecutive linear trends from 1900 till 2015, i.e. from 1900-2015 till 1995-2015:
I hope you see now that it makes few sense to compare lifetime trends with sat era trends when looking at tide gauges. In a previous sea level thread, I posted two links to uploaded pdf files, each containing, for over 300 gauges, life time and sat era trends.
*
What makes me suspicious is not the NOAA data for the satellite period!
It is the entire data, you see it when comparing the right plot with the rest.
NOAA’s evaluation is way above all others (which keep pretty near, except a peak of Frederikse – blue plot – between 1980 and 1990).
My guess: the difference between NOAA and the rest might be due to their anomaly construction technique.
Links to the PSMSL tide gauge trend lists (generated during my ‘Bin’ evaluations)
Lifetime
https://drive.google.com/file/d/1jIAhx1OifHrLF4Pf5YUqCwRenw26Ev3u/view
Sat altimetry era
https://drive.google.com/file/d/19dXIBq8Q7_ZtQm_V7tfcAPCmvEvHiY1P/view
A second way to look at the data is to compute the trends for the four series in the inverse manner, i.e. with the start fixed instead of the end:
You can see that if you were right, the slope of the red plot would have been much steeper from 1995 on.
The difference between NOAA and the rest starts much earlier than the sat era; hence, it can in my opinion not be due to a mix of sat and gauge data in the final series.
I anticipate the reaction:
“They made the past lower to get the present higher”
or so.
Here’s your hoped for response –
“They made the past lower to get the present higher”
Mainly because it’s true –
https://joannenova.com.au/2020/02/acorn-adjustments-robbed-marble-bar-of-its-legendary-world-record-death-valley-now-longest-hottest-place/
Thanks a lot. I’m aware of that Marble Bar story.
People, Andy May has just provided us with an excellent example of how peer review should work. As opposed to the paleo climatological club, Andy provided all his data and methods freely. When people questioned him, he politely answered in full, giving his reasoning. He directed people on different ways to look at the analyses and admitted to the limitations of his study.
We know damned good and well that none of the paleo climatological hockey stick studies revealed their data and methods, but actively fought people seeking such information. None of their studies have been subjected to rigorous statistical analyses and their pal reviewers did not look at nor analyze data and methodologies.
Andy,
Nice analysis of available data sets. This should be the P50 or best case for future forecasts.