The Effect of a Short Observational Record on the Statistics of Temperature Extremes

New open access paper available in Geophysical Research Letters. H/T Judith Curry

Research Letter Open Access creative commons logo

The Effect of a Short Observational Record on the Statistics of Temperature Extremes

Joel ZederSebastian SippelOlivier C. PascheSebastian EngelkeErich M. Fischer

First published: 25 August 2023

Abstract

In June 2021, the Pacific Northwest experienced a heatwave that broke all previous records. Estimated return levels based on observations up to the year before the event suggested that reaching such high temperatures is not possible in today’s climate. We here assess the suitability of the prevalent statistical approach by analyzing extreme temperature events in climate model large ensemble and synthetic extreme value data. We demonstrate that the method is subject to biases, as high return levels are generally underestimated and, correspondingly, the return period of low-likelihood heatwave events is overestimated, if the underlying extreme value distribution is derived from a short historical record. These biases have even increased in recent decades due to the emergence of a pronounced climate change signal. Furthermore, if the analysis is triggered by an extreme event, the implicit selection bias affects the likelihood assessment depending on whether the event is included in the modeling.

Key Points

  • Standard return period estimates of temperature extremes are systematically overestimated in short records under non-stationary conditions
  • The small-sample bias in maximum likelihood estimates is found both for extremes in climate model data and in synthetic data experiments
  • Future analysis should account for the statistical implications of the selection bias if the analysis is triggered by an extreme event

Plain Language Summary

In June 2021, the Pacific Northwest experienced a record-breaking heatwave event. Based on historical data, the scientific community has applied statistical models to understand how likely this event was to occur. However, due to the record-shattering nature of this particular heatwave, the model suggested that reaching such high temperatures should not have been possible. In this study, we evaluate the accuracy of these statistical models in describing the occurrence probability of extreme events. We find that the current models tend to underestimate the occurrence probability and that the bias has become more pronounced in recent years due to climate change. Finally, we assess how the way extreme events are included in the model can also affect the accuracy of estimates.

1 Introduction

The heat wave in late June and early July 2021 in the Pacific Northwest (PNW), with temperatures well above previous records, had substantial impacts on mortality, infrastructure, and the ecosystems in the densely populated region (White et al., 2023). This heatwave shattered the long-standing Canadian temperature record by a margin of 4.6°C, and local temperature measurements exceeded previous long-term records by several degrees centigrade even 1 month before temperatures usually peak in this area (Philip et al., 2022). Both the unprecedented event intensity (Thompson et al., 2022) and the severe impacts with several hundred reported excess deaths (Henderson et al., 2022) stimulated exceptional scientific interest in the meteorological drivers (Bartusek et al., 2022; Mo et al., 2022; Neal et al., 2022; Qian et al., 2022; Schumacher et al., 2022; Wang et al., 2023) and the predictability (Emerton et al., 2022; Lin et al., 2022; Oertel et al., 2023).

There is consensus that anthropogenic climate change is the key driver for aggravating hot extremes globally (IPCC, 2021), which also increases the probability of previous temperature records being broken by large margins (Fischer et al., 2021). The extreme event attribution (EEA) study by Philip et al. (2022) concludes that the first-order estimate of the PNW 2021 event frequency is on the order of “once in 1,000 years under current climate conditions,” and that climate change increased the probability by a factor of at least 150. Such statements require estimates of the exceedance probability p1 of the extreme event under current climate conditions, as well as the counterfactual exceedance probability p0 without anthropogenic warming (Allen, 2003; Stott et al., 2016). Estimating the exceedance probability from past observational or reanalysis data is a central step in the EEA protocol (Philip et al., 2020; van Oldenborgh et al., 2021), which usually entails fitting a non-stationary probability distribution whose parameters are a function of global warming. For heatwave attribution studies, the usual choice is fitting a generalized extreme value distribution (GEV) to annual temperature maxima, with the location parameter being linearly dependent on a global mean surface temperature (GMST) covariate. Thereby the full distribution is shifted in line with GMST changes. For annual temperature maxima, the fitted GEV distribution often has an upper bound. This is a consequence of a negative shape parameter, which determines the tail characteristics of the GEV. For the 2021 PNW heatwave, fitting this GEV model to annual temperature maxima prior to the event generally results in an infinite return period estimate, both using gridded reanalysis spatial mean temperature data (Bartusek et al., 2022; Philip et al., 2022) or individual station data (Bercos-Hickey et al., 2022). Figure S1 in Supporting Information S1 shows how the event intensity of 2021 exceeded the estimated upper bound for 2021 by a large margin. By including the event in the GEV fit, Philip et al. (2022) and Bartusek et al. (2022) obtained a finite return period estimate (which is an inherent consequence of the estimation method). Whether or not to include the event is an unresolved question in EEA (Philip et al., 2020), which would require addressing the inherent selection bias, as the analysis is conducted due to the event itself. Also, risk assessment studies are often triggered by record-shattering extreme events that tend to represent outliers, thus they are subject to the same selection bias. The statistical implications of an extreme event trigger on metrics relevant for risk assessment are discussed by Barlow et al. (2020) and Miralles and Davison (2023).

This study aims at assessing the robustness of event probability estimates based on observational data in light of the challenges posed by the PNW 2021 heatwave attribution effort, and which has been observed for several other recent events. To this end, we evaluate the approach by using climate model large ensemble (LE) and synthetic GEV data as a test bed. In Sections 2 and 3, we provide information on the data sets and the statistical procedure used for the evaluation. In Section 4, return level and return period estimates are assessed against the reference values obtained from the pooled ensemble data set, and the implications of the selection bias are discussed.

2 Data

We here use climate model data of an 84 member initial condition LE of the fully coupled Community Earth System Model version 1.2 (CESM1.2; Hurrell et al., 2013) and a 90 member LE of CESM version 2 (CESM2; Danabasoglu et al., 2020). The CESM1.2 ensemble consists of 21 members that cover the historical and future period from 1850 to 2099, of which three additional members each are branched off in 1940. All members follow an RCP8.5 forcing scenario after 2006. The CESM2 ensemble members cover the period from 1850 to 2100 and are forced with an SSP3-7.0 scenario after 2015. The results in the main text are primarily based on CESM1.2 data due to data availability in the initial project phase.

From these models, annual maximum daily average temperature (Tx1d) was retrieved for a domain of 45–50°N, 122.5–120°W (Figure S2 in Supporting Information S1), which is a spatial subset of the domain defined by Philip et al. (2022), but is adjusted to the shared model resolution of 2.5° after interpolation onto a common grid. The GMST covariate, expressed in anomalies against the 1981–2010 average, is obtained from the respective climate models. In accordance with the EEA protocol (Philip et al., 2020), a 4-year running mean low-pass filter is applied to smooth unforced inter-annual variability (Figure S3 in Supporting Information S1). Text S1 in Supporting Information S1 summarizes the respective implementation.

3 Methods

We evaluate 100-year return levels estimated from individual realizations against a reference 100-year return level, which is estimated from the pooled data of all ensemble members or explicitly known in the synthetic data experiments. Given the non-stationarity of the data and the model in Equation 1, the return level estimates are always conditional on the corresponding GMST covariate. Unless further specified, in the following text, “(reference) return level” refers to the 100-year (reference) return level, and “return period” to the estimated return period of the 100-year reference return level.

3.1 Statistical Model

Throughout the study, we model the annual temperature maxima yTx1d as realizations of a non-stationary GEV distribution

grl66326-math-0001 (1)

where the location parameter μ is a linear function of the smoothed GMST covariate xGMST

grl66326-math-0002 (2)

Following the EEA protocol for heatwave attribution, we here assume a stationary scale parameter σ and shape parameter ξ (Philip et al., 2020; van Oldenborgh et al., 2021). Amplifying processes like land-atmosphere interactions can further increase the variability in heat extremes, thus for example, Bartusek et al. (2022) additionally assume a non-stationary scale parameter.

For a GEV distribution of annual block maxima, the quantile associated with an exceedance probability p is referred to as the return level zp, which would, on average and under stationary conditions, be exceeded every 1/p years. Therefore, the 100-year return level corresponds to the 99% quantile and is shifting with the full distribution as the location parameter changes with GMST, as expressed in Equation 2. We further estimate the return period r = 1/p for a given 100-year reference return level zref,p = 1% as a function of the estimated GEV parameters:

grl66326-math-0003 (3)

We put specific focus on estimated return periods, as these directly determine the event probability under current and counterfactual climate conditions and, therefore, attribution metrics like the probability ratio PR = p1/p0 (cf. Section 1).

3.2 Evaluation Steps

We assess return level and return period estimates inferred from individual ensemble members, which are, therefore, only subject to (random) sampling uncertainty and differences due to internal climate variability. In the first step, we estimate the reference values from a non-stationary reference GEV model, which is fitted to the entire ensemble data set (pooling all members in a LE), against which the estimates from individual ensemble members are assessed. Text S2 in Supporting Information S1 summarizes the evaluation of the reference return level estimates concerning the questions of (a) whether a block size of 1 year yields stable results and (b) whether the model formulation in Equation 1 sufficiently captures the complexity in the data. In short, the reference return level values of the base model in Equation 1 are found to be consistent with such derived from more complex extreme value models and for larger block sizes.

In a next step, for each individual ensemble member, a GEV model is fitted to either the complete available Tx1d time series (i.e., 1850–2099, referred to as long estimation period), but also to 71-year sub-periods (e.g., 1950–2020, referred to as short estimation period), as shown in Figure 1a. The former is only used to infer the effect on GEV parameter estimates when all historical and future data are available (the impact on the trend parameter μGMST is shown in Figure 1b). The return level and return period estimates are then evaluated against the reference values at the smooth ensemble-mean GMST level of the year following the estimation period (e.g., 2021, see Figure 1c).

Figure 1 PowerPoint (a) Individual CESM1.2 ensemble member Tx1d time series (dots) and estimated, time-varying location parameter (solid lines) and 100-year return levels zp = 1% (dot-dashed lines), and GEV densities for the year 2021. Black color refers to the reference GEV model, dark magenta to long estimation period GEV fits (1850–2100), and light magenta to short estimation period GEV fits (1950–2020). (b) Tx1d data as in (a) but as a function of the covariate xGMST. The line colors again refer to the reference (black), long (dark magenta), and short (light magenta) estimation periods. (c) Enlargement of the dashed box in (a) for all three 2021 densities, showing the 100-year return level (the reference return level zref,p = 1% is marked with a vertical dashed line), and the estimated return period  as shaded area. (d) Ordered maximum likelihood 100-year return level estimates (dots) and non-parametric bootstrap confidence intervals (CI) (shaded area, filled white dots highlight estimates where the CI does not cover the reference return level), and (e) corresponding return period estimates. Box plots summarize the respective distributions (box: 25%–75%, whiskers: 1%–99%). Corresponding Bayesian estimates and long estimation period results, and CESM2 results and are shown in Figures S4 and S5 of the Supporting Information S1.

To assess the effect of the estimation procedure, parameters were estimated with classical maximum likelihood (ML) but also with Bayesian posterior sampling. An excellent introduction to non-stationary extreme value modeling covering both estimation approaches is provided by Coles (2001). For the former, 95% confidence intervals (CI) are retrieved using non-parametric bootstrapping, the default approach suggested in the EEA protocol (Philip et al., 2020). Bayesian point estimates are calculated as posterior means, and central credible intervals (also abbreviated CI), that is, 2.5%–97.5% quantiles of the posterior distribution, are obtained as uncertainty measures for the Bayesian estimates. Text S1 in Supporting Information S1 provides further background on the technical implementation of the estimation procedure.

Additionally, evaluations are conducted in synthetic data experiments, where samples are drawn from a pre-defined GEV distribution as in Equation 1. From these, estimates are obtained by re-fitting the same GEV model, providing an independent data set to assess return level and return period estimates, but for data that follows the assumed GEV model by design. The “static” synthetic data experiment resembles the PNW 2021 attribution study (with a fixed estimation period 1950–2020), with varying GEV reference parameters and GMST covariate time series (resembling the LE setup). Furthermore, two additional synthetic data experiments are conducted to evaluate the quality of ML estimates over time (“transient” experiment) and for increasingly large estimation data sets (“sample size experiment”). Detailed methods and results are provided in Text S3 of the Supporting Information S1 and are referenced throughout the following results section.

4 Results and Discussion

Based on LE climate model data, we first demonstrate a systematic underestimation of the return level and a corresponding overestimation of the return period for an estimation period 1950–2020, analogous to the PNW 2021 attribution study. Later, we expand the analysis to further estimation periods to explore the temporal evolution of the bias. We also outline the effect of a potential selection bias and the consequences of including or not including the extreme or record-breaking event when estimating its return level.

4.1 Systematically Underestimated Return Levels of Temperature Extremes

The right panel in Figure 1c illustrates a GEV model fitted with ML to data of an individual climate model ensemble member for the short estimation period 1950–2020, in which the corresponding return level is lower than the reference return level calculated from all members. Therefore, the estimated return period is much higher than what it should be, that is, 100 years. While this specific ensemble member shown in Figures 1a–1c is just one illustrative case, a key result of our study is that the majority of estimated 100-year return levels fall below the reference return level. This can be seen in Figure 1d, where return levels estimated from individual ensemble members are shown in ascending order along with their actual reference return level (horizontal dashed line, the box plot on the left summarizes the distribution of estimates). Our results show a negative bias in ML estimates of high return levels due to the short estimation period. In principle, for individual members, this result could also arise due to random sampling uncertainty, but the reference value should then fall within the respective CI at the rate of the corresponding confidence level. The CI in Figure 1d, however, show that the fraction of CI not including the reference value amounts to 26.9%, a ratio much larger than the 5% expected for 95% CI, which is referred to as under-coverage. Furthermore, if not covered by the CI, the reference value should fall evenly above and below the CI, which is not the case. Bayesian CI, on the other hand, tend to have too high coverage (Figure S5a in Supporting Information S1), as for all the estimates, the respective 95% CI cover the reference return level.

The systematic underestimation of 100-year return levels in relatively short temperature records implies an underestimation of the exceedance probability of the 100-year reference return level, that is, an overestimation of the return period, as visualized in the right panel of Figure 1c. Figure 1e shows a clear correspondence between return period and the respective 100-year return level estimates (Figure 1d). In 41% of the cases, the return period of an event with a “true” return period of 100 years is estimated to be larger than 1,000 years. In 14% of ensemble members, a 100-year event is even considered to have zero exceedance probability or infinite return period. These results indicate a substantial risk of overestimating the return period when the record of observations used for the statistical analysis is limited. This result has serious implications for adaptation and planning. Temperatures that are estimated to be never reached may actually be exceeded with a 1% probability in a given year (i.e., have a 100-year return period). The large sensitivity of return period estimates is a consequence of the bounded short-tail nature of heat extremes, such that return periods can be overestimated by orders of magnitude if not reaching infinity. In the following sections, we further discuss how this bias evolves for different estimation periods and methods.

4.2 Temporary Bias Amplification

The previous paragraphs outline the underestimation of return levels and the overestimation of return periods for GEV fits in individual ensemble members and the estimation period 1950–2020, evaluated in 2021. In the following, we investigate whether the underestimation of the return level is also influenced by the fact that 2021 falls in a period of rapid warming. Figure 2 shows the distribution of return level (Figure 2a) and return period (Figure 2b) estimates for overlapping 71-year estimation periods (1850–1920, 1851–1921, etc.), aggregated per decade (shaded background), and explicitly for the 1950–2020 estimation period (box plot). We find a roughly constant underestimation in return levels and an overestimation in return periods for estimation periods ending before 1990. For periods ending after 1990, the biases substantially increase temporarily and diminish again toward the end of the century. We further observe a strong increase in the fraction of events whose intensity exceeds the estimated upper bound of the GEV distribution derived from the preceding 71-year estimation period, analogous to the PNW 2021 event (Figure 2c; Text S4 in Supporting Information S1).

Figure 2 PowerPoint Distribution of (a) differences in estimated 100-year return levels relative to the 100-year reference return level and (b) estimated return period of the 100-year reference return level. Box plots show the distribution for the estimation period 1950–2020 across ensemble members (box: 25%–75%, whiskers: 1%–99%). Analogous, the shaded background marks the distribution for estimation periods shown on the abscissa, aggregated per decade (purple shading: 25%–75%, gray shading: 1%–99%). (c) Empirical annual occurrence probability of events exceeding the upper GEV bound estimated from the previous 71-year period, aggregated over 20 years. CESM2 results for (a–c) are shown in Figure S6 of the Supporting Information S1. (d–g) Absolute differences estimates of the shape parameter  (y-axis), and the trend parameter  and scale parameter  (x-axis) relative to their reference value (combined 1950–2020 CESM1.2 and CESM2 estimates are shown). The color of the points refers to the corresponding return period value (as in diamonds mark infinite return periods), the arrows show the direction of the respective gradient. Yellow density contour lines illustrate the distribution of estimates in the “static” synthetic data experiment, with the corresponding yellow gradient vector. Box plots on the side summarize the respective marginal distributions (magenta for the large ensemble data points, yellow for the synthetic data), and the Pearson correlation value R is provided in the bottom left corner .

The time dependence and especially the steep increase of return level and return period biases for periods ending after 1990 are likely to be caused by the fact that a majority of Tx1d data used for the model fit are close to stationary with a weak warming trend. In a period of rapid accelerating warming, non-stationarity and further climate and statistical issues may affect the statistical model and associated biases: Nonetheless, an analysis of the estimated GEV parameters reveals that the temporary overestimation of return periods cannot be attributed to an underestimation of the trend parameter ; first, the estimates are largely unbiased with respect to the reference value (horizontal magenta box plot in Figures 2d and 2f), and second, the effect of  on the estimated return period is negligible, when compared to the effect of an under- or overestimation of the shape parameter  (the return period increases almost exclusively along the y-axis with decreasing values of , visualized by the black gradient arrow in Figures 2d and 2f). In contrast, the scale parameter  is subject to a negative bias (horizontal magenta box plot in Figures 2e and 2g) and an underestimation of  promotes overestimating the return period (tilted gradient arrow in Figures 2e and 2g), but an underestimation of  is often compensated by an overestimation of  (and vice versa, the correlation being R ≈ −0.5). The attribution of the temporary return period overestimation bias to specific GEV parameters is therefore not straightforward, but we conclude that a necessary condition is an underestimation of the shape parameter . The distribution of estimates from the “static” synthetic data experiment (yellow contours) confirms the relationship between estimates (similar correlation values) and the relative effect strength of individual parameters on return period estimates (yellow gradient arrows) found in LE-based estimates.

A temporary return period bias is also not present in synthetic GEV data generated in the “transient” experiment with a non-stationary forced response GMST covariate (Text S3.3 in Supporting Information S1), where the biases are found to be constant over time (Figure S7 in Supporting Information S1). This discrepancy between climate model and synthetic GEV data has two potential explanations; the statistical model is less capable of successfully detecting and accounting for the non-stationarity since the Tx1d data from LE data is not perfectly GEV-distributed. Alternatively, the temporal variation in the biases might be due to model misspecification, as additional forcing agents (modes of internal variability, local effects of volcanic and anthropogenic aerosols, as considered by Risser et al. (2022)) are not accounted for in the GMST-dependent GEV model formulation in Equation 2. Such additional forcing agents could be accounted for by including further covariates, however, this approach comes at the cost of lower regional generalizability (if very specific forcing agents are considered for the respective location) and higher model complexity (more parameters have to be estimated).

In summary, the GEV model correctly identifies the positive trend in Tx1d intensity also in today’s climate with a high warming rate after a period of weaker warming before 1990. Nonetheless, the temporary overestimation of return periods indicates an increased sensitivity to discrepancies between physical climate variables and statistical model specifications. This has implications both for risk assessment and EEA. The underestimation of return levels and the under-coverage of the respective CI can result in an underestimation of the risk associated with extreme heatwave events. The tendency to overestimate the return period of observed extreme heatwave events may fuel the impression that seemingly impossible heatwave extremes are currently clustering at an unprecedented rate.

4.3 Sensitivity to Methodological Choices

In the following, we briefly evaluate whether the bias in return level and return period estimates is sensitive to the underlying estimation method. Comparing the offsets in return level and return period estimates, Figure 2 suggests smaller biases in Bayesian (right panels) compared to ML estimates (left panels). The agreement in ML and Bayesian GEV parameter estimates across ensemble members is extremely high (with correlations above 0.96 for all parameters, as shown in the diagonal panels of Figure S8 in Supporting Information S1), thus there are no fundamental differences between ML and Bayesian model fits. However, Bayesian scale and shape parameters are estimated slightly higher, which partly explains the smaller bias in return level and return period estimates (Figures 2f and 2g).

This investigation is complemented by two synthetic GEV data experiments, where the data follows a GEV distribution by construction. In the “static” synthetic data experiment (Text S3.1 in Supporting Information S1), where ML and Bayesian estimates are directly compared, we also find that ML return period estimates are subject to a systematic offset; 50% of ML return period estimates for the reference 100-year return level are larger than 200 years (Figure S9a in Supporting Information S1). Thus, the biases found in climate model data are clearly not only due to the data not truly following a GEV distribution, but are even found when the underlying data is GEV distributed by construction. Bayesian return period estimates, on the other hand, are largely unbiased, but still, return period estimates of infinity may be reached in a few individual cases. The underlying Bayesian scale and shape parameter estimates are again slightly higher than the corresponding ML estimates (Figures 2f and 2g; Figure S10 in Supporting Information S1), which leads to the alleviated mean bias, and thus confirms the pattern found in estimates derived from climate model data (Figure S8 in Supporting Information S1). Roodman (2018) and Bücher and Zhou (2021) discuss the sub-asymptotic “small-sample” bias of ML derived GEV parameter and return level estimates, which we further investigate in a second “increasing sample size” synthetic data experiment (Text S3.2 in Supporting Information S1). The “small-sample” bias primarily affects the shape parameter, and its direction also depends on the value of the latter, that is, a negative shape is systematically underestimated, and vice versa. This bias propagates and consequently also affects return level and return period estimates (Figure S11 in Supporting Information S1).

4.4 The Implications of the Selection Bias

In the following, we here briefly discuss the implications of the implicit stopping rule, that is, the fact that many scientific papers or risk evaluations are motivated by a very extreme event in observations or model data that needs to be put into context. In such a study, it is often unclear whether or not the respective event should then be used for the fit. This question is particularly relevant for record-breaking or record-shattering events at the end of the observational record, that is, events that are addressed in EEA studies and often initiate a reevaluation of previously estimated climate risk. The GEV analysis assumes that the extreme event in question is an independent and identically distributed sample of the same distributions as the previous observations. However, as the analysis was triggered by the extremeness of the event to be evaluated (a so-called stopping rule is applied), both including or excluding the event will induce a bias (Barlow et al., 2020).

We analyze the effect of including or not including the event of interest at the end of the records by assessing how it affects return level and return period estimates. To this end, we search for events in the LE data sets that have a “true” return period of at least 100 years (based on the reference GEV model). Then we compare the estimated return level and return period estimated from two 71-year estimation periods, where one does and the other does not include the event in question (e.g., for an extreme event in 2051, we compare estimates of the two periods 1980–2050 and 1981–2051). Adding 1 year with a very extreme event strongly changes the GEV fit, as Figure 3 confirms the expectation that including the event results in an overestimation of the return level and an underestimation of the return period. This reversal in the sign of the biases has the interesting effect, that the relative biases in Bayesian estimates are now larger than in ML estimates. This means that Bayesian estimates are not more accurate in both scenarios, and should thus not be considered superior per se. Corresponding estimates from synthetic data of a stationary GEV reveal a similar pattern, but biases are generally smaller.

Figure 3 PowerPoint(a) Differences between the estimated and reference 100-year return level, and (b) the estimated return period of the 100-year reference return level. Box plots differentiate between not including (left) and including (right) the extreme event (with reference return period larger than 100 years) in the GEV fit. Estimates from large ensemble data (pooled CESM1.2 and CESM2) are shown in magenta, and estimates from synthetic data generated by a stationary GEV distribution are shown in yellow.

Barlow et al. (2020) discuss the consequences of the stopping rule in extreme event analysis and propose an adjustment of the likelihood function. With minor adjustments and few considerations regarding the triggering threshold, the method of Barlow et al. (2020) could allow for a more stringent handling of the implicit stopping rule in EEA and risk assessment, as demonstrated by Miralles and Davison (2023) for the PNW 2021 heatwave attribution.

5 Conclusions and Outlook

In this study, we assess different challenges in accurately estimating high return levels and the return period of extreme heatwave events on the basis of short observational records. This evaluation is motivated by the record-shattering PNW 2021 heatwave, where the respective extreme value model derived from observations up to the year before the event suggested an infinite return period or zero probability of reaching the observed event intensity. This raised the question of whether the non-stationary statistical approach widely used in risk assessments, adaptation planning, or EEA, is reliable.

We find that heatwave return levels estimated from limited records are systematically underestimated, which is further aggravated by the fact that the associated CI also underestimate the associated uncertainty range. This bias can result in an underestimation of the associated heatwave risk, which is relevant for adaptation and infrastructure planning. It further translates into an overestimation of the return period of observed extreme events, especially under the strong warming conditions of recent decades. Even though LE climate model data provide a robust test bed, we further verified the evaluation results in targeted synthetic data experiments in which the data is GEV distributed by construction. We identify the ML sub-asymptotic “small-sample” bias in GEV parameter and return level estimates, which only vanishes for sample sizes unattainable in the context of current observational heatwave records.

The offset in Bayesian return level and return period estimates is substantially smaller if the triggering extreme event is not included. Aside from more fundamental considerations (Mann et al., 2017; Shepherd, 2021; Stott et al., 2017), certain practical advantages would call for Bayesian extreme value modeling; more transparency regarding model choices (e.g., priors used for the shape parameter), or the ability to use a joint statistical modeling framework for observational and climate model data (Ribes et al., 2020; Robin & Ribes, 2020) and to avoid artifacts caused by the boundedness of the GEV distribution (Castro-Camilo et al., 2022). We also strongly advocate thoroughly considering the selection bias whenever future analyses are triggered by an extreme event, for example, through implementing the measures suggested by Barlow et al. (2020) and Miralles and Davison (2023). Furthermore, alternative ML-based approaches to derive CI for GEV parameters and return levels should be considered; for example, profile likelihood CI could help to reduce the under-coverage found for return level CI (Coles, 2001).

In summary, our results show that the systematic overestimation of the return periods is largest for short observational periods, in today’s climate when a rapidly warming period follows a period of less or no warming, and for studies focusing on record-breaking and record-shattering temperature extremes. Thereby it affects many recent studies and reports, and it is crucial that those biases are taken into account when putting such events into a climate context, in all fields from adaptation and planning to communication of event attribution.

Acknowledgments

J.Z. and E.M.F. acknowledge funding from the Swiss National Science Foundation within the project “Understanding and quantifying the occurrence of very rare climate extremes in a changing climate” (Grant 200020_178778). S.S. and E.M.F. acknowledge funding from the European Union H2020 project “Extreme events: Artificial intelligence for detection and attribution” (XAIDA; Grant 101003469). O.P. and S.E. acknowledge funding from the Swiss National Science Foundation Eccellenza Grant “Graph structures, sparsity and high-dimensional inference for extremes” (Grant PCEGP2_186858). We thank C. Barnes and F. Otto for the extensive discussion of the results and their constructive and detailed feedback on the model evaluation. We further thank A. Ferreira and L. Belzile for their statistical input on the small-sample bias. The analysis was carried out in R (R Core Team, 2022), thus we thank all contributors for the numerous R packages crucial for this work.

Data Availability Statement

Pre-processed large ensemble data and estimated GEV data are available at https://doi.org/10.3929/ethz-b-000619286. R code (R Core Team, 2022) for pre-processing of data, GEV model estimation from large ensemble data, evaluation of return level and return period estimates, and simulation experiments is available on https://doi.org/10.5281/zenodo.8118283. All original ERA5 reanalysis data (Hersbach et al., 2020) used in this study are publicly available on  https://doi.org/10.24381/cds.adbb2d47.

Supporting Information

FilenameDescription
2023GL104090-sup-0001-Supporting Information SI-S01.pdf 10.5 MBSupporting Information S1

Please note: The publisher is not responsible for the content or functionality of any supporting information supplied by the authors. Any queries (other than missing content) should be directed to the corresponding author for the article.

References

References From the Supporting Information

3.3 6 votes
Article Rating

Discover more from Watts Up With That?

Subscribe to get the latest posts sent to your email.

53 Comments
Inline Feedbacks
View all comments
Scissor
September 4, 2023 2:26 pm

Nine times out of ten, this would have been an important paper. Still, bias shows.

Rud Istvan
September 4, 2023 2:51 pm

This is an obvious conclusion to anyone a bit familiar with math stats.
Or, to explain without any math, for long term trends you need long term data.
Said differently, you cannot estimate long term trends from short term data,

Now why that simple provable truth is oblivious to warmunists is also obvious.

Reply to  Rud Istvan
September 4, 2023 3:50 pm

I am somewhat disturbed by papers like these. They do not use real (observational) data, they presume a consensus view (that the world is warming because the models they use say it is), and they employ obscure statistical methods that are barely understandable to middle-tier scientists (elitist in the real sense of the word).

Having read a download of the paper (and not understood the maths) I’m left with the feeling that the paper is more about the scientists and their mass of graphs, than it is about the weather.

Observational data is likely to show for instance, that maximum temperature depends on rainfall (or lags thereof depending on the timestep); that sites and instruments have not remained the same and that when corrected for inhomogeneities, there is probably no real trend that cannot be explained by rainfall/site changes. Also, frequency analysis of daily data (counts or % within classes) or percentile analysis provides a valid framework for examining the likelihood (and magnitude) of extreme events while also exposing effects of site and instrument changes (and rainfall) on extremes. See for example Figure 6 and Figure 7 and Section 3.2.4 in https://www.bomwatch.com.au/bureau-of-meteorology/charleville-queensland/  

I am surprised that some who visit WUWT do not grab some data for some of these places, which are far away from where I sit in Oz, and have a look and see what long- and short-term weather data say about extremes. Maximum temperature during the 1930s dust-bowl era must have been dreadful for example (https://drought.unl.edu/dustbowl/). Oh wait, they mainly rely on post-1950 off-the-shelf data and then it is not real data.
 
All the best,
Dr. Bill Johnston
http://www.bomwatch.com.au 

Geoff Sherrington
Reply to  Bill Johnston
September 4, 2023 7:06 pm

Well stated, Bill.
Geoff S

don k
Reply to  Bill Johnston
September 5, 2023 4:58 am

Bill: I don’t disagree with your opinions, but I think perhaps a point is being missed here. This is not, (I think) an article about climate. It seems to be an article about statistics that says (I think) something along the line of “Look folks, if you apply statistics to weather phenomena (or anything else in the real world?) using the standard method X, you will underestimate the number and magnitude of outlying values. … And here’s why.” That sounds to me pretty much like what Nassim Taleb (The Black Swan, Fooled by Randomness) told us — albeit more entertainingly and less rigorously. In that context, the use of an aggressive climate model and synthetic data MIGHT be reasonable. If you can’t compute accurate probabilities with synthetic data that you might understand why are you surprised that real data gives even worse results?

That said, I suspect that the article would benefit from a rewrite by someone clever enough to understand the subject matter and possessing better communication skills than the authors. Assuming such an individual exists. That hypothetical communicator is not me. I found large sections of the paper made my eyes glaze over.

Reply to  don k
September 6, 2023 12:10 am

Thanks don k.
 
They start their abstract with “In June 2021, the Pacific Northwest experienced a record-breaking heatwave event”. They did not say that temperatures experienced over x days were the warmest since 19xx, which would provide the reader with some measure of their time-window. Authors are from Switzerland, I’m in OZ, ‘did they consider the 1930s’ immediately came to my mind. Also is there a way of checking for artifact data.
 
Climate data is skewed, rainfall in particular, but temperatures also. Rainfall cannot be less than zero for instance, but the highest rainfall recorded over the last x-years can be exceeded. Although less prominent, distributions of daily maximum temperatures tend to be bimodal and long-tailed.
 
It stands to reason (from the Summary) that “the return period of low-likelihood heatwave events is overestimated, if the underlying extreme value distribution is derived from a short historical record”.
 
Looking at the first two references cited by Zeder (White et al., 2023 [16 Authors] and Philip et al., 2022 [27 Authors]) the paper is clearly about supporting the IPCC in blaming extremes on human emissions. White et al., 2023 present an overview of consequences and refer to data for Lytton BC. I don’t know anything about how the climate of N-America behaves but following the links, daily data for Lytton appear to commence earlier than 1932. I did not persist in trying to develop a composite daily dataset.

White et al say “Lytton, a small town in an arid mountain valley just north of 50°N in the lee of the BC Coast Range …. was tragically destroyed by wildfire on 30 June 2021” which is the day after the record was set. Also, that “The new record temperature was reportedly the hottest worldwide temperature recorded north of 45° latitude, and hotter than any recorded temperature in Europe or South America”. The current weather station (at Lat 50.2245, Long -121.5819) is located in a small clearing on a sand-bench within 160m of the river, and appears to have been sprayed out before 2019.   
 
I’m not saying it was not warm, it clearly was, but was warmth exaggerated by weather station moves and changes and did they consider the 1930s? I could find metadata for Lytton.

Reading through White et al raises the question: to what extent did drought, and wildfires possibly due to arson, lack of preparedness, lack of on-ground hazard reduction etc. contribute to data that was measured at Lytton and other places around that time? Has there been a change in forest management like there has in Australia?
 
Around the world, and in Australia teams of climate scientists have been sitting in front of their models just waiting for the next ‘record’, somewhere, anywhere so they can leap on the wagon. Analysis of Australian whether data using BomWatch protocols convincingly shows that the staged replacement of 230-litre wooden Stevenson screens with 60-litre ones, and those with 60-litre plastic screens, combined with automatic weather stations has resulted in upward bias of daily temperature measurements. The effect is most pronounced on warm days and in some cases while not obvious in averages, the tails of daily data distributions (which is where the records and artifacts are) have shifted warmer.
 
Philip et al 2022 say in their first paragraph “The anomalies relative to the daily maximum temperature climatology for the time of year reached 16 to 20 ◦C (Fig. 1c and d)”. Is it plausible that daily maximum temperatures could exceed the June average for Lytton BC of 24.1 degC by almost 100% (https://www.eldoradoweather.com/canada/climate2/Lytton.html)?
 
Philip et al  say under Observational Data “The dataset used to represent the heat wave is the ERA5 reanalysis (Hersbach et al., 2019) from the European Centre for Medium-range Weather Forecasts (ECMWF) at 0.25◦ resolution”. They also used homogenised data for Portland (1938 to 2021), Seattle (1948 to 2021) and Vancouver (actually New Westminster, which starts in 1875 but contains data gaps in 1882–1893, 1928 and 1980–1993). Also, “As a measure of anthropogenic climate change we use the global mean surface temperature (GMST)…”. Then, they used a double-hand-full of models, and in direct proportion to its complexity the paper goes downhill from there….  
 
As papers like these with impossibly large numbers of Authors, multiple data sources, models, reanalysis, satellites, and >100 references are impossible to digest or verify, in my view, they should be rejected on that basis.
More importantly, could someone really start from first principles, repeat the process and arrive at the same conclusions?
 
All the best,
Bill Johnston
http://www.bomwatch.com.au   

September 4, 2023 3:16 pm

Why would anyone want to post the entire paper, bibliography included, when it is open?

September 4, 2023 3:32 pm

Even an idiot like me realizes that when you make up an arbitrary definition of climate, nature will eventually kick you in the arse. Are you listening Schmidt?

Janice Moore
September 4, 2023 3:32 pm

This paper is a classic example of precision shoving aside ACCURACY.

For all its correct statistics fine-tuning the the computer simulations of climate, its underlying premise,

“There is consensus that anthropogenic climate change is the key driver for aggravating hot extremes globally…,”

is WRONG.

There is no data proving that human CO2 emissions are a controlling cause of meaningful shifts in the climate zones of the earth.

As Richard Feynman would firmly tell us:

“If your hypothesis disagrees with the DATA,

it’s WRONG.”

fah
Reply to  Janice Moore
September 4, 2023 3:41 pm

Feynman was spot on about so many things and said so often. Another quote I like (allegedly from Rutherford, but the attribution is murky), which I will paraphrase a bit for purpose: If your data needs statistics, you ought to get better data.

(The actual alleged quote is “if your experiment needs statistics, you ought to have done a better experiment,”)

Whoever actually said it, they were spot on too.

Rud Istvan
Reply to  fah
September 4, 2023 4:41 pm

Rutherford, not Feynmann

fah
Reply to  Rud Istvan
September 4, 2023 5:04 pm

Yep. That is what I said. It is attributed to Rutherford, but folks have trouble finding a record of him saying exactly that. However, he could have said it easily, given that his most famous experiment likely did not need much in the way of statistics, since he was out to see if the alphas would go straight through the gold or have significant backscatter. Any backscatter at all would have been fairly good demonstration that the plum pudding model of the nucleus was wrong. It gets trickier if the competing theories are closer together and the experiment has a lot of trouble, e.g. the light deflection by the sun experiment, which was tough as an optical observation but onset of radio astronomy fairly conclusively ruled out Newton over Einstein. Even with better experiments, it has been hard to rule out competing gravity theories like Brans-Dicke (or others), that have mushable parameters that can make the results closer to Einsteing.

fah
Reply to  fah
September 4, 2023 5:14 pm

Well, it won’t let me correct the spelling of Einstein. Says I am posting too fast, slow down. My apologies, Albert.

Reply to  fah
September 4, 2023 4:45 pm

Ernest Rutherford 1871–1937
If your experiment needs statistics, you ought to have done a better experiment.

Geoff Sherrington
Reply to  Janice Moore
September 4, 2023 7:06 pm

Well stated, Janice.
Geoff S

Janice Moore
Reply to  Geoff Sherrington
September 4, 2023 7:54 pm

Thank you! 😀

Reply to  Janice Moore
September 5, 2023 11:14 am

Excellent point. Even if the conclusion is correct, there should be validation using valid experimental data.

sherro01
September 4, 2023 3:52 pm

Global ocean surface temperature data show few values above 30C, as if there is a mechanism preventing natural events causing higher temperatures. How does this paper cope with that type of constraint? Should analysis scenarios assume some probability of a cutoff over land as well?
Some researchers like Richard Willoughby are presenting mechanisms that might explain a limit over oceans, bringing reality and speculation closer.
…..
Why do the authors use RCP 8.5 when many authors have described it as unlikely to happen in real life?
……
It would be helpful to test a common assumption that an increase in base temperature (as in many years in the 2000s are reportedly hotter than in the 1900s) will be seen similarly in the magnitude of extremes. (Many stations I have studied do not have such a corresponding rise in extreme values).
……
As always, proper estimates of uncertainty (like Pat Frank has done recently for liquid in glass thermometers) are required for data so filled with errors and adjustments as historic temperatures are.
….
My initial take – this paper is appreciated for advancing understanding, but it might not yet provide a comprehensive analysis. More work needed. Geoff S

Reply to  sherro01
September 4, 2023 4:08 pm

While they say (in plain English) “However, due to the record-shattering nature of this particular heatwave, the model suggested that reaching such high temperatures should not have been possible. They also intimate that “current models tend to underestimate the occurrence probability and that the bias has become more pronounced in recent years due to climate change.
 
My understanding of those words is that, the record (meaning the experience of the heatwave) was ‘real’, and, therefore the models should be ramped-up. They also more-or-less conclude that “the way extreme events are included in the model can also affect the accuracy of estimates”. I believe accuracy of estimates is a misnomer. How can a model produce accurate estimates in the absence of real-data comparisons?
 
Or have I got it totally wrong?

Cheers,

Bill

P.s. going fishing, see you later!

Reply to  Bill Johnston
September 5, 2023 10:07 am

I hate the formulation “due to climate change”, or “caused by climate change,” because it treats climate change like a force that causes the weather to change. I don’t understand why this sloppy usage was ever allowed to take hold — it should have been slapped down the first time it was ever used.

Reply to  sherro01
September 4, 2023 10:08 pm

Why do the authors use RCP 8.5 when many authors have described it as unlikely to happen in real life?

Because if they didn’t use the highly unlikely RCP8.5, they couldn’t see the wood for the trees.

Dave Andrews
Reply to  Redge
September 5, 2023 8:39 am

And note even the BBC acknowledges that RCP 8.5 is unlikely – what better authority do you need? 🙂

September 4, 2023 4:47 pm

So an extreme event in a short observational period is going to cause you to overestimate how soon a similarly extreme event will reoccur ? Whodathunkit? I’m thinking someone experienced would know that…

September 4, 2023 5:42 pm

I don’t understand much of the article; it is way above my meager training. However, since the temperatures were actually measured (some UHI effect certainly) nonetheless, if the models say it was impossible to experience those temperatures, then, obviously, the models were/are wrong. Modelers need to revise their playthings after such an event. No need to Nick Pick, just stating the obvious.

Jeff Alberts
Reply to  John Aqua
September 6, 2023 7:39 pm

I live about 70 miles north of Seattle. Rural area, It wasn’t UHI. It was definitely hot for this area, but it only lasted about 2 days, then was back in the 70s. Yawn.

September 4, 2023 5:43 pm

The June 2021 Northwest heat wave was a stalled high pressure weather system, aka a Heat Dome

With 1100 millibars being high normal, it was trapped within a 1220 millibar region.

These Omega-shaped regions are characterized as having low wind velocities, clear cloudless skies, and a lack of precipitation.

Temperatures always rise within those regions, not because of adiabatic heating, but because the
SO2 aerosol pollution in the atmosphere within the trapped area has time (about 3-5 days) to settle out and temperatures always soar because of the very clean air.

The authors of the paper are clueless as to what caused the heat wave (just meandering of the jet stream), and are trying to “understand and quantify very rare climate extremes in a changing climate”

They try to estimate return periods and return levels of these events in the future, as climate changes (hopeless), although they do correctly expect that more will occur as our climate warms up..

This is being seen this year, with multiple Heat Domes because of the warm El Nino temperatures. caused by the cleansing of our atmosphere of SO2 aerosols by the moisture fallout of the Hunga Tonga water volcano and by Net-Zero and other SO2 aerosol-reduction activities.

Reply to  BurlHenry
September 5, 2023 5:54 am

So environmental regulations are causing global warming by clearing the air of the pollution that used to reflect the Sun.

Reply to  PCman999
September 5, 2023 7:47 am

PCman999

Yes, since about 1980.

However, Net-Zero activities banning the burning of fossil are also reducing SO2 aerosol emissions. And the cleaner the air, the hotter it will get, and the wilder the weather..

September 4, 2023 5:58 pm

I’ve noticed lately, most recently with the heat experienced across N. America this past summer, that there’s a stereotypical freakout over very high daytime temperatures in the hottest months of the year. My understanding has always been that the strongest anthropogenic/greenhouse warming signal would be observed with increases in nighttime low temperatures during the cold months, particularly in the colder climates. Excessive daytime summer heat is entirely caused by a dominance of continental airmasses trapped under high pressure domes. Weather, not climate, as the warmists always remind us…

September 4, 2023 6:06 pm

Let’s not forget the the long-term climate of the Earth is a 2.56 million-year ice age named the Quaternary Glaciation. The Earth is in a warmer interglacial period between glacial periods but still 20 percent of the Earth’s surface is frozen. The ice age won’t officially end until all of the natural ice on Earth melts.
https://en.wikipedia.org/wiki/Quaternary_glaciation

The Sun is in a Grand Solar Minimum that should start cooling the Earth shortly. Sunspot numbers reflect the amount of heat the Sun in radiating and they were forecast NOAA to drop down from the 100-200 we been having to single digits in 2031 and going zero and starting at zero until at least 2040 when the forecast ends. The sunspot number were forecast to start dropping in 2025 but they may have already starting dropping, they are currently below the forecast level.
https://www.swpc.noaa.gov/products/predicted-sunspot-number-and-radio-flux

A recent study shows that cold weather we have every year causes about 4.6 million deaths a year mainly through increased strokes and heart attacks, compared with about 500,000 deaths a year from hot weather. We don’t protect our lungs from the cold air in the winter and that causes our blood vessels to constrict causing heart attacks and strokes.
‘Global, regional and national burden of mortality associated with nonoptimal ambient temperatures from 2000 to 2019: a three-stage modelling study’
https://www.thelancet.com/journals/lanplh/article/PIIS2542-5196(21)00081-4/fulltext

An article from 2015 says that cold weather kills 20 times as many people as hot weather and that moderately warm or cool weather kills far more people than extreme weather. Increased strokes and heart attacks from cool weather are the main cause of the deaths.
‘Mortality risk attributable to high and low ambient temperature: a multi-country observational study’ https://www.thelancet.com/journals/lancet/article/PIIS0140-6736(14)62114-0/fulltext

Also, climate is now defined as 30 years so it is always changing.
https://en.wikipedia.org/wiki/Climate

Reply to  scvblwxq
September 5, 2023 8:07 am

fcvblwxq:

The Maunder minimum had nothing to do with the LIA cooling, the lack of sunspots will have no climatic effect.

We are entering an era where there will be far more heat-related deaths than from cold, as in the past, due to higher temperatures, droughts, starvation, flooding, fires, etc..

Reply to  BurlHenry
September 7, 2023 7:39 am

Evidence?

September 4, 2023 6:38 pm

https://wattsupwiththat.com/2021/07/06/was-global-warming-the-cause-of-the-great-northwest-heatwave-science-says-no/

Cliff Mass is a meteorologist at ground zero (Seattle) and he even believes in manmade global warming. He says no and gives an analysis showing the special weather event components with a heat dome over the mountains and a strong westerly foehn winds that rushed downslope, compressing and generating the high temperatures.

Your extreme value stats are interesting for a variety of weather events,just not this one! If you want apples to apples data, gather foehn wind cases. In Alberta foehn winds in winter can raise -30°C below weather in the path of what we call a Chinook, up above freezing for hours or a day or more. You would soon find the ‘extreme’ data commonplace.

Reply to  Gary Pearse
September 4, 2023 10:20 pm

Cliff Mass is a meteorologist at ground zero (Seattle) and he even believes in manmade global warming.

Many of us “sceptics” of dangerous climate change accept that the world is warming, or at least is warming according to the official temperature records, and is partly due to man.

Our cities, changes to our landscapes, deforestation etc. are examples of how man has changed the environment with the possible knock-on effect of warming albeit local.

Nature trumps man but is there anyone here who doesn’t accept that the earth is warming and that man affects temperatures?

Reply to  Redge
September 5, 2023 12:18 am

Yep, me.

Data are not of sufficient quality to detect warning in any datasets that I have painstakingly analysed. Simply stated, if it can’t be measured in multiple datasets from across Australia then it can’t be happening.

All the best,

Dr Bill Johnston
http://www.bomwatch.com.au

bobclose
Reply to  Redge
September 5, 2023 4:20 am

I tend to agree with you Redge. There is no doubt the UHIE is real and is affecting our temperature records, so that we have higher temperatures measured accurately, but these affected stations are not representative of the local region as a whole, given the more rural stations are warming more slowly or even cooling. So, the choice of representative stations is critical to the final temperature result, and this can be very subjective and biased to ideology. Human land use changes no doubt affect the environment but tend to local or regional, not global as you inferred or have not yet been measurable from natural variance.

Given the methodology of how global temperatures are computed using varying quality regional land-based measurements and limited sea-surface measurements, it is somewhat surprising that coherent temperature trends are achieved by the different datasets. However, I suspect AGW ideology is affecting current increased warming surface trends, that are trying to match GCM trend predictions. Thus, I think that surface results should only be reported for individual regions or continents, and the only coherent global temperature results are provided by the satellite datasets such as UAH.

Reply to  Redge
September 5, 2023 5:04 am

I accept that the Earth’s climate has warmed in the short-term, but it has also cooled in the short-term, and my expectation is that cycle of warming and cooling will continue into the future.

I don’t see any evidence that humans affect the Earth’s atmospheric temperatures with their CO2 omissions. I do accept that human-created cities increase the local temperatures near them.

I don’t accept that the Earth is experiencing unprecedented warmth from any source. From what I can see, it is no warmer today than in the past and CO2 has had no discerable effect on the Earth’s climate or weather, because there is a lot more CO2 in the air now, yet it is no warmer than in the past with less CO2 in the atmosphere.

Reply to  Redge
September 5, 2023 6:34 am

It would be more accurate, especially when we’re talking about climate, to say that the world is currently on a warming trend since about mid 70’s, at least according to the available data.

It really needs to be qualified like that, to avoid missing the big picture, that the Earth has been getting cooler over the past 10K years, even if there are a few warm peaks.

Reply to  PCman999
September 5, 2023 11:34 am

Be careful about “the world”. There are a lot of places with no warming over their baseline.

Reply to  Redge
September 5, 2023 10:13 am

Define “Earth is warming”. I’ve looked at individual station records (GHCN-Monthly) that show cooling in the TMAX record over the past 30 years. How many stations have to be not-warming to disprove the premise?

Reply to  Gary Pearse
September 5, 2023 4:55 am

Thanks for the reality check on the high temperatures, Gary.

Freak weather events combine sometimes to create high temperatures. It has nothing to do with CO2.

Reply to  Gary Pearse
September 5, 2023 8:13 am

Gary Pearse:

MOST Heat Domes occur where there are no foehn winds

Reply to  BurlHenry
September 6, 2023 4:06 am

Yes, but a combination of the two makes things worse.

September 4, 2023 6:40 pm

Oh, those dear little cherubic humans, couldn’t possibly do anything to disturb the weather….

yeah right.
I’m sorry but we really do have to get this butter-wouldn’t meet attitude that we have about ourselves and stop passing the buck onto Ma Mature and or Gaia and or everybody else.

In the attached picture, hopefully also the link will take you there

What you see is a screenshot from a drive along the Lytton-Lillooet Highway in British Columbia

The points of interest are marked by my scribblings:

  • Red arrow: Pointing to bone dry dead or dormant grasses and are with no trees
  • Green arrow: Pointing to green verdant and very much alive grasses
  • Black arrows: Pointing to ‘stock fences’ erected by farmers so as to manage grazing livestock.

We have just passed ‘Ruddocks Ranch‘ at that point along the highway and judging from the loveliness of Ruddock’s new farmhouse, ranching is doing well in that part the world.

see the difference?
The grazing animals have changed the vegetation – allowing in grasses that cannot survive a normal summer (no wonder if they’re constantly being eaten)

So through the summer time, the grazed areas dry out, the entire landscape dries out and thus, there is no water in the landscape to create any buoyant or rising air.
Read: = Clouds, shade or rain
Thus the landscape aridifies ever more strongly and races into desert wasteland = Just Like Maui

Thanks to the joys of ‘weather’, should any strong vertical convections start anywhere nearby (500 to 1,000 miles), the warm/moist air from there will want to return to the ground at some point after it has cooled and dried.
The Very Best Place to do that is at places where there is no warm/moist rising air.
Such as we see in my screenshot of Ruddock’s Ranch
Thus a Hadley Cell sets itself up, lifting water off Vancouver or the ocean on the other side of Van island – hoisting warm moist air of maybe 25°C over the mountain and cooling it (at 6°C per km) as it heads east and inland.

At 6,000 metres altitude, that air will be bone dry and by being at minus10°C, will be very dense.

If it descends onto Ruddock’s Ranch (why not, nothing to stop it) – Lapse Rate will see it warm by 10°C per km as it falls.
Thus by time that descending air is on the ground at the cattle ranch, it will be at (-10 + 60) = 50°C

You got off lightly, highest temps around there and end-June ’21 were about 40°C

My screenshot tells anyone who is serious about weather, water, landscapes and climate exactly what happened there and what caused it.
i.e Captive Grazing Animals

Is that anyone’s definition of Natural Variation?
Just who are The Deniers – or maybe easier to answer:
Who is NOT a denier?

Explore more here

Pacific North West.JPG
Reply to  Peta of Newark
September 4, 2023 9:03 pm

“If it descends onto Ruddock’s Ranch (why not, nothing to stop it) – Lapse Rate will see it warm by 10°C per km as it falls.
Thus by time that descending air is on the ground at the cattle ranch, it will be at (-10 + 60) = 50°C”

I’m not certain where you are going with your analogy – an area in British Columbia becoming hotter than Death Valley?

Reply to  Peta of Newark
September 5, 2023 6:43 am

So that little dried up pasture is responsible for global warming? For the heat dome over the Pacific Northwest (nevermind the Rocky Mountains…)?

September 5, 2023 3:51 am

Such statements require estimates of the exceedance probability p1 of the extreme event under current climate conditions, as well as the counterfactual exceedance probability p0 without anthropogenic warming (Allen, 2003; Stott et al., 2016). Estimating the exceedance probability from past observational or reanalysis data is a central step in the EEA protocol (Philip et al., 2020; van Oldenborgh et al., 2021), which usually entails fitting a non-stationary probability distribution whose parameters are a function of global warming.”

The whole exercise becomes entirely dependent on the assumptions of (a) “probability p0 without anthropogenic warming” and (b) “a non-stationary probability distribution whose parameters are a function of global warming”.

Neither are known are knowable. These are circular arguments as Koonin succinctly points out in “Unsettled”.

You can put lipstick on a pig…

Reply to  ThinkingScientist
September 5, 2023 11:45 am

I was waiting to see if anyone picked up on some things.

“””There is consensus that anthropogenic climate change is the key driver for aggravating hot extremes globally (IPCC, 2021)”””

If you have to rely on consensus of “anthropogenic climate change”, as the truth of the matter, then something is wrong.

September 5, 2023 7:51 am

Everyone makes mistakes. Climate Chaotisists only make mistakes that support their self-serving narrative.

auto
September 5, 2023 1:32 pm

My takeaway is that Climate models are carp [a technical term] becuase of – umm – Climate Change.
It really is worse than some of us believed.
[Belief – that means I’m OK. Doesn’t it? None of the silly sciencey ideas, evidence, etc.!].

Auto

Geoff Sherrington
September 5, 2023 7:57 pm

For the authors,
Many temperatur values used in many ways are not measured, but are interpolated or extrapolated in time or space from measured values. Otheres are adjusted in various ways.
The simple question is: Should statistical analysis be used on non-measured values?
Geoff S

Reply to  Geoff Sherrington
September 6, 2023 12:15 pm

NO! Without the measured values you simply cannot describe the temperature distributions.

Take for instance the ubiquitous “daily average”. The average is really only meaningful as a statistical descriptor if you have a symmetrical distribution, typically a normal distribution. The daily temperature profile is not normal, it is at least bi-modal with different distributions during the day and night, and it is not symmetrical being sinousoidal during the day and an exponential/polynomial decay at night. Therefore the mid-range value, (Tmax+Tmin)/2, is *NOT* an average of the temperature profile. I’m not even sure what is actually *is*, it’s certainly not eve a good “index” for a day’s temperature profile because multiple Tmax and Tmin values will give the same index value.

Anomalies don’t fix this problem. A 1C anomaly at 20C is a very different percentage difference than a 1C anomaly at 0C. In the first case no one will even notice while in the second you will go from frozen to melting! Trying to compare the two by using them to find an average is like trying to average an apple with an orange. The apple and the orange are both round and that’s about as close as it gets. The two anomalies of 1C are both the same value but that’s about as much of a comparison as you can make.

If climate science *really* wanted to use temperature as a measuring stick they should use an integral of the temperature profile to calculate degree-day values, both heating and cooling. The capability of doing this has been available for at least 40 years but climate science has refused to change to this method while ag science and disciplines like HVAC engineering have moved to this method. It fixes at least two problems. One, different temperature profiles are not likely to give the same degree-day values so you get a better index. Two, since it is not a statistical construct made up by ignoring reality, you get absolute values you can compare directly without anomalies. Third, If you want to track the globe, you just add up all the degree-day values over the time period of interest and see what the difference is.

This doesn’t mean that degree-day values won’t have uncertainties propagated from the measuring device creating the temperature profiles. You’ll still have to propagate those uncertainties into any structure derived from them.