Charles Rotter
Scientific findings gain authority not only from data but from how those data are presented. The 2025 paper by Kong and colleagues, “Intensifying precipitation over the Southern Ocean challenges reanalysis-based climate estimates – Insights from Macquarie Island’s 45-year record,” offers an excellent example of how a small, uncertain result can be transformed into sweeping global statements once it passes through the press-release pipeline.

In the peer-reviewed article, the authors are careful, even tentative: they discuss possible biases, limited spatial coverage, and the assumption-laden nature of their extrapolations. Yet in the Phys.org / The Conversation version of the story—authored by two of the same researchers—the tone hardens into certainty. What was a local analysis becomes a global revelation: “Storms in the Southern Ocean are producing more rain—and the consequences could be global.”

The distance between those two versions is not a matter of semantics. It is the difference between a statistical curiosity in a small dataset and a claimed planetary-scale hydrological transformation.
The tiny foundation
Macquarie Island, a windswept ridge halfway between Tasmania and Antarctica, provides one of the very few long-term meteorological records in the Southern Ocean. The authors analyze 45 years of daily precipitation data (1979–2023). Missing days number fewer than ten, an impressive continuity. But while the temporal coverage is long, the spatial coverage is one point—8 meters above sea level, in one of Earth’s most variable weather regions.
Anyone trained in time-series analysis knows what such a dataset can and cannot reveal. Forty-five annual values offer, at best, a few dozen degrees of freedom for trend estimation. In climate contexts where interannual variability is large, a few anomalous years can tilt a regression slope enough to appear “statistically significant.”
When those same years are subdivided among five clusters of “synoptic regimes,” the effective sample size for each trend falls even further—often to fewer than ten truly independent points once serial correlation is considered. Under those conditions, small random fluctuations can masquerade as meaningful patterns.
The clustering that creates the patterns
To explore meteorological “regimes,” Kong et al. applied K-means clustering to 15 atmospheric variables from the ERA5 reanalysis. They tested between three and eight clusters and decided that five was “optimal,” because the results “were broadly consistent with earlier studies.”
This is not an objective optimization; it is a judgment call. K-means imposes spherical clusters of roughly equal size, a geometry that atmospheric data rarely obey. Changing the number of clusters, or even the random initialization, can alter both the cluster composition and the derived trends. With 45 years of daily data—over 16,000 points—it is almost guaranteed that some partitioning will yield clusters with apparent differences in precipitation intensity that look “significant” at the 0.05 level.
In other words, the procedure is prone to statistical mirages. The apparent precision of numbers such as “a 28 % increase in precipitation intensity under warm-air advection” can stem from how the algorithm slices the data, not from any underlying physical change.
Significance by multiplication
After defining five regimes, the authors fit separate linear trends for each regime in both the observational and ERA5 datasets—over 50 regressions in total. Yet the paper uses the conventional p < 0.05 threshold without correction for multiple testing. At that threshold, roughly one in twenty regressions will look “significant” by chance. With dozens of tests, a handful of p values below 0.05 are statistically inevitable even if no real trends exist.
The tables confirm this: most regimes show p values hovering around 0.05; a few dip below and are labeled “significant.” No adjustment (Bonferroni, Benjamini–Hochberg, or otherwise) is applied. The result is a statistical landscape peppered with chance findings elevated to explanatory status.
Moreover, because the five regime time series are mutually exclusive and collectively exhaustive—every day belongs to exactly one cluster—their annual frequencies are inherently dependent. An increase in one regime must coincide with a decrease in others. Treating them as independent samples exaggerates the apparent certainty of each trend.
Data quantity versus data quality
Even if every regression were perfectly executed, the physical meaning would remain ambiguous. Rainfall at an island in the roaring forties depends on local topography, sea-surface temperature, and wind direction, all of which can vary independently of broad climate trends. ERA5, meanwhile, represents a gridded average over roughly 25 × 25 km. The mismatch between a point gauge and a model grid is substantial. The observed increase of 260 mm per year could reflect local station effects, gauge changes, or random decadal oscillations rather than a true regional trend.
The authors acknowledge these caveats deep in the discussion:
“While MAC offers precious long-term observations … its single-point nature introduces potential scale mismatch with the nearest ERA5 grid-cell mean, which may contribute to the observed biases.” (p. 1655)
That is an important concession. Unfortunately, it disappears entirely from the media retelling.
From statistical artifact to global mechanism
The media version begins with rich description—penguins, elephant seals, mossy slopes—and then declares:
“Our new research confirms [the rainfall increase]—and shows the story goes far beyond one remote UNESCO World Heritage site.”
From there, the logic unfolds by implication rather than evidence. Because the Southern Ocean “plays an enormous role in the global climate system,” any change at Macquarie Island must reflect a larger transformation. The authors-turned-communicators then state:
“If the rainfall intensification we see at Macquarie Island reflects conditions across the Southern Ocean storm belt—as multiple lines of evidence indicate—the consequences are profound.”
That sentence retains a conditional “if,” but what follows discards it:
“Our estimate suggests that in 2023 this additional precipitation equates to roughly 2,300 metric gigatons of additional freshwater per year across the high-latitude Southern Ocean—an order of magnitude greater than recent Antarctic meltwater contributions.”
Now the assumption has become a quantitative global statement, complete with apparent precision and comparison to Antarctic mass loss. To a general reader, it reads as fact. In the scientific paper, the same number is prefaced by “assuming this increase is representative…”—a thought experiment, not an observation.
The illusion of scale
Scaling a single-point trend to an ocean basin is more than an extrapolation; it is a dimensional leap. The error bars on such a calculation are effectively unbounded. Yet those bars vanish in the popular narrative.
The reasoning chain, if written transparently, would read:
- Observation: Macquarie Island’s gauge shows a 28 % rise in annual precipitation since 1979.
- Assumption: The island’s change represents the entire 50–60° S latitude band.
- Computation: Multiply the mean increase by the surface area of that belt.
- Result: ~3,400 gigatons additional freshwater flux.
Each step adds orders of magnitude of uncertainty. By the end, the numerical precision (e.g., “2,300 gigatons”) is meaningless. Yet such numbers acquire rhetorical power precisely because they look precise. The specificity signals confidence even when the calculation is little more than arithmetic atop an assumption.
A pattern of inflation
This is not an isolated misstep; it illustrates a systemic tendency in modern climate communication. Researchers, pressed to demonstrate relevance, extend their conclusions beyond the domain their data can support. Editors and outreach offices favor strong declarative headlines over probabilistic phrasing. The result is a progression of certainty:
| Stage | Source | Character of claim |
|---|---|---|
| 1 | Raw data (rain gauge) | “Rainfall has varied and appears to have increased.” |
| 2 | Peer-reviewed paper | “Rainfall intensity at Macquarie Island increased 28 %, assuming representativeness.” |
| 3 | Press article | “Storms in the Southern Ocean are producing more rain.” |
| 4 | Media amplification | “Southern Ocean storms intensify under climate change.” |
At each step, the confidence grows while the evidential base stays the same.
Statistical caution drowned out
The technical paper’s limitations are extensive and clearly listed. It notes that only one regime frequency trend is statistically significant; that ERA5 assimilates local observations, complicating independence; and that the broader extrapolations are speculative. The 95 % confidence intervals on many regressions overlap zero. The authors even suggest that “more supporting evidence is needed.”
None of this appears in the public version. Instead, readers learn that “the Southern Ocean may be cooling itself by 10–15 % more than it did in 1979—simply through the energy cost of evaporation that fuels the extra rainfall.” That statement implies a quantified, basin-wide energy-budget change—derived not from measurement but from the same single-point rainfall increase multiplied by theoretical latent-heat factors. The line between observation and conjecture is erased.
Why small datasets breed large illusions
Short or localized datasets are especially vulnerable to spurious trends because of the autocorrelation inherent in climate time series. Even with purely random year-to-year variations, ordinary least squares can yield apparently significant slopes if successive values are not independent. Standard p values assume independence; when that assumption fails, the true significance is far weaker.
Add to that the possibility of non-stationarity—periods of higher and lower variance—and the confidence intervals widen further. Without explicit tests for autocorrelation and sensitivity to start and end dates, any 45-year linear trend should be regarded as provisional. The paper mentions none of these tests. Thus, the celebrated “28 % increase” could easily be a statistical artifact of natural multidecadal variability.
Such artifacts are not trivial; they shape the narrative. Once published, each becomes another “data point” in meta-analyses and climate-model validations, potentially reinforcing biases in the very reanalyses the authors critique.
Why this communication pattern persists
Researchers are often caught between two expectations: the academic requirement for caution and the public expectation of clarity and impact. Funding bodies and media outlets favor stories that link local findings to global stakes. The safest way to achieve visibility is to hint at large consequences while retaining formal caveats in the technical version. Those caveats, however, rarely survive the press release.
In this case, the same authors wrote both the scientific and the popular versions, removing the normal filter that might have preserved nuance. The outreach piece’s confident tone thus carries the imprimatur of the original authorship, giving readers the impression that the expanded claims rest on data rather than inference.
The consequences of over-certainty
When modest studies are publicized as global breakthroughs, two harms result. First, the public comes to expect constant discovery of new “climate tipping points,” diluting attention from robust, long-term evidence. Second, when subsequent analyses fail to reproduce the headline results, confidence in climate science as a whole erodes. The credibility loss is collective, not confined to one paper.
Over-confidence also affects science internally. Once a narrative hardens—“Southern Ocean storms are producing more rain”—future studies are subtly pressured to conform. Null results risk appearing contrarian even when they are more accurate.
A disciplined alternative
None of this implies that Kong et al. acted improperly; their underlying analysis is a standard exploratory exercise. The problem lies in conflating exploration with confirmation. A disciplined approach would keep the distinction clear:
- Report the observed local trend and its uncertainty.
- Explicitly test sensitivity to start year, cluster count, and autocorrelation.
- Present basin-scale extrapolations as hypothetical scenarios, not estimates.
- Maintain that conditional framing in every public communication.
Had those practices been followed throughout, the study would still be valuable—as a case study in regional precipitation analysis—without overstating its global significance.
The real lesson
The Macquarie Island record is scientifically interesting precisely because it is rare. Its value lies in ground-truthing remote sensing and model products, not in diagnosing planetary change. The temptation to elevate it into global evidence is understandable; data from the Southern Ocean are scarce, and every dataset feels precious. But scarcity is not a substitute for statistical power.
The genuine message of the paper—how little we actually know about precipitation trends in the Southern Ocean—was inverted in the media coverage. Instead of highlighting uncertainty, the outreach piece turned it into certainty: the ocean “is changing faster and more dramatically than we thought.” The reality may be the opposite; with so little direct observation, we do not yet know whether it is changing at all.
Conclusion
The journey of this study—from one gauge’s time series to a global climate headline—shows how modern science communication can inflate tentative statistical patterns into narratives of planetary transformation. Each step in that journey removes a layer of uncertainty until what remains is no longer a cautious inference but a declarative claim.
Reintroducing that uncertainty is not an act of skepticism for its own sake; it is a restoration of proportion. The data from Macquarie Island may hint at interesting variability, but the confidence intervals are wide, the sample small, and the physical mechanisms unresolved. Treating such a fragment as evidence of global hydrological change risks turning science into storytelling.
If there is one robust trend visible here, it is not in rainfall but in rhetoric: the persistent escalation from modest signal to grand conclusion. Until that trend reverses, the distance between climate data and climate discourse will continue to grow—and with it, public confusion about what the numbers truly mean.
can i pick a location now? pretty please?
No only climate change scientists on the public teat get issued with the official climate darts.
I believe it could help if someone pee-ed on the gauge.
Would it help if we pee-ed on their paper?
Like this.
Lou Brown pee’s on Dorne’s papers (Major League)
A little late for my first chuckle of the day, but there it is (-:
Every great baseball team needs a Lou Brown. And it would help in editing some so-called refereed research. We should all have been free to do a lot of this with a lot of the supposed research from the IPCC over the decades. We could start with MBH98. We could carry on with this way with the papers claiming 97% agreement with the consensus.
Koko’s ever-changing list of those who will not be missed is applicable here. Eric Idle was a comic genius.
The Mikado – The List (Eric Idle).avi
Looking at their rain data, the trend seems to be mainly because of a drier period from 1967-1975.
If you started the trend calculation in say 1981, there would be very little trend.
Even then, evapotranspiration is also increasing, confounding their very holey story.
The study authors involved in the subsequent published distortions are out-and-out liars. They should be sanctioned by the study funders, the journal that was deceived, their academic employers and any relevant scientific society(s). Until those steps are widely employed one must assume all CliSciFi outputs are bogus.
Good article. I believe a good number of media articles concerning medical news are following a similar trajectory. Huge numbers of studies are being published with ridiculously small sample sizes resulting in highly dubious reporting.
I had the privilege of visiting Macquarie Island as a tourist earlier this year. It is a marvel, worthy of any bucket list, even though tourist activity is very constrained.
Our ship was the last of the summer and we picked up a few scientists on their way home.
We were coming back from Antarctica. One fellow on our cruise had worked for years on Macquarie eradicating rats, rabbits and cats (extraordinarily difficult work). It was his first visit back in some years; he said he had never seen the place looking so good.
The ferals were only eradicated for good 10 years ago. The native vegetation has made a remarkable comeback since then; any change in vegetation must surely be attributed to the successful eradication not climate change. Talking to the scientists on the voyage home (they gave short lectures) climate change was not mentioned.
Many years ago I was at a conference where a New Zealand environmental scientist outlined their experience with totally eradicating ferals from some of their islands.
The summary was that removing the last 1% took at least the same time and money as removing the first 99%.
Does “multiple lines of evidence” bring it back into the observation category?
“How a Narrow Dataset Becomes a Global Climate Headline”
Where did it become a global climate headline? AFAICS, some authors managed to write an article in Phys.Org/The Conversation. I googled, and the only other source mentioning it was WUWT.
MSN feed on many people’s browsers.
Well, MSN here is just an agregator, and links to just about anything, here to the Phys.org article. But I can’t see tbhat any major publication has headlined it..
So, Nick, you agree the article is full of CARP and didn’t deserve any publication or notice.
OK.. seems we are all in agreement. 🙂
Sloppy, Nick, very sloppy. Don’t you check what you wrote before you post it?
You can even edit your post, after tapping the “Post Comment” button. Didn’t you know that?
Here’s your chance – correct my intentional errors. Put me in my place.
Well, CoveringClimateNow is just an agregator, and links to just about anything too.
But no criticism ever of what CCN does?
Well Charles, we’re regularly told told that if observations don’t pass the Nick-O- Scanner, they don’t actually exist.
Well, Nick, if you refuse to accept reality, and believe that adding CO2 to air makes thermometers hotter, then you may be clinically insane.
Would you value the opinions of someone you believed to be insane?
I certainly wouldn’t.
AI will present it as fact.
Very nice Charles.
“In the peer-reviewed article, the authors are careful, even tentative: they discuss possible biases, limited spatial coverage, and the assumption-laden nature of their extrapolations. Yet in the Phys.org / The Conversation version of the story—authored by two of the same researchers—the tone hardens into certainty. What was a local analysis becomes a global revelation: “Storms in the Southern Ocean are producing more rain—and the consequences could be global.””
You show clearly what I have tried to say in the past and why I view the journalist and that community as more harmful than the academic/scientist and that community. This is a great example because they are the same people. It is a bad thing.
Do these “scientists” think we live on a static, never-changing planet??
This seems like a base rate fallacy. Someone with better statistical training might wish to comment.
Charles,

Thank you for your thorough demolition.
This paper has even more cherry picking than you list.
The official Bureau of Meteorology rainfall record actually starts in 1948
Here are some graphs of rainfall in mm per year.
Both rainfall and temperature show somewhat different time trends, depending on the time period chosen.

As you note, the paper under discussion makes heroic leaps from a really isolated tiny spot in the vast Southern Ocean, to discussions about global implications.
Every single weather station has particular errors associated with it, errors that are wrongly carried forward in the defective type of analysis used by the authors Kong et al. For example, the rain recorded at the gauge can be affected by the direction of the wind at the time it is recorded. Scenario, westerly winds cause low readings. This error cannot be detected by the method used by these authors, so the data are disallowed for global comparison purposes.
If the paper was not so bad, I would dig out the wind data and examine effects like this, but I cannot be bothered. Why does climate research throw up so many papers showing poor quality science?
Geoff S
“Why does climate research throw up so many papers showing poor quality science”
Because most people who gravitate to “climate science” are activists, as opposed to being actual scientists.
It is not the brightest that go into ‘climate science’.The really clever ones either go to CERN or into astrophysics or disappear and sell their soul in the City.
Funding.
Great analysis.
Academics suffer no consequences for doing sloppy work. Paper after paper, study after study never spend time or paper on analyzing and propagating uncertainty as you point out. When was the last time an uncertainy budget was developed using categories like wind, as you show, to calculate a representative measurement uncertainty for single measurements and presented in the analysis.
Treating a series of single, non-repeatable measurements as having the statistical reliability that allows the use of the standard uncertainty of the mean (√n) as measurement standard uncertainty is a joke.
The result of averaging a series of non-repeatable measurements is NOT a measurement. It is actually determining a property of a phenomena.
Can climate scientists be required to have comprehensive training on statistical analysis, among other topics?
That’s what happens when you put the inmates in charge of the asylum.
And “climate scientists” are insane, if refusal to accept reality is a valid symptom of insanity.
To a man (or woman, sex being no defence against delusion) they think that adding CO2 to air makes thermometers hotter! What a pack of utter loonies!
Well, it is just business as usual. Use one thing, make it a headline to extrapolate linear, global trends and impacts.
It’s alarmism in action.
It’s the headline that counts. Headline= climate change. Just stack them up to make it real. The truth does NOT matter…much.
Data can always be tortured to produce the desired results. The problem is not with the raw data, but with the humans doing the torturing. As such anything beyond the raw data is always suspect.
“is changing faster and more dramatically than we thought.”
Yes, absolutely, “the science is settled” is contradicted again.
I’m having a little trouble figuring out how “2,300 metric gigatons of…freshwater” could be added to the ocean by rainfall, when the rain water came from evaporation of the ocean in the first place. It’s easy enough to do a sanity check, though: 2,300E9 tonnes of fresh water has a volume of 2,300E9 cubic meters. The surface area of the Earth is about 5.1E14 square meters, of which about 71% is ocean, i.e. 3.6E14 square meters. The rise in sea level from adding that much water to that much ocean would be 0.0064 meters, or 6.4 mm – twice the actual value.
If I were to say that The Conversation is trash I would be kind because I can use many cruder words that are appropriate descriptions of this source of supposedly reliable scientific information.