# UAH Global Temperature for August – virtually unchanged from July

The pause continues…

Dr. Roy Spencer writes:

The Version 5.6 global average lower tropospheric temperature (LT) anomaly for August, 2013 is +0.16 deg. C (click for large version):

The global, hemispheric, and tropical LT anomalies from the 30-year (1981-2010) average for the last 20 months are:

YR MON GLOBAL NH SH TROPICS

2012 1 -0.145 -0.088 -0.203 -0.245

2012 2 -0.140 -0.016 -0.263 -0.326

2012 3 +0.033 +0.064 +0.002 -0.238

2012 4 +0.230 +0.346 +0.114 -0.251

2012 5 +0.178 +0.338 +0.018 -0.102

2012 6 +0.244 +0.378 +0.111 -0.016

2012 7 +0.149 +0.263 +0.035 +0.146

2012 8 +0.210 +0.195 +0.225 +0.069

2012 9 +0.369 +0.376 +0.361 +0.174

2012 10 +0.367 +0.326 +0.409 +0.155

2012 11 +0.305 +0.319 +0.292 +0.209

2012 12 +0.229 +0.153 +0.305 +0.199

2013 1 +0.496 +0.512 +0.481 +0.387

2013 2 +0.203 +0.372 +0.033 +0.195

2013 3 +0.200 +0.333 +0.067 +0.243

2013 4 +0.114 +0.128 +0.101 +0.165

2013 5 +0.083 +0.180 -0.015 +0.112

2013 6 +0.295 +0.335 +0.255 +0.220

2013 7 +0.173 +0.134 +0.212 +0.074

2013 8 +0.158 +0.107 +0.208 +0.009

Note: In the previous version (v5.5, still provided to NOAA due to contract with NCDC) the temps are slightly cooler, probably due to the uncorrected diurnal drift of NOAA-18. Recall that in v5.6, we include METOP-A and NOAA-19, and since June they are the only two satellites in the v5.6 dataset whereas v5.5 does not include METOP-A and NOAA-19.

Names of popular data files:

uahncdc_lt_5.6.txt

uahncdc_mt_5.6.txt

uahncdc_ls_5.6.txt

From the UAH online press release by Dr. Phillip Gentry:

Global Temperature Report: August 2013

• Global climate trend since Nov. 16, 1978: +0.14 C per decade August temperatures (preliminary)
• Global composite temp.: +0.16 C (about 0.29 degrees Fahrenheit) above 30-year average for August.
• Northern Hemisphere: +0.11 C (about 0.20 degrees Fahrenheit) above 30-year average for August.
• Southern Hemisphere: +0.21 C (about 0.39 degrees Fahrenheit) above 30-year average for August.
• Tropics: +0.01 C (about 0.02 degrees Fahrenheit) above 30-year average for August.

July temperatures (revised):

Global Composite: +0.17 C above 30-year average

Northern Hemisphere: +0.13 C above 30-year average

Southern Hemisphere: +0.21 C above 30-year average

Tropics: +0.07 C above 30-year average

(All temperature anomalies are based on a 30-year average (1981-2010)

for the month reported.)

Notes on data released Sept. 10, 2013:

Compared to seasonal norms, in August the coolest area on the globe was southern Greenland, where temperatures in the troposphere were about 1.97 C (about 3.55 degrees F) cooler than normal, said Dr. John Christy, a professor of atmospheric science and director of the Earth System Science Center (ESSC) at The University of Alabama in Huntsville. The warmest area was south of New Zealand in the South Pacific, where tropospheric temperatures were 2.82 C (about 5.1 degrees F) warmer than seasonal norms.

Archived color maps of local temperature anomalies are available on-line at:

http://nsstc.uah.edu/climate/

As part of an ongoing joint project between UAHuntsville, NOAA and NASA, Christy and Dr. Roy Spencer, an ESSC principal scientist, use data gathered by advanced microwave sounding units on NOAA and NASA satellites to get accurate temperature readings for almost all regions of the Earth. This includes remote desert, ocean and rain forest areas where reliable climate data are not otherwise available.

The satellite-based instruments measure the temperature of the

atmosphere from the surface up to an altitude of about eight

kilometers above sea level. Once the monthly temperature data is

collected and processed, it is placed in a “public” computer file for

immediate access by atmospheric scientists in the U.S. and abroad.

Neither Christy nor Spencer receives any research support or funding from oil, coal or industrial companies or organizations, or from any private or special interest groups. All of their climate research funding comes from federal and state grants or contracts.

— 30 —

Article Rating
Inline Feedbacks
September 10, 2013 11:12 am

So the lack of climate change is worse than we thought.

Robbin Harrell
September 10, 2013 11:17 am

I was wondering if someone would break the chart down to just the current month. IE: Show nothing but the month of Aug for all of the years in the chart. I think that would useful, to graph challenged readers like me.

Editor
September 10, 2013 11:18 am

And as I always try to provide on this thread, for those interested, here’s a link to the full sea surface temperature update for August 2013 (Reynolds OI.v2 data). I posted it yesterday:
http://bobtisdale.wordpress.com/2013/09/09/august-2013-sea-surface-temperature-sst-anomaly-update/
That unusual warming in the extratropical North Pacific was still underway in August.
Regards

Jeff in Calgary
September 10, 2013 11:19 am

So now we are at 201 months of no warming?

Editor
September 10, 2013 11:21 am
BBould
September 10, 2013 11:25 am

Bob Tisdale: Measuring sea levels is problematic. How reliable is measuring sea temp? More or less confidence than say land?

September 10, 2013 11:25 am

All those trillions spent stopping climate change have worked a treat.

rabbit
September 10, 2013 11:39 am

God I hate graphs with running averages or regression superimposed on them, as if our naked eye is incapable of picking out trends. All it does is obscure the data.

Bill Illis
September 10, 2013 11:39 am

The underlying warming trend (after removing the ENSO, AMO, solar cycle, and volcano influences) of the UAH/RSS average and the lower troposphere provided by the radiosonde balloons/HadAT going back to 1958 is a pretty steady 0.056C per decade.

SkepticGoneWild
September 10, 2013 11:46 am

Bill,
Right. It would be warming except for the cooling. (heavy sigh)

September 10, 2013 11:49 am

In order to have a solar/climate connection show up the solar conditions have to vary by a certain degree of magnitude over a certain duration of time, anything short of that WILL NOT BE ENOUGH ,to show a solar /climate connection.
This is why it is hard to show solar/climate connections since the end of the Dalton , to very recently.
However the sun has gone into a prolonged solar minimum state which is turing out much WEAKER then the conventional forecast thus far ,and IS going to have an impact on the climate going forward if the prolonged solar minimum reaches the many solar parameters I have talked about.
solar flux sub 90 sustained.
solar wind sub350 km/sec. sustained.
UV light off upwards of 50% sustained.
cosmic ray count 6500 or more sustained.
solar irradiance off .015% or more sustained
ap index 5.0 or lower 98+% of the time sustained.
These solar values folowing several years of sub solar activity in general which we have had since year 2005.
Salvatore Del Prete says:
September 10, 2013 at 10:47 am
That is the basic flaw in the reasoning of those who keep trying to say there are no solar/climate relationships.
Salvatore Del Prete says:
September 10, 2013 at 10:50 am
Many are in denial of the climatic response to the last two prolonged solar minimum periods,(Maunder Minimum /Dalton Minimum) and do not accept the concept of thresholds, which require a certain degree of magnitude change and duration of time change in the state of solar activity in order for it to exert an influence on the climate.
The period from 1844-2005 should have shown weak to no solar/climate correlations due to the fact solar activity through out that time was in a steady regular 11 year strong sunspot cycle with peaks and lulls which would masked any potential solar/climate correlations.
To clarify there is not one prolonged solar minimum period during that time frame following several years of sub-solar activity in general , to refer to ,to see if prolonged solar minimum conditions do or do not exert an influence on the climate directly and thru secondary means.
Salvatore Del Prete says:
September 10, 2013 at 10:53 am
The low solar activity associated with solar cycle 14 does not meet my criteria for a prolonged solar minmimum period following many years of sub-solar activity in general an thus a definitive solar/climate correlation.
Mike Silver says:
September 10, 2013 at 11:40 am
rgbatduke says:
“As I said, if the sun does enter a prolonged period of comparative inactivity with extended, weak solar cycles, then no matter what the climate does it will be very useful data for those seeking not to assert certain knowledge that they do not, in fact, possess but to determine what the correct theory is by building constructive, physics-based theories that explain the data as the climate moves through something more than a monotonic behavior of irregular warming, which is pretty much all that has persisted for the last 30 to 40 years (as it did for the similar length period at the beginning of the 20th century from roughly 1910 to 1950).”
The weak solar cycle started late 2008, and we have had highly negative AO/NAO conditions and very low land temperature episodes already. It helps to keep up to date.
“As far as I know, there is no convincing evidence that the climate over the last 2000 years was modulated by solar magnetic activity, and we lack direct observational evidence in the form of sunspot counts to extend the Maunder Minimum assertion back to earlier periods of cooling.”
There are Aurora records, and reconstructions:
http://www.aanda.org/articles/aa/full/2007/19/aa6725-06/img81.gif
http://davidpratt.info/climate/clim9-10.jpg
“There is no convincing evidence I’m aware of that solar magnetic activity had anything to do with the Younger Dryas itself,”
http://www.sciencedirect.com/science/article/pii/S0012821X03007015
“I especially doubt that the climate is a one-trick pony, slaved to solar magnetic activity to the exclusion of all else,”
There are always negative AO/NAO conditions when the solar plasma is slow:
http://www.leif.org/research/Ap-1844-now.png
[snip . . misposting . . mod]

RACookPE1978
Editor
September 10, 2013 11:53 am

Bill Illis says:
September 10, 2013 at 11:39 am

The underlying warming trend (after removing the ENSO, AMO, solar cycle, and volcano influences) of the UAH/RSS average and the lower troposphere provided by the radiosonde balloons/HadAT going back to 1958 is a pretty steady 0.056C per decade.

So, in round numbers, there has been a 0.16 degree change since the baseline temperature of the mid-1970’s, or 0.16 degree in 45+ years, right? Makes it just under 0.4 degrees per century for near on the last 1/2 century? 8<)

September 10, 2013 11:57 am

It went out before I had a chance to edit it, all that additional stuff.. Sorry

September 10, 2013 11:57 am

It’s all on the scale and range of graphs selected. We are on the 6th upswing since the end of the last ice age and this upswing is cooler than the preceding 5 upswings as we continue our downward slide to a return to the normal temps for this epoch which is what we term the ice age.
In the nearer term view we are continuing our exit from the little ice age.
In the even nearer term we might, as Bill Illis has pointed out with no context think we have some minor CO2 based warming going on.
Or we can look at an even nearer term and see we are in a cooling phase.
All of this is make interesting discussion points, while the smart money is buying up condo land on the eastern edge of the future American atlantic shoreline at the edge of our current continental shelf 🙂 (and speculating on where the population of Canada is going to migrate)

Doug Allen
September 10, 2013 12:08 pm

The MET and others have already suggested some cooling as we continue through the cool phase of the PDO. The wild card is the solar cycle. Today we have one, tiny little sunspot that some think would not have been counted in past centuries. In any case, we’re that tiny spot away from a sun spotless day with many likely to follow. If we have the grand solar minimum many or most astronomers now expect, we’ll get a better handle on whether or not it’s just TSI or other solar radiation wavelength that forces global temperature change. Exciting times!

September 10, 2013 12:37 pm

Is there a plan to move back to the replacement AMSU-2?

September 10, 2013 12:58 pm

Maybe we could convince Roy to supply the data in kelvin or supply a climatology in addition to anomalies

Eliza
September 10, 2013 1:30 pm

Wow so Bob Tisdale’s global sea surface temps graph is similar to AMSU satellite data from Roy Spencer’s. The correlation is probably rr2 0.99 or something similar just from looking at both graphs. This confirms Roy Spencer’s and RSS data as a very reliable indicator of Global temperature trends. It s shows that Hadcrut, GISS etc BEST land based thermometer Stevenson boxes are probably way off due to UHI etc.

Gail COmbs
September 10, 2013 1:59 pm

John Mason says: @ September 10, 2013 at 11:57 am
All of this is make interesting discussion points, while the smart money is buying up condo land on the eastern edge of the future American atlantic shoreline at the edge of our current continental shelf 🙂 (and speculating on where the population of Canada is going to migrate)
>>>>>>>>>>>>>>>>>>>>>>>
Just make sue you have an iron clad trust passing on to your distant heirs : >)

jlurtz
September 10, 2013 2:36 pm

The only surprise will be the speed of the Solar effect.

Richard Barraclough
September 10, 2013 3:13 pm

Jeff in Calgary says:
September 10, 2013 at 11:19 am
So now we are at 201 months of no warming
Well, Yes and No, Jeff!
The RSS figures just released can be shown to have a tiny negative slope all the way back to November 1996, which is indeed 201 months.
However, for this UAH V5.6 dataset, you really have to cherry-pick your dates to find any cooling (as a pose to not much warming). The furthest back you can go with a negative slope is July 2008, or only 61 months, but within this 61 months there are several shorter periods ending in August 2013 which have a positive slope, so I think the statisticians among us will pour cold water on the idea that we can call it “cooling”.
To illustrate how confusion can be spread with the misuse of statistics, the latest 61-month period is the warmest such period in the dataset, even though it has a negative slope.
You can do a couple of other analyses on this data which also don’t show any cooling – just a deceleration in the warming. If you work out the 34 annual temperatures from 1979, and fit a least squares trend line through those for each year up to the end of 2012, you can go all the way back to 2005 and find a negative trend, though now the trends from 2006, 2007 and 2008 are all positive. But before you say, “Aha 7 years cooling”, repeat the exercise using “years” which run from September to August, in order to include the very latest data, and your 7-year cooling disappears, leaving only the 2 recent annual periods from Sept 2009 and 2010 with negative trends.
Confusing? And that’s without even mentioning statistical significance, or wondering whether each month can be considered to be independent of the previous one.
Many thanks to the providers of these data sets, without whom we would not be able to look forward to the 10th of each month and pick around these few hundredths of a degree….

philincalifornia
September 10, 2013 3:13 pm

Phillip Bratby says:
September 10, 2013 at 11:12 am
So the lack of climate change is worse than we thought.
————————————————————
It will be for some …………

rgbatduke
September 10, 2013 3:18 pm

“There is no convincing evidence I’m aware of that solar magnetic activity had anything to do with the Younger Dryas itself,”
http://www.sciencedirect.com/science/article/pii/S0012821X03007015

A paper is not “convincing evidence”, not when there are at least two or three competing theories, NONE of them particular compelling. Perhaps it was a small asteroid impact — there is evidence to support it. Perhaps it was the melting of an ice berm, and sudden draining of a meltwater “ocean” that flooded the arctic, interrupted the thermohaline circulation which re-established on a more southerly track and caused a thousand-year hiatus in the warming. Perhaps it was the sun. Perhaps it was whatever the hell caused the Ordovician-Silurian glacial era (my favorite has to be cosmological clouds of leftover darkons or magnetic monopoles or maybe just plain old dust that the sun drifted into, although there are a few issues with the solar wind). There might even be more hypotheses out there, these are all ones I’ve read. It could even be pure chaos, and not even have a “proximate cause” in the form of an actual controlling driver. It could be combinations of several things happening in coincidence.
I’d work through the rest of your quotes (of me), but it doesn’t matter. You included the first quote which is the only one that really matters. One way or another, we’ll improve our scientific understanding of something that we have never before seen with modern instrumentation in orbit, and perhaps learn something about the climate even if what we learn is a null result (that the solar cycle has NO discernible effect on the climate, where “discernible” means outside of the unexplained natural variability and noise, which is actually a rather huge range making it not that unlikely an outcome).
rgb

rgbatduke
September 10, 2013 3:27 pm

Many thanks to the providers of these data sets, without whom we would not be able to look forward to the 10th of each month and pick around these few hundredths of a degree….
Ya mon. It becomes even amusinger if you estimate any sort of reasonable statistical/experimental error and associate it with the data. Throw in a $\pm 0.15$ C error bar, increasing as you work backwards in time, and THEN consider the p-value of the null hypothesis comparing a zero-trend least squares fit to the best trended fit via e.g. Pearson’s $\chi^2$. You could drive a truck through there, and if you let the error bars grow to 0.5 C or better in the early 20th century (as is pretty reasonable) you could drive the truck back into the 19th century and barely be able to assert that it has probably warmed. And yeah, that doesn’t even begin to account for autocorrelation in what is most certainly non-Markovian, nonlinear, chaotic time evolution with multiple autocorrelation timescales ranging from seconds to centuries.
Sigh. Now, give me a CENTURY’S worth of high quality instrumental data obtained with nobody’s thumb on the damn scales, spanning something other than a monotonic (supposed) rise, and maybe we’ll get somewhere.
rgb

September 10, 2013 3:36 pm

Jeff in Calgary says:
September 10, 2013 at 11:19 am
So now we are at 201 months of no warming?
With RSS, it is actually 202 months since it went 1 month forward and 1 month back from last month. So it goes from November 1996 to August 2013. See:
As for UAH, I cannot dispute what Richard Barraclough says about 5.6, however version 5.5 went back to January 2005 last month and the new value should push it back a bit more. I can be more specific once the new numbers come out on WFT.

Bennett In Vermont
September 10, 2013 3:40 pm

rgbatduke is one of my heroes.

rgbatduke
September 10, 2013 3:40 pm

Maybe we could convince Roy to supply the data in kelvin or supply a climatology in addition to anomalies
I think that would be lovely for all of climate science, actually. Sadly, CAGW curves are singularly unimpressive when plotted on an absolute scale, especially if you plot them right next to e.g. CO_2 concentration. The hockey stick wouldn’t look the same plotted as the nearly invisible dust on the slightly dusty but otherwise almost perfectly smooth hockey ice. CO_2 concentration, OTOH, has gone up by 1/3. It sort of reduces the feel that there is actually a meaningful correlation, y’know?
Of course to do this creates lots of problems, not the least of which is that (as NASA points out explicitly) we don’t know HOW to renormalize the anomalies computed by the different ways to absolute temperature, not even within a single degree K. Different models and approaches yield different values ranging from 13 to 15 C to be added to the anomaly, and the various anomalies don’t agree to better than a few tenths of a degree and probably have larger error bars than that. A degree of uncertainty means that we cannot even positively resolve if any warming has occurred since the nineteenth century, although it is plausible and even probable that it has.
A second problem is that presenting anomalies instead of absolute results is an actual chapter in the lovely book How to Lie with Statistics (yes, this is a real book and a real topic in the book). It is a favorite trick of con artists, especially when conjoined with its friends, cherrypicking a (favorable to your con) range, and ignoring probable error (which means that one can always do a statistically neutral recomputation of almost any result and pick out one that is within statistically acceptable bounds but that shows a desired trend).
But hey, Roy is just doing as everyone does. Heaven forfend that anyone should just try to actually compute not anomalies — the second moment of an unknown distribution — but the FIRST moment, which is usually BETTER known. In fact, usually you have to know the first moment in order to compute the second moment (e.g. the variance). But hey, I’m sure everybody knows exactly what to subtract. Oh wait, no they don’t. Not to within a degree.
rgb

RACookPE1978
Editor
September 10, 2013 4:42 pm

rgbatduke says:
September 10, 2013 at 3:40 pm
So, why could a “standard anomaly” not be created (re-created) from the original data? is this not what “they” are doing for those billions we spend each month?
Take the original stations, original data – if that exists, but TOB and other “edits” to the original data may need to be used, since original values may not be available.
For each station, develop all of the temperature differences for high, low, and average for every date where data are recorded. No skips, no misses. Break data strings where changes occur, but then resume with a new string, creating a “new” anomaly for the “station_mod_2.
For an agreed-on interval (1970-1990, 1980-2000, 1990-2010 or what ever) develop that average over the interval and define that as the “average” or the baseline. Calculate the differences from that baseline and define that difference as the “anomaly” for that station.
Rinse, wash, and repeat for every station.
When all station anomalies are listed, report the final averaqe of all differences from each station’s baseline.

Rob
September 10, 2013 4:53 pm

The Invariability of Global Climate

Richard M
September 10, 2013 5:43 pm

Richard Barraclough says:
September 10, 2013 at 3:13 pm
…. Confusing?

Not necessarily. One has to compare apples to apples. Trying to plot trends from La Niña periods to El Niño periods or neutral times, or El Niño periods to anything but another one will give you spurious results. However, if you try to keep to ENSO equal periods it generally makes sense.

Werner Brozek
September 10, 2013 7:33 pm

Version 5.6 is not positive for 62 months since July, 2008.
Version 5.5 is not positive for 107 months since October 2004.
RSS is not positive for 202 months since November 1996.
To the end of August, 2013 on UAH would rank 7th warmest if its average anomaly stayed that way for the rest of the year on version 5.5.
To the end of August, 2013 on RSS would rank 8th warmest if its average anomaly stayed that way for the rest of the year.

Ivan
September 10, 2013 7:52 pm

Any news about Watts et al 2012? It had been announced a year ago that the paper will be “submitted within weeks”. What happened?
[So, how many other threads are you going to repeat this “Oh by-the-way-off-topic-disruption of other people’s conversations? Mod]

That Idiot Driver
September 10, 2013 7:59 pm

To John Mason: Excellent post. To illustrate by way of a graph, “Long Range Weather” has a graph covering 4,500 years showing exactly your point. Vividly the lows from the cold periods are getting deeper. We are in a drift down. Short term still coming out of the LIA. But the future “looks” stark. “The Ice Man Cometh.” Just Google that chart. The best chart I’ve seen that cuts to the bottom line.

September 10, 2013 8:08 pm

Looks like Greenland is cooling faster than other places. Shades of Viking colonisation?

richard verney
September 10, 2013 8:52 pm

rgbatduke says:
September 10, 2013 at 3:18 pm
////////////////////
At the time of the Copenhagen conference and the breaking of climategate, the MET Office has a chart on their site which included error bars. On a review of that chart, one could not say with certainty whether the temperature in 2009 was warmer than that of the 1880s!
Of course that was viewing the upper error of the 1880s temps with the lower error for the 21st century temps (ie., maximising the spread), but it does accord with your assertion that when one properly takes into account the margins of error, one cannot say definitively whether it has warmed since the 1880s, although it is likely that it has.
I am not sure whether the Met Office have taken down that chart since it is not particularly helpful for the message they are overly keen to promote.

September 10, 2013 9:32 pm

Steven Mosher on September 10, 2013 at 12:58 pm:

Maybe we could convince Roy to supply the data in kelvin or supply a climatology in addition to anomalies

True, it is very difficult for you to convert between anomalies in °Celsius to anomalies in Kelvin. Perhaps you could find a subroutine for the “R” programming language that’ll do it, IIRC you’ve played around with it a few times.
I work with the LibreOffice spreadsheet program. It has a powerful function where you can convert between Celsius and Kelvin anomalies by simply changing the parameter designation at the top of a column of data by using the F2 key.

Brian H
September 10, 2013 9:58 pm

Haha. In Mosh’s defense, he clearly means to eliminate the anomalies and show only absolute temps.

Mario Lento
September 10, 2013 10:26 pm

September 10, 2013 at 9:32 pm
Steven Mosher on September 10, 2013 at 12:58 pm:
True, it is very difficult for you to convert between anomalies in °Celsius to anomalies in Kelvin.
++++++++++++++++++++++++++++++
Wait a minute. Isn’t the difference between Kelvin and C, an offset of 273? So the anomalies would be exactly the same… as a delta C is the same as a delta Kelvin.

Mario Lento
September 10, 2013 10:27 pm

Steven Mosher says:
September 10, 2013 at 12:58 pm
Maybe we could convince Roy to supply the data in kelvin or supply a climatology in addition to anomalies
++++++++++++
I’ve read other posts referencing this post. I still don’t know what “supply a climatology” means.

September 10, 2013 11:06 pm

From Mario Lento on September 10, 2013 at 10:27 pm:

I’ve read other posts referencing this post. I still don’t know what “supply a climatology” means.

Note above where it says “The global, hemispheric, and tropical LT anomalies from the 30-year (1981-2010) average for the last 20 months are:”
So 1981-2010 is the reference period for calculating the anomalies. To find the monthly anomaly for August, first find the average of all the monthly August averages for all years in the reference period (average(Aug1981:Aug2010)). Then whatever the absolute average temperature is for any given August, subtract that average of the averages for August, and you get the anomaly.
Translating Mosh mutterings can be hazardous, but I think by “climatology” he was asking for those anomaly-making figures, the average of the monthly averages for every month for the 1981-2010 reference period. With those he can work backwards, add them into the anomalies to get the absolute temperatures.

Jean Meeus
September 10, 2013 11:38 pm

Bill says:
“The underlying warming trend (after removing the ENSO, AMO, solar cycle, and volcano influences)…”
After removing solar cycle?? But, but… the alarmists said that solar cycle has no effect on the global temperature, and that only the increasing CO2 is the culprit.

Kelvin Vaughan
September 11, 2013 1:50 am

Jean Meeus says:
September 10, 2013 at 11:38 pm
Bill says:
“The underlying warming trend (after removing the ENSO, AMO, solar cycle, and volcano influences)…”
After removing solar cycle?? But, but… the alarmists said that solar cycle has no effect on the global temperature, and that only the increasing CO2 is the culprit.
The true cause of global warming is Women. They heat up the world by having hot flashes. As the worlds population increases there are more women and more hot flashes!

Richard Barraclough
September 11, 2013 2:15 am

Werner Brozek
I like the way you always provide us with some figures of your own as soon as the new month’s data are released, so I don’t like to be too picky, but surely July 2008 to August 2013 is 61 months, not 62? You can’t really include both months in the count, unless your analysis goes from the beginning of July to the end of August. I mean, supposing you were comparing July 2013 (0.173) with August 2013 (0.158), you’d say there’d been a drop of 0.015 over 1 month (not 2 months).
With regard to your update of the “half-time scores”, UAH V 5.6 looks as though it’s heading for 4th or 5th warmest for this year, so there are quite a few differences, even between data sets from the same source.
1998 0.420
2010 0.399
2005 0.262
2002 0.220
2009 0.213
2007 0.208
Average anomaly for 2013 so far is 0.215, and for the last 12 months 0.249
Regards
Richard

JKrob
September 11, 2013 3:34 am

From the story…
Global Composite: +0.17 C above 30-year average
Northern Hemisphere: +0.13 C above 30-year average
Southern Hemisphere: +0.21 C above 30-year average
Tropics: +0.07 C above 30-year average
(All temperature anomalies are based on a 30-year average (1981-2010)
for the month reported.)
What is with all these “30-year averages”? (that is a rhetorical question BTW). I want to see the numbers compared to the proper PDO/AMO accounted 60-year average.
Jeff

Jimbo
September 11, 2013 3:54 am

This summer it’s been a cold one in the Arctic. So, where is our Arctic amplification. Greenland is doomed.

Co
September 11, 2013 4:00 am

The Idiot Driver references this chart; http://www.longrangeweather.com/images/gtemps.jpg

Bill Illis
September 11, 2013 5:31 am

12 year anniversary of September 11th today. You remember where you were.
There has been Zero warming in that time, while the climate models have about 0.3C of warming.
Daily UAH temperatures back to September 11, 2001.
http://s12.postimg.org/jgfch4ajx/Daily_UAH_Sept11_2001_Trend.png
Note that temperatures were declining throughout August, 2013 and ended the month at only 0.088C.
———————-
Regarding the solar cycle comments above. There really isn’t a solar cycle signal in the numbers, there is just barely a hint of one at certain times. Others have used it (and used improperly on purpose as in Foster and Rahmstorf 2011) so I leave it in but it doesn’t really affect the results at all).

Richard M
September 11, 2013 6:04 am

I find it interesting that while the overall global temperature provided by RSS and UAH is very close, if you look at the breakdown between NH, SH, etc. there are huge differences. This makes me somewhat wary. Why are the differences so large?

Editor
September 11, 2013 7:08 am

Robbin Harrell says: “I was wondering if someone would break the chart down to just the current month. IE: Show nothing but the month of Aug for all of the years in the chart. I think that would useful, to graph challenged readers like me.”
Robbin, the August 2013 anomaly is the difference between the August 2013 temperature and the 30-year (1981-2010) average of the August temperatures. So the 0.16 deg C anomaly in August 2013 means it’s only 0.16 deg C warmer than the 1981-2010 average.

Editor
September 11, 2013 7:20 am

BBould says: “Bob Tisdale: Measuring sea levels is problematic. How reliable is measuring sea temp? More or less confidence than say land?”
Sorry for the delay, BBould. I’ve been wrapped up preparing a YouTube video intro for my new book.
Compared to land surface temperatures, The Reynolds OI.v2 SST data are much more reliable. The Reynolds OI.v2 dataset is based on satellite measurements and in situ measurements from ship inlets and buoys (fixed and drifting). In an early paper (about 2005 if memory serves), NOAA’s Smith and Reynolds used it as the reference for a reconstruction, calling the Reynolds OI.v2 data the “truth”.
I definitely have no qualms with the Reynolds data. They show the East Pacific Ocean from pole to pole hasn’t warmed in 31 years…and the Pacific Ocean as a whole hasn’t warmed in 20 years…and the North Atlantic (with the AMO) stopped warming around 10 years ago.

Werner Brozek
September 11, 2013 8:49 am

Richard Barraclough says:
September 11, 2013 at 2:15 am
You can’t really include both months in the count, unless your analysis goes from the beginning of July to the end of August. I mean, supposing you were comparing July 2013 (0.173) with August 2013 (0.158), you’d say there’d been a drop of 0.015 over 1 month (not 2 months).
Taking your second sentence first, yes, if I were to compare July to August, then I would assume it was 0.173 on July 15 and 0.158 on August 15 and I would assume the drop was over 1 month.
As for the July to August, I cannot demonstrate that on WFT since it uses version 5.5, but if we look at RSS, it is flat from November 1, 1996 to August 31, 2013. So both November and August are counted and this gives 202 months. You can check this for yourself at this site:
Then click on “Raw data”. See that the firsts 2 months are 1996.83 and 1996.92 which are November and December. At the bottom, note that there are 8 months starting with 2013. As well, it also says “#Number of samples: 202” and gives the “slope = -5.24568e-05 per year” for those 202 months.
By the way, from October 1, the slope is “slope = 0.000142636 per year”. So if we assume a smooth transition from October 1 to November 1, we could say the zero mark is at October 23, giving 202.27 months to August 31.

rgbatduke
September 11, 2013 1:00 pm

I work with the LibreOffice spreadsheet program. It has a powerful function where you can convert between Celsius and Kelvin anomalies by simply changing the parameter designation at the top of a column of data by using the F2 key.
OK, OK, now you’re just being mean. Everybody knows the degree size for Celsius and Kelvin is the same. Or at least, I profoundly hope so.
I think he meant convert from anomalies (Celsius and/or Kelvin) into degrees absolute, and that’s how I framed my reply. If he seriously meant change the NAME of the units, well, gee…
rgb

rgbatduke
September 11, 2013 2:09 pm

Of course that was viewing the upper error of the 1880s temps with the lower error for the 21st century temps (ie., maximising the spread), but it does accord with your assertion that when one properly takes into account the margins of error, one cannot say definitively whether it has warmed since the 1880s, although it is likely that it has.
I am not sure whether the Met Office have taken down that chart since it is not particularly helpful for the message they are overly keen to promote.

Ordinarily this is the sort of thing the R-value of a trended fit is supposed to help with, although the issue is seriously complicated here because of the huge error bars in most of the past even in the thermometric era (and believe me, they are larger still whatever the assertions of the authors in the proxy era preceding that) and the fact that there is substantial autocorrelation so that the measurements are not even close to being independent. To be picky, even if you have 143 years of annualized numbers in the range from 1870 to 2013, you don’t have anything like 143 degrees of freedom in the computation of $\chi^2$ in a maximum likelihood fit. This complicates the unbiased estimate of the probability of getting the temperature record we have if there was no actual warming, in point of fact, in the actual (unknown) global average temperature that all of the estimates are intended to estimate if: a) the error bars are some reasonably accurate measure of probable error in the central number, ideally a gaussian/normal estimator such as a standard deviation; b) the average from 2012 and 2013 are not really independent sample, as the temperature of 2012 is EXPECTED to be comparatively close to that of 2013; c) there are systematic biases and errors in the computations of the annual temperatures.
For example, if we postulate — as is I think generally accepted to be the fact — that the thermometric record is increasingly corrupted by an urban heat island effect (UHI) covering more and more of the areas sampled as populations increase and cities grow and more roads are built and more trees are cut, then it is straightforward to produce 100% of the warming observed by simply underestimating the UHI effect in the station data and not reflecting the uncertainty in the UHI correction in the error estimate assigned to the later points. There are other systematic errors reflected in the older data — sea surface temperature used to be estimated by e.g. throwing a bucket overboard astern and measuring the temperature of the water brought back up, on an unbelievably inadequate and ever-shifting grid tied tightly to the usual shipping lanes. Some of these we can identify and try to fix, others we have no hope of fixing. Given that the ocean alone accounts for 70% of the Earth’s surface, the uncertainty in historical SST data is enormous and profound and most of it we don’t even know how to correct (although that won’t stop people from trying).
Additional systematic errors can come from something as simple as when, and how, people recorded daily temperatures, where their thermometer was physically located at the time it was recorded, who MADE their thermometer (in particular, how accurate and precise it was), whether or not the recorder was sober and industrious or an alcoholic who often forgot to check at all and just filled in numbers that sounded reasonable to get paid.
And finally, there are the really, really big holes, even on land. We have heavily oversampled temperature measurements near older cities, but Antarctica (a whole continent, mind you) was virtually unknown and unexplored, let alone systematically sampled for temperature, well into the 20th century. One could actually make a case for the fact that its temperatures aren’t adequately sampled today. Ditto for much of the sahara, much of australia, large chunks of the americas, large chunks of china, siberia, asia. Between the oceans and these large holes with no systematic measurements at all, one has to interpolate, infill, and commit various other data sins that involve replacing actual data with guesses just so you can apply a consistent algorithm over time. Again, all one really needs to create illusory warming is to select perfectly “reasonable” infilling/interpolating/extrapolating algorithms that erase local trends and replace them with, effectively, the local trends of someplace else, probably someplace with an uncontrolled UHI correction that increases over time (because the oldest stations in your record are those near cities, or situated in the countryside of populated regions that became a lot more populated over a century, of course).
Taking all of this into account, yeah –the world probably has warmed, but has it warmed 0.5 C over 143 years? 1 C over 143 years? 1.5 C over 143 years? That is really difficult to say. Interestingly, there are places on the world where UHI is not much of an issue — they were wild in 1870 and are wild today — that show a lot less trended warming than the average. The places that show the most warming almost invariably have a serious UHI problem. The US is rife with UHI corruption in its record, as Anthony has pointed out in a formal paper, but his paper just scratches the surface. One could literally look at the entire record station by station, and it is almost impossible to tell HOW to correct a station with poor siting.
At Raleigh-Durham airport — airports are a primary source of “official” temperature readings for an area — the weather stations aren’t placed in such a way that they obtain optimally accurate temperature readings for the local CLIMATE, they are placed to give accurate readings of the temperatures over the runways, as those are what planes need to know as they land. They tend to be sitting in open fields, right next to runways, over asphalt, in places where a simple change in the direction of the breeze from over the grass to over the nearby runway can cause temperatures to spike during the day, while those same dry runways can retain heat for much longer than insulating, dew-dampened grass at night. Not to mention that they tend to be located in the middle of a tangle of high-density highway traffic (lots of roadway and concrete) and are constantly bathed in CO_2 and water vapor given off by jets taking off and landing and wafting overhead from the nearby road traffic.
In Durham were I live (still almost in the city) temperatures are consistently 1-2C cooler than the airport, especially high temperatures. The temperatures where I live are consistently as much as another degree warmer than a place REALLY out in the country another five miles, far from any expressway or urban center larger than a cluster of houses on a small road. What is the UHI correction for RDU airport? It should arguably be as much as 3 C, probably at least 1.5 to 2 C, but if one subtracted that much from its reading everything changes, record temperatures disappear, much of the supposed warming of the area disappears.
And then there is the entire “anomaly” issue. They don’t actually use RDUs temperature as measured in degrees K to determine the global anomaly per se anyway. How can they? They haven’t got any good measurements even for that location that reliably stretch back to 1870 — the land was all forest or tobacco field back then, and Raleigh and Durham both were barely what we’d consider to be large towns now. Instead they try to turn it all of the local readings into the difference between what is read and a presumed “normal average” temperature for the location, and then average this anomaly spatially with infilling etc with much statistical juju magic.
In the end, the one thing we can say with near certainty is that the corrections that are constantly made to the global temperature estimates over the last 143 years are biased. We can say that because no matter what you think about methodology, the likelihood of making a systematic error that results in more warming over that stretch and making a systematic errror that results in more cooling over that stretch ought to be about the same. On average, one would expect that altering the computational algorithm used to estimate the temperatures as is done between at least the major releases of e.g. GISS and HADCRUT (and now BEST) would result in increased linear trends as often as decreased linear trends. However they have all resulted in increased linear trends! They have cooled the past, or warmed the present, or both — they never seem to warm the past, cool the present, or both. Given that at this point there are many distinct instances of this occurring, one can actually formulate a probability of getting all warming changes given the null hypothesis of no human bias in the selection of the changes themselves, and reject the null hypothesis with a p-value well below 0.05, probably well below 0.01. In other words it is nearly certain that some fraction of the warming in the “official” temperature databases is not only an artifact, it is a deliberate artifact, one made by biased human choices in what changes to implement. I have no doubt that every change can be “justified” a posteriori somehow, but that does not remove the statistical evidence for bias.
Once again, to fix things at this point is probably impossible. We cannot go back in time and measure the temperature in the middle of antarctica in 1887, and any attempt to infer it from ice cores or other proxies have error bars so large that the estimate is all but useless (and easily corrupted with human bias in its analysis). To even fix the UHI problem — that is in principle fixable — would require an enormous investment, the development of a set of serious and inviolable criteria for weather station siting, regular inspection and corroboration/testing by qualified personnel, the development (almost impossible) of a way of correctly estimating the magnitude of the UHI, and the elimination of anomaly-based measurements and analysis altogether. The first moment of a distribution is almost certainly going to be more accurately known that the second moment — the actual mean is more accurate than the deviation from the mean — and even NASA admits that we do not know the actual mean to better than a range of some 2 C even today.
That is an enormous range. Honestly, it is big enough that it makes it quite possible that proxy derived temperature estimates may be more accurate than the thermometric estimates (and generally, may be a bit more difficult to fiddle with although as dendroclimatolgists everywhere persist in demonstrating, far from impossible).
It’s sad, really, and is the reason that I only really trust the satellite record of the last 33 years, with weak extension and increasing error bars back to maybe 1950. Before that, I think we’re dealing with anecdotal evidence, not the real thing. Suggestive, never conclusive. I’m not sure I trust ARGO even now, but in ten or twenty more years I’m guessing it will be beyond fiddling too. Even these measurements have substantial errors, but it is a lot more difficult to introduces human biases into their results.
We are thus at the infancy of real climate science. We have reasonable evidence that the climate changes over geologic and human historic time. We have anecdotal evidence of the changes, and can probably guess at how the temperatures have varied within a degree or two over decadal to century timescales and beyond. We have at least some things that are simply not susceptible to UHI bias — the LTT estimates, ARGO — that are at least reasonably independent of each other as well. It is now much more difficult to introduce a hidden bias into the future climate record. Given a few more decades of reliable data, especially data on both warming and cooling trends instead of monotonic warming, especially data that embraces the many factors that might influence the climate that make the climate more than a one-knob projective pony, we might be able to put together a credible semiempirical theory of climate change that actually works to predict at least a few decades out. I’m not optimistic about being able to do better than that in less time than a full century of observation and refinement.
This is arguably the hardest problem in the world to solve. Well, except for predicting the stock market. That’s harder. But really, really hard. It really isn’t that surprising that we haven’t solved it yet. It would actually be more than a bit surprising if we had.
rgb

Mario Lento
September 11, 2013 6:43 pm

September 10, 2013 at 11:06 pm
++++++++++++
Thank you for translating Mosher for me.

Mario Lento
September 11, 2013 6:46 pm

Steven Mosher on September 10, 2013 at 12:58 pm says:
True, it is very difficult for you to convert between anomalies in °Celsius to anomalies in Kelvin.
++++++++
Was Mosher being sarcastic? I cannot tell. Delta C is exactly the same as delta K… so anomalies would be identical.

beng
September 12, 2013 9:24 am

***
rgbatduke says:
September 11, 2013 at 2:09 pm
***
I imagine Mosher would be squirming/fidgeting if he reads that revealing reply. According to him, it’s all already taken care of.

Ivan
September 12, 2013 4:50 pm

Dear moderator, I am sorry for posting the same OT comment on Watts et al (2012) on two separate threads but I was interested in what happened to the paper and hoped to increase the chances that somebody in the know would read my comments and answer. I did not know whom to ask and how. I sent an email to Anthony before inquiring about the paper, but he never answered (I am not criticizing that, just saying).
[Reply: You are forgiven, my child. Carry on… ~mod.]

geran
September 12, 2013 6:12 pm