Still Going Down the Up Escalator
Posted on 3 February 2012 by dana1981
The Escalator, originally created as a simple debunking to the myth "Global warming stopped in [insert date]", turned out to be a very popular graphic. Going Down the Up Escalator, Part 1 recently surpassed 20,000 pageviews, Part 2 has an additional 4,000+ views, and the graphic itself has been used countless times in other blogs and media articles. Due to its popularity, we have added a link to The Escalator in the right margin of the page, and it also has its own short URL, sks.to/escalator.
The popularity of the graphic is probably due to the fact that (1) it's a simple, stand-alone debunking of the "global warming stopped" myth, and (2) that particular myth has become so popular amongst climate denialists. As The Escalator clearly illustrates, it's easy to cherry pick convenient start and end points to obtain whatever short-term trend one desires, but the long-term human-caused global warming trend is quite clear underneath the short-term noise.
The original Escalator was based on the Berkeley Earth Surface Temperature (BEST) data, which incorporates more temperature station data than any other data set, but is limited to land-only data; additionally the record terminates in early 2010. We originally created the graphic in response to the specific myth that the BEST data showed that global warming had stopped.
It is interesting to apply the same analysis to a current global (land-ocean) temperature record to determine whether short term trends in the global data can be equally misleading. A global version of the Escalator graphic has therefore been prepared using the NOAA NCDC global (land and ocean combined) data through December 2011 (Figure 1).
Figure 1: Short-term cooling trends from Jan '70 to Nov '77, Nov '77 to Nov '86, Sep '87 to Nov '96, Mar '97 to Oct '02, and Oct '02 to Dec '11 (blue) vs. the 42-year warming trend (Jan '70 to Dec '11; red) using NOAA NCDC land-ocean data.
The Predictable Attacks
On 31 January 2012, John Cook emailed me about several recent uses of The Escalator, including an inquiry from Andrew Dessler, requesting to use it in one of his lectures. In the email, John suggested that the graphic had gained so much popularity, it would likely soon be the target of attacks from fake skeptics.
As if eavesdropping on our conversation, the first such attack on the escalator came the very next day, on 01 February 2012. The graphic had been publshed nearly 3 months earlier, and John predicted the fake skeptic response within a day's margin.
The Escalator was recently used by a number of sources in response to the denialist plea for climate inaction published in the Wall Street Journal, including Media Matters, Climate Crocks, Huffington Post, and Phil Plait at Discover Magazine's Bad Astronomy. Statistician William Briggs took issue with Phil Plait's use of the graphic. Specifically, he criticized the lack of error bars on the data used in The Escalator, making some rather wild claims about the uncertainty in the data.
"...the models that gave these dots tried to predict what the global temperature was. When we do see error bars, researchers often make the mistake of showing us the uncertainty of the model parameters, about which we do not care, we cannot see, and are not verifiable. Since the models were supposed to predict temperature, show us the error of the predictions.
I’ve done this (on different but similar data) and I find that the parameter uncertainty is plus or minus a tenth of degree or less. But the prediction uncertainty is (in data like this) anywhere from 0.1 to 0.5 degrees, plus or minus."
As tamino has pointed out, calculating an area-weighted average global temperature can hardly be considered a "prediction" and as he and Greg Laden both pointed out, BEST has provided the uncertainty range for their data, and it is quite small (see it graphically here and here). Plait has also responded to Briggs here.
The Escalating Global Warming Trend
Briggs takes his uncertainty inflation to the extreme, claiming that we can't even be certain the planet has warmed over the past 70 years.
"I don’t know what the prediction uncertainty is for Plait’s picture. Neither does he. I’d be willing to bet it’s large enough so that we can’t tell with certainty greater than 90% whether temperatures in the 1940s were cooler than in the 2000s."
It's difficult to ascertain what Briggs is talking about here. We're not using the current trend to predict (hindcast) the global temperature in 1940. We have temperature station measurements in 1940 to estimate the 1940 temperature, and data since then to estimate the warming trend. Once again, we're producing estimates, not predictions here.
Moreover, the further back in time we go and the more data we use, the smaller the uncertainty in the trend. For example, see this post by tamino, which shows that the global warming trend since 1975 is roughly 0.17 +/- 0.04°C per decade in data from NASA GISS (Figure 2). The shorter the timeframe, the larger the uncertainty in the trend. This is why it's unwise to focus on short timeframes, as the fake skeptics do in their "global warming stopped in [date]" assertions. As tamino's post linked above shows, when we limit ourselves to a decade's worth of data, the uncertainty in the trend grows to nearly +/- 0.2°C per decade (Figure 2).
Figure 2: The estimated global temperature trends through July 2011 (black dots-and-lines), upper and lower limits of the 95% confidence interval (black dashed lines), and the estimated trend since 1975 (red dashed line) using GISS land and ocean temperature data (created by tamino)
Foster and Rahmstorf (2011) also showed that when the influences of solar and volcanic activity and the El Niño Southern Oscillation are removed from the temperature data, the warming trend in the NCDC data shown in the updated Escalator is 0.175 +/- 0.012°C per decade. Quite simply, contrary to Briggs' claims, the warming trend is much larger than the uncertainty in the data. In fact, when applying the Foster and Rahmstorf methodology, the global warming trend in each of the major data sets is statistically significant since 2000, let alone 1940.
Ultimately Briggs completely misses the point of The Escalator.
"...just as the WSJ‘s scientists claim, we can’t say with any certainty that the temperatures have been increasing this past decade."
This is a strawman argument. The claim was not that we can say with certainty that surface temperatures have increased over the past decade (although global heat content has). The point is that focusing on temperatures over the past decade (as the fake skeptics constantly do) is pointless to begin with, and that we should be examining longer, statistically significant trends.
Briggs' post was of course hailed by the usual climate denial enablers (i.e. here and here), despite the rather obvious inflation of the data uncertainty, and utter lack of support for that inflation. Despite the fake skeptic struggles to go the wrong way down, The Escalator unfortunately continues ever-upward.
Also, I was coincidentally using the graphic on Jan. 30th in another forum (responding to the Daily Mail/Rose claim of no warming since 1997), and instantly caught some denialist flak over it. I didn't realize the picture was getting spread around of late, but it's just so great for beating down arguments of short-term "no warming" crap. Funny how things seem to happen to everybody at once sometimes.
It's amazing how people will argue themselves into incoherent knots when they attempt to argue against clear & compelling visualisations of clear & compelling evidence.
[dana1981] Thanks, corrected
"We don't see things as they are, we see things as we are" Anais Nin.
via Frances Moore Lappe's "Ecomind"
As it happens, just today I wrote a program to help with an "escalator" graph using GISS data. I'm not done with the graph, but if you wish I can send you a spreadsheet showing all the negative-slope regressions from each month of 1970-2005 for periods of 60 thru 131 months out. I zeroed out the positive slopes so the negative ones jump right out at you; it's a cherry-picker's dream.
Perhaps someone can superimpose the two onto a single graphic?
Try looking here and hereto get a better idea of what Briggs is saying.
Plait stated that the WSJ authors were "dead wrong" when they claimed that there had been no warming over the last 10 years, and illustrated this point with a static version of the escalator graph. Briggs says they weren't, and that this particular graphic cannot tell us that they were wrong. The goal wasn't to "get" the purpose of the escalator graph.
WheelsOC
No, I think Briggs' biggest problem is that people don't understand what he's talking about. Sloppy and inconsistent terminology abounds, and the Plait critique was apparently written with long-time readers or those intimately familiar with Bayesian predictive techniques solely in mind.
And Briggs never "backpedaled." He's remained consistent in his description of averages as models.
Defending Briggs by saying he has been consistent about describing measured results as models is absurd. Measured results are measured results. Briggs is attempting to raise doubt (doubt is our product!) by calling the measurements models. This alone proves he is not serious.
Yes, a common tactic with fake-skeptics - no matter how ridiculous their assertion is, they never admit they are wrong. No claim is too ridiculous to be defended by them it seems.
Consistency is also a common tactic with people who are correct. Who'd have thunk it?
To this DeepClimate responded at Open Mind:
That seems to me sufficient rebutal, but Briggs was "consistent", and wrote,
So DeepClimate did:
This may seem like a trivial matter, but accusations of cherry-picking, even veiled accusations, are not trivial. It is relevant in this case, however, because it shows that Briggs "consistency" is in significant degree merely the inability to admit error. Don't pretend otherwise when the disproof is so straightforward.
Briggs says of the dark grey band that it is "...the classical parametric prediction interval for these new proxy measurements." Earlier he had mentioned that, "The 95% parametric prediction interval for this model happens to be 17.6oC to 20.4oC." Ergo the "prediction interval" shown is the 95% prediction interval, ie, the interval inside which we would expect 95% of values to fall.
The problem is, in the graph shown, just 22.6% (7 of 31) of observed values fall within the 95% prediction interval. How is it possible for a measure of the 95% prediction interval fail to include close to 95% of the data from which the prediction interval is calculated? To my mind that only 22.6% the data from which the 95% prediction interval is derived falls inside the 95% prediction interval seems like a contradiction in terms.
On the face of it, Briggs has simply miscalculated the 95% prediction interval. His calculated prediction interval performs so badly because it is simply the wrong interval. And given that, he is able to show his calculated interval radically underestimates uncertainty simply because his error causes it to radically underestimate the uncertainty.
I am interested in hearing why we should trust a statistician who makes so fundamental error (if he has); or alternatively just why statisticians so distort the language that a 95% prediction interval means (as it must if Briggs has made no error) that interval in which approx 25% of the data will fall.
If the man was using any of the commonly accepted meanings of these words then what he said is flat out wrong. If he wasn't then there really isn't any way to tell what he was actually saying.
In any event, if any part of what Briggs said were true it would ironically also contradict all of the 'no warming since XyZ' claims... because those are all based on the 'model predictions' (aka, thermometer measurements) Briggs is saying can't be used that way. Indeed, he is claiming that the entire ~130 year period is too short to establish a trend line... so 10 years would just be pathetic.
I'm not sure what Briggs' thinking was, but DC does indeed provide a convincing rebuttal to that point. Despite what you appear to think, I don't believe Briggs to be the Second Coming, and I won't attempt to defend every statement he has made. But many of the criticisms leveled at him over the past few days come from a simple lack of understanding of precisely what he is saying.
In particular, Tamino's entire critique revolves around the idea that Briggs is simply using the words "model" and "prediction" as some sort of semantics ploy in an effort to evoke the instinctual rejection by denialists of any argument which contain those things. This is not true.
The planet is not perfectly sampled at every point on its surface, so in creating an average global temperature, you're attempting to "predict" the temperature at unsampled points using data from sampled points. These "predictions" result in uncertainty in the end product, which Briggs alleges is only properly accounted for under Bayes theorem. He fleshes this process out in more detail in the second link I give above. Such terminology is fairly common in certain fields, but not in others.
Michael,
I had written a response to you earlier, but it disappeared. Let's try again: What exactly do you mean by measured results? Can we measure the temperature at every point on the Earth's surface? If no, how can we combine the data we do have to create a coherent record of global temperature? Is this method completely without error?
Your post was deleted as trolling, perhaps this one will be also.
Averaging results as described in published papers is not predicting. You claim that in order to know something we must measure it perfectly everywhere. If that were true, then nothing can be measured. Every measurement is an average of several others.
You also asked for the error bars which are linked in the OP. If you cannot bother to read the OP, why comment? A brief glance at the error bars shows that the error is much less than the trend. No analysis is needed for such an obvious observation. Briggs is wrong about the error also. It is common to leave off the error bars to make the graph more readable for a general audience. Briggs is trying to artificially manufacture doubt
which made me think of this:
Note I am not saying that Briggs is a screw-up (he isn't), merely that being consistent in promulgating an incorrect argument is not necessarily a good thing.
His point seems to be about including the uncertainty in computing the global mean temperature into computing the error bars on the trend is perfectly reasonable. However what he should have done is first demonstrate that it makes a meaningful difference to the conclusion, rather than just hint that it might to spread baseless uncertainty (the uncertainty in computing the means looks pretty small to me compared to the uncertainty in the trend due to weather noise - is internal climate variability)
As the regression line is not being used to predict or extrapolate the temperature anywhere, the "prediction" uncertainty is irrelevant.
The credible interval for the observations themselves is a combination of the credible interval for the regression line, plus a component due to the noise in the data (i.e. the spread of data around the regression line).
I am not surprised that this misunderstanding should ocurr, Briggs' articles seem rather ambiguous and opaque from what I have read so far.
We do not expect all the data to fall on the regression line, and there is no reason why it should. Failure of the data to fall within the error bars of the regression line is therefore, irrelevant, contrary to Briggs' claim.
I remain uncertain as to whether Briggs is confusing the error of the "predicted" regression line with the error of the predicted data, or is simply hopelessly confusing in his discussion. However, it is not a good model for to great a confidence in temperature indices in that the confidence intervals shown for temperature indices are for the value, not for the regression. Hence absent good reason to think otherwise I shall disregard that blog as irrelevant.
I think it would be a good idea to include the dark area on plots of regression lines (I'd plot the most likely line as well) as it would show very clearly that the trend estimate for short timescales were very uncertain. This would prevent skeptics from claiming that the trend was flat as the error bars would show that they could be flat, or cooling or warming, and there isn't enough data to decide which.
Which error bars you choose depends on whether you are trying to predict the observations or trying to estimate the rate of warming. In this case it is the latter.
Therefore, he has either described the error margins of the regression as the "prediction interval" which is simply false, or if your conjecture is incorrect, he has not properly calculated the error margins at all (as shown by the fact that less than 25% of the data used in calculating the regression (and hence the prediction) falls within the "prediction interval". In either case he has made an error, and his argument that classical statistics underestimates errors is based firmly on the consequences of that mistake.
This goes directly to RobertS's defense of Briggs. In essence that defense is that Brigg's has been misunderstood because he has been precise. This example shows that the mere use of technical language does not make what you say precise, let alone accurate. If Briggs has been misunderstood, it is because he has been unclear, not because he has been precise.
As interesting and useful of discussions of forced versus unforced “variability” are, they are irrelevant utterly to whether claims (A) and (B) are true. It matters not how the data arose to answer the claims of whether the temperature is higher or lower over the last decade.
However claims A and B relate to the statement
The lack of warming for more than a decade—indeed, the smaller-than-predicted warming over the 22 years since the U.N.’s Intergovernmental Panel on Climate Change (IPCC) began issuing projections—suggests that computer models have greatly exaggerated how much warming additional CO2 can cause.
and the amount of additional warming that CO2 can cause is concerned with the forced component and only the forced component.
The key to success in statistics is a willingness to immerse yourself in the data and understand the data generating process and the purpose of the study. If you just wade in thinking you know what is important and what isn't and not listening to those who know the data, it is a recipe for disaster. Well I did try...
That said, it would certainly be possible to construct an 'escalator' for the satellite temperature record(s). However, it would likely only have a few steps since the satellite record is comparatively short.
You are very right, but for the wrong reasons. It is much harder to generate negative trends from the UAH data (I didn't try RSS). The tropospheric data is much cleaner, not so much because of the global sampling (it's not global, BTW, it reaches far towards the poles but does miss them -- see the gray areas on the RSS images here – the satellite orbits used are not set up to give a good, downward looking view of the earth at the poles), I think, but because the warming is less impacted by noise.
In fact, I couldn't create a negative trend up to the end with the UAH data. Maybe someone else can try. You have to sit and peck around for exactly the right
cherry flavoredend points to use. It's very hard to do what with all of the consistent and unequivocal warming that's occurred in the past fifteen years.I was able to get an ending downward trend for the RSS data, thanks to the recent La Niña and by selecting 2002 as a start date.
[dana1981] Thanks for your feedback, Dr. Thorne. The purpose of The Escalator is to show how 'skeptics' and realists actually view global temperatures. 'Skeptics' constantly look for flattening or cooling trends in short-term data, whereas realists (by definition) don't look for rapid warming trends in short-term data, but rather examine long-term trends.
However, your point is taken that we could do a 'running up the escalator' graph to show that short-term data can be manipulated in the opposite way, as opposed to the way it's actually manipulated, which is what the current Escalator shows. We'll take this suggestion under consideration.
True, satellites still are not a 100% perfect sampling, but the entirety of Earth's surface is sampled below a latitude of ~80 deg. While not every single point is measured at infinite resolution, the temperature of every single point combined with its nearest neighbors is, which means that no extrapolation to unsampled areas is necessary to get a global temperature average except at the very poles.
Also, I'm not surprised at the result (more going up the down escalator!), but now the pseudo-skeptics can't claim that we're cherry-picking GISS temp because it's warmer or some other such bogus argument.....
Um, _no_. That is estimation, not prediction. A prediction is a statement about what will happen in the future, possibly conditionally dependent upon various alternatives, while an estimation is an approximate calculation of measurements in the face of uncertainty or subsampling.
Briggs conflation of the two indicates either a need for a dictionary in his reference materials or perhaps an attempt to cast temperature estimations as models so that he could take advantage of the 'skeptic' disdain for models in general.
I'm still confused about the prediction interval of the global average... the global average is what we want to compare...
Bert
As you can see a horizontal(ish) line just about fits into the credible interval, but the bulk of the credible interval suggests we can be confident that there has been some warming. When I get a moment I'll include the uncertainties in the estimates of the anomalies and see if it makes much of a difference.
It certainly appears to me that Briggs' arguments on this matter simply don't make any sense. I suspect it's because, as Dikran has noted, he has not bothered to understand basic climate science concepts before trying to analyze the data.
His 'prediction uncertainty' also seems like nonsense, because as discussed in the post above, the groups putting together the global surface temperature estimates include uncertainty ranges, which are not even remotely as large as Briggs suggests.
Taking temperatures at separated stations can be described as a statistical sampling of the population of temperatures as opposed to getting 'all' the temperatures. Thus one uses a "model" (extrapolation or interpolation) of sorts to infer that the separated stations do provide information about the space around them. Some of this collapses when you use anomolies instead of temperatures, I think...the rest could be demolished by several experiments (probably available already if one looked) showing that a finer grid didn't change the anomoly in any significant degree.
Furthermore, if he is a Bayesian, his priors whould include prior information on the likely parameter values given preexisting knowledge about the system state and the physics of CO2. He should do a real analysis with error structures defined etc. I'd bet 10 bucks that once he explcitly isolated sources of error in the data that there would be even stronger evidence for that trend. I think he has the whole thing backwards.
Bottom line? Let's see a real analysis rather than word games and obfuscation.
Turn it over to him and the end result of the time, effort, and money, would probably be 'can't really say'. He deals with mathematical precision, not Large Systems Effects.
His main problem is not the statistics, it is that he doesn't understand the climatology.
I guess it would depend on how you structured the error model, for sure, but this is exactly the kind of detail we are not getting from Briggs. That's beside the point that a non-informative prior in this case is basically cheating. What's the point of a Bayesian model if you're not going to include prior information?
Also, how does the "escalator" compare with IPCC AR4, WG1, Ch3, p253?
Stephen Baines@45 The more uncertainty in the estimates, the more uncertainty there will be in the linear regression model also. I think it is likely that it will make little difference, as the uncertainties in recent anomalies are small, but there is only one way to find out!
When I give a talk on Bayesianism, I often use that Rumsfeld quote about there being things we know we know, things we know we don't know, and things we don't know we don't know. The first of these is easy to deal with, the last is impossible to deal with (other than giving the caveat that the almost certainly exist). The real advantage of Bayesianism is that it gives you a sound way of using the expert knowledge that you don't know something. The main advantage is that it makes the conclusions of the anlaysis less certain, and is helpful in avoiding jumping to conclusions. The real problem is that sometimes they are too vague (e.g. a flat prior on the probability of a biased coin coming up heads - in the real world you can only make a slightly biased coin, unless it is one with a head on both sides!).
IPCC AR4 Ch.3