Climate Science Glossary

Term Lookup

Enter a term in the search box to find its definition.

Settings

Use the controls in the far right panel to increase or decrease the number of terms automatically displayed (or to completely turn that feature off).

Term Lookup

Settings


All IPCC definitions taken from Climate Change 2007: The Physical Science Basis. Working Group I Contribution to the Fourth Assessment Report of the Intergovernmental Panel on Climate Change, Annex I, Glossary, pp. 941-954. Cambridge University Press.

Home Arguments Software Resources Comments The Consensus Project Translations About Support

Bluesky Facebook LinkedIn Mastodon MeWe

Twitter YouTube RSS Posts RSS Comments Email Subscribe


Climate's changed before
It's the sun
It's not bad
There is no consensus
It's cooling
Models are unreliable
Temp record is unreliable
Animals and plants can adapt
It hasn't warmed since 1998
Antarctica is gaining ice
View All Arguments...



Username
Password
New? Register here
Forgot your password?

Latest Posts

Archives

Recent Comments

Prev  915  916  917  918  919  920  921  922  923  924  925  926  927  928  929  930  Next

Comments 46101 to 46150:

  1. Climate's changed before

    Mark @345 - there's nothing directly wrong with 'the skeptic argument' as articulated by Lindzen here.

    So the "#1 global warming myth" is not really a myth at all...?

  2. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    I should add that the 5-95% uncertainty range is the dark red region and lighter red region combined.

  3. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    @ karellen

    The 25-75% uncertainty range, means that the actual temperature is expected to be within that range 50% of the time (75-25=50). The 5-95% uncertainty range means that the actual temperature is expected to be within that range 90% of the time (95-5=90).

    Another way to think about it is that the actual temperature has a 50% probability of landing in the dark red region and a 90% probability of landing in the red or lighter red region.

  4. The anthropogenic global warming rate: Is it steady for the last 100 years?

    Nice work DS

    I couldn't even show her the difference between a negative temperature trend and 'statistically significant cooling'.

  5. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    Karellen@21


    My layman's understanding is that the uncertainty ranges are also called "confidence intervals." The 5-95% range is saying that the value being tracked will fall within the wider band 95% of the time. Similarly, the 25-75% uncertainty range is saying that there is a 75% confidence that the value will fall within the narrower band. I'm sure someone with more technical expertise will chime in with a better explanation. That said, between you and me and everyone else, I would like to see the Climate Science Glossary under the Resources tab get a bit better about explaining this kind of thing.

  6. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    Sorry if I'm being stupid here, but I don't understand how the uncertainty ranges on figure 3 are supposed to work.

    At the blue line, the average value looks to be (approx) +0.6°, with the "25-75% uncertainty range" going from +0.5 to +0.7, and "5-95% uncertainty range" going from +0.3 to +0.9.

    So, the outer band is both more uncertain (95% > 75%) *and* less uncertain (5% < 25%) than the inner band? That doesn't make any sense to me.

    Surely if your average value is +0.6, you might be 50% certain that the value will be between +0.5 and +0.7, and 90% certain that the value will be between +0.3 and +0.9. But that would make your inner band "50-100% uncertain" (0-50% certain), not 25-75%, and your outer band "10-50% uncertain" (50-90% certain), not 5-95%.

    I mean, I'm pretty sure I understand what the graph is telling me. I think the graph, and the bands, make sense. It's just the labelling that I can't figure out. The Y axis is measuring temperature, not certainty. Certainty is an expected deviation from a value, with zero uncertainty being zero deviation. Zero uncertainty is not as far as you can go below the expected value, it *is* the expected value.

    Isn't it?

    Clearly, everyone else has no problem with this graph, and is reading it differently from me, but presumably you're all reading it the same way as each other. Therefore, the most obvious conclusion is that I'm the one reading it wrong. But how?

    Can someone please tell me which angle I need to tilt my head at, in order to look at this in the right way so it actually makes sense for me?

  7. K.a.r.S.t.e.N at 07:52 AM on 19 April 2013
    Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Thanks Tom!

    In order to be more transparent, I uploaded the principal figure from Lewis 2013 (Fig 3). It shows the posterior PDFs in comparison with Forest et al. 2006 (which I erroneously referred to as Forster et al. 2006 in #14):

    Figure caption: "Marginal posterior PDFs for the three climate system parameters by diagnostics employed, method used and either upper-air diagnostic EOF truncation parameter (panels a, b and c) or surface diagnostic EOF truncation parameter (panels d, e and f). Panels a, b and c use the F06 diagnostics; panels d, e, and f use revised surface and deep-ocean diagnostics, and no upper-air diagnostic. Panels show marginal posterior PDFs for, from the top, Seq (ECS), Kv (effective vertical deep-ocean diffusivity) and Faer (total aerosol forcing). In panels a, b and c, the solid red lines show marginal PDFs using the new objective Bayesian method and kua=12 (number of EOFs for upper air (ua) and surface (sfc) diagnostics), while marginal PDFs using the F06 method are shown at kua=12 (solid green lines) and at kua=14, as used in F06, (dashed blue lines) , and the dotted black lines show the published F06 PDFs. In panels d, e and f, marginal PDFs using the new objective Bayesian method are shown with ksfc=16 (solid red lines) and ksfc=14 (dotted blue lines), while corresponding marginal PDFs using the F06 method are shown with respectively solid green and dotted black lines. The box plots indicate boundaries, to the nearest fine grid value, for the percentiles 5–95 (vertical bar at ends), 25–75 (box-ends), and 50 (vertical bar in box). Parameter ranges are given by plot boundaries."

    His prior assumptions for Seq, Kv and Faer are:  0.5 to 15 K, 0 to 64 cm2/s, -1.5 to 0.5 W/m2.

    He explicitly states that: "Faer represents net forcing (direct and indirect) during the 1980s relative to pre-1860 levels, and implicitly includes omitted forcings with patterns similar to those of sulfate aerosols." As aerosol forcing seems the most constrained parameter, I wonder what a small shift in the prior to a more plausible range of -1.5 to 0.0 W/m2 would do to the final results. Given that the F06 aerosol forcing mode isn't too far off the one in Lewis 2013, it seems that the method is very sensitive to the prior choice. I might very well be wrong in this regard.

    Note that there the paper comes with Supporting Information, which isn't available yet.

  8. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Much to ponder.

    Without going so far as to suggest that this study is intended as a distratction, there is no doubt that it will be used as such in the 'debate'.

    Perhaps this bears repeating:

    # 21 shoyemore:

    Albatross #9 gives the best synopsis in my view - the meta-estimate of ECS is still stubbornly in the 2.5-3C range, and I somehow doubt if we will even pin it down more accurately. Deniers are just arguing the decimal points, as Professor Scott Denning told the Heartland Institute in 2011, but of course they did not listen then either.
  9. Models are unreliable

    @Dikran

    1. You asked "Exactly what are the policy decisions that are being made based on predictions of SSIE?".
    2. I answered "The current policy decision is to do nothing".
    3. You stated "it is a pity that you dodged my question".
    4. I stated "perhaps we disagree on what constitutes a policy decision"
    5. You stated "I agree that "do nothing" is a policy decision, nothing I wrote suggests otherwise."

    The only way I could reconcile (3) with (2) was if you didn't consider doing nothing a policy decision. That's where (4) came from.

    Essentially the point is that for Arctic sea ice extent, the observations tell you what is happening, the models predict that the Arctic sea ice will last longer than the observations suggest to the extent that we know that their projections are wrong, and hence nobody should be taking them sufficiently seriously as a basis for policy.

    Now you think that they do have an effect on policy, but do you have any evidence to support that?

    The entire purpose of the IPCC is to inform policymakers. The IPCC's AR5 draft mostly discusses models and their projections. If you think that the projections of those models are wrong, why are these projections in AR5?

    Anyway, this is the last post by me on this matter. I will read answers but won't post more. Except for CBDunkersons link, the discussion here hasn't been very useful for me and I doubt that will change.

  10. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Dana @22 & 23, the mode (unlike other parameters) is easilly estimated from the PDF shown in Fig 6 a.  Using a pixel count, I can confirm the estimate of 1.6 C for the mode of the main analysis.

    Contrary to your main article, this is not "simply a misrepresentation" of Aldrin et al.  It is misleading in that he does not point out the difference between the modal value and the central estimate commonly quoted for the IPCC, and that they are not directly comparable.  However, it is not incorrect on any matter of fact.  That is, the modal value of Aldrin et al is the value Lewis quotes, and he does specify that it is the modal value he is citing.  Consequently, as it stands, your section on "Misrepresenting Aldrin" is more misleading than is Lewis' citation of Aldrin.

    Further, Aldrin et al do not exclude cloud and aerosol effects in general.  The nine forcings included in the main analysis are specified as:

    "long-lived greenhouse gases (CO2, CH4, N2O and halocarbons), tropospheric ozone, stratospheric ozone, stratospheric H2O, the direct aerosol effect, the cloud albedo effect (indirect aerosol effect), surface albedo because of land use changes, solar radiation and volcanoes".

    (My emphasis)

    Aldrin et al do exclude cloud lifetime effects from the main analysis and other cloud effects (ie, other than the aerosol indirect effect and cloud life time effects) from all analyses.  As these are not defined as forcings in the IPCC AR4, they may or may not be included in other estimates of climate sensitivity.  De facto, they will be included in purely observational estimates, but their inclusion in model based estimates (or hybrid estimates such as Lewis 2013) may well be doubtfull.

    For what it is worth, by eyeball, inclusion of a cloud lifetime effect of -0.25 W/m^2 lifts Aldrin's modal estimate of climate sensitivity to about 1.8 C, and including a feedback of -0.5 W/m^2 lifts it to about 2.2 C.  Aldrin et al cite confidence interval for the effect of -0 to -1.5 C from Isaksen et al (2009), and from consulting the figures for Isaksen et al (it is pay walled), the mean estimate is about -0.3 W/m^2.

  11. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    I should also note again that he's only looking at the Aldrin sensitivity distribution that excludes cloud and aerosol effects when he uses that 1.6°C mode figure.

  12. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Tom @16 - to be precise, Lewis is quoting what he thinks the mode of the Aldrin PDF is.  I'm fairly sure that Aldrin never reports that mode.  That being said, the abstract does say he's comparing the modes of the two studies.

  13. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Tom Curtis, #17,

    My Master's thesis supervisor always emphasised that in Bayesian Methods, the parameters took on probabilities, and the data was regarded as a static "given", somewhat the reverse of Frequentist Methods. Lewis' work does seem to be a departure from that, and possibly has philosphical implications?

    The key moves seem to be "whitening" the observations by an "optimal fingerprint transformation", using Bayes to derive a PDF for the data. Another move is a "dimensionally reducing change of variables" in the parameter space. I am just summarising the abstract, but this is claimed to result in the tightly constrained estimates. The suspicion is that the "tight constraint" is the subjective motivation rather than the objective result.

    Unfortunately, my own knowledge is insufficient to properly critique his methods, but I hope somebody does. After all, the statisticians flocked around to "help" Michael Mann! If they can do it for a physicist, they can do it for a retired businessman.

    Albatross #9 gives the best synopsis in my view - the meta-estimate of ECS is still stubbornly in the 2.5-3C range, and I somehow doubt if we will even pin it down more accurately. Deniers are just arguing the decimal points, as Professor Scott Denning told the Heartland Institute in 2011, but of course they did not listen then either.

  14. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    BBD @19, I also haven't read the paper.  However, based on the abstract, the main result, ie, that which is discussed first, is the change in methodology and its effect on the climate sensitivity estimate.  Of course, Lewis may want to put his revised low CS estimate up front in discussion, but that should not be assumed to be the reason the paper was published.

  15. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Tom Curtis # 17

    By happy coincidence I have just said the exact same thing as your first paragraph in a comment elsewhere ;-)

    I haven't read the paper (only the blog posts NL has made elsewhere) but I suspect that what you go on to say is correct. Must reserve judgement for now, though.

  16. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Moderator [JH]

    Unless you have just added the defintion of PDF to your glossary, it is already there. The term is underscored in text and the definition appears in a pop-up when the cursor is passed over it.

    Moderator Response:

    [JH] I may have been viewing it on the rolling summary of all comments. It appears that our nifty glossary does not do its magic in that venue.  

  17. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    BBD @15, the main result of Lewis 2013 is the different approach to bayesian estimates of climate sensitivity, ie, the application of uniform priors to the observations rather than to the parameters.  Regardless of the merits of the final estimate of CS, if that method is not transparently flawed it should be placed before the scientific community so that its merits can be discussed and analyzed.

    As to the estimate of climate sensitiivty itself, it uses the output of a single 2 dimensional model to quantify likilihood of the three paramaters.  Consequently it is no better than the accuraccy of that model.  Further, the final quoted value depends on the addition of a number of La Nina effected years to the data.  Given the use of a 2-D climate model, there is no way the effects of El Nino on temperature are modelled, and in consequence the resultant estimate of climate sensitivity is certainly too low.  The 2.4 K mode quoted for the change of method only is therefore more likely to be accurate than Lewis' final result.  Ideally, given that the model does not account for ENSO, it should be used against ENSO adjusted temperature data, such as that from Foster and Rahmstorf.  Doing so will likely give a mode between 1.6 and 2.4 K, but closer to the later than the former; and is likely to give a mean value (the value cited by the IPCC and therefore the correct comparison with the IPCC range) slightly higher than that. 

  18. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Dana, Aldrin et al quote a mean CS of 2 C.  Lewis quotes the mode of the Aldrin PDF on climate sensitivity as 1.6 C, which is correct.  The confusion arises because he is citing a different value.

    Lewis, in fact, has a penchant for quoting the mode of climate sensitivity analyses because it is, as he says, the maximum likilihood result.  Personally I do not think it is the best value to quote.  Rather the median (where it can be determined), ie, the 50/50 mark is far more informative if you must quote a single number.  Of course, with a long tailed PDF, the mode will always be less than both the median and the mean of the PDF.  Without going so far as to say that is Lewis' reason for preferring it, that does make it a tremendously convenient number for "skeptics" to quote.

  19. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    K.a.r.S.t.e.n

    # 10; # 14


    I remember your discussion with NL and PaulS at James Annan's very well. And most illuminating it was too. What puzzles me (troubles me?) about this is that this study has been accepted by J. Climate (not say, E & E) after an apparently long period under review and yet these questions apparently remain outstanding.

  20. Dumb Scientist at 03:50 AM on 19 April 2013
    The anthropogenic global warming rate: Is it steady for the last 100 years?
    Due to my inexperience, I have found it difficult to answer individual questions, mostly of them are technical in nature. I have tried to explain the technical details, but that did not seem to work. [KK Tung]

    It didn't work because your technical details didn't address my point, as many have noted above. That's why I distilled my point into two yes/no questions. I look forward to the educational answers you've undoubtedly provided in your second post.

    There is no obviously right or wrong answers; this is always the case when the science is unsettled---when the science is settled I will have to move to another field.

    Science can be (loosely) defined as the search for answers which are less wrong than previous answers. That slogan is meaningless because all science has uncertainties.

    One argues for the reasonableness of the assumption using evidence and physical mechanisms, and then proceeds to deduce what that assumption will lead to as consequences. In scientist publications, one always lists the assumption clearly so that others could refute it. We should argue whether the assumption is supported by the available evidence or not. But claiming that the argument is circular simply based on the technical fact that the consequence arose from the assumption is missing the bulk of the arguments in our paper leading to that assumption.

    It's true that all science is based on assumptions, such as conservation of energy. But no study based on the assumption of energy conservation would conclude that energy is conserved. That would be a circular argument.

    You regressed global surface temperatures against the AMO in order to determine anthropogenic warming. Because the AMO is simply linearly detrended N. Atlantic SST, this procedure would only be correct if AGW is linear. Otherwise you'd be subtracting AGW signal, sweeping some AGW into a box you've labelled "natural" called the AMO. So you're assuming that AGW is linear, and you're also concluding that AGW is linear.

    I suspect a lot of that may have to do with the fact that our paper is behind a "pay wall", so that many posting here may not have read more than the abstract. I have posted a free link to the entire paper in the first few lines of my post. A correction: The link to our PNAS paper was deleted in this first post. I hope it survives in my second post, where it is provided again.

    Actually, the first link in my article was a free link to the entire paper Tung and Zhou 2013.

    Dr. Tung, I'd also like to thank you for your participation here. Right now I'm also trying to explain Antarctic ice mass balance at Jo Nova's, and the disappointing responses made me appreciate this civilized discussion even more.

  21. K.a.r.S.t.e.N at 03:14 AM on 19 April 2013
    Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    It's not that Nic wouldn't have been aware of the aerosol forcing issue. PaulS and myself, we had an interesting discussion with him over at James Place. His reluctancy to accept the opinion of other (more informed) people on the subject struck me as quite bold. Realizing that he didn't even discuss the uncertainties regarding the aerosol prior in the paper, is something which I consider a huge disappointment. Had he applied a more objective areosol forcing estimate (one may think of an "expert prior"), he wouldn't have come up with such a low ECS (or effective CS for that matter) number. I think he missed the opportunity to demonstrate that his method is superior to Forster et al. 2006 (who were using uniform priors). Too bad ...

  22. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Dana @12,

    Agreed.  The misrepresentation of Aldrin et al by Lewis is egregious and should have been caught. The journal can still correct the paper (the online version is an "early release" and not necessarily the final version), or failing that Lewis should publish a corrigendum. 

  23. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Interesting point, Karsten.  That's an issue with a Bayesian approach - it's important to use accurate informed priors.

    It bothers me a bit that the misrepresentation of Aldrin's results and calling this an estimate of equilibrium sensitivity weren't caught during the peer-review process, or at least weren't fixed in the final version.  It's possible that the issues were raised in the extensive review comments but Lewis disagreed and the editor allowed the errors to be published.  Whatever the reason, those are significant issues with the paper, in my opinion.  He should at least have discussed the relationship between effective and equilibrium sensitivity.  As it stands, the paper simply says it's estimating equilibrium sensitivity, which isn't accurate.  And the Aldrin error being in the abstract is unfortunate too.

  24. The anthropogenic global warming rate: Is it steady for the last 100 years?

    A correction: The link to our PNAS paper was deleted in this first post.  I hope it survives in my second post, where it is provided again.

  25. The anthropogenic global warming rate: Is it steady for the last 100 years?

    Replying to KR: Thank you for the comments.  I appreciated the opportunity to discuss the various issues involved. 

    Due to my inexperience, I have found it difficult to answer individual questions, mostly of them are technical in nature. I have tried to explain the technical details, but that did not seem to work.  Now I appreciate why Gavin Schmidt in Realclimate.org won a prize for communication.

    A general comment not related to any post in particular: There is no obviously right or wrong answers; this is always the case when the science is unsettled---when the science is settled I will have to move to another field.  One argues for the reasonableness of the assumption using evidence and physical mechanisms, and then proceeds to deduce what that assumption will lead to as consequences. In scientist publications, one always lists the assumption clearly so that others could refute it.  We should argue whether the assumption is supported by the available evidence or not. But claiming that the argument is circular simply based on the technical fact that the consequence arose from the assumption is missing the bulk of the arguments in our paper leading to that assumption. I suspect a lot of that may have to do with the fact that our paper is behind a "pay wall", so that many posting here may not have read more than the abstract.  I have posted a free link to the entire paper in the first few lines of my post.

  26. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Hi Karsten @ 10,

    "Correspondingly, he got a median posterior of -0.3 W/m2 with his method. It is certainly a underestimate, no doubt about that. So I don't put much credibility into his most likely ECS values, as they are an underestimate too (perhaps strongly so)."

    You have just raised a huge red flag.  His value for aerosol forcing is undoubtably wrong and as such way too small.

    Remember, peer review is not perfect.  Also, this paper was published in a reputable journal, but even they err from time-to-time.  It is my understanding from what Lewis has been saying that two of the reviewers were very tough.  They allegedly even brought in a forth reviewer to help settle things.

    Regardless, the true test of Lewis starts when it appears in press.  Somehow I very much doubt that it will stand the test of time.  With that all said, kudos to Lewis for submitting his work for peer review; it also shows that there is no censoring going on in journals as "skeptics" claim.

  27. K.a.r.S.t.e.N at 02:16 AM on 19 April 2013
    The anthropogenic global warming rate: Is it steady for the last 100 years?

    Dr Tung, although one might receive my above comment (#30) as destructive (if not offensive), I'd like to stress that I merely expressed my personal opinion ... unfortunately in a rather dismissive and brisk tone for which I humbly apologize. I am just too tired of all these unconvincing attempts to blame the AMO for everything. As I said, I am looking forward to the second post (which is certainly gonna be up soon).

  28. K.a.r.S.t.e.N at 01:58 AM on 19 April 2013
    Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Thanks for putting this together, Dana! Nic Lewis seems unwilling to accept the current (AR5 SOD) aerosol forcing estimate. The total anthropogenic ARF (all direct + indirect effects in terms of adjusted TOA forcing) is -0.9W/m2 since 1750. The most likely range is –1.5 to –0.3 W/m2. Scaling it to 1860 (Lewis' start year) might reduce the aerosol forcing slightly to -0.7 or -0.8 W/m2. However, his prior for the total ARF distribution lies between +0.5 and -1.5 W/m2. Correspondingly, he got a median posterior of -0.3 W/m2 with his method. It is certainly a underestimate, no doubt about that. So I don't put much credibility into his most likely ECS values, as they are an underestimate too (perhaps strongly so). He put an aerosol forcing which he believed to be true and he got a low ECS which he (apparantly) wishes to be true. What a surprise! I am underwhelmed ...

    I really don't understand why no one has bothered to check the assumptions which went into his approach. Why hasn't the unwarrantedly low aerosol forcing prior not been picked up during peer review? I am a bit surprised to see what makes it through peer review as of late. Perhaps just stronger awareness or pure coincidence.

  29. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    While recent research has suggested that an equilibrium climate sensitivity (ECS) above +4.5 C is highly unlikely, a meta analysis of all the evidence continues to stubbornly show a best estimate of ECS to be near +3 C.

    This is pretty worrisome given that on our current trajectory we will probably quadruple CO2 levels before the end of the century (remember that estimates of ECS are traditionally for "only" a doubling of CO2).

    Yet, "skeptics" are trying to reframe the argument by using the Lewis paper (which IMHO looks to have yielded an estimate of transient sensitivity rather than for ECS). "Skeptics" seem to think that a few studies out of literally dozens suggesting a relatively low ECS means that humanity should bargain/bet on ECS being < 2 C so that we can continue on with business as usual. This is both risky and irresponsible.  Engineers design components to fail at the upper end of the range, not the most likely value, nevermind the lower range.

    I doubt many of us would willingly board an aircraft or spaceship that we have been told meets only the minimum safety criteria.  Well, planet earth is our one and only spaceship and we would be prudent to not bet on ECS being < 2 C, especially given the preponderence of evidence to the contrary.

  30. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    For further information on how The Mail "con" their readers on climate change please see:

    http://econnexus.org/the-strange-tale-of-the-mail-and-the-snow-dragon/

    Amongst other things, they "conveniently omit" relevant comments

  31. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Ray - The issues of satellite vs. radiosonde vs. model temperatures have been long and rather complicated; Thorne et al 2011 has a good review. 

    Radiosondes have limited sampling and poor consistency in instrumentation over time, satellite data has multiple identified (corrected) and suspected (not) biases, and it may well be that models are insufficiently handling ozone or other factors. Satellite and model results are the furthest apart, with radiosondes somewhere in between. Right now, given the uncertainties in each of these, all that can be said is that the various numbers are not far enough apart given uncertainties to distinguish between their values. 

    From that report: "...a person with one watch always knows what time it is, but with two watches one is never sure." 

    I would suggest taking further discussion on those topics to the threads on satellite tropospheric measurements or those on the tropospheric 'hot spot'. Given those uncertainties, I would consider sensitivity estimates from tropospheric temperature model comparisons to be far less accurate than other methods right now. 

  32. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    A CS of 1.6C pushes global temperatures >2C above pre-industrial levels in ~100 years at present emission rates.

    So even if Lewis (2013) is accepted at face value, it doesn't support a a case for business as usual emissions.

  33. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Here's another: van Hateren http://link.springer.com/article/10.1007/s00382-012-1375-3

    TCR of 1.5. He gives a 'millenium scale' sensitivity of 2.0, but with heavy caveats on it.

    The big challenge in comparing these is to work out which forcings they've used, and most importantly how big the aerosol term is.

  34. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Dana linked Troy Master's paper at the end of the article, along with the new Hansen paper. I haven't digested the content, but it's very well written.

    We've got a lot of evidence emerging from three stands of evidence on sensitivity - paleo, GCM and recent climate. A review article trying to collate these and detect any patterns would be really interesting. Publishable if done well.

    My impression is that the recent climate results are coming out lower than the other two, even without taking into account the recent black carbon result. The aerosol uncertainty plays a big role too - if Hansen is right about a higher aerosol forcing then some of the discrepancy is resolved, however most sources I've looked at use lower values than Hansen.

    My naive 2-box model gives a TCR of 1.7 with Hansen's forcings (in line with GCMs), dropping to 1.4 for the Potsdam forcings (which are missing the 2nd aerosol indirect effect), and 1.0 if I simply double the black carbon term (I don't know how it should evolve over time though - maybe BC grew earlier, causing the early 20thC warming and plateau).

    One very interesting result from the 2-box model is that including the last decade increases rather than decreases the TCR. Adding a few more years of high temperatures pulls TCR up because the hotness of those years is more significant than the lack of trend within that period. This contradicts James Annan's comments on the subject. It would be interesting to investigate that discrepancy.

  35. Further Comments on The Economist's Take on Climate Sensitivity

    The Economist has now published some letters from their readers on their original article.

    They lead with "current climate-change policy is an expensive waste of time" and relegate "Quantum physics and thermodynamics... yield a baseline climate sensitivity of about 3°C" to second place.

  36. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    An acquaintance alerted me to a post on Bihop Hill's blog that a forthcoming paper by Troy Masters estimates ECS as

    a likely (67 %) range of 1.5–2.9 K for equilibrium climate sensitivity, and a 90 % confidence interval of 1.2–5.1 K.

     

  37. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    Mod,

    Typo in above post - " I think James Annan's point is borne out that values above 4C are increasingly of low probability."

     

  38. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    dana,

    Great post, just what we need. The point about single studies is one that contrarians will ignore of course.

    There is some confusion between the Aldrin(2012) paper and the "unpublished Norwegian study". Some commentators seem to think these are the same.

    There have been other papers published in the last couple of years, on the transient climate sensitivity (TCS) and the equilibrium climate sensitivity (ECS).

    Padilla et al (90% conf 1.3 - 2.6C for TCS)

    Gilett et al (1.3 -1.8C for TCS)

    Bitz et al (1.72C for TCS)

    Schwartz (1.6 to 4.3C for ECS)

    Schmittner et al (1.4 to 3.5C likely range for ECS).

    Hargreaves et al (2.5C for ECS)

    Hansen et al Forthcoming (2 to 4C for ECS)

    Putting these with Aldrin (1 - 4C for ECS, PALEOSENS (2.2 to 4.8 for ECS), Lewis (1 - 3C for ECS) is indeed an outlier. However, I think James Annan's point is borne out that values about 4C are increasingly of low probability. How much we should "bet" on that is another matter. A re-drawing of Figure 1 for more recent estimates of ECS would be useful.

    Contrarians emphasise the whittling back of the long upper tail in earlier estimates as a reduction in the value of ECS, but this is not so.

    PS Does anyone else find the TCS, ECS and now Effective Climate Sensitivity as confusing as I do? I thought I was clear on the TCS and ECS, but now I have to understand another parameter (sigh!). :(

    Links to papers:

    Padilla et al (90% conf 1.3 - 2.6C for TCS)

    http://journals.ametsoc.org/doi/abs/10.1175/2011JCLI3989.1

    Gilett et al (1.3 -1.8C for TCS)

    http://onlinelibrary.wiley.com/doi/10.1029/2011GL050226/abstract

    Bitz et al (1.72C for TCS)

    http://journals.ametsoc.org/doi/abs/10.1175/JCLI-D-11-00290.1

    Schwartz (1.6 to 4.3C for ECS)

    http://www.ecd.bnl.gov/steve/pubs/ObsDetClimSensy.pdf

    Schmittner et al (1.4 to 3.5C likely range for ECS).

    http://people.oregonstate.edu/~schmita2/pdf/S/schmittner11sci_man.pdf

    Hargreaves et al (2.5C for ECS)

    http://onlinelibrary.wiley.com/doi/10.1029/2012GL053872/abstract

    Hansen et al Forthcoming (2 to 4C for ECS)

    http://arxiv.org/vc/arxiv/papers/1211/1211.4846v1.pdf

     

     

  39. Climate Sensitivity Single Study Syndrome, Nic Lewis Edition

    From comparisons of satellite measurements from 1979-2012 of global temperature with estimates of these temperatures from 1975-2025 by 44 GCMs, Roy Spencer has shown divergence of satellite readings from model readings, with satellite readings indicating weaker warming, than the average from the GCMs. He notes that although the average of the models was higher than the satellite measurements, a few models, usually those with lower climate sensitivity, support weaker warming. These comparisons are in accord with Lewis's suggestion of lower climate sensitivity. Spencer's comparisons are here

  40. Real Skepticism About the New Marcott 'Hockey Stick'

    It appears I owe Brandon Shollenberger an apology.  I mistook him as having a legitimate (if overblown) concern that the Romm graph spliced the MIT predictions directly to the Marcott uptick, thereby temporally displacing one or the other by 50 years.  That splice misrepresents the data and hence should not have been done.  The difference it makes in the graphic is so slight, however, that it is unlikely to decieve anybody.

    It turns out that Shollenberger's actual objection to the graph is that when it says it shows Marcott's reconstruction, it actually shows Marcott's reconstruction rather than some truncated version of it.  That is not a legitimate criticism.  If you say that you show the Marcott reconstrution, then you must show the reconstruction shown by Marcott et al in their paper, ie, the one shown by Romm.  Uptick and all.  Doing otherwise would be dishonest.  You should then note that the uptick in the reconstruction is not robust - but there is no basis for not showing it.

    Indeed, the final value in the uptick in the main reconstruction shows a positive anomaly of 0.05 C, compared to the 0.19 C of the 1981-2000 average in the instrumental record.  If the vertical component of the uptick is Shollenberger's concern, that fact shows him to be indulging in shere obfustication.  The instrumental records shows very robustly that the twentieth century uptick is larger than that shown by Marcott.  Marcott's reconstructed uptick is not robust, and is rises too rapidly too soon, but when showing a continuous period through to 2100, it is the displacement on the x-axis, not the y-axis which is the concern.  

    In any event, I do apologize to Brandon for incorrectly understanding him as making a valid though trivial point rather than, as apparently he now inists, demanding that Romm act unethically in preparing his graph.

    I note that Shollenberger says:

    "Removing the spurious uptick from Romm's graph creates a glaringly visible gap where the two lines no longer meet. That's a huge change."

    Bullshit!

    The only way it would have made a "glaringly visible gap" is if all temperatures post 1850 had been excized to create a denier special, ie, by hiding the incline.  If the robust uptick (as shown by the RegEM or Tamino's difference method) is shown the gap is visible, and clearly inconsequential.  Either that is what Shollenberger tried, or (more likely) he in fact did not check his claim at all prior to making it.

  41. Real Skepticism About the New Marcott 'Hockey Stick'

    Furthermore, Brandon, if you feel that the Mann 2008 data is significantly different than the last 150 years of the Marcott data (a difficult argument given the scaling in the opening post graph) - show it. 

    Thou doth protest too much, methinks...

  42. Real Skepticism About the New Marcott 'Hockey Stick'

    Brandon Shollenberger - You have complained repeatedly about the "uptick". but Marcott et al 2013 does not use the last 150 years of their reconstruction for alignment with instrumental temperatures in any way - they align with 1000 years of overlap with Mann 2008, which is then itself aligned with overlapping instrumental data - three steps, not two, those 150 years are not in play. 

    So your various complaints about alignment over that 150 year period are nothing but a strawman argument, completely ignoring the interim step. Furthermore, your repeated assertions of that argument indicate that you have indeed not read (or perhaps understood?) the Marcott paper. 

    And as I pointed out above, the Marcott, Mann, and instrumental data can all be correctly referred to as "Reconstructed" - as stated in the graph. You have, quite frankly, no grounds for your complaints. 

  43. Brandon Shollenberger at 13:39 PM on 18 April 2013
    Real Skepticism About the New Marcott 'Hockey Stick'

    scaddenp @82, glad to hear it. It looks like that's the all the agreement there'll be here.

    Tom Curtis @83, you just created a strawman whereby you claimed the issue I raised deals with the horizontal aspects of the graph when in reality it deals with the vertical. Grossly distorting a person's views is bad enough, but using it to make implicit insults is just pathetic.

    That said, I'm happy to take your word when you say Romm's graph is flawed in more ways than I stated. I don't know why you'd say I think the issue "is not even worthy of a note" though. I didn't check the exact length of the line so I wasn't aware of the issue. I don't think anyone should make much of that.

    If you want to argue I didn't criticize the graph for as many reasons as I could have, you're welcome to. It's perfectly possible Romm did a worse job than I thought.

    KR @84, you can keep saying I should read the paper, but that won't make your claims correct. It probably won't even convince people I haven't read the paper. Removing the spurious uptick from Romm's graph creates a glaringly visible gap where the two lines no longer meet. That's a huge change.

    Moderator Response:

    [JH] You are skating on the thin ice of sloganeering. Please cesase and desist or face the consequences.

  44. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    @sphaerica

    "In a chaotic system, the observed values will wiggle in and out of the range of values from time to time, even in and out of the 95% range." Yes I get that part, but a 5%-95% interval should also have close to the same number of overshoots and undershoots, which is also what the author stated.

    "I think you're getting confused by the thickness of the line for observed temperatures."

    I figured that was probably the case. I think you guys should probably make that figure 3 chart clearer because the current chart looks like it has a lot more dips outside the interval than peaks outside the interval. Just a suggestion.

  45. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    Kayell - "I'm interested to know what specific thermodynamic mechanism is being utilized by the amounting atmospheric CO2 to restrain the release of the absorbed solar heat from the ocean soas to raise the OHC without warming the troposphere and the surface first ...?"

    There's an excellent discussion over at RealClimate titled Why greenhouse gases heat the ocean, and another with rather more detail here at SkS under How Increasing Carbon Dioxide Heats The Ocean. Greenhouse gases slow radiation to space from the atmosphere, (relatively) warmer atmosphere warms the viscous ocean skin layer via downward IR, the decreased thermal gradient slows energy loss from the oceans. 

    Under La Nina conditions (such as recently) the oceans and atmosphere are both cooler than under neutral conditions, but the gradients are still less, and ocean warming more, than they would be without the increased GHGs. John Nielson-Gammon has a very interesting post showing that El Nino, La Nina, and ENSO-neutral years are all trending upwards at ~0.16 C/decade - due to increased GHGs. 

    ---

    I will note that both of these items (ocean skin layer and John N-G's analysis) were discussed in some detail on a previous El Nino thread that you extensively participated in. I am disappointed that you (apparently) did not fully follow that thread.

  46. Real Skepticism About the New Marcott 'Hockey Stick'

    Relative to Tom Curtis's post, the terminology used in the graph is a split between "Reconstructed" and "Predicted" temperatures.

    I would have to say that the Marcott et al 2013 Holocene reconstruction, Mann 2008 paleo work (which as stated in Marcott joins that reconstruction to the present via a 1000 year overlap - not the last 150 years over which there has been some controversy), and recent instrumental records all meet the definition of "Reconstructed" temperatures. 

    As noted before, Brandon's complaints about "...a graph that relies upon the uptick..." could best be addressed by simply reading the Marcott et al paper, since the graph does not rely on that feature of the Marcott data. 

  47. The anthropogenic global warming rate: Is it steady for the last 100 years?

    Dr. Tung - I would also like to express my appreciation for the very interesting discussion. 

  48. Real Skepticism About the New Marcott 'Hockey Stick'

    Brandon Shollenberger @78 & 81,

    First, for everybody's convenience, here is the graph in question:

    A brief examination shows that there are two flaws in the graph.  The first is that, as noted by Brandon, the reconstruction should not lead directly into the projection.  That is because the terminal point of the reconstruction is 1940 (or technically, the 1930-1950 mean), whereas the the initial point of the reconstrution is 1990.  That time seperation represents about one pixel on the graph.  It is an important pixel, however, and the one pixel seperation should be there.  Further, the modern instrumental record should probably have been shown.

    Personally I am not going to fault Romm for that because, at the date when the graph was created (March 8th) preceded extensive discussion of the cause of the uptick by a week.  That missing pixel represents an error of interpretation rather than the misrepresentation of common knowledge Shollenberger presents it to be.  In light of that discussion, however, Romm should have included an update pointing out that issue; and his nearest thing, the follow on post has far more problems in the title than in the graph.

    Confining ourselves to the graph, however, the second problem is the projections.  Romm identifies the projections as being those of the MIT "No Policy" case.  Romm significantly misrepresents that case.  Specifically, he shows a projection of 8.5 F increase relative to the 1960-1990 mean temperature.  As it happens, the MIT median projection is for a 9.2 F increase relative to 1990.  Romm understates the projection by more than 0.7 F.  (More, of course, because the "1990" temperature, more accurately the 1981-2000 mean, is higher than the 1960-1990 mean.)

    This second error makes a 15 pixel difference to the graph.  Now, what I am wondering is what sort of though process was behing Shollenberger's decision that the one pixel difference was worthy of comment and makes a significant difference, whereas the 15 pixel difference is not even worthy of note?

  49. Global Warming is Accelerating, but it's Still Groundhog Day at the Daily Mail

    engineer,

    I think you're getting confused by the thickness of the line for observed temperatures, but you're also missing the point.  In a chaotic system, the observed values will wiggle in and out of the range of values from time to time, even in and out of the 95% range.

  50. Real Skepticism About the New Marcott 'Hockey Stick'

    Okay, Brandon. I see your issue. I was actually confusing Paul's graph with Koomey's graph. I agree that graph at start of the article is better.

Prev  915  916  917  918  919  920  921  922  923  924  925  926  927  928  929  930  Next



The Consensus Project Website

THE ESCALATOR

(free to republish)


© Copyright 2024 John Cook
Home | Translations | About Us | Privacy | Contact Us