Science

 According to Torontos Globe and Mail (Sept. 8), Canadas greenhouse gas emissions have increased at a faster rate than in the United States since 1990.  The figures, agreed by both governments, show that Canadas emissions have grown by 20 percent, while Americas have risen by only 14 percent.  If emissions generated by Canadian energy exports to the U. S. are removed, then the Canadian rise drops to 15 percent.

 Harlan Watson, chief climate negotiator for the U. S. State Department, said the figures showed how difficult emissions reductions are.  He told the Globe and Mail, As we say, actions speak louder than signing pieces of paper.  I don’t want to be harsh on Canada.  We have some of the most self-righteous folks in Europe, of course, and they’re [also] having a great deal of difficulty.  Energy use accounts for something in the order of 80 per cent of the emissions.  Until we get a handle on that, there’s no way to get from here to there, absent an economic collapse.

Canadian researchers Steve McIntyre and Ross McKitrick have published an update on their work that raises even more doubts about the legitimacy of the so-called hockey stick temperature reconstruction of the University of Virginias Michael Mann and colleagues (MBH98).


Nature magazine, in which MBH98 originally appeared, recently published a Corrigendum by Mann, which corrects some of the errors uncovered by McIntyre and McKitrick.  Nature had also agreed to publish a further 800-word criticism by McIntyre and McKitrick of the methodology and results obtained by MBH98.  However, after adding a third reviewer who was much less supportive of their findings than those who recommended publication of the initial submission and after reducing the number of words allowed for the piece to 500, Nature informed McIntyre and McKitrick that they had decided against publishing their article.



Nature said, In the light of this detailed advice, we have regretfully decided that publication of this debate in our Brief Communications Arising section is not justified.  This is principally because the discussion cannot be condensed into our 500-word/1 figure format (as you probably realise, supplementary information is only for review purposes because Brief Communications Arising are published online) and relies on technicalities that do not bring a clear resolution of the underlying issues.



McIntyre and McKitrick comment that, This decision primarily reflected the views of the new reviewer, who stated: Generally, I believe that the technical issues addressed in the comment and the reply are quite difficult to understand and not necessarily of interest to the wide readership of the Brief Communications section of Nature. I do not see a way to make this communication much clearer, particularly with the space requirements, as this comment is largely related to technical details.  In other words, Nature decided that a discussion of fundamental issues of how the hockey stick was arrived at would not be of interest to many people.



McIntyre and McKitrick have posted their draft submissions and correspondence with Nature at http://www.uoguelph.ca/~rmckitri/research/fallupdate04/update.fall04.html.  The site also discusses another aspect of their work, an investigation into various lacunae and additional inaccuracies in MBH98.  In the process, they describe how neither Prof. Mann nor Nature magazine have provided access to important parts of the data on which MBH98 relies.



They comment, Reluctance on the part of Mann et al. and Nature to produce the results for their experiments, and in particular for the AD1400 step, would be one thing if the source code that generated them were available; but the refusal to provide either one is completely unjustifiable, especially since Nature based its decision against our paper, in part, on claims about the RE statistics that can only be verified by looking at the “experiment” results.  We surmise, based on our implementation of the methodology, that the R-squared and Coefficient of Efficiency (as this is defined in paleoclimate studies) statistics fail to reach statistical significance for the AD1400 step. It may also show that there are other problems in MBH98 besides the ones that we have described already.


They conclude, While we are frustrated that the time invested in the Nature process did not result in their willingness to correct the publication record therein, it did at least allow us to clarify several methodological issues, especially the crucial role of the controversial bristlecone pine series.  We will submit a revised article to a peer reviewed publication.  We have also submitted an abstract for a planned presentation at the forthcoming AGU meeting in December.


The global temperature readings for August 2004 from the Earth System Science Center of the University of Alabama in Huntsville, based on satellite readings, show the recent trend of cooler than normal temperatures is continuing.


 The global composite temperature for August was 0.07C (about 0.13F) cooler than the twenty-year average for 1979-1998, caused mainly by a cool Southern hemisphere.  The Northern hemisphere was 0.09C warmer than the average.



UAHs Dr. John Christy said that the data showed that, For the continental U. S., August completed the coolest summer in a dozen years and the second coolest in the 26-year satellite record.  Moreover, the coolest summer, 1992, was a result of factors not present this summer.  That summer, said Dr. Christy, was cool due to the global shading from the Mount Pinatubo eruption.  This summer there isnt anything like that to pin it on.  It was just a cool summer.

It’s been a disappointing summer for global warming alarmists.

Hollywood, Mother Nature and the media just haven’t cooperated. Even with the unusual situation of two successive hurricanes pounding Florida and another bearing down imminently, global warming hysteria seems to be on ice for now.


The summer began with so much promise for the climate control crowd with the release of the global warming disaster movie, “The Day After Tomorrow.” While the movie made plenty of money, global warming activists wanted much more than that. They hoped the movie would foment global warming hysteria in the same way that “The China Syndrome” and “Silkwood” contributed to public sentiment against nuclear power plants.


Instead, the movie was so over-the-top with implausible weather phenomena that no one not even the usually global warming-sympathetic media took it seriously. Then, unlike the movie, the real “day after tomorrow” turned out to be pretty nice.


Across the U.S., summer temperatures were cooler than normal. Aberdeen, S.D., experienced its coolest August in 115 years with an average temperature seven degrees below normal (63.4 vs. 70.5).


Michigan officials attribute a dip in visits to state parks and other outdoor attractions to cooler weather.


“The water temperature along the [Lake Michigan] beach usually is in the 60s from Memorial Day to Labor Day, reaching the low 70s during the hottest days of July and August. Except for a few days in early July and again in mid-August, the water temperature never reached 70 this year,” a Michigan official told The Associate Press.


Portland, Maine’s high temperature of 82 degrees in July was the coldest high temperature ever recorded for the month, and the average daily high was four degrees below normal at 74.6 degrees.


An apparent heat wave in Monterey, Calif., this week “is actually fairly typical, it just seems to stand out because we’ve had a particularly cool summer so far,” a National Weather Service spokesperson told the Monterey County Herald.


It’s hard to get people worked up about “global warming” when it’s too cool to get in the pool.


The final indignity of this summer forced upon the global warmers came this week with media reports on Hurricanes Charley, Frances and the coming Ivan. Though stronger, more frequent hurricanes are the sort of severe weather that activists want us to believe are attributable to global warming, meteorologists and the media just aren’t cooperating.


The Miami Herald interviewed National Oceanic and Atmospheric Administration meteorologist Stanley Goldenberg and reported on Sept. 8: “Research Goldenberg conducted with NOAA scientist Chris Landsea, private expert William Gray and others found distinct patterns of low-activity hurricane periods and high-activity periods, each of which endured for decades. These patterns, unrelated to the current concern over global warming, are caused by regular cycles of oceanic and atmospheric phenomena, such as unusually warm water in hurricane breeding grounds.”


The New York Times reported on Sept. 5: “Global warming is not a significant factor in this year’s storminess, experts said. While some climate models predict that warming might eventually mean somewhat stronger hurricanes, that effect is expected to be very small compared to the natural hurricane cycle.”


The Washington Post began a Sept. 3 article with, “Bad luck, not global warming, is the best explanation for the arrival of two severe hurricanes on the Florida peninsula in three weeks, several experts said yesterday.”


These reports must be particularly bitter pills to swallow for the activists since the Times, Post and Herald don’t typically pass up opportunities to promote the agenda of global warmers. As the summer fades into fall, so too will fade the season for global warming polemics.


Steven Milloy is the publisher of JunkScience.com, an adjunct scholar at the Cato Institute and the author ofJunk Science Judo: Self-Defense Against Health Scares and Scams(Cato Institute, 2001).

UHIE? UHIE who?

by William Yeatman on September 9, 2004

in Science

Not who but what


UHIE is the acronym for Urban Heat Island Effect.

So what is it?


Very simply, UHIE is the result of buildings and pavement absorbing greater amounts of broad spectrum solar radiation than vegetation does. This makes a difference because the atmosphere does not absorb solar radiation across the spectrum but greenhouse gasses (GHGs such as water vapor, carbon dioxide, methane…) absorb infrared. As buildings and pavement re-radiate absorbed broad spectrum radiation in the infrared band absorbable by GHGs we get an increase in the solar energy available to warm the air in cities and towns. With perpendicular walls increasing collector area, acting as trombe walls (storing and re-radiating energy at night) and excluding cooling breezes and with blacktop paved areas reducing surface albedo (reflection), cities and towns have become highly effective accumulators of solar warmth.

Doesn’t this prove global warming is true?


In a word – no. All it does is remind us of something we’ve known and observed for ages – partially enclosed spaces warm up more in the sun than similar open spaces do.

How significant is UHIE?


It can be very significant. For example, here’s a NASA item on “Hotlanta” stating that urban Atlanta can reach 5 to 8 degrees Fahrenheit (~2.8 to 4.4C) or higher than surrounding rural areas. It’s been quite topical too, “HOT-LANTA EFFECT – Urban heat: Growing season, hardiness affected” – “Heat from cities is lengthening the growing season and allowing for more warm-weather plants” (The Atlanta Journal-Constitution) – see Urban Heat Islands Make Cities Greener (NASA GSFC release).

The above release suggests that UHIE exerts an influence over an area almost 2.5 times the size of the city and we surmise it’s likely to be similar for towns and villages too. This suggests that (being generous) up to 10% of the Earth’s non-ice land surface may be affected to some extent – perhaps 3% of the total globe surface. Because the atmospheric column is not uniformly affected, merely the near-surface portion over the city and tapering off on the downwind “plume,” significantly less than 3% of the troposphere will exhibit UHIE.

To what extent this effect is mitigated by increased albedo from croplands where more absorptive forest and scrub once grew is difficult to establish but anyone who has stood in the midst of a few thousand acres of ripe wheat on a hot summer’s day can tell you that the albedo is indeed significant.

How global is the near-surface temperature recording coverage?


According to the NASA Goddard Institute for Space Studies (GISS) Surface Temperature Analysis page, not very:

“The Figures below indicate:
(a) the number of stations with record length at least N years as a function of N ,
(b) the number of reporting stations as a function of time,
(c) the percent of hemispheric area located within 1200km (~750miles) of a reporting station.


Compared with the near-global coverage of NOAA satellite-mounted MSUs (the “satellite record”), reporting stations on a grid of about 1500miles is not particularly good coverage. Given that so many near-surface stations are located in urban settings and even airports it is unavoidable that the near-surface record is affected by UHIE.

Could UHIE create the illusion of “global warming’?


It is difficult to see how it could not. The late John L. Daly explored this problem in a report to the Greening Earth SocietyThe Surface Record: Global Mean Temperature and how it is determined at surface level

Is UHIE the reason for the increasing discrepancy between the MSU record and GISS GISTEMP?


As the above graph indicates, the only time in the last decade that agreement between the records has improved was when the troposphere reacted to the powerful 1997/98 El Nio event. With the MSU record in good agreement with data from radio-sonde balloons and satellite-mounted instruments giving significantly greater global coverage than the GISTEMP near-surface record it seems reasonable to wonder whether most recorded warming is global or merely urban.

The global temperature report for July 2004 from the University of Alabama in Huntsville (USA) Earth System Science Center found that July 2004 was the planet’s coolest month in four and a half years and the coolest July in a dozen years.
 
The data show that the global temperature was 0.21C (about 0.38F) below the 20-year average for July.  This followed on from a June temperature about 0.02C below the average.  Only 3 months in the last 41 had been below this norm.
 
Dr. John Christy of UAH said, This was the coolest July since 1992, when global temperatures were cooled by dust thrown into the atmosphere by the Mount Pinatubo volcano.  A color map of temperature anomalies will be available at http://climate.uah.edu/ .
 
Regardless, new studies purportedly supporting alarmist, regional claims in the U.S. and Europe have been based on outputs from two models, including the Hadley Center Model, which reviewers admitted during the course of the National Assessment on Climate Change performed no better than a table of random numbers in predicting past climate.
 
One such study, Emissions pathways, climate change, and impacts on California, published in the August 24 issue of Proceedings of the National Academy of Sciences garnered considerable press coverage in California and the rest of the nation on August 17.  The Associated Press coverage was typical:
 
Global warming could cause dramatically hotter summers and a depleted snow pack in California, leading to a sharp increase in heat-related deaths and jeopardizing the water supply, according to a study released Monday.
 
Under the most optimistic computer model, periods of extreme heat would quadruple in Los Angeles by the end of the century, killing two to three times more people than in heat waves today; the Sierra Nevada snow pack would decline by 30% to 70%; and alpine forests would shrink 50% to 75%.
 
The most pessimistic model projects five to seven times as many heat-related deaths in Los Angeles, with six to eight times as many heat waves.  Snow pack and high altitude forests would shrink up to 90%.  The scientists’ temperature projections are higher than previous estimates, particularly in summer. Their predictions of an extreme decline in snow pack, alpine forests and the spread of desert areas all exceed earlier projections.
  
In addition to its “random numbers problem”, the model was run on the basis of data from the discredited SRES scenarios of the Intergovernmental Panel on Climate Change that predict countries such as Zimbabwe, Vanuatu and North Korea overtaking the USA in per capita income by 2100.  There was no discussion of the appropriateness or robustness of these data choices in the published paper.

It’s Hurricane Season and the claims are escalating faster than the storms. Already we’ve had Knight Ridder Newspapers claiming “The past nine years, from 1995 through 2003, mark the busiest, most intense nine-year storm period on record” based, apparently, on this Climate Prediction Center graphic. Of course, if you choose your “record” carefully enough you can prove anything – but it is not necessarily instructive to do so.

US_decadal.gif (6847 bytes) The National Hurricane Center provides a somewhat longer perspective online in U.S. Hurricane Strikes by Decade. From that table it is a simple matter to create a pretty graphic. The thumbnail at right links to a stacked column graph of U.S. landfalling hurricanes by Saffir-Simpson Category and decade. Counts are not labeled for category 5 storms, one occurring in each of the 30s, 60s and 90s (with Hurricane Andrew [1992] having been escalated from cat 4 to cat 5 on reanalysis in 2002).

UScount.gif (6531 bytes) On a raw count it appears the 1940s were pretty active while the 1970s were rather quiet. Naturally this doesn’t necessarily mean that the ferocity of these storms are comparable so we need to differentiate by some sort of composite rating.

Ferocity.gif (6654 bytes) A quick check on the numbers of more severe storms (cat 3-5), already tallied for us on the linked NHC page, indicates a similar high/low pattern but it would be nice to have something better than simple raw counts. How about using the storms’ Saffir-Simpson Category as a point system? This will give us an instant weighting system so we can accumulate both the number and severity of storms in any given decade – think of it as a sort of decadal hurricane ferocity score.

Activity.gif (6011 bytes) Perhaps an activity rating would be the more accurate description because the second half of the Twentieth Century does not appear as ferocious as the first.

Granted, this is a very rudimentary analysis but it does suffice to demonstrate the effect of record selection. It is entirely possible, in fact probable, that the Climate Prediction Center has only that length of record with sufficient data to perform their total energy analysis but it does not justify Knight Ridder‘s claim of “busiest, most intense” period on record.

According to The Most Intense Hurricanes in the United States 1900-2000 only 28 of the listed 65 events occurred since 1950. The Deadliest, Costliest, And Most Intense United States Hurricanes From 1900 To 2000 (And Other Frequently Requested Hurricane Facts) indicates that fully half the years when no hurricanes struck mainland U.S. are after 1950 (10 of 19). The most hurricanes to strike in one year were six in 1916 and 1985. There were five in 1933, and four in 1906, 1909, and 1964. Three hurricanes struck the U.S. in one year a total of sixteen times. Ten of these sixteen times occurred during the sixteen years from 1944 to 1959.

It would appear our rough and ready activity rating scheme of allocating and accruing points by Saffir-Simpson Category adequately reflects the relative severity of the hurricane cycle. While not fully represent the energy of the season a total event count obviously gives a rough approximation of the season severity.

The one thing we did not find is any suggestion of increasing hurricane season severity. The most active period within the Twentieth Century record is the 1930s-1960s with something of a lull subsequently. This is not supportive of the hypothesis that the globe is warming catastrophically or that there are more and more severe storms occurring.

The annual report of the U. S. Climate Change Science Program for fiscal years 2004-5, entitled, Our Changing Planet, was released on August 25.  It was immediately hailed as a turn-around in the Bush Administrations position by the media and environmental groups.

 The New York Times in a story by Andrew Revkin on August 25 set the tone, and an editorial the next day called the report a striking shift in the way the Bush administration has portrayed the science of climate change.  Other newspaper editorial columns and environmental groups jumped on this interpretation.



The striking shift is confined to several short passages in a 130-page document that are less qualified and more direct than in the FY 2003 edition.  The statements that attracted the most attention are the following:



Multiple ensemble simulations of the 20th century climate have been conducted using climate models that include new and improved estimates of natural and anthropogenic forcing.  The simulations show that observed globally averaged surface air temperatures can be replicated only when both anthropogenic forcings, e.g., greenhouse gases, as well as natural forcings such as solar variability and volcanic eruptions are included in the model.  These simulations improve on the robustness of earlier work (pages 46-7).


 Comparison of index trends in observations and model simulations shows that North American temperature changes from 1950 to 1999 were unlikely to be due only to natural climate variations.  Observed trends over this period are consistent with simulations that include anthropogenic forcing from increasing atmospheric greenhouse gases and sulfate aerosols.  However, most of the observed warming from 1900 to 1949 was likely due to natural climate variation (page 47).


 Administration officials disputed that the report represents a striking shift in their position.  In a Washington Post article on Aug. 27, White House Science Adviser John Marburger, one of the signatories to the report, was quoted as saying that the findings had no implications for policy.


 Further, a New York Times reporter covering the presidential campaign put the question directly to President Bush (Aug. 27): Asked why the administration had changed its position on what causes global warming, Mr. Bush replied, Ah, we did?  I don’t think so.



The report may be found on the web at http://www.usgcrp.gov/usgcrp/Library/ocp2004-5/default.htm.


 


Tim Lambert of the University of New South Wales, recently discovered (see his web log at http://cgi.cse.unsw.edu.au/~lambert/cgi-in/blog/ science/mckitrick6.html) an error in calculations in a scientific article by Ross McKitrick of the University of Guelph and Patrick Michaels of the University of Virginia.  The article (see the June 9 issue of the newsletter), titled A Test of Corrections for Extraneous Signals in Gridded Surface Temperature Data and which appeared in Climate Research, found that recent temperature records were strongly influenced by socio-economic factors.


 Prof. McKitrick has now corrected the calculations upon which this conclusion was based.   In a draft erratum, available at http://www.uoguelph.ca/~rmckitri/research/gdptemp.html, he writes, The principal effect of the correction is a reduced weight on the constant term and an increased weight on the COSABLAT variable itself. Indeed the correction improves the overall fit and removes the anomalously small cosine-latitude effect. The socioeconomic variables remain significant and the effects carry over from the station data to the gridded data as before.


 Because the main patterns of results persist across the revised tables, the original discussions as worded in our paper need only minor modification, and our overall conclusion, re-stated here, is unaffected:


 Overall, the results of this study support the hypothesis that published temperature data are contaminated with nonclimatic influences that add up to a net warming bias, and that efforts should be made to properly quantify these effects.


Princeton University scientist Robert Socolow recently co-authored a paper (Pacala, S., Socolow, R., Stabilization wedges: solving the climate problem for the next 50 years with current technologies, Science, 305, 968-972) which argues that existing technologies are sufficient to significantly reduce anthropogenic greenhouse gas emissions if widely adopted.  The paper significantly did not consider the costs of adopting these existing technologies.


 In a story based on the article, Dr. Socolow is reported as telling the Washington Post (Aug. 23), If governments fail to actthe concentration of carbon dioxide in the atmosphere will triple in 50 years.  Keeping it below a doubling is a heroic task, he said.


 The Greening Earth Society pointed out the hyperbole involved in this statement (World Climate Alert, Aug. 25):


 Before people began burning fossil fuels to release the energy that powers life as we know it, the atmospheric concentration of carbon dioxide was about 280 parts per million. Its now about 375 ppm an increase of about 34 percent. Twenty-five years ago the concentration was around 330 ppm, or 18 percent above background.  In other words, Socolow is telling Eilperin (apparently with a straight face) that the 16 percent rise in the last twenty-five years will morph into a 300 percent rise in the next fifty if governments fail to act.  This is nonsensical!  To triple the 280 ppm background by 2053, the atmospheres CO2 concentration must increase 1.65 percent per year.


 According to data compiled by the U. S. Energy Information Administration, the amount of carbon dioxide emissions per capita has been dropping worldwide since the 1980s and population (all those capita) isnt increasing at nearly the rate predicted twenty-five years ago.  In 1980, the United Nations predicted a global population of 15 billion by 2050. Their most recent estimate is nine billion. Theyve reduced their population prediction 40 percent.  As companies have competed to produce and deploy more efficient technologies (principally in developed countries), the rate of increase in the atmospheric CO2 concentration has remained much smaller than the required 1.65 percent per year.  In fact, it has changed very little. Over the period for which we have accurate records (1958 to present), the increase has fluctuated between 0.4% per year and 0.45%.