Friday, September 30, 2011

Saudis to Build Nuclear Plants at $7 Billion Each

"[T]he kingdom [of Saudi Arabia] will build 16 nuclear reactors by 2030 at a cost of around $7 billion each." - source.


In an ever-growing list of countries that either are building, or plan to build, nuclear-powered electric power plants, none are building at an affordable cost.  The USA, Finland, China, now Saudi Arabia all publish numbers that indicate a new, 1,000 MW reactor costs anywhere from $7 to $11 billion.  China is building a six-reactor plant for $66 billion, or $11 billion apiece.   The recently-cancelled South Texas Nuclear Project Expansion in the USA was to cost $17 billion, but that was just a dream; no shovel had been turned and no delays had yet started, with the inevitable increase in financing costs.  Fully costed, the STNP expansion would be at least $22 billion, more likely $25 billion.  


At these price levels, electricity must be sold for at least 35 cents per kWh, just to pay for the investment and provide a reasonable return.  


The Saudis indicated that their growing economy requires a 7 percent per year increase in electric power production.  They don't want to burn oil for making power, they would rather sell the oil.  Thus, the need for nuclear power plants.  The Saudis are smart, as I've written before, but they are mistaken on this one.  No economy grows, nor can it grow, at much above 3 percent per year for very long.  A temporary growth spurt might occur of 7 or 8 percent for a year or two, but this is not sustainable.  


Thus, there is no need for the nuclear power plants. The Saudis should, instead, do what the rest of the world does where economics are important: build combined-cycle gas turbine power plants (CCGT).  The Saudis have access to natural gas in the Middle East, and could easily purchase what they don't self-produce.  These CCGT power plants are much more efficient than conventional steam-based power plants, at 59 percent compared to approximately 35 percent.  They also do not use nearly as much water, which is a huge consideration for nuclear power plants.  Where, and how, will the Saudis obtain sufficient cooling water for 16 nuclear power plants?  Nuclear plants require at least twice as much water for cooling, compared to the CCGT plants.  Of course, the nuclear power plants could be built on the coast and use seawater.  This greatly increases the cost of the plant because seawater is more corrosive than fresh water. 


Perhaps the Saudis have another motive, from watching what the Iranians have done in the past several years with their nuclear "power" program.  Perhaps, just perhaps, the Saudis are in a race for parity and do not want the Iranians to have the upper hand, even in nuclear power plants. 


Roger E. Sowell, Esq. 
Marina del Rey, California

Wednesday, September 28, 2011

EPA CO2 Endangerment Finding Review by OIG


The US EPA’s Office of Inspector General (OIG) released today its report titled “Procedural Review of EPA’s Greenhouse Gases Endangerment FindingData Quality Processes.”

It is important to note that this was a Procedural review and not a Substantive review of the underlying science.   Procedural review merely means comparing the procedures that EPA used to what is required under the various laws and regulations.   Substantive review means evaluating the data and science that EPA relied on in forming their Endangerment Finding.   The Procedural rules that EPA must follow depend on whether the Technical Support Document (TSD) is a “highly influential scientific assessment” or not.  OIG considers the TSD to be a highly influential scientific assessment, but EPA did not.   There is a higher standard of care, or procedures that must be followed, for a highly influential scientific assessment.  It is these additional procedures that OIG found lacking in EPA’s work.

Background

For some background, and a description of a TSD:  As the primary scientific basis for EPA’s finding, the [EPA] relied upon assessments conducted by other organizations [the IPCC, National Research Council, and US Global Change Research Program].  EPA summarized the results of these and other scientific assessments in a technical support document (TSD).  There are specified criteria by which a document is to be judged to determine if it is a highly influential scientific assessment.  OIG presents these criteria in its report as:

“A highly influential scientific assessment is a scientific assessment that: 
·       
  • A) Could have a potential impact of more than $500 million in any
    year on either the public or private sector, or

    B) Is novel, controversial, or precedent setting, or has significant
    interagency interest.”


OIG stated the level of peer review for the highly influential scientific assessments, and goes on to say that:

“For highly influential scientific assessments, OMB guidance requires more attention to peer review consideration such as individual versus panel review, timing, scope of the review, selection of reviewers, disclosure and attribution, public participation, and disposition of reviewer comments. If the material to be disseminated falls within OMB’s definition of highly influential scientific assessment, OMB requires the agency to adhere to the peer review procedures identified in Section III of its bulletin.
OMG guidance also requires that agencies certify compliance with the requirements of the bulletin and information quality guidelines when using influential scientific information or highly influential scientific assessments to support a regulatory action. This certification and other relevant materials should be included in the administrative record for the action.”

Next, OIG discussed what the EPA did procedurally.  “EPA had the TSD reviewed by a panel of 12 federal climate change scientists. This review did not meet all [Office of Management and Budget] OMB requirements for peer review of a highly influential scientific assessment primarily because the review results and EPA’s response were not publicly reported, and because 1 of the 12 reviewers was an EPA employee.”

No public reporting of the 12 scientists’ review, no public reporting of EPA’s response to that review, and having an EPA staff member as one of the 12 scientists were cited as procedural errors. This is essentially, for the first two errors, a lack of transparency.  The public does not know what the reviewers found and reported, nor the EPA’s response, if any.  Were the findings unanimous?  Or, was there a split of opinion?  Did the EPA ignore the review panel’s findings?  At this point, we don’t know.   The obvious conflict of interest from the reviewer who is an EPA staff member should have made his or her opinion or vote irrelevant.   OMB requires an external peer review.

Reasons Given by EPA why TSD was not Considered a Highly Influential Scientific Assessment

“They [EPA} noted that the TSD consisted only of science that was previously peer reviewed and that these reviews were deemed adequate under the Agency’s policy. They also stated that, as described in the final Federal Register notice, the Administrator primarily relied upon assessments conducted by other organizations rather than the TSD, which summarizes the conclusions and findings of these other assessments.”

End Results

It appears that the OIG will allow the Endangerment Finding to stand, and is recommending only that EPA revise its procedures for future.   This could be a wrong interpretation, however nowhere in the OIG report is the EPA required to revise or re-issue the missing transparency documents, nor hold a second and independent review by qualified scientists. 

The fact that only procedures were evaluated means that the clearly false statements and conclusions of many of the peer-reviewed papers and documents were considered acceptable by EPA.  As reported earlier on SLB, the EPA accepted such wildly inaccurate statements as glaciers disappearing in the Himalayas.  Also, as the State of Texas wrote in their recent petition, regarding the Climategate emails,  

"[t]he emails do not reflect the work of objective
scientists dispassionately conducting their work and zealously pursuing the truth. Rather
they reveal a cadre of activist scientists colluding and scheming to advance what they
want the science to be—even where the empirical data suggest a different outcome." Also, "to the extent their [these scientists'] objectivity, impartiality, truthfulness, and scientific
integrity are compromised or in doubt, so too is the objectivity, impartiality, truthfulness,
and scientific integrity of the IPCC report, the CRU temperature data, the NOAA
temperature data, and other scientific research that is shown to have relied on their
compromised research."


Texas' petition also shows how the IPCC authors manipulated the climate temperature data, citing the by-now infamous email of using a "trick" to "hide the decline." Also, especially egregious data manipulation is discussed with Russian and New Zealand temperature data. Such manipulation showed undue warming. Also, the IPCC admitted they have lost critical climate data.



Then the real fun begins, with several major discredited claims, using non-peer-reviewed sources. These include Himalayan glaciers receding faster than anyone thought (the aren't). Also, Chinese temperature data was seriously flawed, and had no source documents. They made up the data. Next, the claim that 55 percent of the Netherlands is below sea level, and subject to inundation from sea level rise. This is erroneous, as only 26 percent is below sea level. The fourth and final example included in the Petition for Reconsideration is the wild claim that "up to 40 percent of the Amazonian rain forest could react drastically to even a slight reduction in precipitation." This was from the non-scientific, but wildly agenda-driven World Wildlife Federation, the WWF.

Apparently, these types of "peer-reviewed" scientific conclusions on the impact of man-made CO2 on the planet's climate are acceptable to the US EPA.  


Roger E. Sowell, Esq. 
Marina del Rey, California



Wednesday, September 21, 2011

US Long-Term Temperature Trend from NCDC

There is a problem in the NCDC data (National Climatic Data Center, US Department of Commerce, NOAA Satellite and Information Service) for the United States.    The problem is that the reported average temperature trend for the US does not agree with the mean, nor the area-weighted average, of the 48 contiguous states.   NCDC reports the temperature trend for the 48 contiguous states is 1.2 degrees F per century.  However, the mean of the individual states is 0.78 degrees F per century, and the area-weighted average for the 48 states is 0.74 degrees F per century.

This is a problem.  If the NCDC cannot get it right, how much of their data is wrong, and how many other statements issuing from there are also wrong?

Below (Figure 1) is a simple table, listing each of the contiguous 48 states in the US, alphabetically, with the temperature trend next to each state, in degrees F per century.

Figure 1
US 48 Contiguous States and Long-term Temperature Trend, Deg F/Century
Data from US NCDC

The area-weighted average was computed by weighting each temperature trend by the relative geographical area of each state.  This does not change the average much, but gives a better number because small states (Rhode Island, Connecticut, Delaware, etc.) do not have an undue influence over  large states (Texas, Montana, California, etc.)  

Other things pop out upon closer inspection of this table.  

There is a problem of uneven heating in adjacent states.  As an example, Texas shows a trend of zero degrees F per century, yet its neighboring state to the north, Oklahoma, is warming at 0.7 degrees F per century.  This is not likely, nor is it due to CO2 or any other so-called "greenhouse" gases in the atmosphere.  As I have stated before, how does CO2 know to ignore the entire state of Texas, yet concentrate its radiant heat beams on Oklahoma?   Note that, in earlier posts on SLB, I pointed out that adjacent cities have grossly different warming trends, again showing that CO2 cannot do what climate scientists claim it does.  

This gets even worse when one examines Texas' westerly neighbor, New Mexico.  New Mexico is warming at the rate of 0.9 degrees F per century.  How does CO2 know to focus its beams on New Mexico, yet ignore Texas?

Another example is the pair of states, Oklahoma and Arkansas.  Oklahoma, as stated just above, is warming at just under the national average at 0.7 degrees F per century.  Meanwhile, its neighbor to the east, Arkansas, is cooling at minus 0.3 degrees F per century.   Again, one must question how is this possible, if CO2 is responsible for the warming?  How can Arkansas be cooling?   I've been to Arkansas and can attest to the great lush green growth in that state, as CO2 abounds.  

Yet another example is the adjacent states of North Dakota and South Dakota.  North Dakota is warming at the alarming rate of 2.5 degrees F per century.  Its neighbor to the south, South Dakota, however is warming at half that rate, 1.2 degrees F per century.    How does CO2 know to focus so much energy from its heat rays on North Dakota? 

Yet another example is the adjacent states of Pennsylvania, and New York.  Both are of comparable size and located in the Northeast.  Pennsylvania is warming very slightly at 0.1 degrees F per century.  However, New York to its immediate north is warming at a much higher rate of 1.2 degrees F per century.  Again, how does CO2 know to ignore Pennsylvania and concentrate its heat rays on New York?

Then, there is the entire band of states along the edge of the Gulf of Mexico: Louisiana, Mississippi, Alabama, and Georgia.  Their reported temperature trends are zero for Louisiana, but negative for the others: Mississippi (negative 0.7 degrees F per century), Alabama (negative 0.8 degrees F per century), and Georgia (negative 0.6 degrees F per century).   Contrast those to Florida, immediately south of Georgia, which has a warming of 0.3 degrees F per century.  How could CO2 ignore the southern states but heat up other states? 

There are other curious adjacent states with wide disparities:

California: 0.7 degrees F per century, and Nevada to the east at 2.3 degrees F per century.  

Michigan: 0.1 degrees F per century, and Ohio to the south at 0.7 degrees F per century.  

Finally, the overall trend of 1.1 degrees F per century for the US contiguous 48 states is repeated on the US EPA's website, with the following text:   (note that the EPA website uses 1.1 degrees, while NCDC reports the trend is 1.2 degrees.  Perhaps that is acceptable for government work, and is lost in the rounding error.)  


United States Surface Temperature Trends
Observations compiled by NOAA’s National Climatic Data Center indicate that over the past century, temperatures rose across the contiguous United States at an average rate of 0.11°F per decade (1.1°F per century). Average temperatures rose at an increased rate of 0.56°F per decade from 1979 to 2005. The most recent eight-, nine-, and ten-year periods were the warmest on record.

Warming occurred throughout most of the U.S., with all but three of the eleven climate regions showing an increase of more than 1°F since 1901. The greatest temperature increase occurred in Alaska (3.3°F per century). The Southeast experienced a very slight cooling trend over the entire period (-0.04°F per century), but shows warming since 1979."   (bold emphasis added)

Meanwhile, the coastal regions of the west coast (Washington, Oregon, and California) all show a very sudden and steep temperature decline since 2002.    The average for the coastal areas is negative 21 degrees F per century.    One can only wonder why CO2 has abandoned the warming task set for it by climate scientists.   Perhaps the coastal cooling has more to do with the rapidly cooling Pacific Ocean along the west coast of the US.  

In summary, one can only wonder at what other examples of gross exaggeration are to be found upon close inspection of the data, and the conclusions arrived therefrom by the alarmist climate science community.  Also, the individual states show gross disparities in warming rates, from a high of 2.5 for North Dakota to a low of negative 0.8 for Alabama.  Adjacent states show gross disparities that indicate that CO2 cannot be causing any warming at all.  CO2 cannot act capriciously, but must act uniformly if it is indeed a physical phenomenon and not a figment of imagination. 

Roger E. Sowell, Esq. 
Marina del Rey, California.  Where it is indeed growing colder year by year. 


Sunday, September 11, 2011

From Man-made Global Warmist to Skeptic, My Journey



By Roger E. Sowell, Esq. 
Marina del Rey, California September 11, 2011


Several of my friends have asked me lately how I can be so positive that CO2 is not the evil, planet-killing pollutant that the science community insists that it is.  This missive is a partial response to those friends.  I will have more to add, likely some figures, charts, graphs, links to other sites and such.  But, here is the first effort.  Fair warning:  this is a long, long piece.  It covers a lot of ground.  It is as accurate as I can make it.  I haven't delved into the "why", but concentrated on the "what" and some of the "how."  There are a few "whos" in here, also.   Most of this has been covered by me in one or several earlier posts on SLB.  

Scientists said CO2 emissions are heating the Earth, with dire, even catastrophic consequences about to happen: ice caps melt, sea level rise, shores inundated, massive very strong hurricanes and more of them, heat waves with deadly tropical diseases moving into temperate zones, crop failures due to heat and desertification, human health pandemics from heat-aggravated issues, snow disappearing from the California Sierra Nevada range, fresh water shortages, the list went on and on. 

The solution, the scientists said, was to stop using fossil fuels, e.g. natural gas, coal, and oil.  Instead, we were to conserve and learn to use less electric power, drive electric cars, make gasoline from corn-based ethanol, make diesel from recycled animal fats and seed oils, recycle all our garbage into trash-burning power plants, build wind-turbines and solar panel-farms to generate electricity, collect methane from dairy farms’ manure pits and landfills, then burn the methane for fuel in power plants.  But, until those technologies could carry the load, we had to capture CO2 from power plants and big furnaces, and hide it away forever.  This hiding was named “sequestration.” 

My interest was piqued, to say the least.  The Earth is becoming un-inhabitable?  Millions of climate refugees would be on the move, seeking places to live?  Wars would be fought over food, and fresh water?  Coastlines would flood and be gone forever?  And this is all due to our fossil fuel use?  My industry?  The oil and gas industry?  

I knew that chemical engineers would be involved, and in the thick of it, too.  Chemical engineers are the ones that know how to provide substitutes for oil, for diesel, make ethanol from corn or cellulose.  Chemical engineers also are the ones that know how to design, build, and operate a CO2-capture plant, and find ways to either chemically bond the dangerous CO2 or find ways to permanently store it underground as part of that sequestration. 

So, I began to look into what chemical engineers could do to solve the problem, and seek ways to benefit from my chemical engineering background, and legal expertise as an attorney.  Surely, there would be some opportunities in all this for a guy with my skills.  I had to do my due diligence, and verify the scientific claims.   First, just how does CO2 cause all this warming?   I had worked with CO2 for decades, in many forms and many places.  CO2 is a combustion product (along with water vapor) from burning natural gas, other light hydrocarbons, gasoline, diesel, fuel oil, petroleum coke, wood, coal, even peat and dried animal dung.  I knew more than most, I supposed.  I had designed and installed and operated liquid CO2 storage tanks, equipment to gasify liquid CO2, to compress it, to re-liquefy the compressed gas.  I had also designed and operated process equipment to scrub CO2 out of a furnace’s flue gases, and designed and operated other process equipment that made solid particles out of the CO2.  I figured there would be lots of opportunities for me.

Finding out how much CO2 needed to be removed seemed like a good place to start.  I began by reading blog posts on a website called RealClimate.org, where they claimed “real climate science written by real climate scientists.”   That seemed like a good thing, to get the information right from the experts.   I saw there some charts and graphs,  and I understand charts and graphs.  Chemical engineers know all about such things.   One of those graphs showed the earth’s global average temperature since about 1880 up until 2005.   There was a dramatic and noticeable upward trend from around 1975 until the present.  That trend, if it continued, would certainly appear to make the world hotter, and indeed, perhaps the ice caps would all melt. 



So, being a good engineer, and a lawyer trained to look at all sides of the issue, I looked at the rest of the chart.  It looked a bit odd, to me.  You see, there was a rather flat area from around 1940 to 1975, or perhaps even a slight downward trend in those 35 years.  Hmmm…I wonder what caused that?  Perhaps CO2 was going down in that period?  Made a note to check that out.

Then, the period before 1940 really caught my eye.  From about 1900 to 1940, the graph showed a remarkably similar upward trend, just like the one from 1975 to 2005.  Hmmmm, again…how did that warming trend happen?  Was CO2 rising in those days?  And if it was, why did it stop around 1940?  The world was in a global war in the late 30s and first half of the 40s….did we not use any coal, or oil, or natural gas in those days?   Something seemed not quite right about that, as I distinctly remember from my reading about the oil industry that oil production rose dramatically during World War II, due to all the military machines that needed gasoline and diesel fuel, and the ships that needed fuel oil to run.  I knew that atomic power was not around until well after the war, so all those ships were running on heavy fuel oil, what we referred to as Bunker C oil.   Also, factories across the world were humming at full capacity before and during World War II, turning out munitions, steel, aluminum, war machines, tanks, jeeps, ships, and all the other trappings of war.  Not much conservation going on there, I thought.  Lots of CO2 being emitted, too.  Nobody cared about efficiencies or conservation, or even pollution, the only thought was for more production as fast as possible.  There was a war on, after all. 

I then looked into the amount of CO2 in the atmosphere, thinking it was some great percent, probably 3 or 4 or maybe 5 percent.  The graphs I found looked wrong, at first.  The CO2 was nowhere near 5 percent.  Not even 1 percent.  It was so low it was measured in parts per million!  About 365 ppm (parts per million), I found, and increasing by about 2 ppm per year.  The measurements went back to 1959, and even more amazing to me, this 365 ppm was on a bone-dry basis.  That means, the air sample was desiccated, or dried very thoroughly to remove any water vapor before quantifying the CO2 amount.  That is actually a good practice, because it eliminates any variations due to changes in air humidity.  But on a practical basis, if the atmosphere contains very much water vapor, then the actual CO2 concentration will be somewhat less.  I worked it out, and for air in the tropics at 80 degrees F and 90 percent humidity, air contains about 0.2 pounds of water per pound of dry air.  Roughly, 17 percent of the air we breathe in is water vapor.  That, then, would reduce the CO2 concentration also by 17 percent, so that 365 ppm was actually about 310 ppm.  Wow.  Could CO2 at 310 ppm be causing all that trouble?  I had to see how this worked. 



It didn’t take me long, looking around on the internet to find that scientists were claiming that CO2 absorbs heat from the Earth’s surface, and re-radiates about half of that heat back down to Earth.   The effect was termed the “greenhouse” effect, and CO2 was labeled a “greenhouse” gas.  Hmmm…that’s not how greenhouses work, I knew, because we had designed and built greenhouses, too.  Greenhouses stay warm just like a car stays warm when the windows are rolled up.  Heat from the sun passes through the glass, is absorbed by the dark green plants, and heats up the air inside the greenhouse.  Winds cannot blow the warm air away because the glass walls and roof are present.  In engineering terms, there is no convection heat transfer.   Well, this “greenhouse” gas may be a term of art, and I have certainly encountered many such terms of art in engineering, and in the law.  More investigation was clearly needed.

So, I found the Kyoto Protocol, which had a lot to say about greenhouse gases.  Turns out, there are six of them, not just CO2.  The Kyoto Six included CO2 of course, plus methane, nitrous oxide, sulfur hexafluoride, hydrofluorocarbons, and perfluorocarbons.  

Then, how did the CO2 and others absorb heat and re-radiate it back to Earth?  And, how did that create a warming?  Some things already did not add up, such as increasing CO2 since 1959 but the temperature cooling from 1940 to 1975.  More on that, later.

More investigation led me to Anthony Watts’ blog, Watts Up With That.  It appeared to be a place where articles were posted that questioned the orthodoxy of man-made, CO2-caused global warming.  So, I read and read about Al Gore and his movie An Inconvenient Truth, Dr. James Hansen of NASA who creates the world’s temperature chart, and several other figures.  There was something called “The Team” and I did not know who was on the team, and what game they played.  From the context, though, being a member of The Team seemed like not a good thing, as the term was used somewhat disparagingly. 

But, back to CO2 and heating the Earth.  I read about CO2 absorbing heat.  Some small bells went off way deep in my memory.  I had heard about this somewhere, a long time ago.  I pulled out my ancient handbook of Chemical Engineering, known as Perry’s Handbook.  That book is full of rock-solid, never-wrong science and engineering facts.  If it was in Perry’s, it was a fact.  It could be relied upon.  Chemical engineers use the principles and knowledge found in Perry’s every day, around the world.  So, I looked in Perry’s.  And I found it.  Under Heat Transfer, sub-heading Radiative Heat Transfer, furnace design.  Furnaces usually burn some form of fossil fuel, perhaps coal, or oil, or natural gas, or a mixture of light hydrocarbons if the furnace is in an oil refinery.  Home heating furnaces burn a medium oil similar to diesel fuel, and there are millions of these around the world.  There are a similar number of large, industrial furnaces and boilers in power plants, factories, steel mills, refineries, and chemical plants.  Furnace design is a very mature art, having been practiced and perfected over not just decades, but centuries.  Even railroad locomotives burned wood or coal, and have done so for centuries.  And, sure enough, one of the correction factors that must be included in the furnace design is the effect of CO2 in the combustion gases.  Water vapor also must be accounted for, and its effect is even greater than that of CO2. 

A bit of an aside is in order, here.  When a fossil fuel is burned, a chemical reaction occurs that gives off a great quantity of heat.  A fossil fuel is a hydrocarbon, meaning most of the molecules are either hydrogen, or carbon.  Hence, Hydro-Carbon.  Chemists are not very inventive when naming things, sometimes.  Air is added, and heat, and the oxygen in the air reacts chemically with the carbon, and with the hydrogen.  One carbon atom combines with two oxygen atoms in the air, to form CO2, Carbon Dioxide – meaning one Carbon, and two Oxygens.  Again, not very creative naming.   Similarly, two Hydrogen atoms react with one oxygen atom to form H2O, Di-hydrogen Monoxide, more commonly known as Water.   The water is in the gaseous state, so it is water vapor.   This is important in furnace design, because what flows out of a furnace’s exhaust stack is mostly nitrogen from the air fed into the furnace, very little oxygen because most of it is reacted, and the rest is water vapor and CO2.    How much CO2, and how much water vapor?  Is the CO2 a very low concentration, like that in the atmosphere?  Turns out the answer is no, CO2 is on the order of 19 percent, and water vapor is also about 29 percent.    This is very, very different from the concentration in the atmosphere.   Stated another way, 19 percent is the same as 190,000 ppm.  So, the concentration of CO2 is much, much higher in a furnace.  What else is different?  For one thing, the temperatures are much different.  CO2 in a furnace is glowing, white-hot.  Its temperature is on the order of 1800 degrees F.  Yet, in the atmosphere, CO2 is on the order of 90 degrees F down to - 40 degrees F.  I wondered if that made a difference, and if it did, how much?    The basic answer was that yes, CO2 and water vapor each absorb radiant heat and re-emit that radiant heat.  So, there appeared to be some valid basis for the scientists’ claims that CO2 absorbs heat.  But still, I wondered. 

I then read some more in the scientific literature and found that CO2 only absorbs a small fraction of the radiant heat leaving the Earth’s surface.  A very, very small portion.  Not only a very small portion, but the effect of adding more CO2 to the air has a diminishing effect on how much radiant energy is absorbed.    In effect, the atmosphere’s radiant absorption is the same whether CO2 is present, or not.  The effect is further diluted because water vapor also absorbs radiant energy at the same wavelengths as does CO2.   Water vapor also absorbs radiant energy at other wavelengths, but there is an overlap with CO2.

So, I sat back and pondered over all this, gave it a good thinking through.  The Earth, as I knew, cooled off rapidly and substantially in the clear desert nights, even in the heat of summer.  The cold desert nights are attributed to the very dry air, that is, almost zero water vapor.  This effect is pretty amazing, and one can actually see ice form in a shallow pan at night in the desert when the surrounding air is above freezing.   This is a favorite event for Scouts who go desert camping.  One takes a shallow pan, such as a brownie pan, puts about a quarter inch of fresh water in the pan, and sets the pan down on some insulating material such as Styrofoam or cardboard.  We wrapped a dry towel around the sides, too, to keep the air from warming the pan.  Sure enough, just before sunrise, we checked and there was a layer of ice on the water surface.  Enough heat had escaped from the water via radiation into the black sky above, unimpeded by CO2 or water vapor, to allow the water to chill and create ice.  The ice water was great for filling canteens. 

Yet, one cannot do this at night in a humid climate, such as Houston, Texas where I grew up and also did some camping.  The water vapor in the air, even on a clear night, prevents this. 

So, I wondered and pondered the entire question of CO2 absorbing heat in the atmosphere.  First, the Perry’s handbook made mention of a most important parameter, the “mean beam length,” or MBL.   This refers to the distance from the hot CO2 and water vapor gases to the furnace tubes where the liquid is contained that must be heated.  The greater the distance, the less impact the radiant energy has.  This is rather obvious from everyday experience, also, if anyone has ever built a campfire or lit a candle.  Closer to the flame is much hotter, and far away from the flame is much cooler.  This is common knowledge, except among very young children.  This is also well-known from the planets in the solar system, with planets closest to the sun being very hot such as Mercury, and those farther away growing colder and colder.  Yet, the Sun’s surface is the same temperature.  Clearly, distance has something to do with the amount of radiant energy absorbed.   I wondered just how much energy CO2 could absorb at altitudes of 10,000 feet, 20,000 feet, 30,000 feet, and higher.  Also, as the atmosphere grows thinner and thinner with altitude, I wondered how many CO2 molecules are present at each altitude to absorb whatever heat energy happens to be passing through. 

It was rather obvious that even ancient man knew some of these basic facts, as references to “the cold stars” are common in literature.  Yet, we now know that stars are in fact suns, and some of them are much bigger and far hotter than our sun.  We cannot feel the heat from them, due to the very great distances measured in billions of miles, if not trillions.  Far away means very cold.  Up close means very hot.    

I had not yet formed a conclusion, a firm opinion, on all the scientific claims of CO2 causing the earth to warm, but it was looking pretty shaky to me.  Then I considered my engineering background in process control, and kept thinking about campfires.   The Law of the Campfire is simple, and was stated briefly above:  if you are too hot, move back.  If you are too cold, move closer.   Closer is hotter, every time, for a campfire that has constant heat output. 

And yet, I had seen the chart that showed CO2 was slowly rising, a nice smooth curve.  At the same time, the average temperature for the entire earth had peaked in about 1940 and decreased for 35 years.  Then, the trend reversed, and the earth started warming again.  That is impossible, if CO2 is what is causing the warming.  For CO2 to cause the earth to cool for 35 years, then warm again for the next 35 violates the fundamentals of process control.  A noted PhD chemical engineer, Dr. Pierre Latour, wrote on this same subject in a familiar magazine, Hydrocarbon Processing.  I had my own blog by then, and wrote an article discussing Dr. Latour’s writings.  For CO2 to allow cooling then warming, would be like moving your chair closer to the campfire to cool down on some occasions, but moving away to cool down on other occasions.  I knew right then that CO2 could not do what the scientists say it does.  Not at those low concentrations in the atmosphere, and not at those low temperatures.  But, I wanted to look further, so I kept reading and questioning.

About that time, November of 2009, the Climategate scandal broke when thousands of emails and computer files were released into the internet.  The files were incredibly damning, and damaging to the climate warmists’ cause because they revealed improper actions by some of the scientists at the heart of the climate debate.   In damage-control mode, the scientists at the Hadley Center’s Climate Research Unit of the University of East Anglia, UK, chose to release some of their files on temperature records for about 1,000 locations around the world.  The intent was to show that there was nothing to hide, and in a good faith effort, here was the raw data for all the world to see.   I copied the files onto my computer and had a look. 

First, the so-called raw data was anything but raw.  I know raw data, having acquired reams and reams of raw data as a practicing chemical engineer, in refineries and chemical plants all  over the world over more than 20 years.  What HadCRU (Hadley Climate Research Unit) had released was processed data.  Their release showed the average monthly temperature over a period of several decades for the chosen cities.  The monthly average was created from the daily average temperature.  The daily average was created by averaging the high and low temperature for the day.  The high and low temperature were each sometimes adjusted, or fudged, by accounting for the time of day for that temperature reading.   Also, there was no indication of how missing data was replaced or created.  Instruments are not 100 percent reliable, and sometimes require attention.  They may require cleaning, calibration, parts replacement, or other servicing.  They may be out of service for a period while someone notices the data is missing and fixes the instrument. 

Yet, here was a data set of monthly averages for about a thousand cities.  I decided to look at what was there, for the USA.  There were 87 records, all in the lower 48 states.  The data were for cities all across the USA, not in every state, but in most states, and were fairly evenly distributed.  Some were in great cities like New York, Los Angeles, San Francisco, Miami, some in mid-sized cities like St. Louis, Spokane, and Fresno.  Others were in small cities or large towns, like Abilene, Texas, and Meridian, Mississippi.    I loaded the data for each city into a popular spreadsheet and made graphs of the monthly temperature versus time.  I included a moving average to see what trends were apparent, if any, then added a linear best-fit trend line.  The results were so fascinating that I uploaded all the graphs onto my blog, with some commentary.  What I found confirmed what I had suspected all along. CO2 cannot do what the scientists claim it does.

What the graphs showed was a terrible inconsistency in the warming of cities in the USA.  Some cities did, indeed, show a pronounced warming trend over a roughly 100 year period from 1900 to 2009.   Not all cities had data that extended that far back, though, as only 62 had long-term records.   Yet, other cities among those 62 had cooling trends, or neutral trends.  That seemed odd, because if CO2 was truly warming the earth, then it must be warming every part and not being selective about what to warm and what to ignore.  This is especially true for adjacent cities, or those not separated by great distances north to south.  There would be some difference, supposedly, for cities in the far north compared to those near the equator.  But, the USA is only about 1500 miles from north to south in the lower 48 states.     For the earth’s total distance from pole to pole of approximately 12,000 miles, that is barely more than 10 percent.   For cities that are only one or two hundred miles apart, it seemed very odd to me that CO2 would ignore one and focus its heating rays on the other.  Physics does not work that way.  If a phenomenon is truly a physical effect, it works consistently and equally at all times and places.  Gravity, for instance, has the same downward force in Houston, Texas as it does in Mumbai, India, or Bora Bora.     One can imagine the confusion if travelers had to adjust to different gravity effects depending on what city their plane had landed in.   No, physics does not play whimsical games like that.

Or, one could imagine how chaos would reign if the properties of steel were capricious, like CO2.  Engineers might be building a bridge in Cairo, Egypt, and require hefty steel beams 12 inches wide and 24 inches deep.  But, an identical bridge across the Mediterranean in Rome would require lighter beams of only 4 inches width and 12 inches deep.  Engineers will laugh at this, because that simply does not happen.  A given grade and quality of steel will hold up the identical weight, no matter where in the world it is used.  Purists will note that this is not strictly true, as steel is somewhat affected by temperature.  But, for most purposes, bridges do not get hot enough to weaken the steel noticeably. 

At this point, I looked at adjacent cities and noted that some cities, as I wrote above, showed a cooling or neutral trend.  Abilene, Texas, and Shreveport, Louisiana are two of those.   Abilene shows a slight cooling of 0.19 degrees C per century, while Shreveport shows a very slight cooling of 0.01degrees C per century; essentially no change at all. These cities are only approximately 250 miles apart, east to west.  They are at essentially the same latitude.   At the same time, St. Louis, Missouri, shows a warming of 1 degree C per century.  St. Louis is only approximately 300 miles north and a bit east of Shreveport, and approximately 400 miles from Abilene.    Clearly, something is amiss in the CO2-causes-global-warming science.  How could CO2 know to ignore Shreveport, but focus its beams on St. Louis? 





Another example came to my attention: San Francisco, California, and its neighboring city, Sacramento.  These cities are separated by only about 50 miles, and nearly at the same latitude.  Yet, San Francisco had a warming of 1.5 degrees C per century, while Sacramento cooled by 0.3 degrees C per century. 

One possibility that explains the heating versus cooling or no trend is what I learned was called the Urban Heat Island effect, or UHI.   At first I thought this referred to the University of Hawaii until I finally found what the acronym spelled out.   UHI is a phenomenon that causes cities, or large urban areas, to be hotter during the day, and warmer during the night, compared to more rural areas nearby.  The UHI effect is small for small cities, but grows larger for large cities.  The UHI is due to several factors, including expanses of asphalt and concrete paving, stone or brick or glass-and-steel buildings, great consumption of electricity to heat or cool the buildings, industrial heat from factories and other heavy industries, and large numbers of cars, trucks, buses, and airplanes that consume great quantities of fossil fuel. 
But, even UHI has problems.  For example, Meridian, Mississippi is a small town and it is warming at the identical rate as the large city in Texas, San Antonio.  Both show a modest warming of 0.26 degrees C per century.  How can that be, if UHI is important?

Other small cities show substantial warming, such as Helena, Montana and Duluth, Minnesota, at 2 degrees C per century.    Duluth’s population hasn’t changed much from around 80,000 people since 1930.  It reached 107,000 in 1960 but has been decreasing since then.  Helena has grown from about 12,000 in 1910 to 28,000 in 2010.  

I want to turn briefly to the amazing small town of Eureka, California.  I have never been there, but it is on the coast in northern California between San Francisco and the Oregon border.   When the winter Olympics are held near 2075, Eureka should put in a bid as the host city.  It will soon be covered in snow year-round, and may have a localized ice age if the present cooling trend does not reverse.   Eureka has, starting in about 1990, had a cooling trend of approximately 15 degrees C per century.   Its average temperature currently is about 10 degrees C, so in 65 years the average temperature will be zero C.   One can only wonder why CO2 has ignored the small town of Eureka.  If any town needed some global warming, it would be Eureka. 



Finally, my attention was turned to a published study by Dr. James Goodridge, the former state climatologist for California, now retired.  His work showed that California’s counties could be grouped in three groups according to population, and the average temperature trend for each group computed.  He found that counties with large populations showed a distinct increasing temperature over 80 years, while those with small populations showed essentially no warming at all.  The mid-sized counties showed an intermediate amount of warming.   One must seriously question how CO2 did that, in a state as geographically large and diverse as California.  The large population counties are typically on the coast, with the cities of San Diego, Los Angeles, and San Francisco.  Small population counties are all across the state, including on the remaining coastal areas.    It is highly unlikely that CO2 is smart enough to pick and choose which counties in California will receive its warming beams, and which counties will be ignored by CO2.

  

To summarize the journey to this point, then, I found that scientists claim the earth is warming at an alarming rate, but there was a previous warming of equal magnitude and duration (1910-1940) during a period when atmospheric CO2 was at low concentration.  Also, the earth stopped warming and cooled a bit from 1940 to 1975, then started the warming again.  For CO2 to cause a warming, then a cooling, then a warming again is impossible and violates the fundamentals not only of physics but of process control.  Finally, CO2 ignores completely some cities in the USA, indeed entire counties in California, while warming adjacent cities and counties with large populations at an alarming rate.   CO2 is a simple molecule with one carbon atom and two oxygen atoms, and cannot possibly be that smart.

All the above was more than enough to convince me that the threat man-made global warming is false, it is a hollow threat, and has zero substance.   Yet, if one reads the policy summaries and scientific studies, the premise is that CO2 causes warming and more CO2 causes more and faster warming.  All else follows from that failed premise.

Still, there is more to the story.  I want to describe what I found when I looked at the temperature record itself, the one that shows a warming from 1910 to 1940, a cooling from then until 1975, then a warming again until about 2000.   To preface this, it is important to know that in engineering data collection and analysis, indeed in any scientific data collection and analysis, it is only rarely appropriate to go back and change one’s data.  It is extremely inappropriate to change one’s data over and over again.   That requires some explanation.

The question is, how accurate is any data?   The data here, for climate purposes, and whether the Earth is warming or cooling, is temperature data.  Temperature data can be obtained very accurately and very precisely with modern technology.   Not to get too technical with this, but accuracy and precision are not the same thing.  Accuracy can be considered as how close to the truth is the measurement, while precision is the measure of how many decimal places are believable in the measurement.  In the early days of thermometers, it was difficult to calibrate them and also difficult to obtain a reading within half a degree.  Thus, the thermometer may have read 80 degrees on a fine sunny day but due to mis-calibration, the actual temperature was only 78 degrees.   What is not known in many cases is when a thermometer broke and was replaced.   It is also not known if the replacement thermometer was calibrated to read the same as the earlier one.    Finally, if different observers read the thermometer, one may have judged the reading to be 70.5 degrees, while another would read it to be 70 degrees.   These seem like small differences, and they are.  However, the entire warming over the past century was said to be only 0.7 degrees C, or roughly 1.2 degrees F.

My attention was called to some amazing work by E. M. (Michael) Smith, who runs a blog titled chiefio.wordpress.com.   The pertinent portions of his blog entries are known as The March of the Thermometers.  Michael is rather a whizz at computer programming and data analysis.  He accessed the publicly-available massive data and computer code used at NASA by Dr. James Hansen, known as the GISS code.  I believe GISS stands for Goddard Institute for Space Studies.    Michael unraveled the code, and wrote of his findings in several postings.  The key findings were that the code re-writes the past data each time it is run.  Also, the code makes questionable choices in how missing data is treated, and how discontinuous data is spliced together.   Another and, to me, most important finding was that Hansen deleted major portions of the temperature measurement stations in recent years.  That does not appear to be random but perhaps (likely?) was chosen in a way to show much more warming in recent years.   In effect, the temperature trend that results from NASA GISS is false.    I highly recommend that anyone who is curious about the temperature history of the last 120 years or so read what E.M. Smith wrote about it.   The past data is not only changed, it is changed frequently.   If missing data is discovered, the computer code simply reaches out to an adjacent station that can be 1200 kilometers distant (about 700 miles!!!) and uses that data. 

Then, the entire system spits out a global average temperature based on anomalies from thousands of measuring stations around the world.   Anomalies are another area for creating great mischief.   The problem lies in having some cities in cold locations, and others in warmer locations.   What climate scientists do is assign each month an average temperature, based on some pre-determined period of about 30 years.  Some use 50 years, though, for reasons not clear to me.  Further, the base period is not the same but is updated every ten years or so.  Again, moving targets with constantly-adjusted data.   It reminds me of the ancient three-shell game with the pea, where the mark tries to keep his eye on the shell that has the pea under it while the con-man shuffles them all around.  

The use of anomalies supposedly allows one to merge or blend temperature trends together without concern over where the average temperature was 5 degrees (Alaska) or 25 degrees (Bora Bora).   Still, it is quite disconcerting to see yet another opportunity for data manipulation.    A better method, in my view, is to obtain the data trend for each decade, for each station.  If the station showed a warming trend of 0.1 degree per decade, then that 0.1 goes into the averaging pot.  There would be no anomalies, no base periods, no changing base periods every 10 years, just a simple, one-time calculation of the decadal trend.   That decadal trend would then be golden and not subject to change.

Another very disconcerting revelation was the excellent work by Anthony Watts, who was mentioned earlier in connection with his blog, Watts Up With That.   Anthony also performed a heroic task in assessing the vast majority of the USA’s climate measuring stations.  His assessment focused on how well or how poorly each station was situated, or sited, according to the existing guidelines.  For example, a well-sited station must be a certain distance from trees, from buildings or other structures, must be placed at the correct height over a grass area, and not be subjected to artificial heating or cooling measures such as an air conditioning exhaust.   What Anthony found was appalling.  He wrote up his findings along with several co-authors and had a paper published in 2011 (summary here).  Many stations ranked as the poorest rating, and only a few had the best rating.  Some were indeed mounted next to brick walls, on asphalt parking lots, next to air conditioner condensers, next to barbecue pits, on asphalt rooftops, at airports where they are heated by massive runways and jet exhaust, and other unacceptable locations.  These are the sources of the temperature records for the USA, which is supposed to be the best and most accurate of any country in the world.   Where the siting conditions become important is how the temperatures are affected over time, over a period of years.   A rural setting would likely show lower readings in the early years, but warmer and warmer readings as buildings are built, roads are installed, parking lots are installed, and other such things.  Thus, part of what Anthony did was determine how much of any demonstrated warming trend was due to siting changes.

There are a few more points, and I will finish.

First, sea levels are not rising and oceans are not getting hotter.   This alone disproves the entire CO2-induced global warming nonsense.  By the warmists’ belief, the oceans must grow warmer, and the sea levels must increase as CO2 increases.  Neither is happening. In fact, the opposite is happening.  NASA and NOAA’s own data show this quite clearly.  The chart below is from U. Colorado, and is based on NASA's satellites that measure sea level.   Note on the chart the dramatic decrease in trend starting in about 2005, and the sudden decrease in sea level in early 2010.





First-part B, Sierra Nevada snowpack and snow-water-equivalent (SWE) have not changed significantly in almost 100 years.  Dr. John Christy of University of Alabama, Huntsville, published a paper on this in 2010.  His data ended in 2009.  Since then, there have been near-record snowfalls in the Sierras.  His key graph is shown below, normalized to show deviation from the average.  From his paper, HL refers to a key snow measuring station, Huntington Lake.   The paper is at this link.




Second, a recent peer-reviewed paper from CERN shows that clouds and sunspots are related, with the 20th century having great sunspot activity, but the little ice age had few to zero sunspots.  Recently, in the last 4 years or so, our sun has again gone suddenly very, very quiet and it is getting cooler.  The lack of sunspots came as a complete surprise to scientists all around the globe.  The relationship is that more sunspots equals hotter climate.  The mechanism is that the sun’s magnetic field is immense, and grows larger and more intense as sunspot activity increases.  The magnetic field shields the Earth from Galactic Cosmic Rays, GCRs.  However, when GCRs hit the atmosphere, they create cloud nucleation particles and more clouds form.  More clouds reflect more sunlight away from the Earth, and cooling occurs.   Once again, more evidence that the science is not settled.  Heck, they cannot even predict how many sunspots, nor when they will occur.   The CERN experiment and published paper was only a few weeks ago. 

Third, climate models cannot agree, and their projections do not match the satellite measurements.   This shows that the science is far from settled, and when a model fails to match the measured data, the models are wrong and must be scrapped or improved if possible.  A very recent paper by Spencer and Braswell has caused an uproar in the climate community because it shows very clearly that the climate models are far off in their predictions.  The satellite data doesn’t match the models.

Fourth, hurricanes are not growing more intense and in greater numbers.   In fact, hurricane energy is at a historic low for the entire period of satellites.  Meanwhile, CO2 continues to increase.  Another busted prediction, proving their ideas are totally nonsense.  The chart below, from Dr. Ryan Maue of Florida State University, shows the current status of the world's tropical cyclones measured as Accumulated Cyclone Energy from 1972 until today. The top line is the total for the world, the bottom line is for the Northern Hemisphere.   The total cyclone energy is back to what it was in the middle 1970s, and meanwhile CO2 continues to rise.



Fifth, and finally, the only prediction the warmists have correctly made is the continual reduction in the Arctic ice cap.  However, they have the cause and effect completely wrong, for the following reasons.  First, the warmists maintain that a shrinking ice cap is strong evidence that the Arctic area is warming, and that warming is due to the heat rays beamed down by CO2 in the atmosphere.   In reality, ice acts as an insulator and prevents heat from being released from the ocean into the night sky via radiation.  Ice acts in a similar way on lakes, it keeps the lake from freezing solid unless the lake is very shallow.  The growing and retreating Arctic ice acts as a negative feedback on the ocean’s heat content.  When the oceans are warm, the ice begins to melt at the edges.  There is thus more open water that loses heat due to radiation.  The ice extent is at a minimum usually around mid-September, which allows great amounts of heat loss in the long polar nights.  The oceans then cool, which eventually cools the air, and allows more ice to form in future years.  The system oscillates then between more ice and less ice, with the ocean temperature and heat content also oscillating but slightly out of phase. 

In conclusion, if anyone still believes that CO2 does what scientists claim it does, I suggest you think about that the next time you are at a campfire, or near a candle, or any other fairly stable heat source.  Move toward it, then move back.  Also, find a nice masonry wall that has ample sunshine on it.  Just after dusk, when the sun is no longer shining on the wall, place your hand on the wall and feel its warmth.  Then, move slowly away from the wall and see how long you can continue to feel the warmth.  Think about that little CO2 molecule, having to also feel that warmth, get all excited, absorb the heat, and then re-radiate the heat back out again.    Remember that scientists insist that the Earth warmed from 1910 to 1940 - yet CO2 was very, very low.  

Also, have a look at Anthony's blog and E.M. Smith's blog.   Think about this: if the science is settled and we must act now or lose the Earth's future to a hell of warming, rising oceans, monster hurricanes and all the rest, why did the CERN experiment show that clouds are far more important than CO2?  Why do the satellite temperature measurements show the models' predictions are all wrong?  Why has nothing ever panned out for the climate warmists?  The only thing they can point to is the declining ice in the Arctic, but as I discussed above, they have that completely wrong. 

Finally, have a look at the temperature graphs of the USA's cities on SLB.  See for yourself how many, many cities have zero warming or a slight cooling.  Then, ask yourself how can that be?  How did CO2 get so smart that it can selectively ignore some cities?

CO2 is innocent.  It always has been, and always will be. 

Roger E. Sowell, Esq.