Solid Waste & Recycling

Blog

Sallie Baliunas and the solar data


As a follow-up to my blog entry yesterday, which announced that we will provide occasional coverage of the global warming issue on this website and in Solid Waste & Recycling magazine, I thought I’d provide readers with an update on an interesting and often overlooked dimension, which is the correlation between solar energy output and earth temperatures.
When I first wrote about global warming for the Globe & Mail almost ten years ago, one of the most interesting scientists with whom I spoke was Sallie Baliunas, Ph.D., — Senior Scientist at George C. Marshall Institute. Her team has studied the matter of how solar cycles affect the Earth’s oceans and climate. She is particularly knowledgable about the interaction of the sun’s energy and the El Niño cycles. Her research has made her skeptical of man-made climate change, and in this speech from 2002 she explains why. She also voices her opposition to the Kyoto Accord.
I’ve cut and pasted the speech below.
Baliunas is correct that even the full implementation of Kyoto isn’t going to make much impact on the climate. But, personally, I’ve come to regard Kyoto as sort of a downpayment on an insurance policy, in case the IPCC predictions turn out to be correct. A lot of the energy efficiency things we’d have to implement are things we should be doing anyway, if only to reduce our reliance on oil that comes from despotic regimes in the Middle East that are hostile to the western democracies. The events of 9/11 gave us a non-environmental reason to reduce our oil addiction, to paraphrase George Bush (!).
I’ll give you a couple of quick examples of the sort of stupid things we do that we ought naught if we cared a whit about efficiency, much less the welfare of the planet.
1) In the hot weather, many retail stores turn up the air conditioning, then leave the front door open to create a cold blast of air on the pavement outside. This is supposed to entice customers inside to cool off, and to shop. It’s been shown that if this practice was ended (possibly by being banned), many jurisdictions (e.g., Ontario) could avoid the need to build new power plants. An incredible amount of energy is wasted this way.
2) What is this penis envy between cities to have the world’s tallest building? In today’s paper I noticed a story about plans in Tokyo and Jakarta to build something taller than Toronto’s CN Tower. The worst (and I mean the worst) is the Burj Dubal skyscraper in Dubai, which will soar 700 metres. The problem is not the height — rather it’s the insanity of building a glass and steel tower in the desert, a sealed structure that will have to burn incredible amounts of fossil fuel to air condition, even as it attracts the intense heat of the sun. Yes, they’re sitting atop a lot of oil over there, but from a CO2 and natural resource point of view, buildings like that are monsters, and don’t fit with the topography of that part of the world at all. There’s no cultural tradition or natural heritage at all in this construction, just boastful and wasteful misplaced competitiveness to “be the tallest.” In time these buildings will be abandoned, and will stand as monuments to man’s foolish pride.
3) Metering systems exist, and are being installed in homes in California, that allow people to monitor the precise energy consumption of each appliance in their home, and the display tells then the cost of the energy as it fluctuates throughout the day. These meters encourage people to use certain appliances during non-peak periods when prices are low. New power plants are not often needed for increases in “general” energy consumption — it’s the rising “peak” demand that’s the problem, and these inexpensive devices can smooth out the demand highs and lows to the extent that new plants aren’t required. Why aren’t they being installed across Canada right now?
4) As a more general point, and I wanted to state this yesterday, even if global warming never happens, I still think it’s outrageous that we’re going to burn our way through the world’s readily available fossil fuel inventories in less than two centuries (i.e., from about 1900 to 2100). Yes, I’m sure more oil and gas deposits will be found, but they will become more difficult and expensive to harvest. My point is that these materials took millions of years to form in geological formations, then the industrial revolution begins and — poof! — we use it all up in just a couple of generations. I believe that just as the oceans and the forests are “borrowed” from future generations, so is this precious energy heritage. Imagine in a hundred years time (maybe less) how angry people will be when they look at film and video images of people driving about — one person to a car — to and fro from the mall, eschewing public transit and gorging up the cheap petroleum.
5) There is uncertainty about how our actions may heat up or otherwise affect the planet. But I think it’s fair to say that we’re conducting a dangerous experiment taking several hundred million years worth of accumulated carbon and spewing it into that atmosphere in only a century or two, and just crossing our fingers that everything will be okay, and that somehow the atmosphere will “shed” the excess heat. I’m not in favor of radical surgery at this point, but I believe the patient (us) should at least be put on a diet. There are in fact a lot of easy things we can do to reduce energy waste right now, without having to live like our aboriginal forebearers (at least, not yet).


Warming Up to the Truth: The Real Story About Climate Change
by Sallie Baliunas, Ph.D.
Heritage Lecture #758
August 22, 2002
The Climate Action Report, a periodic report to the United Nations, was issued in early June. A media frenzy claimed that this report somehow contained revelatory new science that changed the debate on global warming.
The report has little new science. But since 1992, when America embarked on the Rio Treaty, a great deal of new science has come forward. The United States is a leader in studying the subject. The U.S. has invested some $45 billion in research funding on this question over the past 10 years.
I wanted to update you on the latest science since 1992 and assure you that what is in the Climate Action Report is really nothing revelatory.
The scientific facts on which everyone agrees are that, as a result of using coal, oil, and natural gas, the carbon dioxide content of the air is increasing. The air’s concentration of other human-produced greenhouse gases, like methane, has also increased. These greenhouse gases absorb infrared radiation from the sun, and they retain some of that energy close to earth.
All computer simulations of climate change say that, based on how we understand climate to work, the low layer of air for one to five miles up (the low troposphere), where the radiation is trapped, should warm. That low layer of air warming should, in turn, warm the surface.
Scientific facts gathered in the past 10 years do not support the notion of catastrophic human-made warming as a basis for drastic carbon dioxide emission cuts.
You probably know that the Kyoto agreement fails to stop the hypothesized human-made global warming. Kyoto would hurt America’s and the world’s workers and the struggling poor and the elderly, owing to the severe cuts in energy use that it entails.
MEASURING SURFACE TEMPERATURE
Now for the science. There are two important records that we’ll look at. I just told you how we think climate operates in the presence of increasing carbon dioxide and greenhouse gases in the air from human activities. The layer of air one to five miles up retains energy and that layer, in turn, heats the surface of the earth. The human-made greenhouse warming component must warm both layers of air, with computer simulations indicating the low troposphere would warm more quickly and to a greater amount than the surface.
Let’s start with the surface temperature records. They are made by thermometers, and go back to about the mid-19th century in locations scattered around the world. For some locations the records go back even further.
Two groups have analyzed these surface temperature records: the Climatic Research Unit in Great Britain, and the NASA Goddard Institute for Space Sciences. They broadly say the same thing: The 19th century was cooler than the 20th century. There may be some disagreement on the exact amount of the warming, but certainly the 20th century was warmer than the 19th.
To see if the 20th-century surface warming is from human activity or not, we begin looking in detail at the surface record. In the 20th century, three trends are easily identified. From 1900 to 1940, the surface warms strongly. From 1940 to about the late 1970s, a slight cooling trend is seen. Then from the late 1970s to the present, warming occurs. Briefly, the surface records show early 20th-century warming, mid-20th-century cooling, and late 20th-century warming.
Most of the increase in the air’s concentration of greenhouse gases from human activities–over 80 percent–occurred after the 1940s. That means that the strong early 20th century warming must be largely, if not entirely, natural.
The mid-20th-century cooling can’t be a warming response owing to the air’s added greenhouse gases. The only portion of this record that could be largely human-made is that of the past few decades. The slope of that trend calculated over the past few decades is about one-tenth of a degree Centigrade per decade.
Now, most all the computer models agree that the human-made warming would be almost linear in fashion. So over a century the extrapolated warming trend expected from continued use of fossil fuels would amount to about 1 degree Centigrade per century. That’s what the surface temperature says would be the upper limit.
But I gave you a scientific test to do early in my remarks. The question is, What happens in the low layer of air from one to five miles up that must warm in response to the increase in greenhouse gas concentrations? The surface warming can be concluded as owing to human-made greenhouse gas emissions only if the low troposphere warms, if the computer simulations are accurate.
One can have surface warming from a variety of reasons. So the key layer of air to look at is the one-to-five-mile up layer of air.
MEASURING AIR TEMPERATURE
NASA launched satellites starting in 1979 to measure this layer of air. The satellites look down and record these measurements daily. I’ve plotted the monthly averages. There are lots of jigs and jags in the data, and they are real.
The air temperature varies not only on a daily basis, on a monthly basis, but also from year to year. A very huge warming spike in 1997-1998 is a strong, natural phenomenon called El Niño, a warming of the Pacific that in turn warms the air. Because the Pacific is so pervasive in the global average, it raises the temperature. But it doesn’t last very long, and after the El Niño subsided, temperatures fell.
El Niños are natural and occur every several years. In 1982, an equally strong El Niño was developing in the Pacific. But then, a volcano erupted. Material lofted by strong volcanic eruptions can temporary cool temperatures. So those two events occurring at nearly the same time meant there was a net cooling just after 1982, instead of an unmasked strong El Niño-driven pulse of warmth.
El Niño is part of a system of ocean and air changes called the El Niño Southern Oscillation, in which the La Niña phase tends toward cooling. Detailed physical understanding of the El Niño Southern Oscillation is lacking.
Again, these phenomena are naturally occurring. They have existed for many millennia prior to human-added greenhouse gases in the air.
I asked the computer to naively draw a linear trend through the data recorded by satellites. This linear trend probably has a bias, an upward bias because of that strong 1997-1998 El Niño warm pulse. Nonetheless, the fitted trend is: positive four-hundredths of a degree Centigrade per decade.
Now, this is the layer of air sensitive to the human-made warming effect, and the layer that must warm at least as much as the surface according to the computer simulations. Yet, the projected warming from human activities can’t be found in the low troposphere in any great degree. The four-hundredths of a degree Centigrade might be entirely due to this El Niño bias. If the small warming trend in the low troposphere were assumed to be entirely human-caused, the trend is much smaller than forecast by any model. Extrapolated over a century, the observed trend indicates a human-made warming trend no greater than four-tenths of a degree Centigrade.
In contrast, the computer models say this very key layer of air must be warming from human activities. The predictions are that the air must be warming at a rate of approximately a quarter of a degree Centigrade per decade.
Comparing what the computer models say should be happening with the actual satellite observations shows a mismatch of around a factor of 6. That is, this layer of air just is not warming the way the computer simulations say it should. There should have been a half a degree Centigrade per decade warming in this layer of air over the period of satellite observations. The human-made warming trend isn’t there.
Now, an argument is often made that the measurements made by satellites looking down on this key layer of air are biased, or that the satellites have instrumental problems.
NASA researchers worked very hard to make these measurements the best possible, and to correct for any of the deficiencies seen in them. But it’s always useful to have an independent set of data, and we have that from NOAA (the National Oceanic and Atmospheric Administration) scientists and from other groups around the world.
Measurements are also made of this layer of air from weather balloons that carry thermometers. Balloons are launched worldwide every day to make the measurements. The balloon data go back to 1957, and importantly, they overlap with the satellite data which began in 1979 and have continued through the present. During the period of overlap, the correlation coefficient between the two data sets, the technical term for how well do these two independent measurements agree, is well over 99 percent.
In other words, the satellite data and the balloon data both say that the records reflect the actual change in this layer of air. Again, as with the satellite record, one can recognize short-term natural variations–El Niño, La Niña, volcanic eruptions–but one does not see the decades-long human-caused warming trend projected by climate models.
Often, one sees these same data from this key layer of air with a linear trend drawn through them. However, because of bias in the record from a natural phenomenon, it is not appropriate to draw a straight line through the four decades of the temperature record. One must work around the natural phenomenon I’m going to tell you about.
Every 20 to 30 years, the Pacific Ocean changes sharply. The sudden shift is called the Pacific Decadal Oscillation, or PDO, and produces an ocean, air, and wind current shift. Fishermen will notice, for example, migrations of fish species along the West Coast.
In 1976-1977 the Pacific Decadal Oscillation shifted, and is labeled the Great Pacific Climate Shift of 1976-1977. As a result, temperatures changed dramatically from their former average (since around 1946), and returned to warmth seen from around 1923 to 1946. So sharp is the shift that the appropriate thing to do is to look for a secular trend (which might be the human-made trend) before 1976-1977, and then after 1976-1977. But drawing a straight line through that natural event should be avoided.
The PDO is natural, because proxy records–of tree growth, for example–detail the oscillation going back several centuries, which is prior to human activities that significantly increase the content of greenhouse gases in the air.
And also known from computer simulations is that the human-made warming trend is supposed to grow steadily over decades. So, a shift all at once in 1976-1977 is ruled out by those two reasons. One, it’s not what the models project; and two, we see this event before the build-up of human-made greenhouse gases, and it is therefore natural.
The satellite data and the balloon data agree when both records coexist, from 1979 to the present. The balloon record reaches back four decades. Neither record sees a meaningful human-made warming trend.
Now, just remember this one thing from this talk, if nothing else: That layer of air cannot be bypassed; that layer of air must warm if computer model projections are accurate in detailing the human-made warming trend from the air’s increased greenhouse gases. But that layer of air is not warming. Thus the human-made effect must be quite small.
Additionally, the recent warming trend in the surface record must not owe to the human-made effect. The surface temperature is warming for some other reason, likely natural influences. The argument here, from NASA and NOAA data, is that this layer of air from one to five miles in altitude is not warming the way computer simulations say it must warm in the presence of human activity. Therefore, the human-made effect is small. The surface data must be warming from natural effects, because the human-made warming trend must appear both in the low troposphere and at the surface. All models are in agreement on that.
SOLAR ACTIVITY
Now, if the surface data are warming for a natural reason, what might that be? Our research team studies changes in the energy output of the sun and its influence on life and the environment of earth.
Records of sunspot activity reach back to the days of Galileo, some 400 years ago. Scientists then could project an image of the sun and draw these dark sunspots that were seen through early telescopes. We know sunspots to be areas of intense magnetic activity, and from NASA satellite measurements in the last 20 years, we know that over time periods of decades, when the magnetism of the sun is strong, the energy output of the sun is also more intense. That is, the sun is a little bit brighter when magnetism is high, and the sun is a bit fainter when magnetism is weaker.
The sharp ups and downs in the sunspot record define the familiar 11-year cycle, or sunspot cycle. The period is not exactly 11 years. It varies between eight and 15 years, and there is no good explanation for the cause of the cycle. But I’m not going to look at the short term, but rather the changing sun over decades to centuries.
Over the past half-century, the sun has become very active, and the sun is more active than it has been for 400 years. Therefore, the sun is likely at its brightest in 400 years.
Also noteworthy is a feature called the Maunder Minimum. In the 17th century, the observations of sunspots show extraordinarily low levels of magnetism on the sun, with little or no 11-year cycle. That phase of low solar activity has not been encountered in modern times (although radiocarbon records indicate that a Maunder-minimum episode occurs for a century every several centuries). The 17th-century Maunder Minimum corresponds with the coldest century of the last millennium.
That may not be a coincidence. If the sun’s energy output had faded, the earth may have cooled in response to that decrease in the sun’s total energy output.
The next step is to look closer at the temperature records on earth, and see if they link to the decadal-to-century changes in the sun’s energy output. Climate scientists believe they can reliably reconstruct Northern Hemisphere land temperature data back to, say, the year 1700.
If changes in the energy output of the sun, drawn from the envelope of that activity of changes in the sun’s magnetism, are superposed on the reconstructed temperature record, then the two records show a good correlation.
The ups and downs of each record match fairly well. The coincident changes in the sun’s changing energy output and temperature records on earth tend to argue that the sun has driven a major portion of the 20th century temperature change. For example, a strong warming in the late 19th century, continuing in the early 20th century, up to the 1940s, seems to follow the sun’s energy output changes fairly well.
The mid-20th century cooling, and some of the latter 20th century warming also seem matched to changes in the sun.
To review: The surface warming that should be occurring from human-made actions, which is predicted to be accompanied by low troposphere warming, cannot be found in modern records from balloon and satellite platforms.
Thus, the recent surface warming trend may owe largely to changes in the sun’s energy output.
ECONOMIC CONSEQUENCES OF THE POLICY DEBATE
Science is the primary tool to understand human-caused global warming. But economic consequences of policies meant to cut greenhouse gas emissions also enter the policy debate.
Kyoto-type greenhouse gas emission cuts are expected to make little impact on the forecast rise in temperature, according to the computer simulations (which seem to give exaggerated warming trends, as discussed). One forecast, from the UK Meteorological Office, underscores the point. Without Kyoto, that model predicts a rise in globally averaged temperature of just about 1 degree Centigrade by the year 2050. Implementing Kyoto, according to that model, would result in a slightly but insignificantly lower temperature trend. The temperature rise avoided by the year 2050–the difference between the two trends–is six-hundredths of a degree. That is insignificant in the course of natural variability of the climate. Another way to look at the averted warming is that the temperature rise expected to occur by 2050 is projected to occur by 2053 if the emission cuts are enacted.
The conclusion is that one Kyoto-type cut in greenhouse gas emissions averts no meaningful temperature rise, as projected by the models. In order to avoid entirely the projected warming, British researchers estimate that 40 Kyoto-type cuts in greenhouse gas emission would be required.
The cost of implementing one Kyoto-type cut is enormous. Fossil fuels supply approximately 85 percent of energy needs in the United States; worldwide the fraction is about 80 percent. International policy discussions propose expensive solutions centered on sharp fossil fuel use cuts and a massive increase in solar and wind power. A cost-effective solution that does not stunt energy use and energy growth is to shut down coal plants, extend the licenses of the 100 nuclear power plants in the United States, and build about 800 more. However, that is not under serious discussion as a solution to what is often described as the most pressing crisis facing the earth.
Renewable energy sources like solar and wind are not only expensive but also environmentally damaging in their vast land coverage. Those renewable energy sources are not foreseen as seriously meeting projected energy and economic growth. For economic growth, fossil fuels will be relied on for the next decade or two.
The cost of engaging in one Kyoto-type greenhouse gas emission cut ranges between $100 billion and $400 billion of lost GDP annually in the United States. For comparison, consider that the Social Security Trustees estimated $407 billion was transferred to retirees in 2001. The $400 billion annual loss in GDP is approximately numerically equal to the total amount of public and private primary and secondary education spending in the United States.
A recent study from Yale University says that over the next 10 years, Kyoto-type cuts would cost about $2.7 trillion in lost GDP in the U.S.
Those costs must be increased if the target of greenhouse gas emission cuts is not one Kyoto-type agreement but 40.
Another possible target for emission cuts is the benchmark of stabilizing the atmosphere at a level of 550 parts per million of equivalent carbon dioxide concentration. That target probably will be discussed at the World Summit on sustainability in Johannesburg. Current discussions imply that developed countries like the United States would be forced to go to zero net carbon emissions by the year 2050. Beyond 2050, the United States would produce net negative carbon emissions, i.e., the United States would not only continue to emit zero net carbon, but also to begin removing carbon from the atmosphere.
In summary, little evidence supports the idea of catastrophic human-made global warming effects. Undertaking a Kyoto-type program would produce little abatement of the forecast risk, while the cost of such a program would divert resources and attention from major environmental, health, and welfare challenges.
In that regard, forecasts are made of the hypothesized impacts of projected human-made global warming effects. For example, one scenario is that hurricanes may increase because more carbon dioxide has been added to the air. This would be a serious economic impact because hurricanes are the costliest natural disaster in the U.S. But hurricanes have not increased in number or severity in the past 50 years. The cost of property damage has increased, because the cost of property has risen along with the rise in U.S. wealth–not because carbon dioxide has been added to the air.
Another scenario is that human-made global warming will see sweeping epidemics of infectious diseases like malaria in the United States. But malaria is endemic to the United States. Malaria strikes were quelled not by controlling the weather, or by controlling the amount of carbon dioxide in the air, but through increased wealth. That the United States became wealthier from fossil fuel use meant people could be protected from malaria by living inside screened or climate-controlled structures, by reducing the disease vector, mosquitoes, and by advancing medical knowledge and care. In contrast, nearly one million people die from malaria each year; many of its victims are children in Africa and other developing nations.
Diminishing the impact of natural disasters is an immediate worldwide need that rests on keeping the U.S. and world economy vibrant. Energy use, that is, fossil fuel use, helped achieve stunning progress for humankind and the environment in the 20th century. For example, life expectancy in the U.S. in the 20th century nearly doubled.
Agricultural experts estimate that technology has improved crop output. But some increase in crop growth, namely about 10 percent, may owe to the added carbon dioxide in the air, that is, the aerial fertilization effect from carbon dioxide. Carbon dioxide is not a toxic pollutant. It is essential to life on earth.
The latest scientific results are good news: The human influence on global climate change is small and will be slow to develop. The conclusion comes from the lack of meaningful warming trends of the low layer of air, in contradiction to the computer simulations that project a strong human effect should already be present. Those results present an opportunity to improve climate theory, computer simulations of climate, and obtain crucial measurements.
The economic consequences of not relying on science but instead on the anti-scientific Precautionary Principle, are considerable, and are not so speculative. The economic impact of significantly cutting fossil fuel use will be hard-felt, and they will be devastating to those on fixed incomes, those in developing countries, and those on the margins of the economy.
For the next several decades, fossil fuel use is key to improving the human condition. Freed from their geologic repositories, fossil fuels have been used for many economic, health, and environmental benefits. But the environmental catastrophes that have been forecast from their use have yet to be demonstrated by their critics.
Sallie Baliunas, Ph.D., is a Senior Scientist at George C. Marshall Institute and co-host of TechCentralStation.com. The views expressed here do not necessarily represent those of any institutions with which she is affiliated.


Print this page

Related Posts



Have your say:

Your email address will not be published. Required fields are marked *

*