Epilogue |
Typically, engineers are becoming more specialized and are less likely to understand the hazards of the space environment. Often we have to relearn things that were known by people who retire or move on. . . . It is also recognized that private companies have a disinclination to release information on their own problems.
—NOAA/SEC Satellite Working Group, 1999
Sometimes we work too hard to make coincidences into real cause and effect: recall the example of the Exxon Valdez accident during the March 1989 space weather event. And sometimes it’s not easy to grasp just how complex space weather issues have become in the last ten years. There are many facets to the story and, like a diamond, the impression you get depends on your perspective. When I first started learning about this subject, I was overwhelmed by the lack of careful documentation, and the impossibility of ever finding it, for many of the outages I had heard about. I was also nervous about the basic issue of simultaneous events masquerading as cause and effect. If you are intent on seeing every mishap as a demonstration that satellites are inherently vulnerable, then there is ample circumstantial evidence to support the claim.
Why is it that satellites are assumed to be unaffected by events like cosmic rays, solar flares, or energetic electrons until such a claim can be “proven” by recovering the “body”? If we know that space is a hostile environment to off-the-shelf and nonradiation hardened systems, why do multinational corporations greet the inevitable failures with suspicion when they do happen? Where did this presumption of innocence begin to enter the professional discussion?
So long as there is any scientific uncertainty about cause and effect, well founded or otherwise, there will be no measurable change in the attitudes of satellite manufacturers. As NASA Goddard space scientist Michael Lauriente notes, “Because of the statistical nature of the risks, apathy reigns among some of the spacecraft community.” Space physicist A. Vampola from the Aerospace Corporation also confronts the irony of the sparring between scientists and satellite owners, “The space environment is hostile. It is not a benign vacuum into which an object can be placed and be expected to operate exactly as it did when it was designed and tested on the ground.”
There are plenty of uniquely interested parties who would denounce each satellite “kill” as simply a technological problem with insulation or some other factor, and there is also ample cause to support this point of view. Scientific data is too sparse for us to look into every cubic yard of space and say, with authority, that “this satellite was killed by that event.” This means that you are completely free to argue that satellites are actually quite invulnerable to any space weather effect, and that the failures are purely a matter of quality control. But, if that is the case, why, then, do satellite owners take out insurance policies? Why do they overdesign solar panels and go on record saying that they expect to lose up to six satellites per year?
A satellite communications corporation may very well be intent on maintaining the status quo in supporting the ragtag effort of space weather research–and may find just cause to believe that this strategy is OK, too. After all, severe solar storms capable of producing a blackout happen very infrequently compared to other kinds of power outages (ice storms, line sags, and the like). Since satellite owners suppress evidence of anomalies traceable to solar storms, no trail of problems remains to require better space weather models. However, NOAA has many satellite owners on confidential lists that receive daily updates about space conditions. In a classic catch-22 situation, these clients would like better forecasts but cannot support such an effort at NOAA or NASA without going public and implying that their service is vulnerable.
While satellite owners and electric power utilities continue to expand their vulnerability, and reap billions of dollars in annual profits, space weather forecasting continues to languish. Both NASA and NOAA have been forced into “flat” or declining budgets for these specific activities. Back in June 1997, NOAA’s Space Environment Center was facing a funding crisis with flat funding or zero growth planned for the years following 1997. According to Ernest Hildner, the director of NOAA’s Space Environment Center in Boulder, Colorado, “Without increased funding, we will be facing the next solar cycle with two-thirds as much staff as we had during the last solar cycle.” A similar route of the National Geophysical Data Center was attempted in the 1970s.
Space physics, and modeling research in support of space weather forecasting, tends to be underfunded given the complexity of the subject. Through much of the 1990s, efforts to develop better forecasting models at NASA were also stuck on the slow track. This occurred at the same time new floods of data entered the NASA data archives. Space physics doesn’t seem to have a constituency the way cosmology and planetary exploration does. A mission like SOHO costs $1.2 billion to build and support, of which $400 million was supplied by NASA and the rest provided by the European Space Agency. According to Art Poland, who was the former NASA project manager and project scientist for SOHO,
SOHO can support 20–25 scientists full time, and has done so for the last 4 years. Including institutional overhead costs, it costs about $150,000 to support each scientist. The total research budget for a mission like SOHO is about $8 million per year. To support the entire ISTP program each year including technicians, engineers and scientists costs about $50 million.
So, despite the billions of dollars that go into building and supporting satellite hardware, in some cases less than a few percent of the total mission cost ever shows up as support for scientists to analyze the incoming windfall of new data. This means that, in some instances, less than $50 million per year supports scientists to develop the forecasting techniques to safeguard $100 billion in satellite assets and $210 billion in electrical power company profits. Researchers may get dazzling images of CMEs from one satellite like SOHO between 1996–2001, solar wind data from another satellite like ACE or WIND from 1999–2002, and geospace data from another satellite such as IMAGE between 2000–2002. This information may not overlap in time–it may not even overlap in wavelength–yet it all has to be knitted together. This makes developing space weather models a frustrating enterprise, because it is easier to decipher the language of space weather if you have complete sentences rather than fragments.
No agency other than NASA is mandated to build the satellites that keep watch on space weather. This means that space weather monitoring activities are forced to piggyback on satellites optimized for pure research. The organizations that benefit most from NASA research satellites receive the space weather data free of charge and, for a variety of internal reasons, are unable to make a significant financial investment in data acquisition themselves. This solution is certainly easier than investing in space weather satellites, but it also means that newer, better forecasting techniques, for example, in the electric power industry, are held hostage to the necessities of ongoing yet uncertain NASA funding.
The National Academy of Sciences evaluated the readiness of NASA, NOAA, DoD, and DOE to develop a comprehensive space weather forecasting model. Although they identified many successes in the existing programs, they also identified glaring problems that work against accomplishing the National Space Weather Program goals anytime soon. NASA, up until 1997, was planning to stop funding its ISTP program after fiscal year 1998; NOAA’s Space Weather Center suffered a 33 percent staff reduction and had stopped supporting the translation of data-based models into forecasting tools. “The lack of commitment at NOAA to this unique and critical role will have a fundamental impact on the success of the NSWP.” A similar critique was leveled against DoD, where the staff turnover time was less than a single eleven-year solar cycle so there was no institutional memory of what was learned during past cycles.
Fortunately, NASA’s Living with a Star program, which was conceived in 1999, promises to keep ISTP fully funded until its spacecraft expire from wear and tear. The program will also set in motion a more aggressive support of space weather research during the next decade. To be ready for the next solar maximum in 2011, missions must be planned today so that the necessary hardware will be in place when the new round of solar activity rises to a climax.
Beyond the interests of the research community, many different elements of our society have recently begun to appreciate the need for a deeper understanding of the space environment. The military and commercial satellite designers, for example, are not happy that they have to use twenty- to thirty-year-old models to predict space weather conditions. This forces them to fly replacement satellites more frequently or overdesign others to withstand even the occasional major solar flare, which may never come at all. Unfortunately, the current trends in satellite design seem to be directed toward increasing satellite vulnerability to disabling radiation damage. As pointed out by William Scott in Aviation Week and Space Technology,
Austere defense budgets also have increased reliance on more affordable, but perhaps less robust, commercial off-the-shelf hardware. . . . Expensive radiation-hardened processors are less likely to be put on some military satellites or communication systems now, than was once the case according to USAF officers. . . . Newer chips are much more vulnerable than devices of 10–15 years ago.
In an age when “cheaper, faster, and smaller” drives both civilian and military satellite design, satellites have become more susceptible to solar storm damage than their less sophisticated but more reliable predecessors during past solar cycles. As more satellites become disabled by “mysterious” events that owners prefer not to discuss openly, old lessons in satellite design need to be rediscovered. In recent advertisements, some satellite manufacturers boast that they employ “advanced composite materials which improve performance while reducing weight.” This also makes for poor shielding because weight, and quite a bit of it for shielding, has proven itself to be a good radiation mitigation strategy.
In all this concern over satellite survivability, there looms another harsh circumstance that may make long satellite life times impossible. Large networks of satellites in LEO are suffering something of a shakeout that has nothing to do with the rise and fall of solar storms. The pace of communications technology, and consumer needs, has begun to change so quickly that by the time a billion-dollar satellite network is in place it is nearly obsolete. Companies have to install their networks within one to two years or run the risk of becoming a technological backwater before the planned life span has been reached. The two companies that have put in place the two largest networks by 1999, Iridium and Globalstar, have already filed for bankruptcy. Iridium was able to sell only ten thousand of the one hundred thousand wireless telephone handsets it planned, and these phones did not have the capability to send data along with voice transmissions. No sooner had Iridium gone on the air than its owners regretted the decision not to include high-speed data channels. Globalstar, meanwhile, had not even completed installing its full satellite system before it followed Iridium LLC into bankruptcy. Yet this kind of record, so soon in the fledgling LEO industry, doesn’t appear to bother some people. Satellite communications is still seen as a highly lucrative business despite lightweight satellites that will inexorably face space weather problems. The tide of technological progress is sweeping the industry at an ever faster pace. A sense of urgency now pervades the industry: make profits as quickly as possible. For example, Bruce Elbert, senior vice president of Hughes Space and Communications International, and the leading manufacturer of geosynchronous communications satellites, suggested with great enthusiasm, “The next millennium may see all land-line communications going wireless. You could wait [to enter this market], but why put off gaining the economic and potentially competitive advantages of using satellite technology today?”
Meanwhile, communication technology has not stopped evolving. New devices and systems are on the rise that may eventually make communication satellites less desirable, at least by the peak of the next solar cycle in 2011.
The first fiber optic cable, TAT-8, entered commercial service in 1988, and since then no fewer than 408,000 miles of fiber have been laid across the oceans by 2000. The current investment in undersea fiber exceeds $30 billion and is expected to surpass $50 billion by 2003. One project alone, Project Oxygen, will be a $14 billion cable tying thirty countries in Europe and North America together. It will take no more than three months to lay this cable and, when it is completed, it will carry twenty-five million phone calls and ten thousand video channels. It can carry all the world’s Internet traffic on one of its four fibers, which can deliver 320 gigabytes per second of capacity. The combined bandwidth of 1.2 terabytes per second is enough to transmit the entire text contents of the Library of Congress in a few seconds. Even so, the signals carried by these fibers have to be amplified, and these electronic units can be vulnerable to space weather effects on the ground and under the ocean.
The driving force behind the spectacular growth in fiber optic technology is the Internet and the insatiable appetite it creates for massive volumes of data delivered immediately. At the same time, the explosive growth in the market for cellular phones has driven the satellite industry to meet traveler’s desires to stay in touch with family and coworkers no matter where they may be on the Earth. The biggest drawback to fiber optic communications is that to take advantage of the high data rates it requires land lines to individual users. Satellites, meanwhile, require that their users need only a portable handphone or a satellite dish to receive their signals directly. Satellites work well when connecting large numbers of rural or off-the-road users. Yet fiber optic cables still have the advantage of the highest data rates, and they do not have to be replaced every five to ten years the way satellites must. Modern cables are not rendered useless as new, faster, technologies arise. Only the equipment at the cable’s end stations has to be upgraded, not the entire undersea fiber cable. In addition, when a cable is damaged, it can be easily repaired underwater using remote-controlled robots. On the other hand, satellite transponders create the bottleneck problem as the capability of this technology increases, and they can only be upgraded by launching a new $200 million satellite.
The pace of fiber optic cable evolution and deployment has been so brisk that even the U.S. National Security Agency has been forced to rethink its intelligence operations. With more and more sensitive data being carried by fiber, the NSA’s powerful satellite eavesdropping systems no longer hear as much of the secret messaging that they used to pick up from satellite or radio transmitters.
Unlike satellites, which largely support our entertainment and financial needs, the pervasive use of electrical power can create far more serious problems than even the most dramatic satellite outage. Electrical power outages can completely shutdown an entire region’s commerce for hours or days, and they can cause death. Geomagnetic storms will continue to bombard the power grid, and there is little we can do to harden the system. The blackouts we will experience in the future will have as much to do with our failure to keep electrical supply ahead of demand as it does with the failure of the electrical power industry to put in place the proper forecasting techniques. Unlike satellite design, there are few easy solutions to the declining operating margins because power plants seem to be universally unpopular. Our vulnerability to the next blackout has become as much a sociological issue as a technological one.
When the ice storm of January 1999 darkened the lights of nearly half a million residents in Virginia, Maryland, and Washington, D.C., no one thought it was especially amusing or trivial, even though only 0.1 percent of the U.S. population was involved. My family and I were among the last to have our electrical service restored after waiting in the cold and dark for five days. The first night was a curious mix of concern and genuine delight at the new experience. We huddled together under down comforters and actually sang campfire songs as our home slowly gave up its latent heat. We were delighted to see the beautiful trees, like sculptures of ice, bent over along the street, which was now an impassable skating rink. Elsewhere, emergency rooms were filling up with people who had broken bones, turned wrists and ankles, or blood streaming down from head wounds and concussions. Hundreds of traffic accidents offered up a handful of fatalities as people died for no reason other than being in the wrong place at the wrong time. By the third night, we had joined thousands of others in Montgomery County, calling the electrical utility company to find out when we might be reconnected. We found ourselves the only street in our community that did not have electrical service. At night, we watched our neighbor’s security light cast ghostly figures of tree limbs on our bedroom walls, by day we frequented local shopping malls to keep warm and ate our meals at restaurants. We felt a temporary kinship with homeless people who endure this daily. It wasn’t the Sun and its mischief that had brought this on, but a common natural occurrence of more mundane origin. Still, the discomfort, expense, physical pain, embarrassment, and even death that it caused is a potent reminder that we can no longer afford to be as tolerant of power and communication outages as we once were. When you gamble with space weather, a major winter storm can occasionally happen at the same time as a large solar storm.
When I first bemoaned how astronomy and astrophysics seldom have practical consequences, I hardly suspected that simply the search for answers to questions in cosmology and galactic astronomy could cause a domino fall that could make blackouts more likely. The very satellites that I backed as a professional astronomer to further my particular area of research and curiosity were in the zero-sum game of modern budget analysis, robbing the space scientists of the tools they would need to improve space weather forecasting. In the future, the next power outage my family and I have to endure may, in some sense, be the consequence of my own professional choices elsewhere in life.
Our vulnerability to solar storms during Cycle 23 follows a pattern that can be seen in other more familiar arenas. Between 1952 and 1986, the average cost of national disasters (earthquakes, tornadoes, hurricanes, and floods) averaged $8 billion every five years. This then doubled in 1987–1991 to $20 billion, and in 1992–1996 the cost skyrocketed to over $90 billion. The reason for this sudden increase in the last five to ten years is that more people, often with higher incomes, are moving in droves to coastal communities–exactly those areas typically battered by earthquakes and hurricanes. As Greg van der Vink and his colleagues at Princeton University’s Department of Geophysics have noted, “We are becoming more vulnerable to natural disasters because of the trends of our society rather than those of nature. In other words, we are placing more property in harm’s way.”
High-cost events, such as a devastating earthquake, simply are not factored into the way our society now plans and develops communities. In fact, economic incentives that would encourage more responsible ways to use land are actively denounced or, worse, stifled. The kinds of institutions and programs that have been established actually subsidize bad choices. For example, they offer low-interest rebuilding loans after an earthquake or help to rebuild beachfronts after a hurricane.
There is much in this terrestrial process that we now see reflected in the way we conduct activities in space. Satellite insurers underwrite bad satellite designs by charging only slightly higher rates (3.7 percent per year compared to 1.2 percent for “good” satellites) for poor or risky designs. They pay no attention to the solar activity cycle in establishing their rates. And once the stage has been set, it is difficult to change old habits. Like homeowners rebuilding property along receding coastlines after a hurricane, satellite owners insist on orbiting low-cost, inherently vulnerable satellites. Electrical utilities, meanwhile, forego investment in even the simplest mitigation technologies and prefer to view GIC problems as a local technical difficulty with a specific piece of equipment.
In the next ten years, the excitement of Cycle 23 will fall behind us, just as the heyday of Cycle 22 has now become a historical footnote. In 2006, we will find ourselves at the beginning of a new round of solar storminess. The great experiments of the LEO communication satellite networks will have run their inevitable courses, either as intermittent successes or as technological dinosaurs. Barring any major blackouts this time around, electrical utilities operating in a deregulated climate will debate on a region by region basis, whether the investment in GIC forecasting makes sense or not. We will also have weathered several years of round-the-clock occupancy of the International Space Station. This latter experience will teach us many lessons—some harsh—about what it really takes to deal with space weather events and what is needed to become permanent occupants of space. At some time, perhaps in the twenty-second century, our present concern and obsession with harmful space weather events will be a thing of the past. There is even the hope that the Sun may, unexpectedly, cease its stormy, cyclical behavior for a century or two as it has before. However, the last time the Sun became inactive in the 1600s, Europe experienced its famous mini ice age for several decades.
But, for now—for this year, for this cycle—we must remain vigilant even in the face of what seems like little cause for outright concern. For every pager that is temporarily silenced, others will work just fine. For every city that is darkened or loses air conditioning, a hundred others will experience just another ordinary day. Space weather is like a shark whose fin we see gliding across the interplanetary waters. We know it is there, but we haven’t completely figured out what to do when it strikes.