Todd Gross, chief meteorologist for WHDH-TV, an NBC affiliate in Boston, Massachusetts, said, "They are getting better and better and better."
"The four-day forecast is now as accurate as the old two-day forecast," he said. "The two-day forecast is as accurate as the old-one daythough after about 72 hours, it's still difficult to get an accurate forecast from these computer models."
Life, Death, and Dollars
When dangerous storms occur, accurate forecasting can have life-saving consequences, particularly for the growing population found in U.S. coastal regions.
"With respect to hurricane landfall it costs about a million dollars a mile to evacuate [people from] the coast," Hoke, the NOAA scientist, said. "So every mile that you don't evacuate saves a considerable amount of money."
The global economy is also increasingly dependent on weather forecasts. As much as one-third of the U.S. gross domestic productthree trillion dollars' worth of goods and servicesis at least partially dependent on weather, according to estimates.
Obvious examples include road, sea, and air transportation. Less obvious, perhaps, are power companies, which depend on temperature forecasts to anticipate consumer demand, and school systems.
"My wife directs a preschool, and the parents and staff are highly weather sensitive," Hoke added. "She pays more attention to the local weather in the Washington, D.C.-area than I do."
In 1965 Gordon Moore, the co-founder of computer chip maker Intel, posited that the number of transistors possible on a chip would double every two years. The prediction, which became known as Moore's Law, has proved a tad conservative.
Such progress has led to dramatic growth in supercomputing power. NOAA's machines pack a serious punch.
But their critical responsibilities keep them a bit off the cutting edge of computer capability, said Dion Rudnicki, an IBM vice president based in Bethesda, Maryland.
"The main characteristic for NOAA is that they live and die by a 24-by-7-by-365 clock[and are] concerned about reliability and availability," Rudnicki said. "This is the system producing the output that they need to be able to run the nation's weather services."
The machines still feature staggering abilities.
The early computers of the 1950s worked at speeds of about a thousand calculations per second, a rate known as a kiloflop.
By 2000 supercomputers processed data at a rate of trillions of operations per seconda speed described in teraflops.
To put that in perspective, think of a kiloflop as a piece of paper with a thousand numbers recorded both sides in the span of a second.
By comparison, "a teraflop [is] a stack of paper of the altitude achieved by Spaceship One[62 miles/100 kilometers], the edge of spaceevery second," explained Dongarra, the University of Tennessee computer expert.
NOAA's top machines currently run at more than four teraflops per second.
Dongarra reports that by 2010 computers will likely achieve speeds measured in petaflops, a rate of data processing represented by a stack of paper 62,140 miles (100,000 kilometers) in elevationor one-fourth the distance to the moon.
As processing speeds continue to soar, weather forecasters and those who count on them stand to benefit.
"The progress in weather prediction is right up there with the progress in, say, medical science," Hoke said. "The improvement is just undreamed of."
Free E-Mail News Updates
Sign up for our Inside National Geographic newsletter. Every two weeks we'll send you our top stories and pictures (see sample).
SOURCES AND RELATED WEB SITES