Supercomputers Take a Cue from Microwave Ovens
Berkeley Researchers Say Co-design May Be the Answer to Modeling Clouds and Other Big Problems
November 30, 2011
Linda Vu, firstname.lastname@example.org, +1 510 495 2402
As sophisticated as modern climate models are, one critical component continues to elude their precision—clouds. Simulating these fluffy puffs of water vapor is so computationally complex that even today’s most powerful supercomputers, working at quadrillions of calculations per second, cannot accurately model them.
“Clouds modulate the climate. They reflect some sunlight back into space, which cools the Earth; but they can also act as a blanket and trap heat,” says Michael Wehner, a climate scientist at the Lawrence Berkeley National Laboratory (Berkeley Lab). “Getting their effect on the climate system correct is critical to increasing confidence in projections of future climate change.”
In order to build the breakthrough supercomputers scientists like Wehner need, researchers are looking to the world of consumer electronics like microwave ovens, cameras and cellphones, where everything from chips to batteries to software is optimized to the device’s application. This co-design approach brings scientists and computer engineers into the supercomputer design process, so that systems are purpose-built for a scientific application, such as climate modeling, from the bottom up.
“Co-design allows us to design computers to answer specific questions, rather than limit our questions by available machines,” says Wehner.
Co-design Test Case: Clouds
In a paper entitled “Hardware/Software Co-design of Global Cloud System Resolving Models,” recently published in Advances in Modeling Earth Systems, Wehner and coauthors argue that the scientific supercomputing community should take a cue from consumer electronics like smart phones and microwave ovens: Start with an application—like a climate model—and use that as a metric for successful hardware and software design.
The paper which uses the climate community’s global cloud resolving models (GCRMs) as a case-study argues that an aggressive co-design approach to scientific computing could increase code efficiency and enable chip designers to optimize the trade-offs between energy efficiency, cost and application performance.
According to coauthor David Donofrio, a co-designed system for modeling climate would contain about 20 million cores (today’s most powerful scientific cluster, Japan’s ‘K Computer’ contains about 705,000 cores) and be capable of modeling climate 1,000 times faster than what is currently possible.
“Most importantly, the system would remain fully programmable so that scientific codes with similar hardware needs to the GCRMs, like seismic exploration, could also benefit from this machine,” says Donofrio, a computer scientist at Berkeley Lab.
“Today when we purchase a general purpose supercomputer, it comes with a lot of operating system functions that science applications don’t need. When you are worried about power, these codes can be very costly,” says John Shalf, who heads Computer and Data Sciences research at Berkeley Lab. “Instead of repurposing a chip designed for another market, the scientific HPC (high performance computing) community should specify what they want on a chip—the intellectual property (IP)—and only buy that.’”
According to Shalf, a co-designed system for modeling climate would use about one quarter to one tenth the energy required for a conventional supercomputer with the same capabilities.
Consumers Pave the Way for Next Generation Supercomputers
Although innovative for scientific supercomputing, the idea of application-driven design is not new. Electronics like cell phones and toaster ovens are built of simpler embedded processor cores optimized for one or a few dedicated functions.
“Because the ultimate goal of the embedded market is to maximize battery life, these technologies have always been driven by maximizing performance-per-watt and minimizing cost. Application-driven design is the key to accomplishing this,” says Shalf. “Today we look at the motherboard as a canvas for building a supercomputer, but in the embedded market the canvas is the chip.”
He notes that the most expensive part of developing a computer chip is designing and validating all of the IP blocks that are placed on the chip. These IP blocks serve different functions, and in the embedded market vendors profit by licensing them out to various product makers. With an application in mind, manufacturers purchase IP block licenses and then work with a system integrator to assemble the different pieces on a chip.
“You can think of these IP blocks as Legos or components of a home entertainment system,” says Donofrio. “Each block has a purpose. You can buy them separately and connect them to achieve a desired result, like surround sound in your living room.”
“The expensive part is designing and verifying the IP blocks, not the cost of the chip. These IP blocks are commodities because the development costs are amortized across the many different licenses for different applications,” says Shalf. “Just as the consumer electronics chip designers choose a set of processor characteristics appropriate to the device at hand, HPC designers should also be able to chose processor characteristics appropriate to a specific application or set of applications, like the climate community’s global cloud resolving model.”
He notes that the resulting machine, while remaining fully programmable, would achieve maximum performance on the targeted set of applications, which were used as the benchmarks in the co-design process. In this sense, Shalf notes that the co-designed machine is less general purpose than the typical supercomputer of today, but much of what is included in modern supercomputers is of little use to scientific computing anyway and so it just wastes power.
“Before this work, if someone asked me when the climate community would be able to compute kilometer scale climate simulations, I would have answered ‘not in my lifetime,’” says Wehner. “Now, with this roadmap I think we could be resolving cloud systems within the next decade.”
Although climate was the focus of this paper, Shalf notes that future co-design studies will explore whether this will also be cost-effective for other compute intensive sciences such as combustion research.
In addition to Shalf, Wehner and Donofrio, other co-authors of the paper include Berkeley Lab's Leonid Oliker, Leroy Drummond, Norman Miller and Woo-Sun Yan; Marghoob Mohiyuddin of the University of California at Berkeley; Celal Konor, Ross Heikes and David Randall of Colorado State University; and Hiroaki Miura of the University of Tokyo.
About Berkeley Lab
Founded in 1931 on the belief that the biggest scientific challenges are best addressed by teams, Lawrence Berkeley National Laboratory and its scientists have been recognized with 16 Nobel Prizes. Today, Berkeley Lab researchers develop sustainable energy and environmental solutions, create useful new materials, advance the frontiers of computing, and probe the mysteries of life, matter, and the universe. Scientists from around the world rely on the Lab’s facilities for their own discovery science. Berkeley Lab is a multiprogram national laboratory, managed by the University of California for the U.S. Department of Energy’s Office of Science.
DOE’s Office of Science is the single largest supporter of basic research in the physical sciences in the United States, and is working to address some of the most pressing challenges of our time. For more information, please visit energy.gov/science.