Wild blue yonder: Engineers tackle challenges of hypersonic flight

Print view

A multi-year collaboration among Stanford engineering departments uses some of the world’s fastest supercomputers to model the complexities of hypersonic flight. Someday, their work may lead to planes that fly at many times the speed of sound.

Aeronautical engineers believe that hypersonic planes flying at seven to fifteen times the speed of sound will someday change the face of air travel here on Earth and out into space. If we can master its ‘known unknowns,” that is.

Hypersonic flight is a particularly intense engineering challenge both in the mechanical forces placed on the structure of the plane and in the physics of the sophisticated engines that must operate without fail in the extremes of the upper atmosphere.

Real-world laboratories can only go so far in reproducing such conditions, however, and test vehicles are often rendered extraordinarily vulnerable. Of the U.S. government’s three most recent tests, two ended in vehicle failure.

Now, thanks to a five-year, $20 million U.S. Department of Energy grant, an interdepartmental, multiyear research effort is underway at Stanford University to use some of the world's fastest super computers to tackle the challenges virtually.  

Hypersonic planes
Mechanical engineering Professor Parviz Moin is director of the Predictive Science Academic Alliance Program (PSAAP). The image behind him is from a computer model of jet fuel injected into a supersonic flow stream within the combustor of a scramjet engine. Photo: John Todd
 

The Stanford Predictive Science Academic Alliance Program (PSAAP) is using computers to model the physical complexities of the hypersonic environment—specifically how fuel and air flow through a hypersonic aircraft engine, known as a ‘scramjet’ engine.

PSAAP is a collaboration of the departments of mechanical engineering, aeronautics and astronautics, computer science, mathematics and Stanford’s Institute for Computational and Mathematical Engineering.

In particular, the program focuses on what is known as the scramjet’s ‘unstart’ problem, said Parviz Moin, the Franklin P. and Caroline M. Johnson Professor in the School of Engineering. He is the founding director of Stanford’s Center for Turbulence Research and faculty director of PSAAP.

“If you put too much fuel in the engine when you try to start it, you get a phenomenon called ‘thermal choking,’ where shock waves propagate back through the engine,” he explained. “Essentially, the engine doesn’t get enough oxygen and it dies. It’s like trying to light a match in a hurricane.”

Complex flows and ‘epistemic uncertainty’

Modeling the unstart phenomenon requires a clear understanding of the physics and then reproducing mathematically the immensely complex interactions that occur at hypersonic speeds.

It is impossible to model the physical world exactly, noted Juan Alonso, a professor of aeronautics and astronautics. “When you base decisions on computations that are in some way imperfect, you make errors,” he said. “Not only that, but these hypersonic vehicles are themselves subject to uncertainties in how they behave in the air.”

As a result, PSAAP principal goal is to try and quantify those uncertainties—the known unknowns—so that scramjet engineers can build the appropriate tolerances into their designs to allow the engines to function in these extraordinary environments.

Measuring what the engineers call ‘epistemic uncertainty’ is common in real-world experimental work, where researchers typically note uncertainties in their measurements in the form of 'bars' that have upper and lower limits. But hard numbers generated by computer models do not have uncertainty bars and therefore can take on an unwarranted, potentially dangerous air of certainty.

“We have asked experimentalists for a long time to give us uncertainty bars on their measurements,” observed Moin. “But I think the time is right to ask the computationalists to do the same.” Quantification of uncertainties in numerical predictions is at the core of PSAAP. Professor Gianluca Iaccarino, assistant professor of mechanical engineering, is leading this effort.

Innovations at the exascale

One reason computational uncertainty quantification is a relatively new science is that, until recently, the necessary computer resources simply didn’t exist.

“Some of our latest calculations run on 163,000 processors simultaneously,” Moin said. “I think they’re some of the largest calculations ever undertaken.”

Thanks to its close relationship with the Department of Energy, however, the Stanford PSAAP team enjoys access to the massive computer facilities at the Lawrence Livermore, Los Alamos and Sandia National Laboratories to run their largest and most complex simulations.

It takes specialized knowledge to get computers of this scale to perform effectively, however. “And that’s not something scientists and engineers should be worrying about,” suggested Alonso, “which is why the collaboration between departments is critical. Mechanical engineers and those of us in aeronautics and astronautics understand the flow and combustion physics of scramjet engines and the predictive tools. We need the computer scientists to help us figure out how to run our tests on these large computers.”

That need will only increase over the next decade as supercomputers move towards the exascale—machines with a million or more processors able to execute a quintillion calculations in a single second.

Thinking ahead to that day, the PSAAP team under the direction of computer science Professor Pat Hanrahan has created LISZT, an entirely new computer language designed specifically to run complex simulations on massive processor sets.

The great virtue of LISZT is its capability to directly express problems in engineering and science through code that is built for exascale architectures, making it equally accessible to experts working in fluid physics, combustion, turbulence and other mathematically intense applications. At the same time, LISZT remains highly computationally efficient.

Indeed, LISZT seems to be one of the most viable methods – and some say the only method – for doing real scientific modeling at the exascale, a distinction that has it attracting widespread international interest.

“It’s something you could never have created unless you put computer scientists, mathematicians, mechanical engineers, and aerospace engineers together in the same room,” argued Alonso. “Do it, though, and you can produce some really magical results.” 

Advances in several dimensions

Beyond progress in the treatment of ‘epistemic uncertainty’ in computer modeling and the creation of LISZT, the Stanford PSAAP team has made other key advances in the specific disciplines that underpin scramjet engineering: combustion, turbulence and fluid flow in general.

And while the challenges of running an actual scramjet engine in a wind tunnel at hypersonic speeds remain daunting, the Stanford PSAAP project is unique in that it has supported advances in physical experimentation.

“We've had strong validation of our computational work against experiments,” said Moin. “In particular, in the High Temperature Gas Dynamics Lab and the Flow Physics group we’ve developed a number of new techniques to validate how we build physical models into our computer code.”

Collectively, these insights will enable the design of safer, more reliable hypersonic engines. But PSAAP’s advances in quantifying uncertainty have other, far broader implications, ventured Alonso.

“These same technologies can be used to quantify flow of air around wind farms, for example, or for complex global climate models," he said. "I was in Los Alamos talking with people who are interested in global climate and guess what? Just like the models for the scramjet, right now their climate models are far from perfect, but it doesn’t stop them from pushing ahead.”

Last modified Fri, 24 Aug, 2012 at 12:18