[House Hearing, 113 Congress] [From the U.S. Government Publishing Office] AMERICA'S NEXT GENERATION SUPERCOMPUTER: THE EXASCALE CHALLENGE ======================================================================= HEARING BEFORE THE SUBCOMMITTEE ON ENERGY COMMITTEE ON SCIENCE, SPACE, AND TECHNOLOGY HOUSE OF REPRESENTATIVES ONE HUNDRED THIRTEENTH CONGRESS FIRST SESSION __________ WEDNESDAY, MAY 22, 2013 __________ Serial No. 113-31 __________ Printed for the use of the Committee on Science, Space, and Technology Available via the World Wide Web: http://science.house.gov U.S. GOVERNMENT PRINTING OFFICE 81-195 WASHINGTON : 2013 ----------------------------------------------------------------------- For sale by the Superintendent of Documents, U.S. Government Printing Office Internet: bookstore.gpo.gov Phone: toll free (866) 512-1800; DC area (202) 512-1800 Fax: (202) 512-2104 Mail: Stop IDCC, Washington, DC 20402-0001 COMMITTEE ON SCIENCE, SPACE, AND TECHNOLOGY HON. LAMAR S. SMITH, Texas, Chair DANA ROHRABACHER, California EDDIE BERNICE JOHNSON, Texas RALPH M. HALL, Texas ZOE LOFGREN, California F. JAMES SENSENBRENNER, JR., DANIEL LIPINSKI, Illinois Wisconsin DONNA F. EDWARDS, Maryland FRANK D. LUCAS, Oklahoma FREDERICA S. WILSON, Florida RANDY NEUGEBAUER, Texas SUZANNE BONAMICI, Oregon MICHAEL T. McCAUL, Texas ERIC SWALWELL, California PAUL C. BROUN, Georgia DAN MAFFEI, New York STEVEN M. PALAZZO, Mississippi ALAN GRAYSON, Florida MO BROOKS, Alabama JOSEPH KENNEDY III, Massachusetts RANDY HULTGREN, Illinois SCOTT PETERS, California LARRY BUCSHON, Indiana DEREK KILMER, Washington STEVE STOCKMAN, Texas AMI BERA, California BILL POSEY, Florida ELIZABETH ESTY, Connecticut CYNTHIA LUMMIS, Wyoming MARC VEASEY, Texas DAVID SCHWEIKERT, Arizona JULIA BROWNLEY, California THOMAS MASSIE, Kentucky MARK TAKANO, California KEVIN CRAMER, North Dakota ROBIN KELLY, Illinois JIM BRIDENSTINE, Oklahoma RANDY WEBER, Texas CHRIS STEWART, Utah VACANCY ------ Subcommittee on Energy HON. CYNTHIA LUMMIS, Wyoming, Chair RALPH M. HALL, Texas ERIC SWALWELL, California FRANK D. LUCAS, Oklahoma ALAN GRAYSON, Florida RANDY NEUGEBAUER, Texas JOSEPH KENNEDY III, Massachusetts MICHAEL T. McCAUL, Texas MARC VEASEY, Texas RANDY HULTGREN, Illinois MARK TAKANO, California THOMAS MASSIE, Kentucky ZOE LOFGREN, California KEVIN CRAMER, North Dakota DANIEL LIPINSKI, Illinois RANDY WEBER, Texas EDDIE BERNICE JOHNSON, Texas LAMAR S. SMITH, Texas C O N T E N T S Wednesday, May 22, 2013 Page Witness List..................................................... 2 Hearing Charter.................................................. 3 Opening Statements Statement by Representative Cynthia Lummis, Chairwoman, Subcommittee on Energy, Committee on Science, Space, and Technology, U.S. House of Representatives...................... 9 Written Statement............................................ 10 Statement by Representative Randy Hultgren, Committee on Science, Space, and Technology, U.S. House of Representatives........... 11 Written Statement............................................ 11 Statement by Representative Eric Swalwell, Ranking Minority Member, Subcommittee on Energy, Committee on Science, Space, and Technology, U.S. House of Representatives.................. 12 Written Statement............................................ 13 Witnesses: Dr. Roscoe Giles, Chairman, Advanced Scientific Computing Advisory Committee Oral Statement............................................... 16 Written Statement............................................ 18 Dr. Rick Stevens, Associate Laboratory Director for Computing, Environment and Life Sciences, Argonne National Laboratory Oral Statement............................................... 32 Written Statement............................................ 34 Ms. Dona Crawford, Associate Director for Computation, Lawrence Livermore National Laboratory Oral Statement............................................... 46 Written Statement............................................ 48 Dr. Daniel Reed, Vice President for Research and Economic Development, University of Iowa Oral Statement............................................... 60 Written Statement............................................ 62 Discussion....................................................... 71 Appendix I: Answers to Post-Hearing Questions Dr. Roscoe Giles, Chairman, Advanced Scientific Computing Advisory Committee............................................. 84 Dr. Rick Stevens, Associate Laboratory Director for Computing, Environment and Life Sciences, Argonne National Laboratory..... 91 Ms. Dona Crawford, Associate Director for Computation, Lawrence Livermore National Laboratory.................................. 95 Dr. Daniel Reed, Vice President for Research and Economic Development, University of Iowa................................ 102 AMERICA'S NEXT GENERATION SUPERCOMPUTER: THE EXASCALE CHALLENGE ---------- WEDNESDAY, MAY 22, 2013 House of Representatives, Subcommittee on Energy Committee on Science, Space, and Technology, Washington, D.C. The Subcommittee met, pursuant to call, at 10:05 a.m., in Room 2318 of the Rayburn House Office Building, Hon. Cynthia Lummis [Chairwoman of the Subcommittee] presiding. [GRAPHIC] [TIFF OMITTED] 81195.001 [GRAPHIC] [TIFF OMITTED] 81195.002 [GRAPHIC] [TIFF OMITTED] 81195.003 [GRAPHIC] [TIFF OMITTED] 81195.004 [GRAPHIC] [TIFF OMITTED] 81195.005 [GRAPHIC] [TIFF OMITTED] 81195.006 [GRAPHIC] [TIFF OMITTED] 81195.007 Chairwoman Lummis. Good morning. The Subcommittee will come to order. And we are delighted to have a terrific panel here this morning, so welcome to our hearing entitled ``America's Next Generation Supercomputer: the Exascale Challenge.'' In front of you are packets containing the written testimonies, biographies, and truth-in-testimony disclosures for today's witness panel. And now, I will recognize myself for five minutes for and opening statement followed by our Ranking Member Mr. Swalwell. The development and expanded availability of supercomputers has enabled society to push the frontiers of nearly every scientific discipline, and accelerate applications of that science in countless fields. It has enabled modeling and simulation necessary to address national security needs. It drives the boundaries of medical research, reduces cost to develop new products, and improves materials design processes, just to name a few. High performance computing has also revolutionized how the energy sector operates. Advanced modeling and simulation techniques, driven by computer algorithms and faster computing speeds, improve the efficiency of energy production and consumption technologies. These advancements ultimately trace back to Federal investments in basic research that provided the foundation for most of today's computing technologies. From the first megaflop supercomputers of the 1960s, the Federal investments have led to push across each landmark thousand-fold speed barrier to gigaflops, teraflops, and petaflops. I always think of floppy- eared rabbits and when I was a kid showing critters in 4H, I should have named them Giga, Tera, and Peta, but I just didn't know about it back then because that proceeded the first megaflop. Throughout this computing age, we have witnesses--we have witnessed yesterday's supercomputers become today's desktop computers and consumer devices often in incredibly short time frames. The spillover benefits to society are countless and immeasurable. The Department of Energy, led by the Advanced Scientific Computing Research program, plays a critical role in driving these computing technology breakthroughs. DOE supports world- class computational science facilities, such as the National Energy Research Scientific Computing Center. Additionally, DOE funds cutting-edge applied mathematics research and next- generation networking activities. DOE's next major computing challenge, constructing an ``exascale'' computer system that is a thousand times faster than current world-leading supercomputers, may be the most daunting. Key scientific and technical obstacles associated with the architecture and energy efficiency of an exascale system must be overcome, and an immense amount of resources and effort will be required. As we head down this inevitable path to exascale computing, it is important we take time to plan and budget thoroughly to ensure a balanced approach that ensures broad buy-in from the scientific computing community. The Federal Government has limited resources and taxpayer funding must be spent on the most impactful projects. We need to ensure DOE efforts to develop an exascale system can be undertaken in concert with other foundational advanced scientific computing activities. This morning, we will hear testimony from expert witnesses regarding how best to achieve this balance. I would like to recognize if he is here, yes, he has come in, a leader in this effort, my colleague on the Energy Subcommittee, Representative Randy Hultgren. I would now like to yield the balance of my time to the gentleman from Illinois to summarize the discussion draft of his bill, ``American High-End Computing Leadership Act.'' [The prepared statement of Mrs. Lummis follows:] Prepared Statement of Subcommittee Chairman Cynthia Lummis Good morning and welcome to today's Energy Subcommittee hearing to examine high performance computing research and development challenges and opportunities. The development and expanded availability of supercomputers has enabled society to push the frontiers of nearly every scientific discipline, and accelerate applications of that science in countless fields. It has enabled modeling and simulation necessary to address national security needs. It drives the boundaries of medical research, reduces cost to develop new products, and improves materials design processes, just to name a few areas. High performance computing has also revolutionized how the energy sector operates. Advanced modeling and simulation techniques, driven by complex algorithms and faster computing speeds, improve the efficiency of energy production and consumption technologies. These advancements ultimately trace back to Federal investments in basic research that provided the foundation for most of today's computing technologies. From the first megaflop supercomputers of the 1960s, Federal investments have led the push across each landmark thousand-fold speed barrier-to gigaflops, teraflops, and petaflops. Throughout this computing age, we have witnessed as yesterday's supercomputers become today's desktop computers and consumer devices often in incredibly short time frames. The spillover benefits to society are countless and immeasurable. The Department of Energy, led by the Advanced Scientific Computing Research program, plays a unique and critical role in driving these computing technology breakthroughs. DOE supports world-class computational science facilities, such as the National Energy Research Scientific Computing Center. Additionally, DOE funds cutting edge applied mathematics research and next generation networking activities. DOE's next major computing challenge-constructing an ``exascale'' computer system that is a thousand times faster than current world- leading supercomputers-may be the most daunting. Key scientific and technical obstacles associated with the architecture and energy efficiency of an exascale system must be overcome, and an immense amount of resources and effort will be required. As we head down this inevitable path to exascale computing, it is important we take time to plan and budget thoroughly to ensure a balanced approach that ensures broad buy-in from the scientific computing community. The Federal government has limited resources and taxpayer funding must be spent on the most impactful projects. We need to ensure DOE efforts to develop an exascale system can be undertaken in concert with other foundational advanced scientific computing activities. This morning, we will hear testimony from expert witnesses regarding how best to achieve this balance. I would like to recognize a leader of this effort, my colleague on the Energy Subcommittee, Representative Randy Hultgren. I would now like to yield the balance of my time to the gentleman from Illinois to summarize the discussion draft of his bill, ``American High-End Computing Leadership Act.'' Mr. Hultgren. Thank you, Madam Chair, for holding this hearing today. Exascale computing represents a brave new world of science for our Nation. The application of the next generation of supercomputers is vast. A thousand-fold increase in processing power will give us the intense computing tools necessary to ensure our national security by better testing our nuclear stockpile, revolutionized our understanding and treatment of complicated healthcare problems like neurological diseases or the genetics underpinning cancer with the ability to model new treatments and ensure our Nation's competitiveness in the big data economy of the 21st century by spilling over knowledge and expertise into industry and academia. And while I can postulate further on some of the applied uses of faster machines, I also know that simply by making these investments in basic science needed to overcome challenges in the immensely massive parallelism, power management, new architecture, and programming models, we will enrich our Nation intellectually and ensure our labor force remains competitive. I think at that point I will yield back, Madam Chair. Let me follow up if I have another minute. Do I? Chairwoman Lummis. Mr. Hultgren, you do. Mr. Hultgren. Madam Chair, let me summarize my bill quickly. Thank you. My bill would amend the existing statute by specifying the need to target the specific challenges and power requirements and parallelism required to make the leap to exascale. It also will instruct the Secretary of Energy to conduct a coordinated research program to develop exascale computing systems and require an integrated strategy and program management plan to ensure the health of existing research activities is not harmed. The bottom line is we do not know all of the ways we will use this next-generation of supercomputers, but given the vast and unpredictable ways that computing technology has already enhanced every part of our lives and given the investments being made in other countries to deploy large-scale systems, it is more important than ever that we make this investment today. I look forward to hearing the witnesses, what they think of this legislative proposal, areas we can improve it, challenges that we will face. And with that, I do thank you. I apologize for my confusion here but I yield back to the Chairwoman. Thank you very much, Madam Chair. [The prepared statement of Mr. Hultgren follows:] Prepared Statement of Representative Randy Hultgren Thank you, Madam Chair, for holding this hearing today. Exascale computing represents an exciting new world of science for our nation. The applications for the next generation of super computers are vast. A thousand fold increase in processing power will give us the intense computing tools necessary to ensure our national security by better testing our nuclear stockpile; revolutionize our understanding and treatment of complicated health care problems like neurological diseases or the genetics underpinning cancer with the ability to model new treatments; and ensure our nation's competitiveness in the big data economy of the 21st century by spilling over knowledge and expertise into industry and academia. And while I can postulate further on some of the applied uses of faster machines; I also know that simply by making these investments in the basic science needed to overcome challenges in immensely massive parallelism, power management, new architectures and programming models, we will enrich our nation intellectually and ensure our labor force remains competitive. Madam Chair, my bill would amend the existing statute by specifying the need to target the specific challenges in power requirements and parallelism required to make the leap to exascale. It would also instruct the Secretary of Energy to conduct a coordinated research program to develop exascale computing systems, and require an integrated strategy and program management plan to ensure the health of existing research activities is not harmed. The bottom line is, we do not know all of the ways we will use the next generation of supercomputers, but given the vast and unpredictable ways that computing technology has already enhanced every part of our lives, and given the investments being made in other countries to deploy large scale systems, it is more important than ever that we make this investment today. I look forward to hearing what the witnesses think of this legislative proposal, areas we can improve it, challenges we face, and with that I thank you and I yield back. Chairwoman Lummis. The gentleman yields back. And I might add on a personal note, today, my daughter is being awarded her master's degree in digital media from Columbia University. I unfortunately cannot be at her graduation because Congress is in session but I get to watch it on the computer, so I will get to see it. And I think to myself, first of all, what is a master's degree in digital media? Somebody my age doesn't even know what that is. And certainly, when I was her age, I could not have even begun to envision the career that would be open to her as of today, and the career that is open to her as of today is due in part to the investment that the people in this room and that the American people have made in computing, for science, and for the benefit of mankind. So this is a very important subject. The fact that it is such an important subject leads me to let you all know that there will be several comings and goings by Committee Members this morning. There are concurrent meetings going on around the buildings. In my case, we have the IRS in front of us down in Oversight and Government Reform and I know there are other Members that may have to come and go from time to time. We deeply appreciate your testimony here today. In my absence, our Vice Chair Mr. Weber will be in the chair, and of course, Mr. Swalwell, who is our Ranking Member, who I will recognize now, the gentleman from California, Mr. Swalwell. Mr. Swalwell. Thank you, Chairman Lummis. And also congratulations to your daughter on this achievement. And thank you for holding this hearing today. And I want to thank the witnesses for being here. I also thank the witnesses who are not from the 15th Congressional District. We welcome you as well but especially welcome Ms. Crawford from Livermore, California. I am excited to learn more about the work that the DOE is doing in partnership with industry and our national laboratories, including both Lawrence Livermore and Berkeley national laboratories in particular and are carrying out to maintain the United States' leadership in the critical area of high-performance computing. As I am sure the witnesses will all describe in more detail, this capability enables our best and brightest minds to gain new insights into societal concerns ranging from Alzheimer's disease to climate change. Other examples of both industrial and academic research that benefit from our advanced, high-end computing capabilities include high- temperature superconductivity to significantly reduce energy losses in transmitting electricity; aerodynamic modeling for aircraft and vehicle design; pharmaceutical development; next- generation nuclear reactor design; fusion plasma modeling; and combustion simulation to guide the design of fuel-efficient clean engines such as work being carried out at the Sandia National Laboratory's combustion research facility. In short, many of the most pressing issues of our time, whether it is how we find our energy resources, how we make our energy resources more efficient, or how we solve the rising cost of healthcare can be solved through investments in high- performance computing. A focus of today's hearing is the development of an exascale computing capability. Now, my understanding is that exascale is often interchangeably used with extreme scale to refer to the next generation of supercomputers in general, but it also refers to a computing system that would be able to carry out a million trillion operations per second. Yes, a million trillion or a 1 with 18 zeros after it. That is about 500 times faster than the world's fastest computer today. Such a system would be critical to meeting the Nation's needs in a number of important research areas like combustion science, climate science, modeling of the human brain, and ensuring the reliability of our nuclear weapons stockpile. That said, as we pursue the next generation of supercomputing capabilities, which I fully support, I want to ensure that the Nation is getting the most bang for buck out of our current world-leading facilities. It is noteworthy that while Lawrence Livermore, Argonne, and Oak Ridge national laboratories are three of the most powerful supercomputing centers in the world, and they are addressing incredibly important scientific issues that really require their advanced computing capabilities. Lawrence Berkeley's National Energy Research Scientific Computing Center actually serves thousands more users with only a fraction of those leadership machines' computing power. The point is not every computational research effort requires the fastest most sophisticated system we can possibly build and I think we also need to work more to make sure that what is sometimes called capacity supercomputing is more accessible to both the academic and industrial research communities that could benefit. I have always believed whether it was as a local city councilman or a sitting Member of Congress that the government works best when we can share our resources with the private sector. It doesn't serve anyone any good if we are just doing the research in the government and not transferring that research out to the private sector, and I think in high- performance computing we have already shown in our laboratories we are transferring it out. The transfer out makes us more efficient, can reduce healthcare costs, and also more importantly, especially in our area, it can create private- sector jobs on top of the thousands of jobs that already exist at our laboratories. So with that, I look forward to discussing these important issues with each of you today and I yield back the balance of my time. [The prepared statement of Mr. Swalwell follows:] Prepared Statement of Subcommittee Ranking Member Eric Swalwell Thank you Chairman Lummis for holding this hearing today, and I also want to thank the witnesses for being here--even the ones from outside of the 15th District of California! I am excited to learn more about the great work that the Department of Energy in partnership with industry and our national laboratories, including both Lawrence Livermore and Lawrence Berkeley National Laboratories in particular, are carrying out to maintain and advance U.S. leadership in the critical area of high performance computing. As I'm sure the witnesses will describe in more detail, this capability enables our best and brightest scientists to gain new insights into societal concerns ranging from Alzheimer's disease to climate change. Other examples of both industrial and academic research that benefit from our advanced high-end computing capabilities include: high temperature superconductivity to significantly reduce energy losses in transmitting electricity; aerodynamic modeling for aircraft and vehicle design; pharmaceutical development; next generation nuclear reactor design; fusion plasma modeling; and combustion simulation to guide the design of fuel-efficient clean engines, such as work being carried out at the Sandia National Laboratories' Combustion Research Facility. A focus of today's hearing is the development of an exascale computing capability. Now, my understanding is that ``exascale'' is often used interchangeably with ``extreme scale'' to refer to the next generation of supercomputers in general, but it also refers to a computing system that would be able to carry out a million trillion operations per second. (Yes, a million trillion, or a 1 with 18 zeros after it.) That's about 500 times faster than the world's fastest computers at today. Such a system would be critical to meeting that nation's needs in a number of important research areas like combustion science, climate science, modeling of the human brain, and ensuring the reliability of our nuclear weapons stockpile. That said, as we pursue the next generation of supercomputing capabilities-which I fully support-I also want to ensure that the nation is getting the most bang per buck out of our current world- leading facilities. It is noteworthy that while Lawrence Livermore, Argonne, and Oak Ridge National Laboratories are 3 of the most powerful supercomputers in the world, and they are addressing incredibly important scientific issues that really require their advanced computing capabilities, Lawrence Berkeley's National Energy Research Scientific Computing Center actually serves thousands of more users with only a fraction of those leadership machines' computing power. The point is, not every computational research effort requires the fastest, most sophisticated system we can possibly build, and I think we also need to do more to make what's sometimes called ``capacity'' supercomputing more accessible to both the academic and industrial research communities that could benefit. With that, I look forward to discussing these important issues with each of you today, and I yield back the balance of my time. Chairwoman Lummis. Thank you, Mr. Swalwell. If there are Members who wish to submit additional opening statements, your statements will be added to the record at this point. Well, at this time I would like to introduce our witnesses, and the fun part today is we have two Members here who have witnesses from their districts. So I will start by introducing Dr. Roscoe Giles, Chairman of the Advanced Scientific Computing Advisory Committee of the Department of Energy and Professor at Boston University. Dr. Giles--and I have that right, don't I, Dr. Giles? Thank you. He has served in a number of leadership roles in the community including Member of the Board of Associated Universities, Inc., Chair of the Boston University Faculty Council, and General Chair of the SC Conference in 2002. He received his Ph.D. in physics from Stanford University in 1975. That is a remarkable record of achievement, Dr. Giles. Thank you for being here. At this time, I would like to yield to the gentleman from Illinois, Mr. Hultgren, to introduce our second witness. Mr. Hultgren. Thank you, Madam Chair. Our second witness is Dr. Rick Stevens, Associate Laboratory Director for Computing, Environment, and Life Sciences at Argonne National Laboratory. He heads Argonne's Computational Genomics Program and co-leads the DOE's laboratory planning effort for exascale computing research. He is also Professor of computer science at the University of Chicago and is involved in several interdisciplinary studies at the Argonne University of Chicago Computation Institute and at the Argonne University of Chicago Institute for Genomics and Systems Biology. He is doing amazing work at Argonne and at the University and the entire Illinois community is proud of his contributions to this cutting edge field of science. We are very glad to have you here, Dr. Stevens. Thank you. I yield back. Chairwoman Lummis. Thank you for your attendance today. That was my field although at a much lower level of academic achievement, Dr. Stevens. We are delighted you are here. Now, I would like to yield to the gentleman from California, Mr. Swalwell, to introduce our third witness. Mr. Swalwell. Thank you, Chairman Lummis. And I have been very eager on this Committee to have a witness from Lawrence Livermore laboratory. Chairwoman Lummis. I can testify to that. Mr. Swalwell. I thank you for allowing this witness to be here today. Lawrence Livermore is the largest employer in my Congressional District and I have to really just commend the laboratory for their advocacy of the issues facing Lawrence Livermore. They are in constant contact with our office and this Committee so I am honored to today introduced Dona Crawford, who is the Associate Director of Computation at Lawrence Livermore National Laboratory. Ms. Crawford is responsible for a staff of roughly 900 to develop and deploy an integrated computing environment for advanced simulations of complex physical phenomena like climate change, clean energy creation, biodefense, and nonproliferation. She has served on Advisory Committees for the National Academies and the National Science Foundation and currently serves as co-Chair of the Council on Competitiveness High-Performance Computing Advisory Committee, and is a member of IBM's Deep Computing Institute External Advisory Board. Ms. Crawford has a master's degree in operations research from Stanford University and a bachelor's degree in mathematics from the University of Redlands, California. Ms. Crawford, thank you for being here today and I yield back the balance of my time. Chairwoman Lummis. Thank you, Mr. Swalwell. And my first exposure to Livermore, I used to walk around the lab. My first job out of college was working for a rodeo company in Northern California, and we were putting on the rodeo at Livermore. Mr. Swalwell. It is the fastest rodeo in the world. Did you know that? Chairwoman Lummis. You know, considering the rodeo company I worked for, I would believe that. Those rodeos ran like that and I used to go for walks around the lab just to get some exercise when I was there at Livermore putting on rodeos. So I know where you are, at least I knew where you are when I was a young college graduate in my first job. Our final witness is Dr. Daniel Reed, Vice President of Research and Economic Development at the University of Iowa. Previously, he served as a Senior Leader at Microsoft serving as Microsoft's Computing Strategist to Corporate Vice President for Extreme Computing, I love that, and Technology Policy. He received his Ph.D. in computer science in 1983 from Purdue University. As our witnesses should know, spoken testimony is limited to five minutes each, after which the Members of the Committee will have five minutes each to ask questions. I now recognize Dr. Giles for five minutes to present his testimony with deep gratitude to all of you for your attendance today. Dr. Giles. TESTIMONY OF DR. ROSCOE GILES, CHAIRMAN, ADVANCED SCIENTIFIC COMPUTING ADVISORY COMMITTEE Dr. Giles. Yes, thank you, Chairman Lummis. And thanks to Members of the Committee for inviting me to testify today. I think the bill you are considering is very, very important for our field and for maintaining the Nation's leadership in computing and computational science. I am testifying today in my role as Chair of the Advisory Committee to ASCR and I will try to reflect that committee's views of some elements of the ASCR program and hope to demonstrate that we are ready to move forward and sort of eager to move forward in this direction. And it is important that we do so. The Office of Advanced Scientific Computing Research has programs and investments that include computer and networking facilities that support DOE's science programs; leadership computing facilities for which the exascale discussion is very directly relevant with unique high-end capabilities made available to DOE and to all the Nation, including industry; applied mathematics research whose results provide the framework for future applications and systems; computer science system and software research, whose results both enable applications of current systems and chart the direction for future systems. And beyond this, ASCR investments--ASCR is the abbreviation for Advanced Scientific Computer Research--we get lost in acronyms sometimes. ASCR investments have also built human expertise in the scientific and technical staff at the labs and through attention to integrating the next generation of computational science leaders into DOE programs and facilities through programs like the Computational Science Graduate Fellows Program, which I also am involved with. It is hard in these few minutes to state the breadth and depth of science productivity that is being enabled by these machines. We now see the initial results of the petascale era. As one measure, we might mention that more than 2,000 peer- reviewed research articles based directly on projects supported by ASCR computing facilities were published in 2012 alone. One I love is a trillion-particle simulation in cosmology, since I started out in the '80s struggling to do a million-particle simulation of molecular dynamics, and to go another factor of a million is astonishing. In 2009 our advisory committee was charged with reviewing ASCR's body of work on exascale computing. We delivered the Rosner report, ``The Opportunities and Challenges of Exascale Computing,'' in fall of 2010. We found the case for exascale computing compelling and recommended the DOE should proceed expeditiously with an exascale initiative so that it continues to lead in using extreme scale computing to meet important national needs. As you have heard mentioned this morning, when we wrote that, we were talking about growing a factor of 1,000 forward in the future. Now, that is a factor of 500. I am glad to see that we are starting to in this bill to really move forward on this. And we have had a sense in the committee that we have been waiting for that forward motion from the system. Some of the--during this time, ASCR has been busy doing foundational research to make this possible, so there--and we will hear more about it, I am sure, from other speakers. But the establishment of co-design centers, computing research, and applied mathematics research and some prototype projects with fast forward and design forward that are bringing us in this direction, and I think we are making progress but not the progress we should be making at the scale we should be making it, and hopefully, the bill will help deal with that issue. Our committee has been asked to review ASCR facility plans for the relatively short-term future of the next ten years, not including exascale deployment, and we found those facility plans to be very sound and compelling that involve enhancements to the petascale systems. We have also recently examined the intersection of big data needs within the Department of Energy and ASCR's exascale program and found them quite convergent. The exascale technologies we are talking about developing will be essential in systems that analyze big data problems of the nature that come to the Department of Energy from both experiment and theory, and we have a--quite a long and detailed report about that. I wanted to just summarize by saying I am very, very glad to see the legislation that we have here. I am very supportive of the direction we are going. I would only ask that the funding level be sure to be sufficient for the scope of our dreams. Thank you. [The prepared statement of Dr. Giles follows:] [GRAPHIC] [TIFF OMITTED] 81195.008 [GRAPHIC] [TIFF OMITTED] 81195.009 [GRAPHIC] [TIFF OMITTED] 81195.010 [GRAPHIC] [TIFF OMITTED] 81195.011 [GRAPHIC] [TIFF OMITTED] 81195.012 [GRAPHIC] [TIFF OMITTED] 81195.013 [GRAPHIC] [TIFF OMITTED] 81195.014 [GRAPHIC] [TIFF OMITTED] 81195.015 [GRAPHIC] [TIFF OMITTED] 81195.016 [GRAPHIC] [TIFF OMITTED] 81195.017 [GRAPHIC] [TIFF OMITTED] 81195.018 [GRAPHIC] [TIFF OMITTED] 81195.019 [GRAPHIC] [TIFF OMITTED] 81195.020 [GRAPHIC] [TIFF OMITTED] 81195.021 Mr. Weber. [Presiding] Thank you, Dr. Giles. Now, I recognize Dr. Stevens to present his testimony. Turn your mike on. TESTIMONY OF DR. RICK STEVENS, ASSOCIATE LABORATORY DIRECTOR FOR COMPUTING, ENVIRONMENT AND LIFE SCIENCES, ARGONNE NATIONAL LABORATORY Dr. Stevens. Oh, thanks. Thank you. Madam Chair, Ranking Member Swalwell, Members of the Subcommittee, I appreciate this opportunity to talk to you about the future of high-performance computing research and development and about the importance of U.S. leadership in the development and deployment of exascale computing. In my own work at Argonne and the University of Chicago I split my time between trying to advance high-performance computing architectures and systems and doing research on how to do computational genomics in the pursuit of problems in energy, the environment, and infectious disease. And those projects have given me insight not only on the underlying technology but on the impact of applications. I believe that advancing American leadership in high- performance computing is vital to our national interest. High- performance computing is a critical technology for the Nation, and it is also the underlying foundation for advancing progress in modeling and simulation and big data. It serves both of these needs. It is also needed by all branches of science and engineering for forward progress. It is used more and more by U.S. industry to maintain a competitive edge in the development of new products and services, and it is emerging as a critical policy tool for government leaders who can rely on simulations to add insight to policy or technical decisions. Today, the United States is the undisputed leader in the development and use of high-performance computing technologies. However, other nations are increasingly aware of the strategic importance of HPC and are creating supercomputing research programs that challenge our leadership. Japan has significant programs for over a decade in this area. They have fielded large-scale machines that are comparable to the machines in the United States. But China is emerging as a serious player as well and Europe has been investing in revitalization of their own high-performance computing sector. So we now have at least three sectors on the planet besides the United States making serious progress. All have set their sights on the development of machines that are 1,000 times faster than those most powerful machines today. Everyone is looking at exascale. And achieving this goal is important. The drive to exascale will have a sustained impact on American competitiveness. It gives companies and researchers the means and the impetus for developing new processes, new services, and new products. For example, we need increased compute power to enable first principle simulations of materials for energy storage that would give us access to a potential 500-mile battery pack for electric cars. We want to build end-to-end simulations of advanced nuclear reactors that are modular, safe, and affordable. We want to revolutionize small business manufacturing and digital fabrication and put in place a digital supply chain that would potentially revolutionize the economy in the United States. We want to model controls for power grids that have significant amount of renewable energy, and we want to increase the resolution of climate models to provide more details on regional impacts. And finally, we want to create a personalized medicine that can incorporate an individual's genomic information into a specific customized plan for prevention or treatment of disease. All of these challenges require machines that are thousands of times faster than the current machines. The development of practical exascale system, however, will also mean affordable petascale systems and broad deployment, broad accessibility. The DOE Office of Science supercomputer centers at Argonne, Berkeley, and Oak Ridge are currently oversubscribed by at least a factor of three. This means that not all of the science that we could be doing on these machines is getting done. With current funding levels, these systems can only be upgraded about once every four to five years. And at current research-- at current levels of research investment, the U.S. vendors are not likely to reach an exascale performance level that we can afford to deploy until considerably after 2020. This is a problem for us if we want to maintain our leadership. Both China and Japan are working on plans to reach the level by 2020 or before. Japan is building a $1.1 billion investment program aiming to deploy exascale machines by 2020, and China has announced a goal to reach exascale before 2020. China is aggressively spending on infrastructure for supercomputing and succeeding in deploying large-scale systems rivaling the largest systems deployed in the United States. It is widely expected they will regain lead on this capability this year, although their designs are mostly based on incorporating U.S. components. In the future, they plan to deploy systems based on Chinese components. I have been working since 2007 building a plan with my colleagues at the laboratories, academia, and DOE, and we identified five hurdles that we must cross in order to reach exascale. We have to reduce systems powered by a factor of 50; we must improve memory performance and cost by a factor of 100; we must improve our ability to program these systems; we must increase the parallelism in our applications; and we must improve reliability. These are not simple tasks but these are very important if we are to reach this goal. And I believe we have a duty to move as swiftly as we can on this objective. Thank you. I would be more than happy to answer your questions. Thank you. [The prepared statement of Dr. Stevens follows:] [GRAPHIC] [TIFF OMITTED] 81195.022 [GRAPHIC] [TIFF OMITTED] 81195.023 [GRAPHIC] [TIFF OMITTED] 81195.024 [GRAPHIC] [TIFF OMITTED] 81195.025 [GRAPHIC] [TIFF OMITTED] 81195.026 [GRAPHIC] [TIFF OMITTED] 81195.027 [GRAPHIC] [TIFF OMITTED] 81195.028 [GRAPHIC] [TIFF OMITTED] 81195.029 [GRAPHIC] [TIFF OMITTED] 81195.030 [GRAPHIC] [TIFF OMITTED] 81195.031 [GRAPHIC] [TIFF OMITTED] 81195.032 [GRAPHIC] [TIFF OMITTED] 81195.033 Mr. Weber. Thank you, Dr. Stevens. I recognize Ms. Crawford for her testimony. TESTIMONY OF MS. DONA CRAWFORD, ASSOCIATE DIRECTOR FOR COMPUTATION, LAWRENCE LIVERMORE NATIONAL LABORATORY Ms. Crawford. I thank you. I thank Chair Lummis and I thank you, Mr. Vice Chairman Weber and Ranking Member Swalwell, for inviting me to be here today. I ask that my full statement as submitted to the Committee made part of the hearing record, and if I may, I will summarize. Mr. Weber. Without objection, so ordered. Ms. Crawford. I am Dona Crawford, Associate Director of Computation at Lawrence Livermore National Laboratory. I will shorten that by saying LLNL or Livermore. Livermore is a national security laboratory of the National Nuclear Security Administration of the Department of Energy and home to Sequoia, one of the fastest computers in the world. Livermore has the responsibility for maintaining the safety, security, and effectiveness of the Nation's strategic nuclear deterrent through the Stockpile Stewardship Program. High-performance computing has been a core competency of the lab to meet this mission need since over 60 years. In fact, the NNSA labs, working in close partnership with U.S. HPC industry, were at the forefront of the last revolutionary design shift in HPC computer architectures and applications development. That is the foundation of today's HPC systems. Over the past 20 years, the NNSA labs learned many valuable lessons, including how to best structure R&D efforts to develop computing architectures that meet our demanding mission requirements while cost-effectively leveraging market-driven technology within industry. These lessons are very valuable in our efforts to develop exascale computing. I applaud the Committee for its determined efforts to sustain U.S. leadership in this vitally important and increasingly competitive arena of high-performance computing. It is imperative that the United States embark on an R&D program to develop new technologies and computer architectures to support exascale computing. My main point of emphasis today is straightforward. This pursuit must be a joint Office of Science/NNSA effort working in tandem through partnership with U.S. HPC industry to ensure system architectures that meet Office of Science and NNSA mission requirements. Working together, the Office of Science and NNSA have combined scarce resources and have already initiated a number of R&D efforts and contracts with industrial partners but lack the resources to invest at the magnitude necessary to assure success over the next decade. Due to the technically challenging nature of developing exascale supporting technologies in computing capability, it is vitally important to ensure there are competitive teams each with Office of Science and NNSA laboratories partnered with U.S. HPC industry collaborators. Equally important is the development of an integrated strategy and program management plan. Current U.S. leadership in HPC is a direct result of the Nation's investment in computational capability to support the Stockpile Stewardship Program. U.S. HPC investment has provided significant computing capability to maintain the U.S. nuclear deterrent and this computing capability enables us to simulate in 2-D at high resolution and high physics fidelity or simulate in 3-D at low resolution. Today, we cannot simulate in 3-D at high resolution and high physics fidelity which will be required for the stockpile mission needs. Therefore, a new architecture enabling exascale computing is required for the NNSA mission. This will not be easy. Development of exascale-class systems cannot be achieved through a straightforward refinement of today's technologies. Surmounting multiple technical issues will require sustained research and development effort. But there is no doubt exascale computing will yield valuable benefits to near-term mission requirements, as well as to U.S. economic competitiveness. Over the last two decades, supercomputers have transformed the way the world conducts scientific research and has enabled discovery and development across a broad set of disciplines. In a 2008 U.S. Council on Competitiveness report, the Council states, ``supercomputing is part of the corporate arsenal to beat rivals by staying one step ahead of the innovation curve. It allows companies to design products and analyze data in ways once unimaginable.'' In one example, Livermore is leveraging its HPC capabilities in the California Energy Systems for the 21st Century Initiative. The California Public Utilities Commission and state investor-owned utilities are collaborating with Livermore to improve and expand energy systems to meet our future energy needs. The owners, operators, regulators, and a joint team of technical experts will use the Nation's most advanced modeling simulation and analytical tools to gain unprecedented insight and generate new data to reduce risk and inform solutions to issues facing 21st-century energy systems such as renewable energy integration and use of smart grid technology. There are many other examples that highlight the importance of supercomputing and reinforce the value of maintaining U.S. HPC leadership. For now, let me close again by saying thank you and I look forward to working with the Committee to ensure continued U.S. HPC leadership. [The prepared statement of Ms. Crawford follows:] [GRAPHIC] [TIFF OMITTED] 81195.034 [GRAPHIC] [TIFF OMITTED] 81195.035 [GRAPHIC] [TIFF OMITTED] 81195.036 [GRAPHIC] [TIFF OMITTED] 81195.037 [GRAPHIC] [TIFF OMITTED] 81195.038 [GRAPHIC] [TIFF OMITTED] 81195.039 [GRAPHIC] [TIFF OMITTED] 81195.040 [GRAPHIC] [TIFF OMITTED] 81195.041 [GRAPHIC] [TIFF OMITTED] 81195.042 [GRAPHIC] [TIFF OMITTED] 81195.043 [GRAPHIC] [TIFF OMITTED] 81195.044 [GRAPHIC] [TIFF OMITTED] 81195.045 Mr. Weber. Thank you, Ms. Crawford. Dr. Reed, I recognize you for your testimony. TESTIMONY OF DR. DANIEL REED, VICE PRESIDENT FOR RESEARCH AND ECONOMIC DEVELOPMENT, UNIVERSITY OF IOWA Dr. Reed. Thank you. Chair Lummis, Vice Chair Weber, Ranking Member Swalwell, Members of the Subcommittee, my name is Dan Reed and I am the Vice President for Research and Economic Development at the University of Iowa. Thank you for the opportunity to share my perspectives on exascale computing and to respond to your questions regarding the American High- End Computing Leadership Act. Today, I would like to make four points regarding the exascale and high-performance computing program followed by a set of specific recommendations for the future. They are drawn from my nearly 30 years of experience in high-performance computing as a researcher, as an academic and corporate leader, as a Director of the National Science Foundation Supercomputing Center, and as a participant in national science and technology policy. First of all, as others have noted, high-performance computing is unique among scientific instruments. It is distinguished by its universality as an intellectual amplifier. New, more powerful supercomputers and computational models yield insights across all scientific and engineering disciplines. Advanced computing is also essential for analyzing the torrent of experimental data produced by scientific instruments and sensors, but it is about more than science. With advanced computing, real-time data fusion, and powerful numerical models, we have the potential to predict the tracks of devastating tornadoes such as the recent one in Oklahoma, saving lives, and ensuring the future of our children. My second point is that we face an uncertain future of computing and in particular high-performance computing leadership in this country. As others have noted, today, HPC systems from Oak Ridge, Lawrence Livermore, and Argonne National Laboratories occupy the first, second, and fourth place on the list of the world's fastest computers. From this, one might surmise that all is well, yet U.S. leadership in both deployed HPC capability and in the technologies needed to create future systems is under challenge. Also, as others have noted, other nations are investing strategically in high-performance computing to advance national priorities. And the U.S. research community has repeatedly warned of the potential and actuality of eroding U.S. leadership in computing and in high-performance computing and emphasized the need for sustained and strategic investment. I have had the privilege of chairing many of those studies personally as a member PITAC, of PCAST, of National Academies' boards, and yet many of these warnings have been largely unheeded. This brings me to my third point: the deep interdependence a basic research of vibrant U.S. computing industry and high- performance computing capability. It has long been axiomatic that the United States is the world leader in information technology. Our global leadership is not a birthright. As Andy Grove, the former CEO of Intel, noted in his famous aphorism ``only the paranoid survive.'' U.S. leadership has been repeatedly earned and hard-fought based on continued Federal Government commitment to basic research, translation of that research into technological innovations, and the creation of new products by vibrant U.S. industry. This brings me to my fourth point. Computing is in deep transition to a new era with profound implications for the future of U.S. industry and HPC. My colleague Mr. Stevens touched on many of the issues around energy management and low- power devices and they are key to this topic. U.S. consumers and businesses are an increasingly small minority of the global market for mobile devices and for cloud services. We live in a post-PC world, as we all know, where U.S. companies compete in a global device ecosystem. Unless we are vigilant, these economic and technical changes could further shift the center of enabling technology R&D away from the United States with profound implications for our future HPC capability. Given this, what are my recommendations for the future? First and most importantly, we need to change our model for HPC research and deployment if the United States is to maintain its leadership. This must include deep and sustained interagency collaborations defined by a regularly updated strategic R&D plan and associated, verifiable metrics, commensurate budget allocations, and accountability to realize the plan's goals. DOE's partners--it needs the National Science Foundation, the Department of Defense, NIST and NIH, and other agencies to fulfill their important and complementary roles to DOE as engaged partners and supporters of basic research in technology development. We also need long-term industry engagement. Second, advanced HPC deployments are crucial, but the computing R&D journey is as important as any single system deployment. A vibrant U.S. ecosystem of talented and trained people and technical innovation is the true lifeblood of sustainable exascale computing. Finally, we must balance and embrace dual-use technology R&D supporting both high-performance computing and ensuring U.S. industry competitiveness. Neither HPC nor big data R&D can be sacrificed to advance the other, nor can hardware R&D dominate investments in algorithms, software, applications, and people. All are crucial. Finally, let me again commend this Committee for its continued commitment to high-performance computing. It has been my privilege to testify here many times. I appreciate the support of the Committee. And like my colleagues, I would be delighted to answer questions at the appropriate time. [The prepared statement of Dr. Reed follows:] [GRAPHIC] [TIFF OMITTED] 81195.046 [GRAPHIC] [TIFF OMITTED] 81195.047 [GRAPHIC] [TIFF OMITTED] 81195.048 [GRAPHIC] [TIFF OMITTED] 81195.049 [GRAPHIC] [TIFF OMITTED] 81195.050 [GRAPHIC] [TIFF OMITTED] 81195.051 [GRAPHIC] [TIFF OMITTED] 81195.052 [GRAPHIC] [TIFF OMITTED] 81195.053 [GRAPHIC] [TIFF OMITTED] 81195.054 Mr. Weber. Thank you, Dr. Reed. And thank you all for your testimony. Man, lots of questions come to mind. And, you know, I guess I am an old-timer. I grew up back in the '60s and we didn't have computers, actually, we did. There was a flat table, you put a quarter in, and you chased a little Pac-Man around. Those were our computers. So I have a question here, and I think you kind of alluded to it, Dr. Reed, but I will ask this maybe starting with Dr. Giles. Is it Giles? Dr. Giles. Giles actually. Mr. Weber. Giles, there you go. Thank you. In December 2011 Congress directed DOE to provide a strategic roadmap relating to the development of an exascale computing system. However, it is my understanding that after 15 months of the mandated completion date, the report is not yet finalized. Are you aware of this report? Dr. Giles. I am aware of it but my position is as an external representative of the community relative to ASCR so I am actually not an insider and I have not seen the report. Mr. Weber. Okay. Dr. Giles. My understanding is exactly what you said. Mr. Weber. Okay. Dr. Giles. But I don't have anything unfortunately to add to that. Mr. Weber. Nothing that you want to admit here publicly? Dr. Giles. No, actually nothing that I know. Anything I know, I will tell you. Mr. Weber. Of course, our goal is to get it. Dr. Stevens, how about you? Dr. Stevens. Well, I am aware of the report. I think it is a fine plan. I think that the internal process of getting that report out is what has blocked it, and I hope it reaches you quickly. Mr. Weber. Okay. Any help you can give us in that endeavor? Dr. Stevens. I don't have any specific recommendation except to just reemphasize that this is a critical plan that must be delivered and must be understood and articulated and executed. Mr. Weber. Okay. Thank you. And Ms. Crawford, I don't mean to put you on the hot seat, but you are on the hot seat. Ms. Crawford. I have nothing to add to what Dr. Stevens said. We are--we work at the laboratories and we are not part of the formal process between the DOE and the OMB to get that report out. I do support what is written in the report. The labs had a lot of input. I have not seen the final report. Mr. Weber. And Dr. Reed, since you came to us with four points followed by recommendations, and I love that by the way. One of the things you said in your recommendation was the Department of Energy needs partners and long-term industry engagement. How do we expedite this? How do we make this happen? Dr. Reed. Well, I think there are several points relevant. One is to recognize that, as I said, it is a false dichotomy to pit investment and some of these big data issues against high- performance computing, and I think frankly that is the root of some of the issues that we have to resolve in terms of moving forward. In terms of the agencies, I believe, as I pointed out in my written testimony, that they each fulfill and historically have fulfilled important and complementary roles. The Department of Energy has been crucial in terms of advanced prototyping and deploying of the largest scale systems. The other agencies, though, provide support for enabling technology research. The National Science Foundation is one of the key enablers of that long-term research. What is important is that all those players be at the table and be engaged in supporting this integrated agenda. I think from the industry's perspective to sort of answer your specific question, that is where industry--and I speak now again from my industry experience--it is important that the government be a committed and not fickle partner because the cost of money and the time planning for companies to execute is really crucial. And as I was saying, that combination is key to the future of the U.S. industry not just for high-performance computing but for how much information technology means to the U.S. economy. Mr. Weber. All right. Thank you. And Ms. Crawford, let me come back to you. I think you said that this exascale computing either can't or won't be achieved through refinement. What did you mean by that? Ms. Crawford. What I mean by that is the current system architectures today can't simply be scaled up to produce a usable and cost-effective system. In principle, one could scale it up and you would have a system that would fill the room and would take 100 megawatts of power, so that is not a cost- effective system. So the technologies have to change and we have to change in memory, in processors, in storage and networking and the programming models. And so that is what I mean by we can't simply scale-up the programs of today. Mr. Weber. Let me send that over to Dr. Stevens. And you mentioned about more or less power, I guess explain, you said less power. Dr. Stevens. Right. We need to develop processors and memory and network components that consume considerably less power than current systems in order to scale-up. Right now, if we took a current kind of 10 petaflop system and scaled it up to an exascale, it would consume nearly a gigawatt of power, which is not feasible from a physical infrastructure standpoint or a---- Mr. Weber. Right. Dr. Stevens. --cost standpoint. So we need much more power- efficient devices. We also need better programming models because we are going to have to have a lot more parallelism inside these machines, 1 million--or 1,000 times more parallelism than we have now and we need ways of accessing that parallelism easily for programmers. So we have a lot of work to do. We know what to do. The DOE's plan includes all of these activities so it is--I think the United States has a good position to do this; we just need the resources and the long- term commitment. Mr. Weber. All right. Thank you. And I just want to make an observation before I yield to the Ranking Member and that is that I am glad to hear you say that national security is involved in this and tied up in this. That is very crucially important. And I think it will carry a lot of weight with Congress. Hopefully, it will. So I thank you for your testimony. And with that, I yield to Mr. Swalwell. Mr. Swalwell. Thank you, Mr. Chair. And my questions will principally be for Ms. Crawford. First, does research in high-performance computing require the United States Government to make investments? And what I mean is why can't we simply rely on the private sector to innovate and invent the next supercomputing architecture and software and then the government can just buy off-the-shelf technology? Ms. Crawford. The short answer is, yes, the U.S. Government does need to invest in order to shape the exascale architectures for our mission needs. I can use an old example. When we started the Accelerated Strategic Computing Initiative in the mid-1990s for the Stockpile Stewardship Program, industry and the consumer base was driving computing in a direction that would not meet our needs. And without our investment and our sustained investment and focused on cooperation and developing those processors that would meet our needs, we wouldn't have had the computers and the computing capability that we have today. And so today, it is essential that we work together with the Office of Science laboratories and the NNSA laboratories to meet this mission needs. A shorter answer perhaps is that we are going to follow industry technologies. We can't afford our own, you know, brand-new fab or our brand-new machines. What we want to do is pay on the margins to make those machines viable for our particular applications, which is mimicking the, you know, physical phenomena around us. Mr. Swalwell. And when we look at our global competitors-- Japan, China, India, Brazil, Russia--are they allowing or relying solely upon the private market or are they also having government investment at the table as well? Ms. Crawford. There is strong government investment in Japan, China, Russia, the European Union. It is about $1.1 billion of investment in Japan. I would have to do the translation but the Ministry of Science and Technology five- year plan within China is investing and again not just in the hardware technologies but they are investing in the low-level software and the applications and making sure that they have the ecosystem in order to be able to deploy these systems effectively to make a difference to their underlying national security and economic competitiveness. So---- Mr. Swalwell. So it sounds like---- Ms. Crawford. --they are going to be large investments. Mr. Swalwell. It sounds like for the United States to keep its edge in high-performance computing, we will need to continue to have the Federal Government make investments in these programs, is that right? Ms. Crawford. Absolutely. Mr. Swalwell. You talked a little bit about the joint partnership that must take place between NNSA and the Office of Science. Why is exascale capability so critical to DOE's National Nuclear Security Administration? Ms. Crawford. So then I will take a more focused view on just what is going on within the NNSA laboratories. It is our duty to assess the state of the stockpile on an annual basis, and the stockpile is being decreased in the numbers of weapons and the types of weapons. That makes each single weapon remaining in the stockpile critically important to understand what is going on---- Mr. Swalwell. Going toward a more leaner and meaner model, right? Ms. Crawford. Leaner and meaner, and so those systems, as they age, they are being modernized as parts begin to fail, and so there are a number of things that we need to understand, you know, physically. You know, nuclear weapons are very complex. Think about parking your car in the garage and not turning it on but then wanting to be able to use it when you have to. You know, there are special materials that are changing over time and all kinds of things that go on just sitting there. We need high fidelity 3-D simulations to understand, you know, the initial conditions, the engineering features, safety features, the security features, and today, we cannot simulate at that high fidelity. So we have a number of--what we do is look at the kinds of calculations we are going to do and the kinds of computing that is required to do those calculations and so--for stockpile assessment, for the life extension programs for materials aging, for safety and surety, we have a range of exascale needs for the kinds of calculations that will have to go on in high fidelity, high-resolution 3-D, and they range from half-an-exascale to 1,000 exascales over the period of the next 10 years. Mr. Swalwell. And Ms. Crawford---- Ms. Crawford. Starting in about 2018. Mr. Swalwell. Can you tell me more about Livermore's work to address industrial and medical research needs, for example, your groundbreaking simulation of the human heart and your recent work with the California Energy Commission to improve energy management throughout the State and how exascale and HPC have affected our ability to do this? Ms. Crawford. I would be glad to. Having developed these capabilities for our mission drivers, then they are applicable, as Dr. Reed has said, to many other activities. Last year, we worked with IBM to develop a code called cardioid and it does-- it models the electrical signals of the heart and it has the potential to be used to test drugs or medical devices, the code ran in nearly real-time across our 20 petaflop machine at Livermore beating an astonishing 60 beats per minute, so this is almost, you know, 12 percent of real-time. This calculation ran at 59 percent of peak of this machine, and that is--you know, it is very incredible and amazing thing to take a new code and put it on a new machine and run at this scale. It runs in a time to solution over 1,200 times faster than the previous state-of-the-art and this work shows promise for what advanced computing can do for understanding the human body. But it also demonstrated the extreme level of specificity and technical acuity required to achieve this result. And of course, these insights that we gain there will then be applied back to the stockpile. Mr. Swalwell. Great. Well, thank you so much, Ms. Crawford. And thank you again to our other witnesses. Thank you, Mr. Chair. Mr. Hultgren. [Presiding] Thank you. And I will recognize myself for five minutes for a few questions. Part of our challenge as a Subcommittee is certainly to understand the right thing to do but also to present it to the larger Committee and even beyond that to Members of Congress, so a couple of questions. Just if you have been messaging or how to present how important this is and why this is so important so I would address this first question to Dr. Stevens and Ms. Crawford. Wonder if you could just discuss the expected breadth of applications for the exascale computing. Is this something that could be used for a wide range of important disciplines from material to chemistry to medicine to nuclear science similar to the current supercomputers or is the expected range of disciplines more narrow such as climate science modeling or for weapons development? Dr. Stevens. So the range of applications for exascale are no less broad than the current machines. In fact, there are many problems that haven't been tried in the past, particularly in biomedical science where we were just afraid to try them. We didn't have enough compute power. This idea of trying to build, say, detailed models in the human body, not just the heart but now include the lungs, include the nervous system, include the gastrointestinal system and build a virtual human, that is a problem that will require 1,000 times current machines. It is not really feasible so people haven't tried it. So my sense is that we will find more and more applications as we build more capable systems. We are also going to increase the ability for these systems to deal with data, and so a new class of applications that is emerging in both national security and in engineering research is this idea of doing modeling simulation with uncertainty quantification, this idea that not only will you get a result, you will get some confidence measure on that result. And that is something that requires hundreds of times more compute power than the current capability which means you can only do one simulation. Ms. Crawford. I second everything that Dr. Stevens said. And it is limitless. Computing is so foundational. Anything that--any physical process that you can represent mathematically, which are most of them, you can then model in the computer with great fidelity. And the greater the fidelity we have, the better we can understand the world around us. And so I can just go on and on and on but, you know, we work at our laboratory in a number of areas with industry, with other national laboratories, with academia to make sure that we are applying these to the breadth of possibilities. Mr. Hultgren. Well, Ms. Crawford, if I can get into just a little bit more specific and really following up on the Ranking Member's discussion and also on the Vice Chairman's of what does speak to Members of Congress and inspire us to make a commitment, especially a financial commitment at a time like this, and certainly, one of those is national security. So I wondered if you could just talk briefly. Is exascale computing considered critical to advancing national security, and if so, has the National Nuclear Security Administration gone on record to say that? If so, how is the NNSA prepared to financially contribute to this effort and what would be an appropriate percentage contribution to an exascale computer from NNSA, would you say? Ms. Crawford. There is a lot of questions there so---- Mr. Hultgren. Yes. Ms. Crawford. --let's see if I can remember them. Mr. Hultgren. The first thing is have they gone on record of saying that this is a key component? And then basically then what kind of commitment should we expect from them? Ms. Crawford. Computing is the integrating element of maintaining the safety, security, and reliability of the nuclear weapons stockpile without returning to underground tests. So by integrating element, what I mean is we have the old test data, we have aboveground small experiments that we are doing, and we have a lot of theory and we have our new models. And we are bringing this all together in the computer. So this is an integrating element and this is the only way that we know to understand what is going on in the nuclear weapon. And so for that reason, we believe that it is extremely important. The NNSA is making an investment in the Advanced Scientific Computing Program. To maintain leadership, you need to have a base program. You need to have, you know, sort of a near-term program and you need to have a far-reaching program. Currently, the Office of Science and the NNSA both have a very strong base program. We have heard about the wonderful facilities at the laboratories, and of course it is not just the computer hardware itself but it is the applications that help us understand the world around us. We are investing with the Office of Science in some near- term research with industrial partners to look at some of those long lead time technology changes that need to be made. We need to make additional investments that are not in our current budgets in the programming environments for the exascale computing and in the math libraries so that we can actually use this billion-way parallelism. Mr. Hultgren. Okay. I see my time is expired. At this point, I hope that we will have an opportunity to have a second round of questioning as well. Mr. Swalwell. I don't have any objection. Mr. Hultgren. We can talk about that. Well, let's go ahead and we will recognize Mr. Veasey from Texas. Okay. Then Mr. Lipinski from Illinois is recognized for five minutes. Mr. Lipinski. Thank you, Mr. Chairman. I wanted to ask everyone on the panel a question about international partnerships. You know, obviously this cuts both ways. You can reduce the cost of reaching exascale capabilities with international partnerships but then there is the issue of, you know, damaging our Nation's economic competitiveness, potentially our national security, because we are not doing this on our own. Now, where do you come down on this? Is it worthwhile and how far should we go in international partnerships and at what point is it still an advantage? At what point does it become a disadvantage for us economically, giving up our lead on high-end computing? So whoever wants to start with that one. Dr. Stevens? Dr. Stevens. I will start. So I think the primary opportunity in international collaboration is in software, and in particular, the components of software that are open source that right now most of the software that runs on these machines other than the applications is built on--based on open source technologies developed largely in the United States. That is a significant lift to move all of that software to next- generation platforms, and international collaboration can help there provided that the software is--stays in the open. I think where we don't want to go at least in the near term is in deep hardware partnerships internationally. I think that is a place where we want to maintain our competitive edge. We have significant advantages with the U.S. vendor community and we want to maintain that as long as we can. Mr. Lipinski. Thank you. Anyone else? Ms. Crawford? Ms. Crawford. Yes. I would like to add that it is very important that the United States maintain the key intellectual property for the next supercomputer levels. If we control that, we have the high ground for the standards space base that will make all the decisions in the coming decade, and I would not want to cede that to another country. I cannot trust the U.S. nuclear weapons technology to a system built in China, say. That is untenable. I would like to not consider that those low- power technologies are developed ahead of time in other countries that we will use embedded in our intelligence systems. To me, it is very important that the United States take a very strong leadership position in this technology arena. Mr. Lipinski. Thank you. Dr. Reed? Dr. Reed. Yes, if I might add to that. It is part of the reason in my testimony I spoke very specifically to the importance of U.S. industry engagement. And as we move into this increasingly mobile device, low-power world, which is one of the key enabling technologies for future exascale systems, it is really crucial that the U.S. vendors maintain the competitive edge and strike a balance, as we do in terms of investment, between the global market and maintaining the unique capabilities for U.S. national security. Now, that is part of the role of the Federal Government in terms of, as Ms. Crawford said in her testimony, about shaping the direction of industry to ensure that we have the technology capability that we need. And I would echo that there are other uses as well. As we have talked about the rise of data analytics and its importance for national security and signal intelligence and other domains, that is another area where we must think carefully about many of the enabling technologies of which hardware is one, but the algorithms and other pieces need careful scrutiny also. Mr. Lipinski. Thank you. Dr. Giles. Yes, I would agree with what has been said. Just two points: I think it would be truly shameful for us to give up the elements of leadership that we have. And one of the things we pointed out and we asked in our exascale report was the criticality of time and of seizing the opportunity that in some way is presented uniquely to us to advance this field. But many, many countries will want to do that and we have a little bit of a time advantage because of our starting place. The other point, which is--it goes sort of in the direction of the open source software is the observation that a lot of the open science that is done in the world is done with international collaboration and with international connections. And we would, I think, like to still be in the position of having a lot of influence on the under-layer of that on which we will all build. But there certainly is international collaboration in science and I wouldn't want to minimize that-- the importance of that for the open science community. Mr. Lipinski. Thank you very much. And I want to ask a question if the Chairman would give me just a few extra seconds here. I just want to also echo what I know some of my colleagues have stated. I know exascale computing is important but we have to make sure that we don't pursue that at the expense of other important R&D activities that ASCR is doing. And I yield back. Mr. Hultgren. The gentleman yields back. We will go through a second round of questioning if anyone would have other questions, so I will begin by recognizing myself for five minutes. And I would address this first to Dr. Stevens but also ask if any of you would have other thoughts on this and really following up on Mr. Lipinski's questions of timing and competitiveness. And I wondered, Dr. Stevens, if you would have some thought of what level of investment is needed for the United States to maintain global leadership in scientific and technical computing for the next decade? And then something specific of if we maintain current investment, at what point would China surpass us in computing capabilities? And then also just looking at dates, what type of approach and how much investment would be necessary for us to lead to a deployable system by 2020? Dr. Stevens. Okay. So on the first one in terms of the resources required to do this, in the plans developed by the laboratories, we estimated that in addition to the current funding levels that we have, we would need an increment over time of approximately $400 million a year. That would be split between the two partners, the Office of Science and the NNSA. At that funding level, we think it is feasible--not guaranteed but feasible--to deploy a system by 2020. Of course, we made those estimates a few years ago when we had more runway than we have now. And that investment would go to both hardware and software and some applications of them--more applications would be needed by that time. At the current funding level, not including the bill---- Mr. Hultgren. Right. Dr. Stevens. --that is in front of us, it is estimated that we would not reach an exascale capability until middle of the next decade. We don't have accurate estimates of precisely what China will do but my guess is they will probably exceed us by the end of the decade if we were in that scenario. I don't remember your---- Mr. Hultgren. I think that covered it. So really it is, you know, without the investment, it is going to be probably 2025 before we would reach that level? Dr. Stevens. Absolutely. Mr. Hultgren. Do you think with the investment, is it a possible---- Dr. Stevens. We have---- Mr. Hultgren. --expectation to reach exascale levels by 2020? Dr. Stevens. I think it is possible. I think we would have to get moving faster than we are now and of course the industry is ready to do this. Labs are ready to do it; academia is ready to do it. We just need the resources and the commitment and also to do it in a way that doesn't cannibalize the current program. We need the base--we have to build on the base both in the Office of Science and in NNSA, and so this is really, really looking at incremental resources unfortunately to do it. Mr. Hultgren. Okay. Thank you. Do any of the others have any thoughts or disagreement? Ms. Crawford. I would just add that understanding what the sustained commitment is, whatever that dollar level turns out to be, is critical because then we can plan into the future. And not knowing whether, you know, the base budget is cut and the exascale R&D budget is cut and we have got a commitment to do this and then we are--now, we must do that because we have a contract and yet that prevents us from doing something else. So not knowing is really difficult to plan ahead and manage it effectively. So understanding that and sustaining that is very important. Mr. Hultgren. I absolutely agree and it is one of the things I am passionate about. I know other Members of our Subcommittee and Committee are as well of bringing some certainty specifically to research and to science. When we are looking to advance these programs it is so important that we are not budgeting month-to-month, which this place, Washington D.C., has kind of fallen into the habit of doing, but it has incredible detrimental impact, I know, on the great work that you all are doing. So I for one and I know my colleagues on both sides of the aisle would love to see some of that change. We are going to be fighting for that. Let me switch gears just a little bit and address this to Dr. Giles if I could and also to Dr. Stevens. But with respect to achieving an extraordinary number of computations per second, exascale appears to be a somewhat arbitrary goal. With current budgetary constraints, could DOE consider slower systems that would still be by far the fastest in the world or how do you see that fitting into this challenge of kind of keeping up with the rest of the world if DOE were to say, well, you know, we want to do some advancement but we are not going to go for that larger goal. We will just kind of settle for a lesser goal. How do you see that impacting the work that you are doing and the work that other nations are doing? Dr. Giles. Okay. Well, I think the key research to lower power consumption, to identify the pathway that takes us to exascale is one that is defined by that goal but which is a sort of--has a certain integrity of its own. Okay. If you do that--if one does that and makes that commitment to do their research and to do that beginning development, then how far you take it is part of the deployment question of how big a machine you build with the technology that you have done the research for. It--so--at least that is my take on it. I am not the technologist that Rick is and you may have a comment on that. Dr. Stevens. Well, what I can say is that the laboratories are excellent stewards of the Nation's money---- Dr. Giles. Yes. Dr. Stevens. --and we will buy the most capable systems that we can afford to buy when we have to replace and when we can replace the current systems. So I think that the question of, you know, can we settle is really a question of do we want to settle for not being able to do all the science or the most impactful engineering or address the most important national security challenges? We will do the best we can with what is provided to us. There is no question. I think lowering our sights though is not in our DNA. Mr. Hultgren. Right. Dr. Stevens. Right. Mr. Hultgren. No, that is helpful. Thank you. My time is expired. I will recognize the Ranking Member, Mr. Swalwell. Mr. Swalwell. Thank you, Mr. Chair. And I appreciate your comments about providing more certainty to our national laboratories. And we know that it is not just the laboratories who need the certainty but also private industry or any contractors who depend on work from the laboratories. One of the first lessons I learned when I was a planning commissioner years ago on a local sign ordinance issue from a local small business owner was vote for me, vote against me, but just give me certainty and, you know, do not have, you know, month-to-month sign regulations that give us no certainty at all, which now I have learned here, as the Chair said, month-to-month budgets also don't serve our laboratories well or private industry well. And so I join you in hoping that we can find ways to provide more certainty. I was hoping to just go witness by witness briefly and if you could just tell me for my own edification, and I am sure many others are curious, what are the private/public partnerships that you have at your laboratories through the exascale program? Dr. Giles. Well, let's see. I don't run a laboratory. Mr. Swalwell. Sure. Dr. Giles. But I would note things like you do run a lab that does the INCITE program in ASCR that invites researchers from outside DOE and from industry and with the particular emphasis on some industries to use the most advanced facilities that we have, so I think that would be one that I would identify coming out of ASCR. Mr. Swalwell. Great. And Dr. Stevens? Dr. Stevens. Well, just a few that we have done in the recent past. We have got a collaboration with Pratt & Whitney developing more efficient turbine engines, with Procter & Gamble on a variety of improving consumer products, with Cummins in improving diesel engines, and Caterpillar improving their ability to model whole vehicles and including the transmission systems and so forth, with the Mayo Clinic in applying computations and larger-scale problems in metagenomics, and so on. There is a long list. Some of these are collaborations with end-user companies and some are collaborations with companies like IBM or with Intel and with Cray in developing next-generation technologies, and we also work with small businesses. So the laboratories have collaborations on both the end- user component of this technology and the company is developing the technology itself. Mr. Swalwell. And when I hear some of those companies, IBM, Intel, Cummins, Caterpillar, I think of billions and billions and billions of dollars of exports. Those are some of the largest exporters in the United States, and if we are going to truly achieve our goal of doubling our exports over the next five years, making sure that those companies can continue to play a part in reducing that trade deficit--we have about $40 billion every month--is crucial and it sounds like the laboratories are helping them to do that so they can sell their goods and services to the marketplace outside the United States. Dr. Stevens. Absolutely. And we are also working with companies like Dow and DuPont and Johnson Controls. And it is a long list, right? And I think we exactly get this idea of helping American industry be more competitive. Mr. Swalwell. Great. And Ms. Crawford? Ms. Crawford. So rather than going through the long list, let me talk about the barriers for industrial adaptation of advanced computing. There have been a number of studies and there are three main barriers. One is the cost of establishing a supercomputing facility, the computer itself, the computing room, et cetera. The second one is the expertise, you know, having the skilled workforce that understands how to use these computers in a meaningful way for their products. And then the third is the software itself that helps them understand their products and how to improve those products. So the kind of partnerships that Dr. Stevens is talking about and that we have in our laboratory are helping to demonstrate to industry how to overcome those barriers so that they can in fact utilize this. And once they have firsthand demonstration and know the value, then they will start making the investments themselves at a higher level to drive their own productivity and competitiveness. Mr. Swalwell. Great. And Dr. Reed, I mean also just like Dr. Giles I know you do not run a laboratory but any public/ private partnerships you are familiar with that are working right now and also helping the innovation economy? Dr. Reed. Certainly. And I have been in similar roles in the past. As I mentioned, I used to run an NSF supercomputer center and we did very similar things in Illinois when I was there. Advanced manufacturing was certainly a target, logistics and supply chain optimization. But in Iowa now, there are many issues around advanced biological modeling and how we think about the future of healthcare in terms of everything from modeling the characteristics of lungs and what the implications are for drug delivery, how we might work with companies about those issues. I would echo what Ms. Crawford said, though. What is really crucial in those engagements and use of high-performance computing is simplicity of use because the domain experts are interested in advancing either the technology or the science or its applications and less interested in understanding what those of us in the technology business might view as the really cool stuff. Mr. Swalwell. Right. Dr. Reed. It is a means to an end and so those software user interface issues are really important. When I was at Microsoft, I spent a great deal of time working with the community in science on exactly those issues. How do we bring the power of advanced computing into small companies and into individual's hands where, from their perspective, the ease-of-use that they find familiar in their mobile device or their PC extends seamlessly and apparently magically to exploit those advanced capabilities? Mr. Swalwell. Great. Well, thank you. Thank you, Mr. Chair. This has been a great hearing. You know, I didn't pay enough attention to this stuff when I was in high school. I am learning a heck of a lot now in Congress and I could sit here for another few hours but I know our witnesses and our panel have other places to be. But thank you again. Mr. Hultgren. Thank you. Thank you. And I do want to thank each one of you for being here today on a very busy day on Capitol Hill. And with that, I just want to thank you for your valuable testimony and I want to thank the Members for the questions that they have had. The Members of the Committee may have additional questions, especially with competing hearings that were going on at the same time, so we will ask if you would be willing to respond in writing to questions that we would submit. And with that thought, we will keep the record open for two weeks for additional comments and written questions from Members and request for your response to those. With that, I again want to thank you so much for your time and for your wisdom and information today. With that, the witnesses are excused and this hearing is adjourned. Thank you. [Whereupon, at 11:20 a.m., the Subcommittee was adjourned.] Appendix I ---------- Answers to Post-Hearing QuestionsAnswers to Post-Hearing Questions Responses by Dr. Roscoe Giles [GRAPHIC] [TIFF OMITTED] 81195.055 [GRAPHIC] [TIFF OMITTED] 81195.056 [GRAPHIC] [TIFF OMITTED] 81195.057 [GRAPHIC] [TIFF OMITTED] 81195.058 [GRAPHIC] [TIFF OMITTED] 81195.059 [GRAPHIC] [TIFF OMITTED] 81195.060 [GRAPHIC] [TIFF OMITTED] 81195.061 Responses by Dr. Rick Stevens [GRAPHIC] [TIFF OMITTED] 81195.062 [GRAPHIC] [TIFF OMITTED] 81195.063 [GRAPHIC] [TIFF OMITTED] 81195.064 [GRAPHIC] [TIFF OMITTED] 81195.065 Responses by Ms. Dona Crawford [GRAPHIC] [TIFF OMITTED] 81195.066 [GRAPHIC] [TIFF OMITTED] 81195.067 [GRAPHIC] [TIFF OMITTED] 81195.068 [GRAPHIC] [TIFF OMITTED] 81195.069 [GRAPHIC] [TIFF OMITTED] 81195.070 [GRAPHIC] [TIFF OMITTED] 81195.071 [GRAPHIC] [TIFF OMITTED] 81195.072 Responses by Dr. Daniel Reed [GRAPHIC] [TIFF OMITTED] 81195.073 [GRAPHIC] [TIFF OMITTED] 81195.074 [GRAPHIC] [TIFF OMITTED] 81195.075