Preparing the Livermore Computing Center for El Capitan and the exascale era of supercomputers required an entirely new way of thinking about the facility’s mechanical and electrical capabilities.
Topic: HPC Systems and Software
The second article in a series about the Lab's stockpile stewardship mission highlights computational models, parallel architectures, and data science techniques.
The first article in a series about the Lab's stockpile stewardship mission highlights the roles of computer simulations and exascale computing.
The new oneAPI Center of Excellence will involve the Center for Applied Scientific Computing and accelerate ZFP compression software to advance exascale computing.
LLNL participates in the CMD-IT/ACM Richard Tapia Celebration of Diversity in Computing Conference (Tapia2022) on September 7–10.
Livermore builds an open-source community around its award-winning HPC package manager.
The Advanced Technology Development and Mitigation program within the Exascale Computing Project shows that the best way to support the mission is through open collaboration and a sustainable software infrastructure.
LLNL has signed a memorandum of understanding with HPC facilities in Germany, the United Kingdom, and the U.S., jointly forming the International Association of Supercomputing Centers.
LLNL's Greg Becker spoke with HPC Tech Shorts to explain how Spack's binary cache works. The video “Get your HPC codes installed and running in minutes using Spack’s Binary Cache” runs 15:11.
Computer scientist Kathryn Mohror is among LLNL's recipients of the Department of Energy’s Early Career Research Program awards.
Learn how to use LLNL software in the cloud. In August, we will host tutorials in collaboration with AWS on how to install and use these projects on AWS EC2 instances. No previous experience necessary.
An LLNL team will be among the first researchers to perform work on the world’s first exascale supercomputer—Oak Ridge National Laboratory’s Frontier—when they use the system to model cancer-causing protein mutations.
Since 2018, software developer Trevor Smith has been putting his education and computing skills to good use supporting the Lab's HPC environment. He helps develop, deploy, and manage systems software that enables effective and secure use of computing resources.
The Lab's upcoming exascale-capable supercomputer will see an implementation of a converged accelerated computing unit, or APU, hybrid CPU-GPU compute engine.
In a presentation delivered to the 79th HPC User Forum at Oak Ridge National Laboratory, LLNL's Terri Quinn revealed that AMD’s forthcoming MI300 APU would be the computational bedrock of El Capitan, which is slated for installation at LLNL in late 2023.
The utility-grade infrastructure project massively upgraded the power and water-cooling capacity of the adjacent Livermore Computing Center, preparing it to house next generation exascale-class supercomputers for NNSA.
After 30 years, the High Performance Storage System (HPSS) collaboration continues to lead and adapt to the needs of the time while honoring its primary mission of long-term data stewardship of the crown jewels of data for government, academic and commercial organizations around the world.
Livermore’s archive leverages a hierarchical storage management application that runs on a cluster architecture that is user-friendly, extremely scalable, and lightning fast.
This year marks the 30th anniversary of the High Performance Storage System (HPSS) collaboration, comprising five DOE HPC national laboratories: LLNL, Lawrence Berkeley, Los Alamos, Oak Ridge, and Sandia, along with industry partner IBM.
The latest issue of the Lab's Science & Technology Review magazine highlights Todd Gamblin among other EMCR Program awardees.
As the U.S. welcomed the world’s first “true” exascale supercomputer, three predecessor machines for LLNL's future exascale system El Capitan managed to rank highly on the latest Top500 List of the world’s most powerful supercomputers.
LLNL and Amazon Web Services (AWS) have signed a memorandum of understanding to define the role of leadership-class HPC in a future where cloud HPC is ubiquitous.
LLNL participates in the International Parallel and Distributed Processing Symposium (IPDPS) on May 30 through June 3.
Winning the best paper award at PacificVis 2022, a research team has developed a resolution-precision-adaptive representation technique that reduces mesh sizes, thereby reducing the memory and storage footprints of large scientific datasets.
LLNL participates in the ISC High Performance Conference (ISC22) on May 29 through June 2.