This open-source file system framework accelerates hierarchical HPC I/O operations with effective, efficient use of node-local storage.
Topic: HPC Systems and Software
Discover how the software architecture and storage systems that will drive El Capitan’s performance will help LLNL and the NNSA Tri-Labs push the boundaries of computational science.
Unveiled at the International Supercomputing Conference in Germany, the June 2024 Top500 lists three systems with identical components as registering 19.65 petaflops on the High Performance Linpack benchmark, ranking them among the world’s 50 fastest.
In a groundbreaking development for addressing future viral pandemics, a multi-institutional team involving LLNL researchers has successfully combined an AI-backed platform with supercomputing to redesign and restore the effectiveness of antibodies whose ability to fight viruses has been compromi
LLNL participates in the ISC High Performance Conference (ISC24) on May 12–16.
LLNL researchers have achieved a milestone in accelerating and adding features to complex multiphysics simulations run on GPUs, a development that could advance HPC and engineering.
The Tools Working Group delivers debugging, correctness, and performance analysis solutions at an unprecedented scale.
Compilers translate human-programmable source code into machine-readable code. Building a compiler is especially challenging in the exascale era.
The El Capitan Center of Excellence provides a conduit between national labs and commercial vendors, ensuring that the exascale system will meet everyone’s needs.
Backed by Spack’s robust functionality, the Packaging Working Group manages the relationships between user software and system software.
The advent of accelerated processing units presents new challenges and opportunities for teams responsible for network interconnects and math libraries.
As a high school baseball coach, LLNL software developer Chris Moussa understands the value of having teammates you can rely on, which is why he’s especially grateful for his “fantastic” team at work.
The high performance computing publication HPCwire has selected LLNL computer scientist Todd Gamblin as one of its “People to Watch” in HPC for 2024. The program recognized 12 HPC professionals who “play leading roles in driving innovation within their particular fields, making significant contributions to society as a whole.”
The system will enable researchers from the National Nuclear Security Administration weapons design laboratories to create models and run simulations, previously considered challenging, time-intensive or impossible, for the maintenance and modernization of the United States’ nuclear weapons stockpile.
An LLNL-led team has developed a method for optimizing application performance on large-scale GPU systems, providing a useful tool for developers running on GPU-based massively parallel and distributed machines.
Johannes Doerfert, a computer scientist in the Center for Applied Scientific Computing, was one of three researchers awarded the honor at SC23 in Denver.
A record number of attendees—more than 14,000—experts, researchers, vendors and enthusiasts in the field of HPC descended on the Mile High City for the 2023 International Conference for High Performance Computing, Networking, Storage and Analysis, colloquially known as SC23.
Leading HPC publication HPCwire presented Spack developers with the Editor's Choice Award for Best HPC Programming Tool or Technology at SC23.
The debut of the NNSA Commodity Technology Systems-2 computing clusters Dane and Bengal on the Top500 List of the world’s most powerful supercomputers brings the total of LLNL-sited systems on the list to 11, the most of any supercomputing center in the world.
LLNL is participating in the 35th annual Supercomputing Conference (SC23), which will be held both virtually and in Denver on November 12–17, 2023.
Merlin is an open-source workflow orchestration and coordination tool that makes it easy to build, run, and process large-scale workflows.
Over several years, teams have prepared the infrastructure for El Capitan, designing and building the computing facility’s upgrades for power and cooling, installing storage and compute components and connecting everything together. Once all the pieces are in place, the life of El Cap as world-class supercomputer begins.
Alpine/ZFP addresses analysis, visualization, data reduction needs for exascale science applications
The Data and Visualization efforts in the DOE’s Exascale Computing Project provide an ecosystem of capabilities for data management, analysis, lossy compression, and visualization.
Quandary is an open-source C++ package for optimal control of quantum systems on classical high performance computing platforms.
The Center for Efficient Exascale Discretizations has developed innovative mathematical algorithms for the DOE’s next generation of supercomputers.