Splitting memory resources in high performance computing between local nodes and a larger shared remote pool can help better support diverse applications.
Topic: HPC Systems and Software
As CTO of Livermore Computing, de Supinski is responsible for formulating, overseeing, and implementing LLNL’s large-scale computing strategy, requiring managing multiple collaborations with the HPC industry and academia.
Lori Diachin will take over as director of the DOE’s Exascale Computing Project on June 1, guiding the successful, multi-institutional high performance computing effort through its final stages.
Livermore CTO Bronis de Supinski joins the Let's Talk Exascale podcast to discuss the details of LLNL's upcoming exascale supercomputer.
Variorum provides robust, portable interfaces that allow us to measure and optimize computation at the physical level: temperature, cycles, energy, and power. With that foundation, we can get the best possible use of our world-class computing resources.
The addition of the spatial data flow accelerator into LLNL’s Livermore Computing Center is part of an effort to upgrade the Lab’s cognitive simulation (CogSim) program.
The Lab was already using Elastic components to gather data from its HPC clusters, then investigated whether Elasticsearch and Kibana could be applied to all scanning and logging activities across the board.
Updating a compiler can affect how code runs, leading to inconsistencies in outputs and creating problems for scientists. A new tool automatically finds the sources of these inconsistencies.
LLNL participates in the ISC High Performance Conference (ISC23) on May 21–25.
An LLNL Distinguished Member of Technical Staff, Gokhale is considered an expert in her field, and continues to enjoy the fast pace of innovation and change in computing.
Supercomputers broke the exascale barrier, marking a new era in processing power, but the energy consumption of such machines cannot run rampant.
LC’s adaptation of OpenZFS software provides high performance parallel file systems with better performance and scalability.
LLNL’s archives recount the contributions of women who developed code during the Lab's early decades.
UCLA's Institute for Pure & Applied Mathematics hosted LLNL's Erik Draeger for a talk about the challenges and possibilities of exascale computing.
“I am delighted to be recognized by HPCwire,” Quinn said. “I feel the recognition has as much to do with the stature of Livermore Computing as the opportunity I’ve had to contribute. "
LLNL’s archives provide a glimpse into the career and contributions of a computing pioneer.
This year, the DOE honored 44 teams including LLNL's Exascale Computing Facility Modernization Project team for significant power and cooling upgrades to support upcoming exascale supercomputers.
LLNL's popular lecture series, “Science on Saturday,” runs February 4–25. The February 18 lecture is titled "Supersizing Computing: 70 Years of HPC."
Computer scientist Johannes Doerfert was recognized as a 2023 BSSw fellow. He plans to use the funding to create videos about best practices for interacting with compilers.
A multidecade, multi-laboratory collaboration evolves scalable long-term data storage and retrieval solutions to survive the march of time.
High performance computing was key to the December 5 breakthrough at the National Ignition Facility.
Two supercomputers powered the research of hundreds of scientists at Livermore’s NNSA National Ignition Facility, which recently achieved ignition.
ASC’s Advanced Memory Technology research projects are developing technologies that will impact future computer system architectures for complex modeling and simulation workloads.
LLNL is home to the world’s largest Spectra TFinityTM system, which offers the speed, agility, and capacity required to take LLNL into the exascale era.
Combining specialized software tools with heterogeneous HPC hardware requires an intelligent workflow performance optimization strategy.