Flux, next-generation resource and job management software, steps up to support emerging use cases.
Topic: HPC Systems and Software
A Laboratory-developed software package management tool, enhanced by contributions from more than 1,000 users, supports the high performance computing community.
Livermore builds an open-source community around its award-winning HPC package manager.
Learn how to use LLNL software in the cloud. Throughout August, join our tutorials on how to install and use several projects on AWS EC2 instances. No previous experience necessary.
Anna Maria Bailey, LLNL’s Chief Engineer for HPC and a Distinguished Member of Technical Staff, has enjoyed her “many careers” at the Lab and the ability to jump around to follow her interests.
A research team from Oak Ridge and Lawrence Livermore national labs won the first IPDPS Best Open-Source Contribution Award for the paper “UnifyFS: A User-level Shared File System for Unified Access to Distributed Local Storage.”
LLNL CTO Bronis de Supinski talks about how the Lab deploys novel architecture AI machines and provides an update on El Capitan.
Splitting memory resources in high performance computing between local nodes and a larger shared remote pool can help better support diverse applications.
As CTO of Livermore Computing, de Supinski is responsible for formulating, overseeing, and implementing LLNL’s large-scale computing strategy, requiring managing multiple collaborations with the HPC industry and academia.
Lori Diachin will take over as director of the DOE’s Exascale Computing Project on June 1, guiding the successful, multi-institutional high performance computing effort through its final stages.
Livermore CTO Bronis de Supinski joins the Let's Talk Exascale podcast to discuss the details of LLNL's upcoming exascale supercomputer.
Variorum provides robust, portable interfaces that allow us to measure and optimize computation at the physical level: temperature, cycles, energy, and power. With that foundation, we can get the best possible use of our world-class computing resources.
The addition of the spatial data flow accelerator into LLNL’s Livermore Computing Center is part of an effort to upgrade the Lab’s cognitive simulation (CogSim) program.
The Lab was already using Elastic components to gather data from its HPC clusters, then investigated whether Elasticsearch and Kibana could be applied to all scanning and logging activities across the board.
Updating a compiler can affect how code runs, leading to inconsistencies in outputs and creating problems for scientists. A new tool automatically finds the sources of these inconsistencies.
LLNL participates in the ISC High Performance Conference (ISC23) on May 21–25.
An LLNL Distinguished Member of Technical Staff, Gokhale is considered an expert in her field, and continues to enjoy the fast pace of innovation and change in computing.
Supercomputers broke the exascale barrier, marking a new era in processing power, but the energy consumption of such machines cannot run rampant.
LC’s adaptation of OpenZFS software provides high performance parallel file systems with better performance and scalability.
LLNL’s archives recount the contributions of women who developed code during the Lab's early decades.
UCLA's Institute for Pure & Applied Mathematics hosted LLNL's Erik Draeger for a talk about the challenges and possibilities of exascale computing.
“I am delighted to be recognized by HPCwire,” Quinn said. “I feel the recognition has as much to do with the stature of Livermore Computing as the opportunity I’ve had to contribute. "
LLNL’s archives provide a glimpse into the career and contributions of a computing pioneer.
This year, the DOE honored 44 teams including LLNL's Exascale Computing Facility Modernization Project team for significant power and cooling upgrades to support upcoming exascale supercomputers.
LLNL's popular lecture series, “Science on Saturday,” runs February 4–25. The February 18 lecture is titled "Supersizing Computing: 70 Years of HPC."