Our research projects vary in size, scope, and duration, but they share a focus on developing tools and methods that help LLNL deliver on its missions to the nation and, more broadly, advance the state of the art in scientific HPC. Projects are organized here in three ways: Active projects are those currently funded and regularly updated. Legacy projects are no longer actively developed. The A-Z option sorts all projects alphabetically, both active and legacy.

Active | A-Z | Legacy


Advanced Data Analytics for Proliferation Detection

ADAPD—Advanced Data Analytics for Proliferation Detection—integrates subject-matter expertise from the DOE’s national laboratories to create new capabilities for analyzing growing global data…


Analytics and Informatics Management Systems

AIMS (Analytics and Informatics Management Systems) develops integrated cyberinfrastructure for big climate data discovery, analytics, simulations, and knowledge innovation.


Improving Simulation Workflows with Machine Learning and Big Data

Simulation workflows for Arbitrary Lagrangian–Eulerian (ALE) methods are highly complex and often require a manual tuning process. There is an urgent need to semi-automate this process to reduce…


Fast, Lightweight, Dynamic Tuning for Data-Dependent Code

Apollo, an auto-tuning extension of RAJA, improves performance portability in adaptive mesh refinement, multi-physics, and hydrodynamics codes via machine learning classifiers.

Application-Level Resilience

Efficient Algorithmic Fault Tolerance

Application-level resilience is emerging as an alternative to traditional fault tolerance approaches because it provides fault tolerance at a lower cost than traditional approaches.


Scaling Up Transport Sweep Algorithms

LLNL researchers are testing and enhancing a neutral particle transport code and the algorithm on which the code relies to ensure that they successfully scale to larger and more complex computing…

ASC Proxy Apps

Prepare for Testing and Porting Applications

Proxy apps serve as specific targets for testing and simulation without the time, effort, and expertise that porting or changing most production codes would require.



Diagnosing Performance and Correctness Faults

AutomaDeD is a tool that automatically diagnoses performance and correctness faults in MPI applications. It has two major functionalities: identifying abnormal MPI tasks and code regions and…



High-Performance Language Interoperability

Babel is a high-performance language interoperability tool. The project is mainly developed at the Center for Applied Scientific Computing (CASC) LLNL. Babel started as an internal Lab Directed…


High-Order Finite Element Hydrodynamics

Through research funded at LLNL, scientists have developed BLAST, a high-order finite element hydrodynamics research code that improves the accuracy of simulations, provides a path to extreme…


Build, Link, and Test

BLT software supports HPC software development with built-in CMake macros for external libraries, code health checks, and unit testing.


Solving the Software Complexity Puzzle

This project aims to tackle the complexities of HPC software integration with dependency compatibility models, binary analysis tools, efficient logic solvers, and configuration optimization…


Application Introspection System

A comprehensive understanding of the performance behavior of large-scale simulations requires the ability to compile, analyze, and compare measurements and contexts from many independent sources.…


Copy Hiding Application Interface

A new software model helps move million-line codes to various hardware architectures by automating data movement in unique ways.

Cluster Management Tools

Flexible Support for Our Linux Ecosystem

Large Linux data centers require flexible system management. At Livermore Computing, we are committed to supporting our Linux ecosystem at the high end of commodity computing.


Running Millions of Concurrent MPI Jobs

Cram lets you easily run many small MPI jobs within a single, large MPI job by splitting MPI_COMM_WORLD up into many small communicators to run each job in the cram file independently.

CT Image Enhancement

Novel Processing Pipeline for Threat Detection

Researchers are developing enhanced computed tomography image processing methods for explosives identification and other national security applications.

Data-Intensive Computing Solutions

Addressing Growing Demands

New platforms are improving big data computing on Livermore’s high performance computers.

Derived Field Generation

Execution Strategies for Multiple Hardware Architectures

Livermore computer scientists have helped create a flexible framework that aids programmers in creating source code that can be used effectively on multiple hardware architectures.


Supporting Climate Research Collaboration

The Earth System Grid Federation is a web-based tool set that powers most global climate change research.


Enabling Technologies for High-Order Simulations

The Enabling Technologies for High-Order Simulations (ETHOS) project performs research of fundamental mathematical technologies for next-generation high-order simulations algorithms.



Extreme Resilient Discretization

The Extreme Resilient Discretization project (ExReDi) was established to address these challenges for algorithms common for fluid and plasma simulations.


Fast Global File Status

Fast Global File Status (FGFS) is an open-source package that provides scalable mechanisms and programming interfaces to retrieve global information of a file.


GPU Portability for FORTRAN Codes

FGPU provides code examples that port FORTRAN codes to run on IBM OpenPOWER platforms like LLNL's Sierra supercomputer.