libROM is a library designed to facilitate Proper Orthogonal Decomposition (POD) based Reduced Order Modeling (ROM).
Topic: Data Science
A new component-wise reduced order modeling method enables high-fidelity lattice design optimization.
Highlights include MFEM community workshops, compiler co-design, HPC standards committees, and AI/ML for national security.
High-precision numerical data from computer simulations, observations, and experiments is often represented in floating point and can easily reach terabytes to petabytes of storage.
LLNL is participating in the 34th annual Supercomputing Conference (SC22), which will be held both virtually and in Dallas on November 13–18, 2022.
In a time-trial competition, participants trained an autonomous race car with reinforcement learning algorithms.
After 10 years and 33 hackathons, nothing can stop this beloved tradition.
The Earth System Grid Federation is a web-based tool set that powers most global Earth system research.
Angeline Lee simultaneously serves as a group leader, contributes to programmatic projects, and studies for her bachelor’s degree.
Winning the best paper award at PacificVis 2022, a research team has developed a resolution-precision-adaptive representation technique that reduces mesh sizes, thereby reducing the memory and storage footprints of large scientific datasets.
LLNL participates in the International Parallel and Distributed Processing Symposium (IPDPS) on May 30 through June 3.
Kevin McLoughlin has always been fascinated by the intersection of computing and biology. His LLNL career encompasses award-winning microbial detection technology, a COVID-19 antiviral drug design pipeline, and work with the ATOM consortium.
LivIT tackles challenges of workforce safety, telecommuting, cyber security protocols, National Ignition Facility software updates, and more.
From molecular screening, a software platform, and an online data to the computing systems that power these projects.
LLNL’s cyber programs work across a broad sponsor space to develop technologies addressing sophisticated cyber threats directed at national security and civilian critical infrastructure.
This project advances research in physics-informed ML, invests in validated and explainable ML, creates an advanced data environment, builds ML expertise across the complex, and more.
Highlights include power grid challenges, performance analysis, complex boundary conditions, and a novel multiscale modeling approach.
LLNL is participating in the 33rd annual Supercomputing Conference (SC21), which will be held both virtually and in St. Louis on November 14–19, 2021.
Brian Gallagher works on applications of machine learning for a variety of science and national security questions. He’s also a group leader, student mentor, and the new director of LLNL’s Data Science Challenge.
New research debuting at ICLR 2021 demonstrates a learning-by-compressing approach to deep learning that outperforms traditional methods without sacrificing accuracy.
Highlights include scalable deep learning, high-order finite elements, data race detection, and reduced order models.
BUILD tackles the complexities of HPC software integration with dependency compatibility models, binary analysis tools, efficient logic solvers, and configuration optimization techniques.
Our researchers will be well represented at the virtual SIAM Conference on Computational Science and Engineering (CSE21) on March 1–5. SIAM is the Society for Industrial and Applied Mathematics with an international community of more than 14,500 individual members.
Three papers address feature importance estimation under distribution shifts, attribute-guided adversarial training, and uncertainty matching in graph neural networks.
StarSapphire is a collection of scientific data mining projects focusing on the analysis of data from scientific simulations, observations, and experiments.
 
        