When scientists pushed the world's fastest supercomputer to its limits, they found those limits stretched beyond even their biggest expectations.
The Frontier supercomputer at the Department of Energy's Oak Ridge National Laboratory set a new ceiling in performance when it debuted in 2022 as the first exascale system in history—capable of more than 1 quintillion calculations per second. Now researchers are learning just what heights of scientific discovery Frontier's computational power can help them achieve.
In the latest milestone, a team of engineers and scientists used Frontier to simulate a system of nearly half a trillion atoms—the largest system ever modeled and more than 400 times the size of the closest competition—in a potential gateway to new insights across the scientific spectrum.
"It's like test-driving a car with a speedometer that registers 120 miles per hour, but you press the gas pedal and find out it goes past 200," said Nick Hagerty, a high-performance computing engineer for ORNL's Oak Ridge Leadership Computing Facility, which houses Frontier.
"Nobody runs simulations at a scale this size because nobody's ever tried. We didn't know we could go this big."
The results hold promise for scientific studies at a scale and level of detail not yet seen.
"Nobody on Earth has done anything remotely close to this before," said Dilip Asthagiri, an OLCF senior computational biomedical scientist who helped design the test. "This discovery brings us closer to simulating a stripped-down version of a biological cell, the so-called minimal cell that has the essential components to enable basic life processes."
Hagerty and his team sought to max out Frontier to set criteria for the supercomputer's successor machine, still in development. Their mission: Push Frontier as far as it could go and see where it stopped.
The team used Frontier with the Large-scale Atomic and Molecular Massively Parallel Simulator software module, or LAMMPS, to simulate a system of room-temperature water molecules at the atomic level as they gradually increased the number of atoms.
"Water is a great test case for a machine like Frontier because any researcher studying a biological system at the atomic level will likely need to simulate water," Hagerty said. "We wanted to see how big of a system Frontier could really handle and what limitations are encountered at this scale.
"As one of the first benchmarking efforts to use more than a billion atoms with long-range interactions, we would periodically find bugs in the LAMMPS source code. We worked with the LAMMPS developers, who were highly engaged and responsive, to resolve those bugs, and this was critical to our scaling success."
Frontier tackles complex problems via parallelism, which means the supercomputer breaks up the computational workload across its 9,408 nodes, each a self-contained computer capable of around 200 trillion calculations per second. With each increase in problem size, the simulation demanded more memory and processing power. Frontier never blinked.
Citation: Breaking benchmarks: Frontier supercomputer sets new standard in molecular simulation (2024, June 20) retrieved 20 June 2024 from https://techxplore.com/news/2024-06-benchmarks-frontier-supercomputer-standard-molecular.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.