The National Energy Research Scientific Computing Center (NERSC) Lawrence Berkeley National Laboratory (Berkeley Lab) this week announced its new supercomputer that will combine deep learning and simulation computing capabilities. The Perlmutter system will use AMD's top-of-the-range 64-core EPYC 7763 processors as well as Nvidia's A100 compute GPUs to push out up to 180 PetaFLOPS of 'standard' performance and up to four ExaFLOPS of AI performance. All told, that makes it the second-fastest supercomputer in the world behind Japan's Fugaku.
"Perlmutter will enable a larger range of applications than previous NERSC systems and is the first NERSC supercomputer designed from the start to meet the needs of both simulation and data analysis," said Sudip Dosanjh, director or NERSC.
NERSC's Perlmutter supercomputer relies on HPE's heterogeneous Cray Shasta architecture and is set to be delivered in two phases:
Phase 1 features 12 heterogenous cabinets comprising 1,536 nodes. Each node packs one 64-core AMD EPYC 7763 'Milan' CPU with 256GB of DDR4 SDRAM and four Nvidia A100 40GB GPUs connected via NVLink. The system uses a 35PB all-flash storage subsystem with 5TB/s of throughput.
The first phase of NERSC's Perlmutter can deliver 60 FP64 PetaFLOPS of performance for simulations, and 3.823 FP16 ExaFLOPS of performance (with sparsity) for analysis and deep learning. The system was installed earlier this year and is now being deployed.
While 60 FP64 PetaFLOPS puts Perlmutter into the Top 10 list of the world's most powerful supercomputers, the system does not stop there.
Phase 2 will add 3,072 AMD EPYC 7763-based CPU-only nodes with 512GB of memory per node that will be dedicated to simulation. FP64 performance for the second phase will be around 120 PFLOPS.
When the second phase of Perlmutter is deployed later this year, the combined FP64 performance of the supercomputer will total 180 PFLOPS, which will put it ahead of Summit, the world's second most powerful supercomputer. However, it will still trail Japan's Fugaku, which weighs in at 442 PFLOPs. Meanwhile, in addition to formidable throughput for simulations, Perlmutter will offer nearly four ExaFLOPS of FP16 throughput for AI applications.
The Link LonkMay 28, 2021 at 06:02PM
https://ift.tt/3fu33yl
AMD and Nvidia to Power Four ExaFLOPS Supercomputer - Tom's Hardware
https://ift.tt/2ZDueh5
AMD
No comments:
Post a Comment