|Sponsors||U.S. Department of Energy|
|Architecture||9,216 POWER9 22-core CPUs|
27,648 NVIDIA Tesla V100 GPUs
|Operating system||Red Hat Enterprise Linux (RHEL)|
|Speed||200 petaFLOPS (peak)|
Summit or OLCF-4 is a supercomputer developed by IBM for use at Oak Ridge National Laboratory, capable of 200 petaFLOPS, making it the second fastest supercomputer in the world (it held the number 1 position from November 2018 to June 2020.) Its current LINPACK benchmark is clocked at 148.6 petaFLOPS. As of November 2019, the supercomputer is also the 5th most energy efficient in the world with a measured power efficiency of 14.668 gigaFLOPS/watt. Summit was the first supercomputer to reach exaflop (a quintillion operations per second) speed, achieving 1.88 exaflops during a genomic analysis and is expected to reach 3.3 exaflops using mixed-precision calculations.
The United States Department of Energy awarded a $325 million contract in November 2014 to IBM, NVIDIA and Mellanox. The effort resulted in construction of Summit and Sierra. Summit is tasked with civilian scientific research and is located at the Oak Ridge National Laboratory in Tennessee. Sierra is designed for nuclear weapons simulations and is located at the Lawrence Livermore National Laboratory in California. Summit is estimated to cover 873 square meters and require 219 kilometers of cabling. Researchers will utilize Summit for diverse fields such as cosmology, medicine and climatology.
In 2015, the project called Collaboration of Oak Ridge, Argonne and Lawrence Livermore (CORAL) included a third supercomputer named Aurora and was planned for installation at Argonne National Laboratory. By 2018, Aurora was re-engineered with completion anticipated in 2021 as an exascale computing project along with Frontier and El Capitan to be completed shortly thereafter.
The Summit supercomputer provides scientists and researchers the opportunity to solve complex tasks in the fields of energy, artificial intelligence, human health and other research areas. It has been used in Earthquake Simulation, Extreme Weather simulation using AI, Material science, Genomics and in predicting the lifetime of Neutrinos in physics. 
This section may be too technical for most readers to understand.(May 2020)
Each one of its 4,608 nodes (with 2 IBM POWER9 CPUs and 6 Nvidia Tesla GPUs in each node) has over 600 GB of coherent memory (96 GB HBM2 plus 512 GB DDR4 SDRAM) which is addressable by all CPUs and GPUs plus 800 GB of non-volatile RAM that can be used as a burst buffer or as extended memory. The POWER9 CPUs and Nvidia Volta GPUs are connected using NVIDIA's high speed NVLink. This allows for a heterogeneous computing model. To provide a high rate of data throughput, the nodes will be connected in a non-blocking fat-tree topology using a dual-rail Mellanox EDR InfiniBand interconnect for both storage and inter-process communications traffic which delivers both 200Gb/s bandwidth between nodes and in-network computing acceleration for communications frameworks such as MPI and SHMEM/PGAS.
Edited: 2021-06-18 18:58:23