{"id":3255,"date":"2023-11-13T21:42:58","date_gmt":"2023-11-13T21:42:58","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2023\/11\/13\/new-class-of-accelerated-efficient-ai-systems-mark-the-next-era-of-supercomputing\/"},"modified":"2023-11-13T21:42:58","modified_gmt":"2023-11-13T21:42:58","slug":"new-class-of-accelerated-efficient-ai-systems-mark-the-next-era-of-supercomputing","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2023\/11\/13\/new-class-of-accelerated-efficient-ai-systems-mark-the-next-era-of-supercomputing\/","title":{"rendered":"New Class of Accelerated, Efficient AI Systems Mark the Next Era of Supercomputing"},"content":{"rendered":"<div id=\"bsf_rt_marker\">\n<p>NVIDIA today unveiled at SC23 the next wave of technologies that will lift scientific and industrial research centers worldwide to new levels of performance and energy efficiency.<\/p>\n<p>\u201cNVIDIA hardware and software innovations are creating a new class of AI supercomputers,\u201d said Ian Buck, vice president of the company\u2019s high performance computing and hyperscale data center business, in a special address at the conference.<\/p>\n<p>Some of the systems will pack memory-enhanced <a href=\"https:\/\/developer.nvidia.com\/blog\/nvidia-hopper-architecture-in-depth\/\">NVIDIA Hopper accelerators<\/a>, others a new <a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-grace-hopper-superchip-powers-jupiter-defining-a-new-class-of-supercomputers-to-propel-ai-for-scientific-discovery\">NVIDIA Grace Hopper systems architecture<\/a>. All will use the expanded parallelism to run a full stack of accelerated software for <a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/data-science\/generative-ai\/\">generative AI<\/a>, HPC and hybrid <a href=\"https:\/\/blogs.nvidia.com\/blog\/what-is-quantum-computing\/\">quantum computing<\/a>.<\/p>\n<p>Buck described the new <a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-supercharges-hopper-the-worlds-leading-ai-computing-platform\">NVIDIA HGX H200<\/a> as \u201cthe world\u2019s leading AI computing platform.\u201d<\/p>\n<figure id=\"attachment_68062\" aria-describedby=\"caption-attachment-68062\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/H200-image.jpg\"><\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/H200-image-672x411.jpg\" alt=\"Image of H200 GPU system\" width=\"672\" height=\"411\"><\/p>\n<p><\/a><figcaption id=\"caption-attachment-68062\" class=\"wp-caption-text\">NVIDIA H200 Tensor Core GPUs pack HBM3e memory to run growing generative AI models.<\/figcaption><\/figure>\n<p>It packs up to 141GB of HBM3e, the first AI accelerator to use the ultrafast technology. Running models like GPT-3, NVIDIA H200 Tensor Core GPUs provide an 18x performance increase over prior-generation accelerators.<\/p>\n<p><span>Among other <\/span><a href=\"https:\/\/github.com\/NVIDIA\/TensorRT-LLM\/blob\/release\/0.5.0\/docs\/source\/blogs\/H200launch.md\"><span>generative AI benchmarks<\/span><\/a><span>, they zip through 12,000 tokens per second on a Llama2-13B large language model (<\/span><a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/data-science\/large-language-models\/\"><span>LLM<\/span><\/a><span>). <\/span><\/p>\n<p>Buck\u00a0also revealed a server platform that links four <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/grace-hopper-superchip\/\">NVIDIA GH200 Grace Hopper Superchips<\/a> on an <a href=\"https:\/\/blogs.nvidia.com\/blog\/what-is-nvidia-nvlink\/\">NVIDIA NVLink<\/a> interconnect. The quad configuration puts in a single compute node a whopping 288 Arm Neoverse cores and 16 petaflops of AI performance with up to 2.3 terabytes of high-speed memory.<\/p>\n<figure id=\"attachment_68065\" aria-describedby=\"caption-attachment-68065\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Quad-Grace-Hopper-node-configuration.jpg\"><\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Quad-Grace-Hopper-node-configuration-672x405.jpg\" alt=\"Image of quad GH200 server node\" width=\"672\" height=\"405\"><\/p>\n<p><\/a><figcaption id=\"caption-attachment-68065\" class=\"wp-caption-text\">Server nodes based on the four GH200 Superchips will deliver 16 petaflops of AI performance.<\/figcaption><\/figure>\n<p>Demonstrating its efficiency, one GH200 Superchip using the <a href=\"https:\/\/github.com\/NVIDIA\/TensorRT-LLM\">NVIDIA TensorRT-LLM<\/a> open-source library is 100x faster than a dual-socket x86 CPU system and nearly 2x more energy efficient than an X86 + H100 GPU server.<\/p>\n<p>\u201cAccelerated computing is sustainable computing,\u201d Buck said. \u201cBy harnessing the power of accelerated computing and generative AI, together we can drive innovation across industries while reducing our impact on the environment.\u201d<\/p>\n<h2><b>NVIDIA Powers 38 of 49 New TOP500 Systems<\/b><\/h2>\n<p>The latest TOP500 list of the world\u2019s fastest supercomputers reflects the shift toward accelerated, energy-efficient supercomputing.<\/p>\n<p>Thanks to new systems powered by <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/h100\/\">NVIDIA H100 Tensor Core GPUs<\/a>, NVIDIA now delivers more than 2.5 exaflops of HPC performance across these world-leading systems, up from 1.6 exaflops in the May rankings. NVIDIA\u2019s contribution on the top 10 alone reaches nearly an exaflop of HPC and 72 exaflops of AI performance.<\/p>\n<p>The new list contains the highest number of systems ever using NVIDIA technologies, 379 vs. 372 in May, including 38 of 49 new supercomputers on the list.<\/p>\n<p>Microsoft Azure leads the newcomers with its Eagle system using H100 GPUs in NDv5 instances to hit No. 3 with 561 petaflops. Mare Nostrum5 in Barcelona ranked No. 8, and NVIDIA Eos \u2014 which recently set <a href=\"https:\/\/blogs.nvidia.com\/blog\/scaling-ai-training-mlperf\/\">new AI training records<\/a> on the MLPerf benchmarks \u2014 came in at No. 9.<\/p>\n<p>Showing their energy efficiency, NVIDIA GPUs power 23 of the top 30 systems on the Green500. And they retained the No. 1 spot with the H100 GPU-based Henri system, which delivers 65.09 gigaflops per watt for the Flatiron Institute in New York.<\/p>\n<h2><b>Gen AI Explores COVID<\/b><\/h2>\n<p><span>Showing what\u2019s possible<\/span>, the Argonne National Laboratory used <a href=\"https:\/\/www.nvidia.com\/en-gb\/gpu-cloud\/bionemo\/\">NVIDIA BioNeMo<\/a>, a generative AI platform for biomolecular LLMs, to develop <a href=\"https:\/\/blogs.nvidia.com\/blog\/generative-ai-covid-genome-sequences\/\">GenSLMs<\/a>, a model that can generate gene sequences that closely resemble real-world variants of the coronavirus. Using NVIDIA GPUs and data from 1.5 million COVID genome sequences, it can also rapidly identify new virus variants.<\/p>\n<p>The work<a href=\"https:\/\/blogs.nvidia.com\/blog\/genomic-large-language-model-predicts-covid-variants\/\"> won the Gordon Bell special prize<\/a> last year and was trained on supercomputers, including Argonne\u2019s<a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-turbocharges-extreme-scale-ai-for-argonne-national-laboratorys-polaris-supercomputer\"> Polaris<\/a> system, the U.S. Department of Energy\u2019s<a href=\"https:\/\/blogs.nvidia.com\/blog\/nersc-perlmutter-ai-supercomputer\/\"> Perlmutter<\/a> and NVIDIA\u2019s<a href=\"https:\/\/blogs.nvidia.com\/blog\/making-selene-pandemic-ai\/\"> Selene<\/a>.<\/p>\n<p>It\u2019s \u201cjust the tip of the iceberg \u2014 the future is brimming with possibilities, as generative AI continues to redefine the landscape of scientific exploration,\u201d said Kimberly Powell, vice president of healthcare at NVIDIA, in the special address.<\/p>\n<h2><b>Saving Time, Money and Energy<\/b><\/h2>\n<p>Using the latest technologies, accelerated workloads can see an order-of-magnitude reduction in system cost and energy used, Buck said.<\/p>\n<p>For example, Siemens teamed with Mercedes to analyze aerodynamics and related acoustics for its new electric EQE vehicles. The simulations that took weeks on CPU clusters ran significantly faster using the latest NVIDIA H100 GPUs. In addition, Hopper GPUs let them reduce costs by 3x and reduce energy consumption by 4x (below).<\/p>\n<p><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Sustainable-computing-chart.jpg\"><\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Sustainable-computing-chart-672x351.jpg\" alt=\"Chart showing the performance and energy efficiency of H100 GPUs\" width=\"672\" height=\"351\"><\/p>\n<p><\/a><\/p>\n<h2><b>Switching on 200 Exaflops Beginning Next Year<\/b><\/h2>\n<p>Scientific and industrial advances will come from every corner of the globe where the latest systems are <a href=\"https:\/\/blogs.nvidia.com\/blog\/gh200-grace-hopper-superchip-powers-ai-supercomputers\/\">being deployed<\/a>.<\/p>\n<p>\u201cWe already see a combined 200 <a href=\"https:\/\/blogs.nvidia.com\/blog\/what-is-an-exaflop\/\">exaflops<\/a> of AI on Grace Hopper supercomputers going to production 2024,\u201d Buck said.<\/p>\n<p>They include the massive <a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-grace-hopper-superchip-powers-jupiter-defining-a-new-class-of-supercomputers-to-propel-ai-for-scientific-discovery\">JUPITER supercomputer<\/a> at Germany\u2019s J\u00fclich center. It can deliver 93 exaflops of performance for AI training and 1 exaflop for HPC applications, while consuming only 18.2 megawatts of power.<\/p>\n<figure id=\"attachment_68071\" aria-describedby=\"caption-attachment-68071\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Chart-on-350-exaflops-of-AI-performance-deployed.jpg\"><\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Chart-on-350-exaflops-of-AI-performance-deployed-672x403.jpg\" alt=\"Chart of deployed performance of supercomputers using NVIDIA GPUs through 2024\" width=\"672\" height=\"403\"><\/p>\n<p><\/a><figcaption id=\"caption-attachment-68071\" class=\"wp-caption-text\">Research centers are poised to switch on a tsunami of GH200 performance.<\/figcaption><\/figure>\n<p>Based on Eviden\u2019s BullSequana XH3000 liquid-cooled system, JUPITER will use the NVIDIA quad GH200 system architecture and <a href=\"https:\/\/www.nvidia.com\/en-us\/networking\/quantum2\/\">NVIDIA Quantum-2 InfiniBand<\/a> networking for climate and weather predictions, drug discovery, hybrid quantum computing and digital twins. JUPITER quad GH200 nodes will be configured with 864GB of high-speed memory.<\/p>\n<p>It\u2019s one of several new supercomputers using Grace Hopper that NVIDIA announced at SC23.<\/p>\n<p>The HPE Cray EX2500 system from Hewlett Packard Enterprise will use the quad GH200 to power many AI supercomputers coming online next year.<\/p>\n<p>For example, HPE uses the quad GH200 to power OFP-II, an advanced HPC system in Japan shared by the University of Tsukuba and the University of Tokyo, as well as the DeltaAI system, which will triple computing capacity for the U.S. National Center for Supercomputing Applications.<\/p>\n<p>HPE is also building the <a href=\"https:\/\/blogs.nvidia.com\/blog\/special-address-isc-2022-hpc\/\">Venado<\/a> system for the Los Alamos National Laboratory, the first GH200 to be deployed in the U.S. In addition, HPE is building GH200 supercomputers in the Middle East, Switzerland and the U.K.<\/p>\n<h2><b>Grace Hopper in Texas and Beyond<\/b><\/h2>\n<p>At the Texas Advanced Computing Center (TACC), Dell Technologies is building the Vista supercomputer with NVIDIA Grace Hopper and Grace CPU Superchips.<\/p>\n<p>More than 100 global enterprises and organizations, including NASA Ames Research Center and Total Energies, have already purchased Grace Hopper early-access systems, Buck said.<\/p>\n<p>They join previously announced GH200 users such as <a href=\"https:\/\/blogs.nvidia.com\/blog\/computex-keynote-generative-ai\/\">SoftBank<\/a> and the <a href=\"https:\/\/blogs.nvidia.com\/blog\/uk-largest-ai-supercomputer\/\">University of Bristol<\/a>, as well as the massive <a href=\"https:\/\/blogs.nvidia.com\/blog\/supercomputing-ai-eurohpc\/\">Leonardo<\/a> system with 14,000 NVIDIA A100 GPUs that delivers 10 exaflops of AI performance for Italy\u2019s Cineca consortium.<\/p>\n<h2><b>The View From Supercomputing Centers<\/b><\/h2>\n<p>Leaders from supercomputing centers around the world shared their plans and work in progress with the latest systems.<\/p>\n<p>\u201cWe\u2019ve been collaborating with MeteoSwiss ECMWP as well as scientists from ETH EXCLAIM and NVIDIA\u2019s Earth-2 project to create an infrastructure that will push the envelope in all dimensions of big data analytics and extreme scale computing,\u201d said Thomas Schultess, director of the Swiss National Supercomputing Centre of work on the Alps supercomputer.<\/p>\n<p>\u201cThere\u2019s really impressive energy-efficiency gains across our stacks,\u201d Dan Stanzione, executive director of TACC, said of Vista.<\/p>\n<p>It\u2019s \u201creally the stepping stone to move users from the kinds of systems we\u2019ve done in the past to looking at this new Grace Arm CPU and Hopper GPU tightly coupled combination and \u2026 we\u2019re looking to scale out by probably a factor of 10 or 15 from what we are deploying with Vista when we deploy Horizon in a couple years,\u201d he said.<\/p>\n<h2><b>Accelerating the Quantum Journey<\/b><\/h2>\n<p>Researchers are also using today\u2019s accelerated systems to pioneer a path to tomorrow\u2019s supercomputers.<\/p>\n<p>In Germany, JUPITER \u201cwill revolutionize scientific research across climate, materials, drug discovery and quantum computing,\u201d said Kristel Michelson, who leads Julich\u2019s research group on quantum information processing.<\/p>\n<p>\u201cJUPITER\u2019s architecture also allows for the seamless integration of quantum algorithms with parallel HPC algorithms, and this is mandatory for effective quantum HPC hybrid simulations,\u201d she said.<\/p>\n<h2><b>CUDA Quantum Drives Progress<\/b><\/h2>\n<p>The special address also showed how <a href=\"https:\/\/developer.nvidia.com\/cuda-quantum\">NVIDIA CUDA Quantum<\/a> \u2014 a platform for programming CPUs, GPUs and quantum computers also known as <a href=\"https:\/\/blogs.nvidia.com\/blog\/what-is-a-qpu\/\">QPUs<\/a> \u2014 is advancing research in quantum computing.<\/p>\n<p>For example, researchers at <a href=\"https:\/\/blogs.nvidia.com\/blog\/basf-cuda-quantum-momentum\/\">BASF<\/a>, the world\u2019s largest chemical company, pioneered a new hybrid quantum-classical method for simulating chemicals that can shield humans against harmful metals. They join researchers at Brookhaven National Laboratory and HPE who are separately pushing the frontiers of science with CUDA Quantum.<\/p>\n<p>NVIDIA also announced a collaboration with Classiq, a developer of quantum programming tools, to create a life sciences research center at the Tel Aviv Sourasky Medical Center, Israel\u2019s largest teaching hospital.\u00a0 The center will use Classiq\u2019s software and CUDA Quantum running on an <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/dgx-h100\/\">NVIDIA DGX H100 system<\/a>.<\/p>\n<p>Separately, Quantum Machines will deploy the first <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/dgx-quantum\/\">NVIDIA DGX Quantum<\/a>, a system using Grace Hopper Superchips, at the Israel National Quantum Center that aims to drive advances across scientific fields. The DGX system will be connected to a superconducting QPU by Quantware and a photonic QPU from ORCA Computing, both powered by CUDA Quantum.<\/p>\n<p><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/quantum-computing-partners.jpg\"><\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/quantum-computing-partners-672x402.jpg\" alt=\"Logos of NVIDIA CUDA Quantum partners\" width=\"672\" height=\"402\"><\/p>\n<p><\/a><\/p>\n<p>\u201cIn just two years, our NVIDIA quantum computing platform has amassed over 120 partners [above], a testament to its open, innovative platform,\u201d Buck said.<\/p>\n<p>Overall, the work across many fields of discovery reveals a new trend that combines accelerated computing at data center scale with NVIDIA\u2019s full-stack innovation.<\/p>\n<p>\u201cAccelerated computing is paving the path for sustainable computing with advancements that provide not just amazing technology but a more sustainable and impactful future,\u201d he concluded.<\/p>\n<p>Watch NVIDIA\u2019s<a href=\"http:\/\/www.nvidia.com\/sc23\"> SC23 special address<\/a> below.<\/p>\n<\/p>\n<p>\u00a0<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/efficient-ai-supercomputers-sc23\/<\/p>\n","protected":false},"author":0,"featured_media":3256,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3255"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=3255"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3255\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/3256"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=3255"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=3255"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=3255"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}