{"id":3279,"date":"2023-11-28T19:43:07","date_gmt":"2023-11-28T19:43:07","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2023\/11\/28\/embracing-transformation-aws-and-nvidia-forge-ahead-in-generative-ai-and-cloud-innovation\/"},"modified":"2023-11-28T19:43:07","modified_gmt":"2023-11-28T19:43:07","slug":"embracing-transformation-aws-and-nvidia-forge-ahead-in-generative-ai-and-cloud-innovation","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2023\/11\/28\/embracing-transformation-aws-and-nvidia-forge-ahead-in-generative-ai-and-cloud-innovation\/","title":{"rendered":"Embracing Transformation: AWS and NVIDIA Forge Ahead in Generative AI and Cloud Innovation"},"content":{"rendered":"<div id=\"bsf_rt_marker\">\n<p>Amazon Web Services and NVIDIA will bring the latest generative AI technologies to enterprises worldwide.<\/p>\n<p>Combining AI and cloud computing, NVIDIA founder and CEO Jensen Huang joined AWS CEO Adam Selipsky Tuesday on stage at AWS re:Invent 2023 at the Venetian Expo Center in Las Vegas.<\/p>\n<p>Selipsky said he was \u201cthrilled\u201d to announce the expansion of the partnership between AWS and NVIDIA with more offerings that will deliver advanced graphics, machine learning and generative AI infrastructure.<\/p>\n<p>The two announced that AWS will be the first cloud provider to adopt the latest NVIDIA GH200 NVL32 Grace Hopper Superchip with new multi-node NVLink technology, that AWS is bringing <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/dgx-cloud\/\">NVIDIA DGX Cloud<\/a> to AWS, and that AWS has integrated some of NVIDIA\u2019s most popular software libraries.<\/p>\n<p>Huang started the conversation by highlighting the integration of key NVIDIA libraries with AWS, encompassing a range from <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/products\/ai-enterprise\/\">NVIDIA AI Enterprise<\/a> to <a href=\"https:\/\/developer.nvidia.com\/cuquantum-sdk\">cuQuantum<\/a> to <a href=\"https:\/\/blogs.nvidia.com\/blog\/bionemo-on-aws-generative-ai-drug-discovery\/\">BioNeMo<\/a>, catering to domains like data processing, quantum computing and digital biology.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Slide1.png\" alt=\"\" width=\"2048\" height=\"1152\"><\/p>\n<p>The partnership opens AWS to millions of developers and the nearly 40,000 companies who are using these libraries, Huang said, adding that it\u2019s great to see AWS expand its cloud instance offerings to include NVIDIA\u2019s new L4, L40S and, soon, H200 GPUs.<\/p>\n<p>Selipsky then introduced the AWS debut of the NVIDIA GH200 Grace Hopper Superchip, a significant advancement in cloud computing, and prompted Huang for further details.<\/p>\n<p>\u201cGrace Hopper, which is GH200, connects two revolutionary processors together in a really unique way,\u201d Huang said. He explained that the GH200 connects NVIDIA\u2019s Grace Arm CPU with its H200 GPU using a chip-to-chip interconnect called NVLink, at an astonishing one terabyte per second.<\/p>\n<p>Each processor has direct access to the high-performance HBM and efficient LPDDR5X memory. This configuration results in 4 petaflops of processing power and 600GB of memory for each superchip.<\/p>\n<p>AWS and NVIDIA connect 32 Grace Hopper Superchips in each rack using a new NVLink switch. Each 32 GH200 NVLink-connected node can be a single Amazon EC2 instance. When these are integrated with AWS Nitro and EFA networking, customers can connect GH200 NVL32 instances to scale to thousands of GH200 Superchips<\/p>\n<p>\u201cWith AWS Nitro, that becomes basically one giant virtual GPU instance,\u201d Huang said.<\/p>\n<p>The combination of AWS expertise in highly scalable cloud computing plus NVIDIA innovation with Grace Hopper will make this an amazing platform that delivers the highest performance for complex generative AI workloads, Huang said.<\/p>\n<p>\u201cIt\u2019s great to see the infrastructure, but it extends to the software, the services and all the other workflows that they have,\u201d Selipsky said, introducing NVIDIA DGX Cloud on AWS.<\/p>\n<p>This partnership will bring about the first DGX Cloud AI supercomputer powered by the GH200 Superchips, demonstrating the power of AWS\u2019s cloud infrastructure and NVIDIA\u2019s AI expertise.<\/p>\n<p>Following up, Huang announced that this new DGX Cloud supercomputer design in AWS, codenamed Project Ceiba, will serve as NVIDIA\u2019s newest AI supercomputer as well, for its own AI research and development.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/11\/Slide2.png\" alt=\"\" width=\"3840\" height=\"2160\"><\/p>\n<p>Named after the majestic Amazonian Ceiba tree, the Project Ceiba DGX Cloud cluster incorporates 16,384 GH200 Superchips to achieve 65 exaflops of AI processing power, Huang said.<\/p>\n<p>Ceiba will be the world\u2019s first GH200 NVL32 AI supercomputer built and the newest AI supercomputer in NVIDIA DGX Cloud, Huang said.<\/p>\n<p>Huang described Project Ceiba AI supercomputer as \u201cutterly incredible,\u201d saying it will be able to reduce the training time of the largest language models by half.<\/p>\n<p>NVIDIA\u2019s AI engineering teams will use this new supercomputer in DGX Cloud to advance AI for graphics, LLMs, image\/video\/3D generation, digital biology, robotics, self-driving cars, <a href=\"https:\/\/www.nvidia.com\/en-us\/high-performance-computing\/earth-2\/\">Earth-2 climate prediction<\/a> and more, Huang said.<\/p>\n<p>\u201cDGX is NVIDIA\u2019s cloud AI factory,\u201d Huang said, noting that AI is now key to doing NVIDIA\u2019s own work in everything from computer graphics to creating digital biology models to robotics to climate simulation and modeling.<\/p>\n<p>\u201cDGX Cloud is also our AI factory to work with enterprise customers to build custom AI models,\u201d Huang said. \u201cThey bring data and domain expertise; we bring AI technology and infrastructure.\u201d<\/p>\n<p>In addition, Huang also announced that AWS will be bringing four Amazon EC2 instances based on the NVIDIA GH200 NVL, H200, L40S, L4 GPUs, coming to market early next year.<\/p>\n<p>Selipsky wrapped up the conversation by announcing that GH200-based instances and DGX Cloud will be available on AWS in the coming year.<br \/><b><br \/><\/b><i>You can catch the discussion and Selipsky\u2019s entire keynote on AWS\u2019s YouTube channel.\u00a0<\/i><\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/aws-nvidia\/<\/p>\n","protected":false},"author":0,"featured_media":3280,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3279"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=3279"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3279\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/3280"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=3279"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=3279"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=3279"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}