{"id":3007,"date":"2023-05-29T06:06:31","date_gmt":"2023-05-29T06:06:31","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2023\/05\/29\/live-from-taipei-nvidia-ceo-unveils-gen-ai-platforms-for-every-industry\/"},"modified":"2023-05-29T06:06:31","modified_gmt":"2023-05-29T06:06:31","slug":"live-from-taipei-nvidia-ceo-unveils-gen-ai-platforms-for-every-industry","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2023\/05\/29\/live-from-taipei-nvidia-ceo-unveils-gen-ai-platforms-for-every-industry\/","title":{"rendered":"Live From Taipei: NVIDIA CEO Unveils Gen AI Platforms for Every Industry"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2023\/05\/28\/computex-keynote-generative-ai\/\" data-title=\"Live From Taipei: NVIDIA CEO Unveils Gen AI Platforms for Every Industry\" data-hashtags=\"\">\n<p>In his first live keynote since the pandemic, NVIDIA founder and CEO Jensen Huang today kicked off the COMPUTEX conference in Taipei, announcing platforms that companies can use to ride a historic wave of generative AI that\u2019s transforming industries from advertising to manufacturing to telecom.<\/p>\n<p>\u201cWe\u2019re back,\u201d Huang roared as he took the stage after years of virtual keynotes, some from his home kitchen. \u201cI haven\u2019t given a public speech in almost four years \u2014 wish me luck!\u201d<\/p>\n<p>Speaking for nearly two hours to a packed house of some 3,500, he described <a href=\"https:\/\/blogs.nvidia.com\/blog\/2021\/09\/01\/what-is-accelerated-computing\/\">accelerated computing<\/a> services, software and systems that are enabling new business models and making current ones more efficient.<\/p>\n<p>\u201cAccelerated computing and AI mark a reinvention of computing,\u201d said Huang, whose travels in his hometown over the past week have been tracked daily by local media.<\/p>\n<p>In a demonstration of its power, he used the massive 8K wall he spoke in front of to show a text prompt generating a theme song for his keynote, singable as any karaoke tune. Huang, who occasionally bantered with the crowd in his native Taiwanese, briefly led the audience in singing the new anthem.<\/p>\n<p>\u201cWe\u2019re now at the tipping point of a new computing era with accelerated computing and AI that\u2019s been embraced by almost every computing and cloud company in the world,\u201d he said, noting 40,000 large companies and 15,000 startups now use NVIDIA technologies with 25 million downloads of CUDA software last year alone.<\/p>\n<h2><b>Top News Announcements From the Keynote<\/b><\/h2>\n<h2><b>A New Engine for Enterprise AI<\/b><\/h2>\n<p>For enterprises that need the ultimate in AI performance, he unveiled <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/dgx-gh200\/\">DGX GH200<\/a>, a large-memory AI supercomputer. It uses <a href=\"https:\/\/blogs.nvidia.com\/blog\/2023\/03\/06\/what-is-nvidia-nvlink\/\">NVIDIA NVLink<\/a> to combine up to 256 <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/grace-hopper-superchip\/\">NVIDIA GH200 Grace Hopper Superchips<\/a> into a single data-center-sized GPU.<\/p>\n<p>The GH200 Superchip, which Jensen said is <a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-grace-hopper-superchips-designed-for-accelerated-generative-ai-enter-full-production\">now in full production<\/a>, combines an energy-efficient <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/grace-cpu\/\">NVIDIA Grace CPU<\/a> with a high-performance <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/h100\/\">NVIDIA H100 Tensor Core GPU<\/a> in one superchip.<\/p>\n<p>The DGX GH200 packs an exaflop of performance and 144 terabytes of shared memory, nearly 500x more than in a single NVIDIA DGX A100 320GB system. That lets developers build <a href=\"https:\/\/blogs.nvidia.com\/blog\/2022\/10\/10\/llms-ai-horizon\/\">large language models<\/a> for generative AI chatbots, complex algorithms for <a href=\"https:\/\/blogs.nvidia.com\/blog\/2022\/09\/20\/grace-hopper-recommender-systems\/\">recommender systems<\/a>, and <a href=\"https:\/\/blogs.nvidia.com\/blog\/2022\/10\/24\/what-are-graph-neural-networks\/\">graph neural networks<\/a> used for fraud detection and data analytics.<\/p>\n<p>Google Cloud, Meta and Microsoft are among the first expected to gain access to the DGX GH200, which can be used as a blueprint for future hyperscale generative AI infrastructure.<\/p>\n<figure id=\"attachment_64424\" aria-describedby=\"caption-attachment-64424\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/dgx-gh200-press-computex23-1920x1080-1.jpg\"><img decoding=\"async\" loading=\"lazy\" class=\"size-large wp-image-64424\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/dgx-gh200-press-computex23-1920x1080-1-672x257.jpg\" alt=\"NVIDIA DGX GH200\" width=\"672\" height=\"257\"><\/a><figcaption id=\"caption-attachment-64424\" class=\"wp-caption-text\">NVIDIA\u2019s DGX GH200 AI supercomputer delivers 1 exaflop of performance for generative AI.<\/figcaption><\/figure>\n<p>\u201cDGX GH200 AI supercomputers integrate NVIDIA\u2019s most advanced accelerated computing and networking technologies to expand the frontier of AI,\u201d Huang told the audience in Taipei, many of whom had lined up outside the hall for hours before the doors opened.<\/p>\n<p>NVIDIA is building its own massive AI supercomputer, NVIDIA Helios, coming online this year. It will use four DGX GH200 systems linked with <a href=\"https:\/\/www.nvidia.com\/en-us\/networking\/quantum2\/\">NVIDIA Quantum-2 InfiniBand<\/a> networking to supercharge data throughput for training large AI models.<\/p>\n<p>The DGX GH200 forms the pinnacle of hundreds of systems announced at the event. Together, they\u2019re bringing generative AI and accelerated computing to millions of users.<\/p>\n<p>Zooming out to the big picture, Huang announced <a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-grace-hopper-superchips-designed-for-accelerated-generative-ai-enter-full-production\">more than 400 system configurations<\/a> are coming to market powered by NVIDIA\u2019s latest <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/technologies\/hopper-architecture\/\">Hopper<\/a>, Grace, <a href=\"https:\/\/www.nvidia.com\/en-us\/geforce\/ada-lovelace-architecture\/\">Ada Lovelace<\/a> and <a href=\"https:\/\/www.nvidia.com\/en-us\/networking\/products\/data-processing-unit\/\">BlueField<\/a> architectures. They aim to tackle the most complex challenges in AI, data science and high performance computing.<\/p>\n<h2><b>Acceleration in Every Size<\/b><\/h2>\n<p>To fit the needs of data centers of every size, Huang announced <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/products\/mgx\/\">NVIDIA MGX<\/a>, a modular reference architecture for creating accelerated servers. System makers will use it to quickly and cost-effectively build more than a hundred different server configurations to suit a wide range of AI, HPC and <a href=\"https:\/\/www.nvidia.com\/en-us\/omniverse\/\">NVIDIA Omniverse<\/a> applications.<\/p>\n<p>MGX lets manufacturers build CPU and accelerated servers using a common architecture and modular components. It supports NVIDIA\u2019s full line of GPUs, CPUs, data processing units (<a href=\"https:\/\/blogs.nvidia.com\/blog\/2020\/05\/20\/whats-a-dpu-data-processing-unit\/\">DPUs<\/a>) and network adapters as well as x86 and Arm processors across a variety of air- and liquid-cooled chassis.<\/p>\n<p>QCT and Supermicro will be the first to market with MGX designs appearing in August. Supermicro\u2019s ARS-221GL-NR system announced at COMPUTEX will use the Grace CPU, while QCT\u2019s S74G-2U system, also announced at the event, uses Grace Hopper.<\/p>\n<p>ASRock Rack, ASUS, GIGABYTE and Pegatron will also use MGX to create next-generation accelerated computers.<\/p>\n<h2><b>5G\/6G Calls for Grace Hopper<\/b><\/h2>\n<p>Separately, Huang said NVIDIA is helping shape future 5G and 6G wireless and video communications. A demo showed how AI running on Grace Hopper will transform today\u2019s 2D video calls into more lifelike 3D experiences, providing an amazing sense of presence.<\/p>\n<p>Laying the groundwork for new kinds of services, Huang announced NVIDIA is working with telecom giant SoftBank to build a distributed network of data centers in Japan. It will deliver 5G services and generative AI applications on a common cloud platform.<\/p>\n<p>The data centers will use NVIDIA GH200 Superchips and <a href=\"https:\/\/www.nvidia.com\/content\/dam\/en-zz\/Solutions\/Data-Center\/documents\/datasheet-nvidia-bluefield-3-dpu.pdf\">NVIDIA BlueField-3 DPUs<\/a> in modular MGX systems as well as <a href=\"https:\/\/www.nvidia.com\/en-us\/networking\/ethernet-switching\/\">NVIDIA Spectrum Ethernet switches<\/a> to deliver the highly precise timing the 5G protocol requires. The platform will reduce cost by increasing spectral efficiency while reducing energy consumption.<\/p>\n<p>The systems will help SoftBank explore 5G applications in autonomous driving, AI factories, augmented and virtual reality, computer vision and digital twins. Future uses could even include 3D video conferencing and holographic communications.<\/p>\n<h2><b>Turbocharging Cloud Networks<\/b><\/h2>\n<p>Separately, Huang unveiled <a href=\"https:\/\/www.nvidia.com\/en-us\/networking\/spectrumx\/\">NVIDIA Spectrum-X<\/a>, a networking platform purpose-built to improve the performance and efficiency of Ethernet-based AI clouds. It combines Spectrum-4 Ethernet switches with BlueField-3 DPUs and software to deliver 1.7x gains in AI performance and power efficiency over traditional Ethernet fabrics.<\/p>\n<p><a href=\"https:\/\/developer.nvidia.com\/blog\/turbocharging-ai-workloads-with-nvidia-spectrum-x-networking-platform\/\">NVIDIA Spectrum-X<\/a>, Spectrum-4 switches and BlueField-3 DPUs are available now from system makers including Dell Technologies, Lenovo and Supermicro.<\/p>\n<figure id=\"attachment_64421\" aria-describedby=\"caption-attachment-64421\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/ethernet-switches-spectrum-x-press-1920x1080-1.jpg\"><img decoding=\"async\" loading=\"lazy\" class=\"size-large wp-image-64421\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/ethernet-switches-spectrum-x-press-1920x1080-1-672x340.jpg\" alt=\"NVIDIA Spectrum-X for Ethernet AI clouds\" width=\"672\" height=\"340\"><\/a><figcaption id=\"caption-attachment-64421\" class=\"wp-caption-text\">NVIDIA Spectrum-X accelerates AI workflows that can experience performance losses on traditional Ethernet networks.<\/figcaption><\/figure>\n<h2><b>Bringing Game Characters to Life<\/b><\/h2>\n<p>Generative AI impacts how people play, too.<\/p>\n<p>Huang announced <a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-ace-for-games-sparks-life-into-virtual-characters-with-generative-ai\">NVIDIA Avatar Cloud Engine (ACE) for Games<\/a>, a foundry service developers can use to build and deploy custom AI models for speech, conversation and animation. It will give non-playable characters conversational skills so they can respond to questions with lifelike personalities that evolve.<\/p>\n<p>NVIDIA ACE for Games includes AI <a href=\"https:\/\/blogs.nvidia.com\/blog\/2023\/03\/13\/what-are-foundation-models\/\">foundation models<\/a> such as <a href=\"https:\/\/www.nvidia.com\/en-us\/ai-data-science\/products\/riva\/\">NVIDIA Riva<\/a> to detect and transcribe the player\u2019s speech. The text prompts <a href=\"https:\/\/www.nvidia.com\/en-us\/gpu-cloud\/nemo-llm-service\/\">NVIDIA NeMo<\/a> to generate customized responses animated with <a href=\"https:\/\/www.nvidia.com\/en-us\/omniverse\/apps\/audio2face\/\">NVIDIA Omniverse Audio2Face<\/a>.<\/p>\n<figure id=\"attachment_64418\" aria-describedby=\"caption-attachment-64418\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/NVIDIA-ACE-for-Games.jpg\"><img decoding=\"async\" loading=\"lazy\" class=\"size-large wp-image-64418\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/NVIDIA-ACE-for-Games-672x286.jpg\" alt=\"NVIDIA ACE for Games\" width=\"672\" height=\"286\"><\/a><figcaption id=\"caption-attachment-64418\" class=\"wp-caption-text\">NVIDIA ACE for Games provides a tool chain for bringing characters to life with generative AI.<\/figcaption><\/figure>\n<h2><b>Accelerating Gen AI on Windows<\/b><\/h2>\n<p>Huang described how NVIDIA and Microsoft are <a href=\"https:\/\/blogs.nvidia.com\/blog\/2023\/05\/28\/computex-generative-ai-rtx\/\">collaborating to drive innovation<\/a> for Windows PCs in the generative AI era.<\/p>\n<p>New and enhanced tools, frameworks and drivers are making it easier for PC developers to develop and deploy AI. For example, the Microsoft Olive toolchain for optimizing and deploying GPU-accelerated AI models and new graphics drivers will boost DirectML performance on Windows PCs with NVIDIA GPUs.<\/p>\n<p>The collaboration will enhance and extend an installed base of 100 million PCs sporting RTX GPUs with Tensor Cores that boost performance of more than 400 AI-accelerated Windows apps and games.<\/p>\n<h2><b>Digitizing the World\u2019s Largest Industries<\/b><\/h2>\n<p>Generative AI is also spawning new opportunities in the $700 billion digital advertising industry.<\/p>\n<p>For example, WPP, the world\u2019s largest marketing services organization, is working with NVIDIA to build a first-of-its kind generative AI-enabled content engine on <a href=\"https:\/\/www.nvidia.com\/en-us\/omniverse\/cloud\/\">Omniverse Cloud<\/a>.<\/p>\n<p>In a demo, Huang showed how creative teams will connect their 3D design tools such as Adobe Substance 3D, to build digital twins of client products in NVIDIA Omniverse. Then, content from generative AI tools trained on responsibly sourced data and built with<a href=\"https:\/\/www.nvidia.com\/en-us\/gpu-cloud\/picasso\/\"> NVIDIA Picasso<\/a> will let them quickly produce virtual sets. WPP clients can then use the complete scene to generate a host of ads, videos and 3D experiences for global markets and users to experience on any web device.<\/p>\n<p>\u201cToday ads are retrieved, but in the future when you engage information much of it will be generated \u2014 the computing model has changed,\u201d Huang said.<\/p>\n<h2><b>Factories Forge an AI Future<\/b><\/h2>\n<p>With an estimated 10 million factories, the $46 trillion manufacturing sector is a rich field for industrial digitalization.<\/p>\n<p>\u201cThe world\u2019s largest industries make physical things. Building them digitally first can save billions,\u201d said Huang.<\/p>\n<p>The keynote showed how electronics makers including Foxconn Industrial Internet, Innodisk, Pegatron, Quanta and Wistron are forging digital workflows with NVIDIA technologies to realize the vision of an entirely digital smart factory.<\/p>\n<p>They\u2019re using Omniverse and generative AI APIs to connect their design and manufacturing tools so they can build digital twins of factories. In addition, they use <a href=\"https:\/\/developer.nvidia.com\/isaac-sim\">NVIDIA Isaac Sim<\/a> for simulating and testing robots and <a href=\"https:\/\/www.nvidia.com\/en-us\/autonomous-machines\/intelligent-video-analytics-platform\/\">NVIDIA Metropolis<\/a>, a vision AI framework, for automated optical inspection.<\/p>\n<p>The latest component, <a href=\"https:\/\/developer.nvidia.com\/metropolis-for-factories\">NVIDIA Metropolis for Factories<\/a>, can create custom quality-control systems, giving manufacturers a competitive advantage. It\u2019s helping companies develop state-of-the-art AI applications.<\/p>\n<h2><b>AI Speeds Assembly Lines<\/b><\/h2>\n<p>For example, Pegatron \u2014 which makes 300 products worldwide, including laptops and smartphones \u2014 is creating virtual factories with Omniverse, Isaac Sim and Metropolis. That lets it try out processes in a simulated environment, saving time and cost.<\/p>\n<p>Pegatron also used the <a href=\"https:\/\/developer.nvidia.com\/deepstream-sdk\">NVIDIA DeepStream software development kit<\/a> to develop intelligent video applications that led to a 10x improvement in throughput.<\/p>\n<p>Foxconn Industrial Internet, a service arm of the world\u2019s largest technology manufacturer, is working with NVIDIA Metropolis partners to automate significant portions of its circuit-board quality-assurance inspection points.<\/p>\n<figure id=\"attachment_64430\" aria-describedby=\"caption-attachment-64430\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/Crowds-lined-up-scaled.jpg\"><img decoding=\"async\" loading=\"lazy\" class=\"size-large wp-image-64430\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/Crowds-lined-up-672x401.jpg\" alt=\"Computex 2023 keynote\" width=\"672\" height=\"401\"><\/a><figcaption id=\"caption-attachment-64430\" class=\"wp-caption-text\">Crowds lined up for the keynote hours before doors opened.<\/figcaption><\/figure>\n<p>In a video, Huang showed how Techman Robot, a subsidiary of Quanta, tapped NVIDIA Isaac Sim to optimize inspection on the Taiwan-based giant\u2019s manufacturing lines. It\u2019s essentially using simulated robots to train robots how to make better robots.<\/p>\n<p>In addition, Huang <a href=\"https:\/\/blogs.nvidia.com\/blog\/2023\/05\/28\/isaac-amr-nova-orin-autonomous-mobile-robots\">announced<\/a> a new platform to enable the next generation of autonomous mobile robot (AMR) fleets. <a href=\"https:\/\/www.developer.nvidia.com\/isaac\/amr\">Isaac AMR<\/a> helps simulate, deploy and manage fleets of autonomous mobile robots.<\/p>\n<p>A large partner ecosystem \u2014 including ADLINK, Aetina, Deloitte, Quantiphi and Siemens \u2014 is helping bring all these manufacturing solutions to market, Huang said.<\/p>\n<p>It\u2019s one more example of how NVIDIA is helping companies feel the benefits of generative AI with accelerated computing.<\/p>\n<p>\u201cIt\u2019s been a long time since I\u2019ve seen you, so I had a lot to tell you,\u201d he said after the two-hour talk to enthusiastic applause.<\/p>\n<p>To learn more, watch the <a href=\"https:\/\/www.nvidia.com\/en-us\/events\/computex\/\">full keynote<\/a>.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/2023\/05\/28\/computex-keynote-generative-ai\/<\/p>\n","protected":false},"author":0,"featured_media":3008,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3007"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=3007"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3007\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/3008"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=3007"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=3007"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=3007"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}