{"id":3389,"date":"2024-03-18T23:44:10","date_gmt":"2024-03-18T23:44:10","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2024\/03\/18\/we-created-a-processor-for-the-generative-ai-era-nvidia-ceo-says\/"},"modified":"2024-03-18T23:44:10","modified_gmt":"2024-03-18T23:44:10","slug":"we-created-a-processor-for-the-generative-ai-era-nvidia-ceo-says","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2024\/03\/18\/we-created-a-processor-for-the-generative-ai-era-nvidia-ceo-says\/","title":{"rendered":"\u201cWe Created a Processor for the Generative AI Era,\u201d NVIDIA CEO Says"},"content":{"rendered":"<div id=\"bsf_rt_marker\">\n<p>Generative AI promises to revolutionize every industry it touches \u2014 all that\u2019s been needed is the technology to meet the challenge.<\/p>\n<p>NVIDIA founder and CEO Jensen Huang on Monday introduced that technology \u2014 the company\u2019s new Blackwell computing platform \u2014 as he outlined the major advances that increased computing power can deliver for everything from software to services, robotics to medical technology and more.<\/p>\n<p>\u201cAccelerated computing has reached the tipping point \u2014 general purpose computing has run out of steam,\u201d Huang told more than 11,000 GTC attendees gathered in-person \u2014 and many tens of thousands more online \u2014 for his keynote address at Silicon Valley\u2019s cavernous SAP Center arena.<\/p>\n<p>\u201cWe need another way of doing computing \u2014 so that we can continue to scale so that we can continue to drive down the cost of computing, so that we can continue to consume more and more computing while being sustainable. Accelerated computing is a dramatic speedup over general-purpose computing, in every single industry.\u201d<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2024\/03\/MC1_4926-scaled.jpg\" alt=\"\" width=\"2048\" height=\"1365\"><\/p>\n<p>Huang spoke in front of massive images on a 40-foot tall, 8K screen the size of a tennis court to a crowd packed with CEOs and developers, AI enthusiasts and entrepreneurs, who walked together 20 minutes to the arena from the San Jose Convention Center on a dazzling spring day.<\/p>\n<p>Delivering a massive upgrade to the world\u2019s AI infrastructure, Huang introduced the NVIDIA Blackwell platform to unleash real-time generative AI on trillion-parameter large language models.<\/p>\n<p>Huang presented NVIDIA NIM \u2014 a reference to NVIDIA inference microservices \u2014 a new way of packaging and delivering software that connects developers with hundreds of millions of GPUs to deploy custom AI of all kinds.<\/p>\n<p>And bringing AI into the physical world, Huang introduced Omniverse Cloud APIs to deliver advanced simulation capabilities.<\/p>\n<\/p>\n<p>Huang punctuated these major announcements with powerful demos, partnerships with some of the world\u2019s largest enterprises and more than a score of announcements detailing his vision.<\/p>\n<p><a href=\"https:\/\/www.nvidia.com\/gtc\/\">GTC<\/a> \u2014 which in 15 years has grown from the confines of a local hotel ballroom to the world\u2019s most important AI conference \u2014 is returning to a physical event for the first time in five years.<\/p>\n<p>This year\u2019s has over 900 sessions \u2014 including a panel discussion on transformers moderated by Huang with the eight pioneers who first developed the technology, more than 300 exhibits and 20-plus technical workshops.<\/p>\n<p>It\u2019s an event that\u2019s at the intersection of AI and just about everything. In a stunning opening act to the keynote, Refik Anadol, the world\u2019s leading AI artist, showed a massive real-time AI data sculpture with wave-like swirls in greens, blues, yellows and reds, crashing, twisting and unraveling across the screen.<\/p>\n<p>As he kicked off his talk, Huang explained that the rise of multi-modal AI \u2014 able to process diverse data types handled by different models \u2014 gives AI greater adaptability and power. By increasing their parameters, these models can handle more complex analyses.<\/p>\n<p>But this also means a significant rise in the need for computing power. And as these collaborative, multi-modal systems become more intricate \u2014 with as many as a trillion parameters \u2014 the demand for advanced computing infrastructure intensifies.<\/p>\n<p>\u201cWe need even larger models,\u201d Huang said. \u201cWe\u2019re going to train it with multimodality data, not just text on the internet, we\u2019re going to train it on texts and images, graphs and charts, and just as we learned watching TV, there\u2019s going to be a whole bunch of watching video.\u201d<\/p>\n<h2>The Next Generation of Accelerated Computing<\/h2>\n<p>In short, Huang said \u201cwe need bigger GPUs.\u201d The Blackwell platform is built to meet this challenge. Huang pulled a Blackwell chip out of his pocket and held it up side-by-side with a Hopper chip, which it dwarfed.<\/p>\n<p>Named for David Harold Blackwell \u2014 a University of California, Berkeley mathematician specializing in game theory and statistics, and the first Black scholar inducted into the National Academy of Sciences \u2014 the new architecture succeeds the NVIDIA Hopper architecture, launched two years ago.<\/p>\n<p>Blackwell delivers 2.5x its predecessor\u2019s performance in FP8 for training, per chip, and 5x with FP4 for inference. It features a fifth-generation NVLink interconnect that\u2019s twice as fast as Hopper and scales up to 576 GPUs.<\/p>\n<p>And the <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/gb200-nvl72\/\">NVIDIA GB200 Grace Blackwell Superchip<\/a> connects two Blackwell <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/b200\/\">NVIDIA B200 Tensor Core GPUs<\/a> to the NVIDIA Grace CPU over a 900GB\/s ultra-low-power NVLink chip-to-chip interconnect.<\/p>\n<p>Huang held up a board with the system. \u201cThis computer is the first of its kind where this much computing fits into this small of a space,\u201d Huang said. \u201cSince this is memory coherent, they feel like it\u2019s one big happy family working on one application together.\u201d<\/p>\n<p>For the highest AI performance, GB200-powered systems can be connected with the NVIDIA Quantum-X800 InfiniBand and Spectrum-X800 Ethernet platforms, also <a href=\"https:\/\/nvidianews.nvidia.com\/news\/networking-switches-gpu-computing-ai\">announced today<\/a>, which deliver advanced networking at speeds up to 800Gb\/s.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2024\/03\/VIVY0652-1-scaled.jpg\" alt=\"\" width=\"2048\" height=\"1365\"><\/p>\n<p>\u201cThe amount of energy we save, the amount of networking bandwidth we save, the amount of wasted time we save, will be tremendous,\u201d Huang said. \u201cThe future is generative \u2026 which is why this is a brand new industry. The way we compute is fundamentally different. We created a processor for the generative AI era.\u201d<\/p>\n<p>To scale up Blackwell, NVIDIA built a new chip called NVLink Switch. Each can connect four NVLink interconnects at 1.8 terabytes per second and eliminate traffic by doing in-network reduction.<\/p>\n<p>NVIDIA Switch and GB200 are key components of what Huang described as \u201cone giant GPU,\u201d the <a href=\"https:\/\/developer.nvidia.com\/blog\/nvidia-gb200-nvl72-delivers-trillion-parameter-llm-training-and-real-time-inference\/\">NVIDIA GB200 NVL72<\/a>, a multi-node, liquid-cooled, rack-scale system that harnesses Blackwell to offer supercharged compute for trillion-parameter models, with 720 petaflops of AI training performance and 1.4 exaflops of AI inference performance in a single rack.<\/p>\n<p>\u201cThere are only a couple, maybe three exaflop machines on the planet as we speak,\u201d Huang said of the machine, which packs 600,000 parts and weighs 3,000 pounds. \u201cAnd so this is an exaflop AI system in one single rack. Well let\u2019s take a look at the back of it.\u201d<\/p>\n<p>Going even bigger, NVIDIA today also announced its next-generation AI supercomputer \u2014 the <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/dgx-superpod-gb200\/\">NVIDIA DGX SuperPOD powered by NVIDIA GB200 Grace Blackwell Superchips<\/a> \u2014 for processing trillion-parameter models with constant uptime for superscale generative AI training and inference workloads.<\/p>\n<p>Featuring a new, highly efficient, liquid-cooled rack-scale architecture, the new DGX SuperPOD is built with NVIDIA DG GB200 systems and provides 11.5 exaflops of AI supercomputing at FP4 precision and 240 terabytes of fast memory \u2014 scaling to more with additional racks.<\/p>\n<p>\u201cIn the future, data centers are going to be thought of \u2026 as AI factories,\u201d Huang said. \u201cTheir goal in life is to generate revenues, in this case, intelligence.\u201d<\/p>\n<p>The industry has already embraced Blackwell.<\/p>\n<p>The <a href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-blackwell-platform-arrives-to-power-a-new-era-of-computing\">press release announcing Blackwell<\/a> includes endorsements from Alphabet and Google CEO Sundar Pichai, Amazon CEO Andy Jassy, Dell CEO Michael Dell, Google DeepMind CEO Demis Hassabis, Meta CEO Mark Zuckerberg, Microsoft CEO Satya Nadella, OpenAI CEO Sam Altman, Oracle Chairman Larry Ellison, and Tesla and xAI CEO Elon Musk.<\/p>\n<p>Blackwell is being adopted by every major global cloud services provider,\u00a0 pioneering AI companies, system and server vendors, and regional cloud service providers and telcos all around the world.<\/p>\n<p>\u201cThe whole industry is gearing up for Blackwell,\u201d which Huang said would be the most successful launch in the company\u2019s history.<\/p>\n<h2>A New Way to Create Software<\/h2>\n<p>Generative AI changes the way applications are written, Huang said.<\/p>\n<p>Rather than writing software, he explained, companies will assemble AI models, give them missions, give examples of work products, review plans and intermediate results.<\/p>\n<p>These packages \u2014 NVIDIA NIMs \u2014 are built from NVIDIA\u2019s accelerated computing libraries and generative AI models, Huang explained.<\/p>\n<p>\u201cHow do we build software in the future? It is unlikely that you\u2019ll write it from scratch or write a whole bunch of Python code or anything like that,\u201d Huang said. \u201cIt is very likely that you assemble a team of AIs.\u201d<\/p>\n<p>The microservices support industry-standard APIs so they are easy to connect, work across NVIDIA\u2019s large CUDA installed base, are re-optimized for new GPUs, and are constantly scanned for security vulnerabilities and exposures.<\/p>\n<p>Huang said customers can use NIM microservices off the shelf, or NVIDIA can help build proprietary AI and copilots, teaching a model specialized skills only a specific company would know to create invaluable new services.<\/p>\n<p>\u201cThe enterprise IT industry is sitting on a goldmine,\u201d Huang said. \u201cThey have all these amazing tools (and data) that have been created over the years. If they could take that goldmine and turn it into copilots, these copilots can help us do things.\u201d<\/p>\n<p>Major tech players are already putting it to work. Huang detailed how NVIDIA is already helping Cohesity, NetApp, SAP, ServiceNow and Snowflake build copilots and virtual assistants. And industries are stepping in, as well.<\/p>\n<p>In telecom, Huang announced the NVIDIA 6G Research Cloud, a generative AI and Omniverse-powered platform to advance the next communications era. It\u2019s built with NVIDIA\u2019s Sionna neural radio framework, NVIDIA Aerial CUDA-accelerated radio access network and the NVIDIA Aerial Omniverse Digital Twin for 6G.<\/p>\n<p>In semiconductor design and manufacturing, Huang announced that, in collaboration with TSMC and Synopsys, NVIDIA is bringing its breakthrough computational lithography platform, cuLitho, to production. This platform will accelerate the most compute-intensive workload in semiconductor manufacturing by 40-60x.<\/p>\n<p>Huang also announced the NVIDIA Earth Climate Digital Twin. The cloud platform \u2014 available now \u2014 enables interactive, high-resolution simulation to accelerate climate and weather prediction.<\/p>\n<p>The greatest impact of AI will be in healthcare, Huang said, explaining that NVIDIA is already in imaging systems, in gene sequencing instruments and working with leading surgical robotics companies.<\/p>\n<p>NVIDIA is launching a new type of biology software. NVIDIA today launched more than <a href=\"https:\/\/nvidianews.nvidia.com\/news\/generative-ai-microservices-for-developers\">two dozen new microservices<\/a> that allow healthcare enterprises worldwide to take advantage of the latest advances in generative AI from anywhere and on any cloud. They offer advanced imaging, natural language and speech recognition, and digital biology generation, prediction and simulation.<\/p>\n<h2>Omniverse Brings AI to the Physical World<\/h2>\n<p>The next wave of AI will be AI learning about the physical world, Huang said.<\/p>\n<p>\u201cWe need a simulation engine that represents the world digitally for the robot so that the robot has a gym to go learn how to be a robot,\u201d he said. \u201cWe call that virtual world Omniverse.\u201d<\/p>\n<p>That\u2019s why NVIDIA today announced that <a href=\"https:\/\/www.nvidia.com\/en-us\/omniverse\/\">NVIDIA Omniverse Cloud<\/a> will be available as APIs, extending the reach of the world\u2019s leading platform for creating industrial digital twin applications and workflows across the entire ecosystem of software makers.<\/p>\n<p>The five new Omniverse Cloud application programming interfaces enable developers to easily integrate core Omniverse technologies directly into existing design and automation software applications for digital twins, or their simulation workflows for testing and validating autonomous machines like robots or self-driving vehicles.<\/p>\n<p>To show how this works, Huang shared a demo of a robotic warehouse \u2014 using multi-camera perception and tracking \u2014 watching over workers and orchestrating robotic forklifts, which are driving autonomously with the full robotic stack running.<\/p>\n<p>Huang also announced that NVIDIA is bringing Omniverse to Apple Vision Pro, with the new Omniverse Cloud APIs letting developers stream interactive industrial digital twins into the VR headsets.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2024\/03\/MJC_2441-2-scaled.jpg\" alt=\"\" width=\"2048\" height=\"1365\"><\/p>\n<p>Some of the world\u2019s largest industrial software makers are embracing Omniverse Cloud APIs, including Ansys, Cadence, Dassault Syst\u00e8mes for its 3DEXCITE brand, Hexagon, Microsoft, Rockwell Automation, Siemens and Trimble.<\/p>\n<h2>Robotics<\/h2>\n<p>Everything that moves will be robotic, Huang said. The automotive industry will be a big part of that. NVIDIA computers are already in cars, trucks, delivery bots and robotaxis.<\/p>\n<p>Huang announced that BYD, the world\u2019s largest autonomous vehicle company, has selected NVIDIA\u2019s next-generation computer for its AV, building its next-generation EV fleets on DRIVE Thor.<\/p>\n<p>To help robots better see their environment, Huang also announced the Isaac Perceptor software development kit with state-of-the-art multi-camera visual odometry, 3D reconstruction and occupancy map, and depth perception.<\/p>\n<p>And to help make manipulators, or robotic arms, more adaptable, NVIDIA is announcing Isaac Manipulator \u2014 a state-of-the-art robotic arm perception, path planning and kinematic control library.<\/p>\n<p>Finally, Huang announced Project GR00T, a general-purpose foundation model for humanoid robots, designed to further the company\u2019s work driving breakthroughs in robotics and embodied AI.<\/p>\n<p>Supporting that effort, Huang unveiled a new computer, Jetson Thor, for humanoid robots based on the NVIDIA Thor system-on-a-chip and significant upgrades to the NVIDIA Isaac robotics platform.<\/p>\n<p>In his closing minutes, Huang brought on stage a pair of diminutive NVIDIA-powered robots from Disney Research.<\/p>\n<p>\u201cThe soul of NVDIA \u2014 the intersection of computer graphics, physics, artificial intelligence,\u201d he said. \u201cIt all came to bear at this moment.\u201d<\/p>\n<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/2024-gtc-keynote\/<\/p>\n","protected":false},"author":0,"featured_media":3390,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3389"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=3389"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3389\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/3390"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=3389"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=3389"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=3389"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}