{"id":700,"date":"2020-12-16T22:28:40","date_gmt":"2020-12-16T22:28:40","guid":{"rendered":"https:\/\/machine-learning.webcloning.com\/2020\/12\/16\/nvidia-chief-scientist-highlights-new-ai-research-in-gtc-keynote\/"},"modified":"2020-12-16T22:28:40","modified_gmt":"2020-12-16T22:28:40","slug":"nvidia-chief-scientist-highlights-new-ai-research-in-gtc-keynote","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2020\/12\/16\/nvidia-chief-scientist-highlights-new-ai-research-in-gtc-keynote\/","title":{"rendered":"NVIDIA Chief Scientist Highlights New AI Research in GTC Keynote"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2020\/12\/14\/gtc-china-keynote-dally-ai\/\" data-title=\"NVIDIA Chief Scientist Highlights New AI Research in GTC Keynote\">\n<p>NVIDIA researchers are defining ways to make faster AI chips in systems with greater bandwidth that are easier to program, said Bill Dally, NVIDIA\u2019s chief scientist, in <a href=\"https:\/\/www.nvidia.com\/en-us\/gtc\/keynote\/\">a keynote<\/a> released today for a virtual GTC China event.<\/p>\n<p>He described three projects as examples of how the 200-person research team he leads is working to stoke Huang\u2019s Law \u2014 the prediction named for NVIDIA CEO Jensen Huang that GPUs will double AI performance every year.<\/p>\n<p>\u201cIf we really want to improve computer performance, Huang\u2019s Law is the metric that matters, and I expect it to continue for the foreseeable future,\u201d said Dally, who helped direct research at NVIDIA in AI, ray tracing and fast interconnects.<\/p>\n<figure id=\"attachment_48180\" aria-describedby=\"caption-attachment-48180\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/12\/Huangs-Law-slide-11-jpg.jpg\"><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/12\/Huangs-Law-slide-11-jpg-672x386.jpg\" alt=\"Huang's Law slide 11 jpg\" width=\"672\" height=\"386\"><\/a><figcaption id=\"caption-attachment-48180\" class=\"wp-caption-text\">NVIDIA has more than doubled performance of GPUs on AI inference every year.<\/figcaption><\/figure>\n<h2><b>An Ultra-Efficient Accelerator<\/b><\/h2>\n<p>Toward that end, NVIDIA researchers created a tool called MAGNet that generated an AI inference accelerator that hit 100 tera-operations per watt in a simulation. That\u2019s more than an order of magnitude greater efficiency than today\u2019s commercial chips.<\/p>\n<p>MAGNet uses new techniques to orchestrate the flow of information through a device in ways that minimize the data movement that burns most of the energy in today\u2019s chips. The research prototype is implemented as a modular set of tiles so it can scale flexibly.<\/p>\n<p>A separate effort seeks to replace today\u2019s electrical links inside systems with faster optical ones.<\/p>\n<h2><b>Firing on All Photons<\/b><\/h2>\n<p>\u201cWe can see our way to doubling the speed of our<a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/nvlink\/\"> NVLink<\/a> [that connects GPUs] and maybe doubling it again, but eventually electrical signaling runs out of gas,\u201d said Dally, who holds more than 120 patents and chaired the computer science department at Stanford before joining NVIDIA in 2009.<\/p>\n<p>The team is collaborating with researchers at Columbia University on ways to harness techniques telecom providers use in their core networks to merge dozens of signals onto a single optical fiber.<\/p>\n<p>Called dense wavelength division multiplexing, it holds the potential to pack multiple terabits per second into links that fit into a single millimeter of space on the side of a chip, more than 10x the density of today\u2019s interconnects.<\/p>\n<p>Besides faster throughput, the optical links enable denser systems. For example, Dally showed a mockup (below) of a future NVIDIA DGX system with more than 160 GPUs.<\/p>\n<figure id=\"attachment_48183\" aria-describedby=\"caption-attachment-48183\" class=\"wp-caption alignleft\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/12\/GPU-tray-with-optical-links-slide-73.jpg\"><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/12\/GPU-tray-with-optical-links-slide-73-400x275.jpg\" alt=\"GPU tray with optical links slide 73\" width=\"400\" height=\"275\"><\/a><figcaption id=\"caption-attachment-48183\" class=\"wp-caption-text\">Optical links help pack dozens of GPUs in a system.<\/figcaption><\/figure>\n<p>In software, NVIDIA\u2019s researchers have prototyped a new programming system called Legate. It lets developers take a program written for a single GPU and run it on a system of any size \u2014 even a giant supercomputer like<a href=\"https:\/\/blogs.nvidia.com\/blog\/2020\/08\/14\/making-selene-pandemic-ai\/\"> Selene<\/a> that packs thousands of GPUs.<\/p>\n<p>Legate couples a new form of programming shorthand with accelerated software libraries and an advanced runtime environment called Legion. It\u2019s already being put to the test at U.S. national labs.<\/p>\n<h2><b>Rendering a Vivid Future<\/b><\/h2>\n<p>The three research projects make up just one part of Dally\u2019s keynote, which describes NVIDIA\u2019s domain-specific platforms for a variety of industries such as healthcare, self-driving cars and robotics. He also delves into data science, AI and graphics.<\/p>\n<p>\u201cIn a few generations our products will produce amazing images in real time using path tracing with <a href=\"http:\/\/www.pbr-book.org\/\">physically based rendering<\/a>, and we\u2019ll be able to generate whole scenes with AI,\u201d said Dally.<\/p>\n<p>He showed the first public demonstration (below) that combines NVIDIA\u2019s conversational AI framework called <a href=\"https:\/\/developer.nvidia.com\/nvidia-jarvis\">Jarvis<\/a> with <a href=\"https:\/\/www.nvidia.com\/en-us\/research\/ai-playground\/\">GauGAN<\/a>, a tool that uses generative adversarial networks to create beautiful landscapes from simple sketches. The demo lets users instantly generate photorealistic landscapes using simple voice commands.<\/p>\n<\/p>\n<p>In an interview between recording sessions for the keynote, Dally expressed particular pride for the team\u2019s pioneering work in several areas.<\/p>\n<p>\u201cAll our current ray tracing started in NVIDIA Research with prototypes that got our product teams excited. And in 2011, I assigned [NVIDIA researcher] Bryan Catanzaro to work with [Stanford professor] Andrew Ng on a project that became <a href=\"https:\/\/developer.nvidia.com\/cuDNN\">CuDNN<\/a>, software that kicked off much of our work in deep learning,\u201d he said.<\/p>\n<h2><b>A First Foothold in Networking<\/b><\/h2>\n<p>Dally also spearheaded a collaboration that led to the first prototypes of<a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/nvlink\/\"> NVLink and NVSwitch<\/a>, interconnects that link GPUs running inside some of the world\u2019s largest supercomputers today.<\/p>\n<p>\u201cThe product teams grabbed the work out of our hands before we were ready to let go of it, and now we\u2019re considered one of the most advanced networking companies,\u201d he said.<\/p>\n<p>With his passion for technology, Dally said he often feels like a kid in a candy store. He may hop from helping a group with an AI accelerator one day to helping another team sort through a complex problem in robotics the next.<\/p>\n<p>\u201cI have one of the most fun jobs in the company if not in the world because I get to help shape the future,\u201d he said.<\/p>\n<p>The keynote is just one of more than 220 sessions at <a href=\"https:\/\/www.nvidia.cn\/gtc\/\">GTC China<\/a>. All the sessions are free and most are conducted in Mandarin.<\/p>\n<h2><b>Panel, Startup Showcase at GTC China<\/b><\/h2>\n<p>Following the keynote, <a href=\"https:\/\/www.nvidia.com\/en-us\/gtc\/keynote\/\">a panel of senior NVIDIA executives<\/a> will discuss how the company\u2019s technologies in AI, data science, healthcare and other fields are being adopted in China.<\/p>\n<p>The event also includes a showcase of a dozen top startups in China, hosted by<a href=\"https:\/\/www.nvidia.com\/en-us\/deep-learning-ai\/startups\/\"> NVIDIA Inception<\/a>, an acceleration program for AI and data science startups.<\/p>\n<p>Companies participating in GTC China include Alibaba, AWS, Baidu, ByteDance, China Telecom, Dell Technologies, Didi, H3C, Inspur, Kuaishou, Lenovo, Microsoft, Ping An, Tencent, Tsinghua University and Xiaomi.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>http:\/\/feedproxy.google.com\/~r\/nvidiablog\/~3\/SQNH4lsHcGU\/<\/p>\n","protected":false},"author":0,"featured_media":701,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/700"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=700"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/700\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/701"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=700"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=700"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=700"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}