{"id":3005,"date":"2023-05-29T03:47:35","date_gmt":"2023-05-29T03:47:35","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2023\/05\/29\/nvidia-brings-new-generative-ai-capabilities-groundbreaking-performance-to-100-million-windows-rtx-pcs-and-workstations\/"},"modified":"2023-05-29T03:47:35","modified_gmt":"2023-05-29T03:47:35","slug":"nvidia-brings-new-generative-ai-capabilities-groundbreaking-performance-to-100-million-windows-rtx-pcs-and-workstations","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2023\/05\/29\/nvidia-brings-new-generative-ai-capabilities-groundbreaking-performance-to-100-million-windows-rtx-pcs-and-workstations\/","title":{"rendered":"NVIDIA Brings New Generative AI Capabilities, Groundbreaking Performance to 100 Million Windows RTX PCs and Workstations"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2023\/05\/28\/computex-generative-ai-rtx\/\" data-title=\"NVIDIA Brings New Generative AI Capabilities, Groundbreaking Performance to 100 Million Windows RTX PCs and Workstations\" data-hashtags=\"\">\n<p><a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/data-science\/generative-ai\/\">Generative AI<\/a> is rapidly ushering in a new era of computing for productivity, content creation, gaming and more. Generative AI models and applications \u2014 like NVIDIA NeMo and DLSS 3 Frame Generation, Meta LLaMa, ChatGPT, Adobe Firefly and Stable Diffusion \u2014 use neural networks to identify patterns and structures within existing data to generate new and original content.<\/p>\n<p>When optimized for GeForce RTX and NVIDIA RTX GPUs, which offer up to 1,400 Tensor TFLOPS for AI inferencing, generative AI models can run up to 5x faster than on competing devices. This is thanks to Tensor Cores \u2014 dedicated hardware in RTX GPUs built to accelerate AI calculations \u2014 and regular software improvements. <a href=\"https:\/\/blogs.nvidia.com\/blog\/2023\/05\/23\/microsoft-build-nvidia-ai-windows-rtx\/\">Enhancements introduced last week<\/a> at the Microsoft Build conference doubled performance for generative AI models, such as Stable Diffusion, that take advantage of new DirectML optimizations.<\/p>\n<p><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/nvidia-microsoft-annoucement-nv-blog-header-preview-1280x680-1.jpg\"><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-large wp-image-64227\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/nvidia-microsoft-annoucement-nv-blog-header-preview-1280x680-1-672x357.jpg\" alt=\"\" width=\"672\" height=\"357\"><\/a><\/p>\n<p>As more AI inferencing happens on local devices, PCs will need powerful yet efficient hardware to support these complex tasks. To meet this need, RTX GPUs will add Max-Q low-power inferencing for AI workloads. The GPU will operate at a fraction of the power for lighter inferencing tasks, while scaling up to unmatched levels of performance for heavy generative AI workloads.<\/p>\n<p>To create new AI applications, developers can now access a complete RTX-accelerated AI development stack running on Windows 11, making it easier to develop, train and deploy advanced AI models. This starts with development and fine-tuning of models with optimized deep learning frameworks available via Windows Subsystem for Linux.<\/p>\n<p>Developers can then move seamlessly to the cloud to train on the same NVIDIA AI stack, which is available from every major cloud service provider. Next, developers can optimize the trained models for fast inferencing with tools like the <a href=\"https:\/\/devblogs.microsoft.com\/directx\/optimize-directml-performance-with-olive\/\">new Microsoft Olive<\/a>. And finally, they can deploy their AI-enabled applications and features to an install base of over 100 million RTX PCs and workstations\u00a0 that have been optimized for AI.<\/p>\n<p>\u201cAI will be the single largest driver of innovation for Windows customers in the coming years,\u201d said Pavan Davuluri, corporate vice president of Windows silicon and system integration at Microsoft. \u201cBy working in concert with NVIDIA on hardware and software optimizations, we\u2019re equipping developers with a transformative, high-performance, easy-to-deploy experience.\u201d<\/p>\n<p>To date, over 400 RTX AI-accelerated apps and games have been released, with more on the way.<\/p>\n<p>During his <a href=\"https:\/\/www.nvidia.com\/en-us\/events\/computex\/\">keynote address<\/a> kicking off COMPUTEX 2023, NVIDIA founder and CEO Jensen Huang introduced a new generative AI to support game development, <a href=\"https:\/\/developer.nvidia.com\/omniverse\/ace\">NVIDIA Avatar Cloud Engine (ACE) for Games<\/a>.<\/p>\n<p>This custom AI model foundry service transforms games by bringing intelligence to non-playable characters through AI-powered natural language interactions. Developers of middleware, tools and games can use ACE for Games to build and deploy customized speech, conversation and animation AI models in their software and games.<\/p>\n<h2><b>Generative AI on RTX, Anywhere<\/b><\/h2>\n<p>From servers to the cloud to devices, generative AI running on RTX GPUs is everywhere. NVIDIA\u2019s accelerated AI computing is a low-latency, full-stack endeavor. We\u2019ve been optimizing every part of our hardware and software architecture for many years for AI, including fourth-generation Tensor Cores \u2014 dedicated AI hardware on RTX GPUs.<\/p>\n<p>Regular driver optimizations ensure peak performance. The most recent NVIDIA driver, combined with Olive-optimized models and updates to DirectML, delivers significant speedups for developers on Windows 11. For example, Stable Diffusion performance is improved by 2x compared to the previous interference times for developers taking advantage of <a href=\"https:\/\/github.com\/microsoft\/Olive\/tree\/main\/examples\/directml\/stable_diffusion\">DirectML <\/a>optimized paths.<\/p>\n<p>And with the latest generation of RTX laptops and mobile workstations built on the NVIDIA Ada Lovelace architecture, users can take generative AI anywhere. Our next-gen mobile platform brings new levels of performance and portability \u2014 in form factors as small as 14 inches and as lightweight as about three pounds. Makers like Dell, HP, Lenovo and ASUS are pushing the generative AI era forward, backed by RTX GPUs and Tensor Cores.<\/p>\n<p>\u201cAs AI continues to get deployed across industries at an expected annual growth rate of over 37% now through 2030, businesses and consumers will increasingly need the right technology to develop and implement AI, including generative AI. Lenovo is uniquely positioned to empower generative AI spanning from devices to servers to the cloud, having developed products and solutions for AI workloads for years. Our NVIDIA RTX GPU-powered PCs, such as select Lenovo ThinkPad, ThinkStation, ThinkBook, Yoga, Legion and LOQ devices, are enabling the transformative wave of generative AI for better everyday user experiences in saving time, creating content, getting work done, gaming and more.\u201d \u2014 Daryl Cromer, vice president and chief technology officer of PCs and Smart Devices at Lenovo<\/p>\n<p>\u201cGenerative AI is transformative and a catalyst for future innovation across industries. Together, HP and NVIDIA equip developers with incredible performance, mobility and the reliability needed to run accelerated AI models today, while powering a new era of generative AI.\u201d \u2014\u00a0 Jim Nottingham, senior vice president and general manager of Z by HP<\/p>\n<p>\u201cOur recent work with NVIDIA on Project Helix centers on making it easier for enterprises to build and deploy trustworthy generative AI on premises. Another step in this historic moment is bringing generative AI to PCs. Think of app developers looking to perfect neural network algorithms while keeping training data and IP under local control. This is what our powerful and scalable Precision workstations with NVIDIA RTX GPUs are designed to do. And as the global leader in workstations, Dell is uniquely positioned to help users securely accelerate AI applications from the edge to the datacenter.\u201d \u2014 Ed Ward, president of the client product group at Dell Technologies<\/p>\n<p>\u201cThe generative AI era is upon us, requiring immense processing and fully optimized hardware and software. With the NVIDIA AI platform, including NVIDIA Omniverse, which is now preinstalled on many of our products, we are excited to see the AI revolution continue to take shape on ASUS and ROG laptops.\u201d \u2014 Galip Fu, director of global consumer marketing at ASUS<\/p>\n<p>Soon, laptops and mobile workstations with RTX GPUs will get the best of both worlds. AI inference-only workloads will be optimized for Tensor Core performance while keeping power consumption of the GPU as low as possible, extending battery life and maintaining a cool, quiet system. The GPU can then dynamically scale up for maximum AI performance when the workload demands it.<\/p>\n<p>Developers can also learn how to optimize their applications end-to-end to take full advantage of GPU-acceleration via the <a href=\"https:\/\/developer.nvidia.com\/ai-for-creative-applications\">NVIDIA AI for accelerating applications developer site<\/a>.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/2023\/05\/28\/computex-generative-ai-rtx\/<\/p>\n","protected":false},"author":0,"featured_media":3006,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3005"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=3005"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3005\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/3006"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=3005"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=3005"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=3005"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}