{"id":2995,"date":"2023-05-23T16:59:02","date_gmt":"2023-05-23T16:59:02","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2023\/05\/23\/nvidia-and-microsoft-drive-innovation-for-windows-pcs-in-new-era-of-generative-ai\/"},"modified":"2023-05-23T16:59:02","modified_gmt":"2023-05-23T16:59:02","slug":"nvidia-and-microsoft-drive-innovation-for-windows-pcs-in-new-era-of-generative-ai","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2023\/05\/23\/nvidia-and-microsoft-drive-innovation-for-windows-pcs-in-new-era-of-generative-ai\/","title":{"rendered":"NVIDIA and Microsoft Drive Innovation for Windows PCs in New Era of Generative AI"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2023\/05\/23\/microsoft-build-nvidia-ai-windows-rtx\/\" data-title=\"NVIDIA and Microsoft Drive Innovation for Windows PCs in New Era of Generative AI\" data-hashtags=\"\">\n<p>Generative AI \u2014 in the form of large language model (LLM) applications like ChatGPT, image generators such as Stable Diffusion and Adobe Firefly, and game rendering techniques like NVIDIA DLSS 3 Frame Generation \u2014 is rapidly ushering in a new era of computing for productivity, content creation, gaming and more.<\/p>\n<p>At the <a href=\"https:\/\/build.microsoft.com\/en-US\/home?wt.mc_id=FP_NVIDIA_blog_reg\">Microsoft Build<\/a> developer conference, NVIDIA and Microsoft today showcased a suite of advancements in Windows 11 PCs and workstations with NVIDIA RTX GPUs to meet the demands of <a href=\"https:\/\/www.nvidia.com\/en-us\/glossary\/data-science\/generative-ai\/\">generative AI<\/a>.<\/p>\n<p>More than 400 Windows apps and games already employ AI technology, accelerated by dedicated processors on RTX GPUs called Tensor Cores. Today\u2019s announcements, which include tools to develop AI on Windows PCs, frameworks to optimize and deploy AI, and driver performance and efficiency improvements, will empower developers to build the next generation of Windows apps with generative AI at their core.<\/p>\n<p>\u201cAI will be the single largest driver of innovation for Windows customers in the coming years,\u201d said Pavan Davuluri, corporate vice president of Windows silicon and system integration at Microsoft. \u201cBy working in concert with NVIDIA on hardware and software optimizations, we\u2019re equipping developers with a transformative, high-performance, easy-to-deploy experience.\u201d<\/p>\n<h2><b>Develop Models With Windows Subsystem for Linux<\/b><\/h2>\n<p>AI development has traditionally taken place on Linux, requiring developers to either dual-boot their systems or use multiple PCs to work in their AI development OS while still accessing the breadth and depth of the Windows ecosystem.<\/p>\n<p>Over the past few years, Microsoft has been building a powerful capability to run Linux directly within the Windows OS, called Windows Subsystem for Linux (WSL). NVIDIA has been working closely with Microsoft to deliver GPU acceleration and support for the entire NVIDIA AI software stack inside WSL. Now developers can use Windows PC for all their local AI development needs with support for GPU-accelerated <a href=\"https:\/\/developer.nvidia.com\/deep-learning-frameworks\">deep learning frameworks<\/a> on WSL.<\/p>\n<p>With NVIDIA RTX GPUs delivering <a href=\"https:\/\/www.nvidia.com\/en-us\/design-visualization\/desktop-graphics\/\">up to 48GB of RAM<\/a> in desktop workstations, developers can now work with models on Windows that were previously only available on servers. The large memory also improves the performance and quality for local fine-tuning of AI models, enabling designers to customize them to their own style or content. And because the same NVIDIA AI software stack runs on NVIDIA data center GPUs, it\u2019s easy for developers to push their models to Microsoft Azure Cloud for large training runs.<\/p>\n<h2><b>Rapidly Optimize and Deploy Models<\/b><\/h2>\n<p>With trained models in hand, developers need to optimize and deploy AI for target devices.<\/p>\n<p>Microsoft released the <a href=\"https:\/\/github.com\/microsoft\/OLive\">Microsoft Olive<\/a> toolchain for optimization and conversion of PyTorch models to ONNX, enabling developers to automatically tap into GPU hardware acceleration such as RTX Tensor Cores. Developers can optimize models via Olive and ONNX, and deploy Tensor Core-accelerated models to PC or cloud. Microsoft continues to invest in making PyTorch and related tools and frameworks work seamlessly with WSL to provide the best AI model development experience.<\/p>\n<h2><b>Improved AI Performance, Power Efficiency<\/b><\/h2>\n<p>Once deployed, generative AI models demand incredible inference performance. RTX Tensor Cores deliver up to 1,400 Tensor TFLOPS for AI inferencing. Over the last year, NVIDIA has worked to improve DirectML performance to take full advantage of RTX hardware.<\/p>\n<p>On May 24, we\u2019ll release our latest optimizations in Release 532.03 drivers that combine with Olive-optimized models to deliver big boosts in AI performance. Using an <a href=\"https:\/\/devblogs.microsoft.com\/directx\/dml-stable-diffusion\">Olive-optimized version of the Stable Diffusion<\/a> text-to-image generator with the popular Automatic1111 distribution, performance is improved over 2x with the new driver.<\/p>\n<figure id=\"attachment_64217\" aria-describedby=\"caption-attachment-64217\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/Stable_Diffusion_Performance-532.03_Driver.png\"><img decoding=\"async\" loading=\"lazy\" class=\"size-large wp-image-64217\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2023\/05\/Stable_Diffusion_Performance-532.03_Driver-672x404.png\" alt=\"Chart showing performance improvements in Stable Diffusion with updated NVIDIA drivers.\" width=\"672\" height=\"404\"><\/a><figcaption id=\"caption-attachment-64217\" class=\"wp-caption-text\">Stable Diffusion performance tested on GeForce RTX 4090 using Automatic1111 and Text-to-Image function.<\/figcaption><\/figure>\n<p>With AI coming to nearly every Windows application, efficiently delivering inference performance is critical \u2014 especially for laptops. Coming soon, NVIDIA will introduce new Max-Q low-power inferencing for AI-only workloads on RTX GPUs. It optimizes Tensor Core performance while keeping power consumption of the GPU as low as possible, extending battery life and maintaining a cool, quiet system.\u00a0 The GPU can then dynamically scale up for maximum AI performance when the workload demands it.<\/p>\n<p><b>Join the PC AI Revolution Now<\/b><\/p>\n<p>Top software developers \u2014 like Adobe, DxO, ON1 and Topaz \u2014 have already incorporated NVIDIA AI technology with more than 400 Windows applications and games optimized for RTX Tensor Cores.<\/p>\n<p>\u201cAI, machine learning and deep learning power all Adobe applications and drive the future of creativity. Working with NVIDIA we continuously optimize AI model performance to deliver the best possible experience for our Windows users on RTX GPUs.\u201d \u2014 Ely Greenfield, CTO of digital media at Adobe<\/p>\n<p>\u201cNVIDIA is helping to optimize our WinML model performance on RTX GPUs, which is accelerating the AI in DxO DeepPRIME, as well as providing better denoising and demosaicing, faster.\u201d \u2014 Renaud Capolunghi, senior vice president of engineering at DxO<\/p>\n<p>\u201cWorking with NVIDIA and Microsoft to accelerate our AI models running in Windows on RTX GPUs is providing a huge benefit to our audience. We\u2019re already seeing 1.5x performance gains in our suite of AI-powered photography editing software.\u201d \u2014 Dan Harlacher, vice president of products at ON1<\/p>\n<p>\u201cOur extensive work with NVIDIA has led to improvements across our suite of photo- and video-editing applications. With RTX GPUs, AI performance has improved drastically, enhancing the experience for users on Windows PCs.\u201d \u2014 Suraj Raghuraman, head of AI engine development at Topaz Labs<\/p>\n<p>NVIDIA and Microsoft are making several resources available for developers to test drive top generative AI models on Windows PCs. An Olive-optimized version of the Dolly 2.0 large language model is available on Hugging Face. And a PC-optimized version of <a href=\"https:\/\/www.nvidia.com\/en-us\/ai-data-science\/generative-ai\/nemo-framework\/\">NVIDIA NeMo<\/a> large language model for conversational AI is coming soon to Hugging Face.<\/p>\n<p>Developers can also learn how to optimize their applications end-to-end to take full advantage of GPU-acceleration via the <a href=\"https:\/\/developer.nvidia.com\/ai-for-creative-applications\">NVIDIA AI for accelerating applications developer site<\/a>.<\/p>\n<p>The complementary technologies behind Microsoft\u2019s Windows platform and NVIDIA\u2019s dynamic AI hardware and software stack will help developers quickly and easily develop and deploy generative AI on Windows 11.<\/p>\n<p>Microsoft Build runs through Thursday, May 25. Tune into to learn more on <a href=\"https:\/\/build.microsoft.com\/en-US\/sessions\/8aab36d1-d27d-46dd-81ec-eb3f49cfee6a?source=sessions\">shaping the future of work with AI<\/a>.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/2023\/05\/23\/microsoft-build-nvidia-ai-windows-rtx\/<\/p>\n","protected":false},"author":0,"featured_media":2996,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/2995"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=2995"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/2995\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/2996"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=2995"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=2995"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=2995"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}