{"id":1249,"date":"2021-11-23T08:29:38","date_gmt":"2021-11-23T08:29:38","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2021\/11\/23\/paint-me-a-picture-nvidia-research-shows-gaugan-ai-art-demo-now-responds-to-words\/"},"modified":"2021-11-23T08:29:38","modified_gmt":"2021-11-23T08:29:38","slug":"paint-me-a-picture-nvidia-research-shows-gaugan-ai-art-demo-now-responds-to-words","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2021\/11\/23\/paint-me-a-picture-nvidia-research-shows-gaugan-ai-art-demo-now-responds-to-words\/","title":{"rendered":"\u2018Paint Me a Picture\u2019: NVIDIA Research Shows GauGAN AI Art Demo Now Responds to Words"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2021\/11\/22\/gaugan2-ai-art-demo\/\" data-title=\"\u2018Paint Me a Picture\u2019: NVIDIA Research Shows GauGAN AI Art Demo Now Responds to Words\" data-hashtags=\"\">\n<p>A picture worth a thousand words now takes just three or four words to create, thanks to GauGAN2, the latest version of NVIDIA Research\u2019s wildly popular AI painting demo.<\/p>\n<p>The deep learning model behind <a href=\"https:\/\/blogs.nvidia.com\/blog\/2019\/03\/18\/gaugan-photorealistic-landscapes-nvidia-research\/\">GauGAN<\/a> allows anyone to channel their imagination into photorealistic masterpieces \u2014 and it\u2019s easier than ever. Simply type a phrase like \u201csunset at a beach\u201d and AI generates the scene in real time. Add an additional adjective like \u201csunset at a <i>rocky <\/i>beach,\u201d or swap \u201csunset\u201d to \u201cafternoon\u201d or \u201crainy day\u201d and the model, based on <a href=\"https:\/\/blogs.nvidia.com\/blog\/2017\/06\/08\/ai-podcast-an-argument-in-a-bar-led-to-the-generative-adversarial-networks-revolutionizing-deep-learning\/\">generative adversarial networks<\/a>, instantly modifies the picture.<\/p>\n<p>With the press of a button, users can generate a segmentation map, a high-level outline that shows the location of objects in the scene. From there, they can switch to drawing, tweaking the scene with rough sketches using labels like sky, tree, rock and river, allowing the smart paintbrush to incorporate these doodles into stunning images.<\/p>\n<p>The new GauGAN2 text-to-image feature can now be experienced on <a href=\"https:\/\/www.nvidia.com\/en-us\/research\/ai-demos\/\">NVIDIA AI Demos<\/a>, where visitors to the site can experience AI through the latest demos from NVIDIA Research. With the versatility of text prompts and sketches, GauGAN2 lets users create and customize scenes more quickly and with finer control.<\/p>\n<\/p>\n<h2><b>An AI of Few Words<\/b><\/h2>\n<p>GauGAN2 combines segmentation mapping, inpainting and text-to-image generation in a single model, making it a powerful tool to create photorealistic art with a mix of words and drawings.<\/p>\n<p>The demo is one of the first to combine multiple modalities \u2014 text, semantic segmentation, sketch and style \u2014 within a single GAN framework. This makes it faster and easier to turn an artist\u2019s vision into a high-quality AI-generated image.<\/p>\n<p>Rather than needing to draw out every element of an imagined scene, users can enter a brief phrase to quickly generate the key features and theme of an image, such as a snow-capped mountain range. This starting point can then be customized with sketches to make a specific mountain taller or add a couple trees in the foreground, or clouds in the sky.<\/p>\n<p>It doesn\u2019t just create realistic images \u2014 artists can also use the demo to depict otherworldly landscapes.<\/p>\n<p>Imagine for instance, recreating a landscape from the iconic planet of Tatooine in the <i>Star Wars<\/i> franchise, which has two suns. All that\u2019s needed is the text \u201cdesert hills sun\u201d to create a starting point, after which users can quickly sketch in a second sun.<a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2021\/11\/tattoine.png\"><\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2021\/11\/tattoine.png\" alt=\"\" width=\"512\" height=\"296\"><\/p>\n<p><\/a><\/p>\n<p>It\u2019s an iterative process, where every word the user types into the text box adds more to the AI-created image.<\/p>\n<p>The AI model behind GauGAN2 was trained on 10 million high-quality landscape images using the <a href=\"https:\/\/www.nvidia.com\/en-us\/on-demand\/session\/supercomputing2020-sc2019\/\">NVIDIA Selene supercomputer<\/a>, an <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/dgx-superpod\/\">NVIDIA DGX SuperPOD system<\/a> that\u2019s among the world\u2019s 10 most powerful supercomputers. The researchers used a neural network that learns the connection between words and the visuals they correspond to like \u201cwinter,\u201d \u201cfoggy\u201d or \u201crainbow.\u201d<\/p>\n<p>Compared to state-of-the-art models specifically for text-to-image or segmentation map-to-image applications, the neural network behind GauGAN2 produces a greater variety and higher quality of images.<\/p>\n<p>The GauGAN2 research demo illustrates the future possibilities for powerful image-generation tools for artists. One example is the <a href=\"https:\/\/blogs.nvidia.com\/blog\/2021\/06\/23\/studio-canvas-app\/\">NVIDIA Canvas<\/a> app, which is based on GauGAN technology and available to download for anyone with an NVIDIA RTX GPU.<\/p>\n<p><a href=\"https:\/\/www.nvidia.com\/en-us\/research\/\">NVIDIA Research<\/a> has more than 200 scientists around the globe, focused on areas including AI, computer vision, self-driving cars, robotics and graphics. Learn more about <a href=\"https:\/\/www.youtube.com\/watch?v=S7k-uWYqLt0\" target=\"_blank\" rel=\"noopener\">their work<\/a>.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/2021\/11\/22\/gaugan2-ai-art-demo\/<\/p>\n","protected":false},"author":0,"featured_media":1250,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/1249"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=1249"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/1249\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/1250"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=1249"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=1249"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=1249"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}