{"id":2215,"date":"2022-07-21T15:39:12","date_gmt":"2022-07-21T15:39:12","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2022\/07\/21\/researchers-use-gpus-to-give-earbud-users-a-mute-button-for-background-noise\/"},"modified":"2022-07-21T15:39:12","modified_gmt":"2022-07-21T15:39:12","slug":"researchers-use-gpus-to-give-earbud-users-a-mute-button-for-background-noise","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2022\/07\/21\/researchers-use-gpus-to-give-earbud-users-a-mute-button-for-background-noise\/","title":{"rendered":"Researchers Use GPUs to Give Earbud Users a \u2018Mute Button\u2019 for Background Noise"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2022\/07\/21\/mute-button-clearbuds\/\" data-title=\"Researchers Use GPUs to Give Earbud Users a \u2018Mute Button\u2019 for Background Noise\" data-hashtags=\"\">\n<p>Thanks to earbuds you can have calls anywhere while doing anything. The problem: those on the other end of the call hear it all, too, from your roommate\u2019s vacuum cleaner to background conversations at the cafe you\u2019re working from.<\/p>\n<p>Now, work by a trio of graduate students at the University of Washington who spent the pandemic cooped up together in a noisy apartment, lets those on the other end of the call hear just you \u2014 rather than all the stuff going on around you.<\/p>\n<p>Users found that the system, dubbed \u201cClearBuds\u201d \u2014 presented last month at the ACM International Conference on Mobile Systems, Applications, and Services \u2014 improved background noise suppression much better than a commercially available alternative.<\/p>\n<\/p>\n<p>\u201cYou\u2019re removing your audio background the same way you can remove your visual background on a video call,\u201d explained Vivek Jayaram, a doctoral student in the Paul G. Allen School of Computer Science &amp; Engineering.<\/p>\n<p>Outlined in a <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3498361.3538933\">paper<\/a> co-authored by the three roommates, all computer science and engineering graduate students at the University of Washington \u2014 <a href=\"https:\/\/www.maruchikim.com\/\">Maruchi Kim<\/a>, <a href=\"https:\/\/ishanchatterjee.weebly.com\/\">Ishan Chatterjee<\/a>, and Jayaram \u2014 ClearBuds are different from other wireless earbuds in two big ways.<\/p>\n<figure id=\"attachment_58335\" aria-describedby=\"caption-attachment-58335\" class=\"wp-caption alignleft\">\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2022\/07\/clearbuds-earbuds.jpg\" alt=\"\" width=\"233\" height=\"155\"><figcaption id=\"caption-attachment-58335\" class=\"wp-caption-text\">The ClearBuds hardware (round disk) in front of the 3D printed earbud enclosures. Credit: Raymond Smith, University of Washington<\/figcaption><\/figure>\n<p>First, ClearBuds use two microphones per earbud.<\/p>\n<p>While most earbuds use two microphones on the same earbud, ClearBuds uses a microphone from both earbuds and creates two audio streams.<\/p>\n<p>This creates higher spatial resolution for the system to better separate sounds coming from different directions, Kim explained. In other words, it makes it easier for the system to pick out the earbud wearer\u2019s voice.<\/p>\n<p>Second, the team created a neural network algorithm that can run on a mobile phone to process the audio streams to identify which sounds should be enhanced and which should be suppressed.<\/p>\n<p>The researchers relied on two separate neural networks to do this.<\/p>\n<p>The first neural network suppresses everything that isn\u2019t a human voice.<\/p>\n<p>The second enhances the speaker\u2019s voice. The speaker can be identified because it\u2019s coming from microphones in both earbuds at the same time.<\/p>\n<p>Together, they effectively mask background noise and ensure the earbud wearer is heard loud and clear.<\/p>\n<figure id=\"attachment_58339\" aria-describedby=\"caption-attachment-58339\" class=\"wp-caption aligncenter\">\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2022\/07\/clearbuds-network.png\" alt=\"\" width=\"607\" height=\"371\"><figcaption id=\"caption-attachment-58339\" class=\"wp-caption-text\">ClearBuds isolate a user\u2019s voice from background noise by performing voice separation using a pair of wireless, synchronized earbuds. Source: Maruchi Kim, University of Washington<\/figcaption><\/figure>\n<p>While the software the researchers created was lightweight enough to run on a mobile device, they relied on an NVIDIA TITAN desktop GPU to train the neural networks. They used both synthetic audio samples and real audio. Training took less than a day.<\/p>\n<p>And the results, users reported, were dramatically better than commercially available earbuds, results that are winning recognition industrywide.<\/p>\n<p>The team took second place for best paper at last month\u2019s ACM MobSys 2022 conference. In addition to Kim, Chatterjee and Jayarm, the paper\u2019s co-authors included <a href=\"https:\/\/www.irakemelmacher.com\/\">Ira Kemelmacher-Shlizerman<\/a>, an associate professor at the Allen School; <a href=\"https:\/\/www.cs.washington.edu\/people\/faculty\/shwetak\">Shwetak Patel<\/a>, a professor in both the Allen School and the electrical and computer engineering department; and <a href=\"https:\/\/homes.cs.washington.edu\/~gshyam\/\">Shyam Gollakota<\/a> and <a href=\"https:\/\/www.smseitz.com\/\">Steven Seitz<\/a>, both professors in the Allen School.<\/p>\n<p><b>Read the full paper here: <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3498361.3538933\">https:\/\/dl.acm.org\/doi\/10.1145\/3498361.3538933<\/a><\/b><\/p>\n<p>To be sure, the system outlined in the paper can\u2019t be adopted instantly. While many earbuds have two microphones per earbud, they only stream audio from one earbud. Industry standards are just catching up to the idea of processing multiple audio streams from earbuds.<\/p>\n<p>Nevertheless, the researchers are hopeful their work, which is open source, will inspire others to couple neural networks and microphones to provide better quality audio calls.<\/p>\n<p>The ideas could also be useful for isolating and enhancing conversations taking place over smart speakers by harnessing them for ad hoc microphone arrays, Kim said, and even tracking robot locations or search and rescue missions.<\/p>\n<p>Sounds good to us.<\/p>\n<p><em>Featured image credit: \u00a0Raymond Smith, University of Washington<\/em><\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/2022\/07\/21\/mute-button-clearbuds\/<\/p>\n","protected":false},"author":0,"featured_media":2216,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/2215"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=2215"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/2215\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/2216"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=2215"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=2215"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=2215"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}