{"id":208,"date":"2020-09-12T16:01:34","date_gmt":"2020-09-12T16:01:34","guid":{"rendered":"https:\/\/machine-learning.webcloning.com\/2020\/09\/12\/the-great-ai-bake-off-recommendation-systems-on-the-rise\/"},"modified":"2020-09-12T16:01:34","modified_gmt":"2020-09-12T16:01:34","slug":"the-great-ai-bake-off-recommendation-systems-on-the-rise","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2020\/09\/12\/the-great-ai-bake-off-recommendation-systems-on-the-rise\/","title":{"rendered":"The Great AI Bake-Off: Recommendation Systems on the Rise"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2020\/09\/10\/ai-recommendation-systems-challenge\/\" data-title=\"The Great AI Bake-Off: Recommendation Systems on the Rise\">\n<p>If you want to create a world-class <a href=\"https:\/\/blogs.nvidia.com\/blog\/2020\/05\/14\/whats-a-recommender-system\/\">recommendation system<\/a>, follow this recipe from a global team of experts: Blend a big helping of GPU-accelerated AI with a dash of old-fashioned cleverness.<\/p>\n<p>The proof was in the pudding for a team from NVIDIA that won this year\u2019s <a href=\"http:\/\/www.recsyschallenge.com\/2020\/\">ACM RecSys Challenge<\/a>. The competition is a highlight of an annual gathering of more than 500 experts who present the latest research in recommendation systems, the engines that deliver personalized suggestions for everything from restaurants to real estate.<\/p>\n<p>At the Sept. 22-26 <a href=\"https:\/\/recsys.acm.org\/recsys20\/\">online event<\/a>, the team will describe its dish, already available as open source code. They\u2019re also sharing lessons learned with colleagues who build NVIDIA products like <a href=\"https:\/\/developer.nvidia.com\/rapids\">RAPIDS<\/a> and <a href=\"https:\/\/developer.nvidia.com\/nvidia-merlin\">Merlin<\/a>, so customers can enjoy the fruits of their labor.<\/p>\n<p><a href=\"https:\/\/www.nvidia.com\/en-us\/deep-learning-ai\/education\/?iactivetab=certification-tabs-2#RecommenderSystems\"><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/09\/dli-recommender-course-devzone-banner-280x240-1.jpg\" alt=\"\" width=\"280\" height=\"240\"><\/a><\/p>\n<p>In an effort to bring more people to the table, NVIDIA will donate the contest\u2019s $15,000 cash prize to <a href=\"https:\/\/blackinai.github.io\/\">Black in AI<\/a>, a nonprofit dedicated to mentoring the next generation of Black specialists in machine learning.<\/p>\n<h2><b>GPU Server Doles Out Recommendations<\/b><\/h2>\n<p>This year\u2019s contest, sponsored by Twitter, asked researchers to comb through a dataset of 146 million tweets to predict which ones a user would like, reply or retweet. The NVIDIA team\u2019s work led a field of 34 competitors, thanks in part to a system with four <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/v100\/\">NVIDIA V100 Tensor Core GPUs<\/a> that cranked through hundreds of thousands of options.<\/p>\n<p>Their numbers were eye-popping. GPU-accelerated software engineered in less than a minute features that required nearly an hour on a CPU, a 500x speedup. The four-GPU system trained the team\u2019s AI models 120x faster than a CPU. And GPUs gave the group\u2019s end-to-end solution a 280x speedup compared to an initial implementation on a CPU.<\/p>\n<p>\u201cI\u2019m still blown away when we pull off something like a 500x speedup in feature engineering,\u201d said Even Oldridge, a Ph.D. in machine learning who in the past year quadrupled the size of his group that designs NVIDIA Merlin, a framework for recommendation systems.<\/p>\n<figure id=\"attachment_46760\" aria-describedby=\"caption-attachment-46760\" class=\"wp-caption aligncenter\"><a href=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/09\/RecSys-Challenge-chart-x-1280.png\"><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/09\/RecSys-Challenge-chart-x-1280-672x398.png\" alt=\"Recommendation systems on GPUs\" width=\"672\" height=\"398\"><\/a><figcaption id=\"caption-attachment-46760\" class=\"wp-caption-text\">GPUs and frameworks such as UCX provided up to 500x speedups compared to CPUs.<\/figcaption><\/figure>\n<h2><b>Competition Sparks Ideas for Software Upgrades\u00a0\u00a0<\/b><\/h2>\n<p>The competition spawned work on data transformations that could enhance future versions of <a href=\"https:\/\/developer.nvidia.com\/blog\/accelerating-etl-for-recsys-on-gpus-with-nvtabular\/\">NVTabular<\/a>, a Merlin library that eases engineering new features with the spreadsheet-like tables that are the basis of recommendation systems.<\/p>\n<p>\u201cWe won in part because we could prototype fast,\u201d said Benedikt Schifferer, one of three specialists in recommendation systems on the team that won the prize.<\/p>\n<p>Schifferer also credits two existing tools. DASK, an open-source scheduling tool, let the team split memory-hungry jobs across multiple GPUs. And cuDF, part of NVIDIA\u2019s RAPIDS framework for accelerated data science, let the group run the equivalent of the popular Pandas library on GPUs.<\/p>\n<p>\u201cSearching for features in the data using Pandas on CPUs took hours for each new feature,\u201d said Chris Deotte, one of a handful of data scientists on the team who have earned the title Kaggle grandmaster for their prowess in competitions.<\/p>\n<p>\u201cWhen we converted our code to RAPIDS, we could explore features in minutes. It was life changing, we could search hundreds of features and that eventually led to discoveries that won that competition,\u201d said Deotte, one of only two grandmasters who hold that title in all four Kaggle categories.<\/p>\n<p>More enhancements for recommendation systems are on the way. For example, customers can look forward to improvements in text handling on GPUs, a key data type for recommendation systems.<\/p>\n<h2><b>An Aha! Moment Fuels the Race<\/b><\/h2>\n<p>Deotte credits a colleague in Brazil, Gilberto Titericz, with an insight that drove the team forward.<\/p>\n<p>\u201cHe tracked changes in Twitter followers over time which turned out to be a feature that really fueled our accuracy \u2014 it was incredibly effective,\u201d Deotte said.<\/p>\n<p>\u201cI saw patterns changing over time, so I made several plots of them,\u201d said Titericz, who ranked as the top Kaggle grandmaster worldwide for a couple years.<\/p>\n<p>\u201cWhen I saw a really great result, I thought I made a mistake, but I took a chance, submitted it and to my surprise it scored high on the leaderboard, so my intuition was right,\u201d he added.<\/p>\n<p>In the end, the team used a mix of complementary AI models designed by Titericz, Schifferer and a colleague in Japan, Kazuki Onodera, all based on XGBoost, an algorithm well suited for recommendation systems.<\/p>\n<p>Several members of the team are part of an elite group of Kaggle grandmasters that NVIDIA founder and CEO Jensen Huang dubbed KGMON, a playful takeoff on Pokemon. The team won dozens of competitions in the last four years.<\/p>\n<h2><b>Recommenders Getting Traction in B2C<\/b><\/h2>\n<p>For many members, including team leader Jean-Francois Puget in southern France, it\u2019s more than a 9-to-5 job.<\/p>\n<p>\u201cWe spend nights and weekends in competitions, too, trying to be the best in the world,\u201d said Puget, who earned his Ph.D. in machine learning two decades before deep learning took off commercially.<\/p>\n<p>Now the technology is spreading fast.<\/p>\n<p>This year\u2019s ACM RecSys includes three dozen papers and talks from companies like Amazon and Netflix that helped establish the field with recommenders that help people find books and movies. Now, consumer companies of all stripes are getting into the act including IKEA and Etsy, which are presenting at ACM RecSys this year.<\/p>\n<p>\u201cFor the last three or four years, it\u2019s more focused on delivering a personalized experience, really understanding what users want,\u201d said Schifferer. It\u2019s a cycle where \u201ccustomers\u2019 choices influence the training data, so some companies retrain their AI models every four hours, and some say they continuously train,\u201d he added.<\/p>\n<p>That\u2019s why the team works hard to create frameworks like Merlin to make recommendation systems run easily and fast at scale on GPUs. Other members of NVIDIA\u2019s winning team were Christof Henkel (Germany), Jiwei Liu and Bojan Tunguz (U.S.), Gabriel De Souza Pereira Moreira (Brazil) and Ahmet Erdem (Netherlands).<\/p>\n<p>To get tips on how to design recommendation systems from the winning team, tune in to an online tutorial <a href=\"https:\/\/recsys.acm.org\/recsys20\/tutorials\/\">here<\/a> on Friday, Sept. 25.<\/p>\n<h2><strong>Our Recommendation? Learn How to Build Intelligent Recommendation Systems<\/strong><\/h2>\n<p>The\u00a0<a href=\"https:\/\/www.nvidia.com\/en-us\/deep-learning-ai\/education\/?iactivetab=certification-tabs-2#RecommenderSystems\">NVIDIA Deep Learning Institute<\/a>\u00a0offers instructor-led, hands-on training on the fundamental tools and techniques for building highly effective recommender systems. Taught by an expert, this in-depth, 8-hour-long workshop instructs participants in how to:<\/p>\n<ul>\n<li>Build a content-based recommender system using the open-source cuDF library and Apache Arrow<\/li>\n<li>Construct a collaborative filtering recommender system using alternating least squares and CuPy<\/li>\n<li>Design a wide and deep neural network using TensorFlow 2 to create a hybrid recommender system<\/li>\n<li>Optimize performance for training and inference using large, sparse datasets<\/li>\n<li>Deploy a recommender model as a high-performance web service<\/li>\n<\/ul>\n<p>Earn a DLI certificate to demonstrate subject-matter competency and accelerate your career growth. Take this workshop at\u00a0<a href=\"https:\/\/www.nvidia.com\/en-us\/gtc\/training\/\">GTC<\/a>\u00a0this fall or\u00a0<a href=\"https:\/\/www.nvidia.com\/en-us\/deep-learning-ai\/education\/?iactivetab=certification-tabs-2#RecommenderSystems\">request a workshop<\/a>\u00a0for your organization.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>http:\/\/feedproxy.google.com\/~r\/nvidiablog\/~3\/w101VVk0bLY\/<\/p>\n","protected":false},"author":0,"featured_media":209,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/208"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=208"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/208\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/209"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=208"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=208"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=208"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}