{"id":492,"date":"2020-11-03T11:01:22","date_gmt":"2020-11-03T11:01:22","guid":{"rendered":"https:\/\/machine-learning.webcloning.com\/2020\/11\/03\/nvidia-a100-launches-on-aws-marking-dawn-of-next-decade-in-accelerated-cloud-computing\/"},"modified":"2020-11-03T11:01:22","modified_gmt":"2020-11-03T11:01:22","slug":"nvidia-a100-launches-on-aws-marking-dawn-of-next-decade-in-accelerated-cloud-computing","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2020\/11\/03\/nvidia-a100-launches-on-aws-marking-dawn-of-next-decade-in-accelerated-cloud-computing\/","title":{"rendered":"NVIDIA A100 Launches on AWS, Marking Dawn of Next Decade in Accelerated Cloud Computing"},"content":{"rendered":"<div data-url=\"https:\/\/blogs.nvidia.com\/blog\/2020\/11\/02\/nvidia-a100-launches-on-aws\/\" data-title=\"NVIDIA A100 Launches on AWS, Marking Dawn of Next Decade in Accelerated Cloud Computing\">\n<p>Amazon Web Services\u2019 first GPU instance debuted 10 years ago, with the NVIDIA M2050. At that time, <a href=\"https:\/\/www.nvidia.com\/en-us\/gpu-accelerated-applications\/\" target=\"_blank\" rel=\"noopener noreferrer\">CUDA-based applications<\/a> were focused primarily on accelerating scientific simulations, with the rise of AI and deep learning still a ways off.<\/p>\n<p>Since then, AWS has added to its stable of cloud GPU instances, which has included the K80 (p2), K520 (g3), M60 (g4), V100 (p3\/p3dn) and T4 (g4).<\/p>\n<p>With its new P4d instance <a href=\"https:\/\/press.aboutamazon.com\/news-releases\/news-release-details\/aws-announces-general-availability-amazon-ec2-p4d-instances-ec2\" target=\"_blank\" rel=\"noopener noreferrer\">generally available today<\/a>, AWS is paving the way for another bold decade of accelerated computing powered with the latest <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/a100\/\" target=\"_blank\" rel=\"noopener noreferrer\">NVIDIA A100 Tensor Core GPU<\/a>.<\/p>\n<p>The P4d instance delivers AWS\u2019s highest performance, most cost-effective GPU-based platform for machine learning training and high performance computing applications. The instances reduce the time to train machine learning models by up to 3x with FP16 and up to 6x with TF32 compared to the default FP32 precision.<\/p>\n<p>They also provide exceptional inference performance. NVIDIA A100 GPUs just last month <a href=\"https:\/\/blogs.nvidia.com\/blog\/2020\/10\/21\/inference-mlperf-benchmarks\/\" target=\"_blank\" rel=\"noopener noreferrer\">swept the MLPerf Inference benchmarks<\/a> \u2014 providing up to 237x faster performance than CPUs.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2020\/11\/aws-p4d.png\" alt=\"\" width=\"374\" height=\"186\"><\/p>\n<p>Each P4d instance features eight NVIDIA A100 GPUs and, with AWS UltraClusters, customers can get on-demand and scalable access to over 4,000 GPUs at a time using AWS\u2019s Elastic Fabric Adaptor (EFA) and scalable, high-performant storage with Amazon FSx. P4d offers 400Gbps networking and uses NVIDIA technologies such as <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/nvlink\/\" target=\"_blank\" rel=\"noopener noreferrer\">NVLink<\/a>, <a href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/nvlink\/\" target=\"_blank\" rel=\"noopener noreferrer\">NVSwitch<\/a>, <a href=\"https:\/\/developer.nvidia.com\/nccl\" target=\"_blank\" rel=\"noopener noreferrer\">NCCL<\/a> and <a href=\"http:\/\/docs.nvidia.com\/cuda\/gpudirect-rdma\/index.html\" target=\"_blank\" rel=\"noopener noreferrer\">GPUDirect RDMA<\/a> to further accelerate deep learning training workloads. NVIDIA GPUDirect RDMA on EFA ensures low-latency networking by passing data from GPU to GPU between servers without having to pass through the CPU and system memory.<\/p>\n<p>In addition, the P4d instance is supported in many AWS services, including Amazon Elastic Container Services, Amazon Elastic Kubernetes Service, AWS ParallelCluster and Amazon SageMaker. P4d can also leverage all the optimized, containerized software available from <a href=\"https:\/\/www.nvidia.com\/en-us\/gpu-cloud\/\" target=\"_blank\" rel=\"noopener noreferrer\">NGC<\/a>, including HPC applications, AI frameworks, pre-trained models, Helm charts and inference software like <a href=\"https:\/\/developer.nvidia.com\/tensorrt\" target=\"_blank\" rel=\"noopener noreferrer\">TensorRT<\/a> and <a href=\"https:\/\/developer.nvidia.com\/nvidia-triton-inference-server\" target=\"_blank\" rel=\"noopener noreferrer\">Triton Inference Server<\/a>.<\/p>\n<p>P4d instances are now available in US East and West, and coming to additional regions soon. The instances can be purchased as On-Demand, with Savings Plans, with Reserved Instances, or as Spot Instances.<\/p>\n<p>The first decade of GPU cloud computing has brought over 100 exaflops of AI compute to the market. With the arrival of the Amazon EC2 P4d instance powered by NVIDIA A100 GPUs, the next decade of GPU cloud computing is off to a great start.<\/p>\n<p>NVIDIA and AWS are making it possible for applications to continue pushing the boundaries of AI across a wide array of applications. We can\u2019t wait to see what customers will do with it.<\/p>\n<p>Visit AWS and get started with <a href=\"https:\/\/aws.amazon.com\/ec2\/instance-types\/p4\" target=\"_blank\" rel=\"noopener noreferrer\">P4d instances<\/a> today.<\/p>\n<\/div>\n","protected":false},"excerpt":{"rendered":"<p>http:\/\/feedproxy.google.com\/~r\/nvidiablog\/~3\/Q-wpBepriAw\/<\/p>\n","protected":false},"author":0,"featured_media":493,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/492"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=492"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/492\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/493"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=492"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=492"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=492"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}