{"id":3965,"date":"2025-04-15T18:42:23","date_gmt":"2025-04-15T18:42:23","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2025\/04\/15\/thousands-of-nvidia-grace-blackwell-gpus-now-live-at-coreweave-propelling-development-for-ai-pioneers\/"},"modified":"2025-04-15T18:42:23","modified_gmt":"2025-04-15T18:42:23","slug":"thousands-of-nvidia-grace-blackwell-gpus-now-live-at-coreweave-propelling-development-for-ai-pioneers","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2025\/04\/15\/thousands-of-nvidia-grace-blackwell-gpus-now-live-at-coreweave-propelling-development-for-ai-pioneers\/","title":{"rendered":"Thousands of NVIDIA Grace Blackwell GPUs Now Live at CoreWeave, Propelling Development for AI Pioneers"},"content":{"rendered":"<div>\n\t\t<span class=\"bsf-rt-reading-time\"><span class=\"bsf-rt-display-label\"><\/span> <span class=\"bsf-rt-display-time\"><\/span> <span class=\"bsf-rt-display-postfix\"><\/span><\/span><\/p>\n<p>CoreWeave today became one of the first cloud providers to bring NVIDIA GB200 NVL72 systems online for customers at scale, and AI frontier companies Cohere, IBM and Mistral AI are already using them to train and deploy next-generation AI models and applications.<\/p>\n<p>CoreWeave, the first cloud provider to make <a href=\"https:\/\/blogs.nvidia.com\/blog\/blackwell-coreweave-gb200-nvl72-instances-cloud\/\">NVIDIA Grace Blackwell generally available<\/a>, has already shown <a target=\"_blank\" href=\"https:\/\/www.coreweave.com\/blog\/coreweave-delivers-breakthrough-ai-performance-with-nvidia-gb200-and-h200-gpus-in-mlperf-inference-v5-0\" rel=\"noopener\">incredible results<\/a> in MLPerf benchmarks with <a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/gb200-nvl72\/\" rel=\"noopener\">NVIDIA GB200 NVL72<\/a> \u2014 a powerful rack-scale accelerated computing platform designed for reasoning and AI agents. Now, CoreWeave customers are gaining access to thousands of <a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/data-center\/technologies\/blackwell-architecture\/\" rel=\"noopener\">NVIDIA Blackwell GPUs<\/a>.<\/p>\n<p>\u201cWe work closely with NVIDIA to quickly deliver to customers the latest and most powerful solutions for training AI models and serving inference,\u201d said Mike Intrator, CEO of CoreWeave. \u201cWith new Grace Blackwell rack-scale systems in hand, many of our customers will be the first to see the benefits and performance of AI innovators operating at scale.\u201d<\/p>\n<figure id=\"attachment_79580\" aria-describedby=\"caption-attachment-79580\" class=\"wp-caption aligncenter\"><img decoding=\"async\" loading=\"lazy\" class=\"wp-image-79580 size-full\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/3025\/04\/coreweave-2.png\" alt=\"\" width=\"587\" height=\"352\"><figcaption id=\"caption-attachment-79580\" class=\"wp-caption-text\">Thousands of NVIDIA Blackwell GPUs are now turning raw data into intelligence at unprecedented speed, with many more coming online soon.<\/figcaption><\/figure>\n<p>The ramp-up for customers of cloud providers like CoreWeave is underway. Systems built on NVIDIA Grace Blackwell are in full production, transforming cloud data centers into <a href=\"https:\/\/blogs.nvidia.com\/blog\/ai-factory\/\">AI factories<\/a> that manufacture intelligence at scale and convert raw data into real-time insights with <a href=\"https:\/\/blogs.nvidia.com\/blog\/blackwell-mlperf-inference\/\">speed<\/a>, accuracy and efficiency.<\/p>\n<p>Leading AI companies around the world are now putting GB200 NVL72\u2019s capabilities to work for AI applications, <a href=\"https:\/\/blogs.nvidia.com\/blog\/what-is-agentic-ai\/\">agentic AI<\/a> and cutting-edge model development.<\/p>\n<h2><b>Personalized AI Agents<\/b><\/h2>\n<p><a target=\"_blank\" href=\"https:\/\/cohere.com\/\" rel=\"noopener\">Cohere<\/a> is using its Grace Blackwell Superchips to help develop secure enterprise AI applications powered by leading-edge research and model development techniques. Its enterprise AI platform, <a target=\"_blank\" href=\"https:\/\/cohere.com\/north\" rel=\"noopener\">North<\/a>, enables teams to build personalized AI agents to securely automate enterprise workflows, surface real-time insights and more.<\/p>\n<p>With NVIDIA GB200 NVL72 on CoreWeave, Cohere is already experiencing up to 3x more performance in training for 100 billion-parameter models compared with previous-generation NVIDIA Hopper GPUs \u2014 even without Blackwell-specific optimizations.<\/p>\n<p>With further optimizations taking advantage of GB200 NVL72\u2019s large unified memory, FP4 precision and a 72-GPU NVIDIA NVLink domain \u2014 where every GPU is connected to operate in concert \u2014 Cohere is getting dramatically higher throughput with shorter time to first and subsequent tokens for more performant, cost-effective inference.<\/p>\n<p>\u201cWith access to some of the first NVIDIA GB200 NVL72 systems in the cloud, we are pleased with how easily our workloads port to the NVIDIA Grace Blackwell architecture,\u201d said Autumn Moulder, vice president of engineering at Cohere. \u201cThis unlocks incredible performance efficiency across our stack \u2014 from our vertically integrated North application running on a single Blackwell GPU to scaling training jobs across thousands of them. We\u2019re looking forward to achieving even greater performance with additional optimizations soon.\u201d<\/p>\n<h2><b>AI Models for Enterprise\u00a0<\/b><\/h2>\n<p>IBM is using one of the first deployments of NVIDIA GB200 NVL72 systems, scaling to thousands of Blackwell GPUs on CoreWeave, to train its next-generation <a target=\"_blank\" href=\"https:\/\/www.ibm.com\/granite\" rel=\"noopener\">Granite models<\/a>, a series of open-source, enterprise-ready AI models. Granite models deliver state-of-the-art performance while maximizing safety, speed and cost efficiency. The Granite model family is supported by a robust partner ecosystem that includes leading software companies embedding large language models into their technologies.<\/p>\n<p>Granite models provide the foundation for solutions like <a target=\"_blank\" href=\"https:\/\/www.ibm.com\/products\/watsonx-orchestrate\" rel=\"noopener\">IBM watsonx Orchestrate<\/a>, which enables enterprises to build and deploy powerful AI agents that automate and accelerate workflows across the enterprise.<\/p>\n<p>CoreWeave\u2019s NVIDIA GB200 NVL72 deployment for IBM also harnesses the <a target=\"_blank\" href=\"https:\/\/www.ibm.com\/products\/storage-scale-system\" rel=\"noopener\">IBM Storage Scale System<\/a>, which delivers exceptional high-performance storage for AI. CoreWeave customers can access the IBM Storage platform within CoreWeave\u2019s dedicated environments and AI cloud platform.<\/p>\n<p>\u201cWe are excited to see the acceleration that NVIDIA GB200 NVL72 can bring to training our Granite family of models,\u201d said Sriram Raghavan, vice president of AI at IBM Research. \u201cThis collaboration with CoreWeave will augment IBM\u2019s capabilities to help build advanced, high-performance and cost-efficient models for powering enterprise and agentic AI applications with IBM watsonx.\u201d<\/p>\n<h2><b>Compute Resources at Scale<\/b><\/h2>\n<p>Mistral AI is now getting its first thousand Blackwell GPUs to build the next generation of open-source AI models.<\/p>\n<p>Mistral AI, a Paris-based leader in open-source AI, is using CoreWeave\u2019s infrastructure, now equipped with GB200 NVL72, to speed up the development of its language models. With models like Mistral Large delivering strong reasoning capabilities, Mistral needs fast computing resources at scale.<\/p>\n<p>To train and deploy these models effectively, Mistral AI requires a cloud provider that offers large, high-performance GPU clusters with NVIDIA Quantum InfiniBand networking and reliable infrastructure management. CoreWeave\u2019s experience standing up NVIDIA GPUs at scale with industry-leading reliability and resiliency through tools such as <a target=\"_blank\" href=\"https:\/\/www.coreweave.com\/mission-control\" rel=\"noopener\">CoreWeave Mission Control<\/a> met these requirements.<\/p>\n<p>\u201cRight out of the box and without any further optimizations, we saw a 2x improvement in performance for dense model training,\u201d said Thimothee Lacroix, cofounder and chief technology officer at Mistral AI. \u201cWhat\u2019s exciting about NVIDIA GB200 NVL72 is the new possibilities it opens up for model development and inference.\u201d<\/p>\n<h2><b>A Growing Number of Blackwell Instances<\/b><\/h2>\n<p>In addition to long-term customer solutions, CoreWeave offers instances with rack-scale NVIDIA NVLink across 72 NVIDIA Blackwell GPUs and 36 NVIDIA Grace CPUs, scaling to up to 110,000 GPUs with NVIDIA Quantum-2 InfiniBand networking.<\/p>\n<p>These instances, accelerated by the NVIDIA GB200 NVL72 rack-scale accelerated computing platform, provide the scale and performance needed to build and deploy the next generation of AI reasoning models and agents.<\/p>\n<\/p><\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/coreweave-grace-blackwell-gb200-nvl72\/<\/p>\n","protected":false},"author":0,"featured_media":3966,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3965"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=3965"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/3965\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/3966"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=3965"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=3965"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=3965"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}