{"id":4355,"date":"2025-11-17T23:39:45","date_gmt":"2025-11-17T23:39:45","guid":{"rendered":"https:\/\/salarydistribution.com\/machine-learning\/2025\/11\/17\/accelerated-computing-networking-drive-supercomputing-in-age-of-ai\/"},"modified":"2025-11-17T23:39:45","modified_gmt":"2025-11-17T23:39:45","slug":"accelerated-computing-networking-drive-supercomputing-in-age-of-ai","status":"publish","type":"post","link":"https:\/\/salarydistribution.com\/machine-learning\/2025\/11\/17\/accelerated-computing-networking-drive-supercomputing-in-age-of-ai\/","title":{"rendered":"Accelerated Computing, Networking Drive Supercomputing in Age of AI"},"content":{"rendered":"<div>\n\t\t<span class=\"bsf-rt-reading-time\"><span class=\"bsf-rt-display-label\"><\/span> <span class=\"bsf-rt-display-time\"><\/span> <span class=\"bsf-rt-display-postfix\"><\/span><\/span><\/p>\n<p>At SC25, NVIDIA unveiled advances across NVIDIA BlueField DPUs, next-generation networking, quantum computing, national research, AI physics and more \u2014 as accelerated systems drive the next chapter in AI supercomputing.<\/p>\n<p>NVIDIA also highlighted storage innovations powered by the <a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/networking\/products\/data-processing-unit\/\" rel=\"noopener\">NVIDIA BlueField-4 data processing unit<\/a>, part of the full-stack BlueField platform that accelerates gigascale AI infrastructure.<\/p>\n<p>More details also came on NVIDIA Quantum-X Photonics InfiniBand CPO networking switches \u2014 enabling AI factories to drastically reduce energy consumption and operational costs \u2014 including that TACC, Lambda and CoreWeave plan to integrate them.<\/p>\n<p>Last month, NVIDIA began shipping DGX Spark, the world\u2019s smallest AI supercomputer. DGX Spark packs a petaflop of AI performance and 128GB of unified memory into a desktop form factor, enabling developers to run inference on models up to 200 billion parameters and fine-tune models locally. Built on the Grace Blackwell architecture, it integrates NVIDIA GPUs, CPUs, networking, CUDA libraries and the full NVIDIA AI software stack.<\/p>\n<p>DGX Spark\u2019s unified memory and NVIDIA NVLink-C2C deliver 5x the bandwidth of PCIe Gen5, enabling faster GPU-CPU data exchange. This boosts training efficiency for large models, reduces latency and supports seamless fine-tuning workflows \u2014 all within a desktop form factor.<\/p>\n<h2><b>NVIDIA Apollo Unveiled as Latest Open Model Family for AI Physics<\/b><\/h2>\n<p><a href=\"https:\/\/blogs.nvidia.com\/blog\/apollo-open-models\">NVIDIA Apollo<\/a>, a family of open models for AI Physics, was also introduced at SC25. Applied Materials, Cadence, LAM Research, Luminary Cloud, KLA, PhysicsX, Rescale, Siemens and Synopsys \u00a0are among the industry leaders adopting these open models to simulate and accelerate their design processes in a broad range of fields \u2014 electronic device automation and semiconductors, computational fluid dynamics, structural mechanics, electromagnetics, weather and more.<\/p>\n<p>The family of open models harness the latest developments in AI physics, incorporating best-in-class machine learning architectures, such as neural operators, transformers and diffusion methods, with domain-specific knowledge. Apollo will provide pretrained checkpoints and reference workflows for training, inference and benchmarking, allowing developers to integrate and customize the models for their specific needs.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"size-medium wp-image-87397 aligncenter\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/11\/industrial-manufacturing-social-cae-ai-physics-open-model-blog-1280x680.jpg-960x510.png\" alt=\"\" width=\"960\" height=\"510\"><\/p>\n<h2 id=\"warp\" class=\"wp-block-heading\"><b>NVIDIA Warp Supercharges Physics Simulations\u200b <a href=\"https:\/\/blogs.nvidia.com\/blog\/accelerated-computing-networking-supercomputing-ai\/#warp\">????<\/a><\/b><\/h2>\n<p><a target=\"_blank\" href=\"https:\/\/developer.nvidia.com\/warp-python\" rel=\"noopener\">NVIDIA Warp<\/a> is a purpose-built open-source Python framework delivering GPU acceleration for computational physics and AI by up to 245x.<\/p>\n<p>NVIDIA Warp provides a structured approach for simulation, robotics and machine learning workloads, combining the accessibility of Python with performance comparable to native CUDA code.<\/p>\n<p>Warp supports the creation of GPU-accelerated 3D simulation workflows that integrate with ML pipelines in PyTorch, JAX, NVIDIA PhysicsNeMo and NVIDIA Omniverse. This allows developers to run complex simulation tasks and generate data at scale without leaving the Python programming environment.<\/p>\n<p>By offering CUDA-level performance with Python-level productivity, Warp simplifies the development of high-performance simulation workflows. It is designed to accelerate AI research and engineering by reducing barriers to GPU programming, making advanced simulation and data generation more efficient and widely accessible.<\/p>\n<p>Siemens, Neural Concept, Luminary Cloud, among others, are adopting NVIDIA Warp.<\/p>\n<figure id=\"attachment_87400\" aria-describedby=\"caption-attachment-87400\" class=\"wp-caption aligncenter\"><img decoding=\"async\" loading=\"lazy\" class=\"size-medium wp-image-87400\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/11\/bluefield-corp-blog-bluefield-4-1280x680-4468150-960x510.png\" alt=\"NVIDIA BlueField-4 DPU powering the OS of AI factories \" width=\"960\" height=\"510\"><figcaption id=\"caption-attachment-87400\" class=\"wp-caption-text\">NVIDIA BlueField-4 DPU: The Processor Powering the Operating System of AI Factories<\/figcaption><\/figure>\n<h2 id=\"bluefield-4\" class=\"wp-block-heading\"><b>Showcasing BlueField-4 for Powering the OS of AI Factories <a href=\"https:\/\/blogs.nvidia.com\/blog\/accelerated-computing-networking-supercomputing-ai\/#bluefield-4\">????<\/a><\/b><\/h2>\n<p>Unveiled at GTC Washington, D.C., <a target=\"_blank\" href=\"https:\/\/www.nvidia.com\/en-us\/networking\/products\/data-processing-unit\/\" rel=\"noopener\">NVIDIA BlueField-4 DPUs<\/a> are powering the operating system of AI factories. By offloading, accelerating and isolating critical data center functions \u2014 networking, storage and security \u2014 they free up CPUs and GPUs to focus entirely on compute-intensive workloads.<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/resources.nvidia.com\/en-us-accelerated-networking-resource-library\/bluefield-4-dpu-datasheet\" rel=\"noopener\">BlueField-4<\/a>, combining a 64-core NVIDIA Grace CPU and <a target=\"_blank\" href=\"https:\/\/resources.nvidia.com\/en-us-accelerated-networking-resource-library\/connectx-9-supernic-datasheet\" rel=\"noopener\">NVIDIA ConnectX-9<\/a> networking, unlocks unprecedented performance, efficiency and zero-trust security at scale. It supports multi-tenant environments, rapid data access, and real-time protection, with native integration of <a target=\"_blank\" href=\"https:\/\/developer.nvidia.com\/networking\/doca\" rel=\"noopener\">NVIDIA DOCA<\/a> microservices for scalable, containerized AI operations. Together, they are transforming data centers into intelligent, software-defined engines for trillion-token AI and beyond.<\/p>\n<p>As AI factories and supercomputing centers continue to scale in size and capability, they require faster, more intelligent storage infrastructure to manage structured, unstructured and AI-native data for large-scale training and inference.<\/p>\n<p>Leading storage innovators \u2014 DDN, VAST Data and WEKA \u2014 are adopting BlueField-4 to redefine performance and efficiency for AI and scientific workloads.<\/p>\n<ul>\n<li><a target=\"_blank\" href=\"https:\/\/www.ddn.com\/press-releases\/ddn-powers-the-next-generation-of-ai-factories-with-nvidia-bluefield-4\/\" rel=\"noopener\">DDN<\/a> is building next-generation AI factories, accelerating data pipelines to maximize GPU utilization for AI and HPC workloads.<\/li>\n<li><a target=\"_blank\" href=\"https:\/\/www.vastdata.com\/blog\/advancing-the-state-of-the-art-with-nvidia-blue-field-4\" rel=\"noopener\">VAST Data<\/a> is advancing the AI pipeline with intelligent data movement and real-time efficiency across large-scale AI clusters.<\/li>\n<li><a target=\"_blank\" href=\"https:\/\/www.weka.io\/company\/weka-newsroom\/press-releases\/weka-announces-new-neuralmesh-architecture-for-nvidia-bluefield-4\/\" rel=\"noopener\">WEKA<\/a> is launching its NeuralMesh architecture on BlueField-4, running storage services directly on the DPU to simplify and accelerate AI infrastructure.<\/li>\n<\/ul>\n<p>Together, these HPC storage leaders are demonstrating how NVIDIA BlueField-4 transforms data movement and management \u2014 turning storage into a performance multiplier for the next era of supercomputing and AI infrastructure.<\/p>\n<figure id=\"attachment_87404\" aria-describedby=\"caption-attachment-87404\" class=\"wp-caption aligncenter\"><img decoding=\"async\" loading=\"lazy\" class=\"size-medium wp-image-87404\" src=\"https:\/\/blogs.nvidia.com\/wp-content\/uploads\/2025\/11\/infiniband-corp-blog-connectx-9-supernic-c9180-1280x680-1-960x510.jpg\" alt=\"NVIDIA ConnectX-9 SuperNIC\" width=\"960\" height=\"510\"><figcaption id=\"caption-attachment-87404\" class=\"wp-caption-text\">NVIDIA ConnectX-9 SuperNIC<\/figcaption><\/figure>\n<h2 id=\"co-packaged-optics\" class=\"wp-block-heading\"><b>Adopting NVIDIA Co-Packaged Optics for Speed and Reliability\u200b<a href=\"https:\/\/blogs.nvidia.com\/blog\/accelerated-computing-networking-supercomputing-ai\/#co-packaged-optics\">????<\/a><\/b><\/h2>\n<p>TACC, Lambda and CoreWeave unveiled that they will integrate NVIDIA Quantum-X Photonics CPO switches into next generation systems as early as next year.<\/p>\n<p>NVIDIA Quantum-X Photonics networking switches enable AI factories and supercomputing centers to drastically reduce energy consumption and operational costs. NVIDIA has achieved this fusion of electronic circuits and optical communications at massive scale.<\/p>\n<\/p>\n<p>As AI factories grow to unprecedented sizes, networks must evolve to keep pace. By eliminating traditional pluggable transceivers, a common cause of job runtime failures, NVIDIA Photonics switch systems not only deliver 3.5x better power efficiency, but also perform with 10x higher resiliency, enabling applications to run 5x longer without interruption.<\/p>\n<p>At GTC 2024 in Silicon Valley, NVIDIA unveiled NVIDIA Quantum-X800 InfiniBand switches, purpose-built to power trillion-parameter-scale generative AI models. These platforms deliver a staggering 800Gb\/s end-to-end throughput \u2014 2x the bandwidth and 9x the in-network compute of their predecessors \u2014 owing to such innovations as SHARPv4 and FP8 support.<\/p>\n<p>As NVIDIA Quantum\u2011X800 continues to be widely adopted to meet the demands of massive-scale AI, NVIDIA Quantum\u2011X Photonics, announced at GTC earlier this year, addresses the critical power, resiliency, and signal-integrity challenges of even larger deployments. By integrating optics directly on the switch, it eliminates failures caused by pluggable transceivers and link flaps, enabling workloads to run uninterrupted at scale and ensuring the infrastructure can support the next generation of compute-intensive applications up to 5x better than with pluggable transceivers.<\/p>\n<p>\u201cNVIDIA Quantum\u2011X Photonics represents the next step in building high-performance, resilient AI networks,\u201d said Maxx Garrison, product manager for cloud infrastructure at Lambda. \u201cThese advances in power efficiency, signal integrity and reliability, will be key to supporting efficient, large-scale workloads for our customers.\u201d<\/p>\n<p>SHARPv4 enables in-network aggregation and reduction, minimizing GPU-to-GPU communication overhead. Combined with FP8 precision, it accelerates training of trillion-parameter models by reducing bandwidth and compute demands \u2014 delivering faster convergence and higher throughput and comes standard with NVIDIA Quantum\u2011X800 and Quantum\u2011X Photonics switches.<\/p>\n<p>\u201cCoreWeave is building the Essential Cloud for AI,\u201d said Peter Salanki, co-founder and chief technology officer at CoreWeave. \u201cWith NVIDIA Quantum-X Photonics, we\u2019re advancing power efficiency, and further improving the reliability CoreWeave is known for in supporting massive AI workloads at scale, helping our customers unlock the full potential of next-generation AI.\u201d<\/p>\n<p>The NVIDIA Quantum-X Photonics platform, anchored by the NVIDIA Quantum Q3450 CPO-based InfiniBand switch and ConnectX-8 SuperNIC, is engineered for the highest-performance environments that also require significantly lower power, higher resiliency and lower latency.<\/p>\n<h2><b>Supercomputing Centers Worldwide Adopting NVQLink<\/b><\/h2>\n<p>More than a dozen of the world\u2019s top scientific computing centers are adopting <a target=\"_blank\" href=\"https:\/\/nvidianews.nvidia.com\/news\/scientific-supercomputing-centers-nvqlink-grace-blackwell-quantum-processors\" rel=\"noopener\">NVQLink<\/a>, a universal interconnect linking accelerated computing to quantum processors.<\/p>\n<p>NVQLink connects quantum processors with NVIDIA GPUs, enabling large\u2011scale workflows powered by the CUDA\u2011Q software platform. NVQLink\u2019s open architecture provides the critical link supercomputing centers need to integrate diverse quantum processors while delivering 40 petaflops of AI performance at FP4 precision.<\/p>\n<p>In the future every supercomputer will draw on quantum processors to expand the problems they can solve and every quantum processor will depend on GPU supercomputers to run correctly.<\/p>\n<p>Quantum computing company Quantinuum\u2019s new Helios QPU was integrated with NVIDIA GPUs through NVQLink, achieving the world\u2019s first real\u2011time decoding of scalable qLDPC quantum error\u2011correction codes. The system maintained 99% fidelity compared with 95% without correction thanks to NVQLink\u2019s microsecond low latencies.<\/p>\n<p>With NVQLink scientists and developers gain a universal bridge between quantum and classical hardware \u2014 making scalable error correction, hybrid applications and real\u2011time quantum\u2011GPU workflows practical.<\/p>\n<p>In the Asia\u2011Pacific region, Japan\u2019s Global Research and Development Center for Business by Quantum-AI technology (G-QuAT) at the National Institute of Advanced Industrial Science and Technology (AIST) and RIKEN Center for Computational Science, Korea\u2019s Korea Institute of Science and Technology Information (KISTI), Taiwan\u2019s National Center for High-Performance Computing (NCHC), Singapore\u2019s National Quantum Computing Hub (a joint initiative of Singapore\u2019s Centre for Quantum Technologies, A*STAR Institute of High Performance Computing, and National Supercomputing Centre Singapore) \u2014 and Australia\u2019s Pawsey Supercomputing Research Centre are among the early adopters.<\/p>\n<p>Across Europe and the Middle East, NVQLink is being embraced by CINECA, Denmark\u2019s DCAI, operator of Denmark\u2019s AI Supercomputer, France\u2019s Grand \u00c9quipement National de Calcul Intensif (GENCI), the Czech Republic\u2019s IT4Innovations National Supercomputing Center (IT4I), Germany\u2019s J\u00fclich Supercomputing Centre (JSC), Poland\u2019s Pozna\u0144 Supercomputing and Networking Center (PCSS), the Technology Innovation Institute (TII), UAE and Saudi Arabia\u2019s King Abdullah University of Science and Technology (KAUST).<\/p>\n<p>In the United States, leading national laboratories including, Brookhaven National Laboratory, Fermi National Accelerator Laboratory, Lawrence Berkeley National Laboratory, Los Alamos National Laboratory, MIT Lincoln Laboratory, National Energy Research Scientific Computing Center, Oak Ridge National Laboratory, Pacific Northwest National Laboratory and Sandia National Laboratories are also adopting NVQLink to advance hybrid quantum\u2011classical research.<\/p>\n<h2><b>Developing Real\u2011World Hybrid Applications<\/b><\/h2>\n<p>Quantinuum\u2019s Helios QPU with NVQLink delivered:<\/p>\n<ul>\n<li>First real\u2011time decoding of qLDPC error\u2011correction codes<\/li>\n<li>~99% fidelity with NVQLink correction vs ~95% without<\/li>\n<li>Reaction time of 60 microseconds, exceeding Helios\u2019 1\u2011millisecond requirement by 16x<\/li>\n<\/ul>\n<p>NVQLink unites quantum processors with GPU supercomputing for scalable error correction and hybrid applications. Scientists can gain a single programming environment through CUDA\u2011Q APIs. Developers can build and test quantum\u2011GPU workflows in real time<\/p>\n<p>With NVQLink the world\u2019s supercomputing centers are laying the foundation for practical quantum\u2011classical systems, connecting diverse quantum processors to NVIDIA accelerated computing at unprecedented speed and scale.<\/p>\n<h2><b>NVIDIA and <\/b><b>RIKEN<\/b><b> Advance Japan\u2019s Scientific Frontiers<\/b><\/h2>\n<p>NVIDIA and RIKEN are building <a target=\"_blank\" href=\"https:\/\/nvidianews.nvidia.com\/news\/nvidia-and-riken-advance-japans-scientific-frontiers-with-new-supercomputers-for-ai-and-quantum-computing\" rel=\"noopener\">two new GPU\u2011accelerated supercomputers<\/a> to expand Japan\u2019s leadership in AI for science and quantum computing. Together the systems will feature 2,140 NVIDIA Blackwell GPUs connected through the GB200 NVL4 platform and NVIDIA Quantum\u2011X800 InfiniBand networking, strengthening Japan\u2019s sovereign AI strategy and secure domestic infrastructure.<\/p>\n<ul>\n<li>AI for Science System: 1,600 Blackwell GPUs will power research in life sciences, materials science, climate and weather forecasting, manufacturing and laboratory automation.<\/li>\n<li>Quantum Computing System: 540 Blackwell GPUs will accelerate quantum algorithms, hybrid simulation and quantum\u2011classical methods.<\/li>\n<\/ul>\n<p>The partnership builds on RIKEN\u2019s collaboration with Fujitsu and NVIDIA to codesign FugakuNEXT, successor to the Fugaku supercomputer, expected to deliver 100x greater application performance and integrate production\u2011level quantum computers by 2030.<\/p>\n<p>The two new RIKEN systems are scheduled to be operational in spring 2026.<\/p>\n<h2 id=\"arm\" class=\"wp-block-heading\"><b>Arm Adopting NVIDIA NVLink Fusion <a href=\"https:\/\/blogs.nvidia.com\/blog\/accelerated-computing-networking-supercomputing-ai\/#arm\">????<\/a><\/b><\/h2>\n<p>AI is reshaping data centers in a once-in-a-generation architectural shift, where efficiency per watt defines success. At the center is Arm Neoverse, deployed in over a billion cores and projected to reach 50% hyperscaler market share by 2025. Every major provider \u2014 AWS, Google, Microsoft, Oracle and Meta \u2014 is building on Neoverse, underscoring its role in powering AI at scale.<\/p>\n<p>To meet surging demand, Arm is extending Neoverse with NVIDIA NVLink Fusion, the high-bandwidth, coherent interconnect first pioneered with Grace Blackwell. NVLink Fusion links CPUs, GPUs, and accelerators into one unified rack-scale architecture, removing memory and bandwidth bottlenecks that limit AI performance. Connected with Arm\u2019s AMBA CHI C2C protocol, it ensures seamless data movement between Arm-based CPUs and partners\u2019 preferred accelerators.<\/p>\n<p>Together, Arm and NVIDIA are setting a new standard for AI infrastructure, enabling ecosystem partners to build differentiated, energy-efficient systems that accelerate innovation across the AI era.<\/p>\n<h2><b>Smarter Power for Accelerated Computing<\/b><\/h2>\n<p>As AI factories scale, energy is becoming the new bottleneck. The <a target=\"_blank\" href=\"https:\/\/catalog.ngc.nvidia.com\/orgs\/nvidia\/collections\/dps\" rel=\"noopener\">NVIDIA Domain Power Service (DPS)<\/a> flips that constraint into an opportunity \u2014 turning power into a dynamic, orchestrated resource. Running as a Kubernetes service, DPS models and manages energy use across the data center, from rack to room to facility. It enables operators to extract more performance per megawatt by constraining power intelligently, improving throughput without expanding infrastructure.<\/p>\n<p>DPS integrates tightly with the <a href=\"https:\/\/blogs.nvidia.com\/blog\/omniverse-dsx-blueprint\/\">NVIDIA Omniverse DSX Blueprint<\/a>, a platform for designing and operating next-generation data centers. It works alongside technologies like Power Reservation Steering to balance workloads across the facility and the Workload Power Profile Solution to tune GPU power to the needs of specific jobs. Together, they form DSX Boost \u2014 an energy-aware control layer that maximizes efficiency while meeting performance targets.<\/p>\n<p>DPS also extends beyond the data center. With grid-facing APIs, it supports automated load shedding and demand response, helping utilities stabilize the grid during peak events. The result is a resilient, grid-interactive AI factory that turns every watt into measurable progress.<\/p>\n<\/p><\/div>\n","protected":false},"excerpt":{"rendered":"<p>https:\/\/blogs.nvidia.com\/blog\/accelerated-computing-networking-supercomputing-ai\/<\/p>\n","protected":false},"author":0,"featured_media":4356,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":[],"categories":[3],"tags":[],"_links":{"self":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/4355"}],"collection":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/comments?post=4355"}],"version-history":[{"count":0,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/posts\/4355\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media\/4356"}],"wp:attachment":[{"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/media?parent=4355"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/categories?post=4355"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/salarydistribution.com\/machine-learning\/wp-json\/wp\/v2\/tags?post=4355"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}