
Mountain View — Google Cloud has unveiled a major networking upgrade for its AI Hypercomputer platform, introducing NCCL/gIB, an enhanced version of NVIDIA’s NCCL designed to particularly improve GPU-to-GPU communication.
The new technology is optimized specifically for Google Cloud’s high-performance infrastructure and is expected to deliver substantial gains for large-scale AI training workloads.
AI Hypercomputer Gains Speed With NCCL/gIB Integration
The introduction of NCCL/gIB marks a strategic step in Google Cloud’s effort to accelerate distributed training across massive GPU clusters.
Early internal benchmarks indicate that the enhanced communication layer can improve throughput by up to 30–40% in multi-node GPU environments, especially for models requiring high-bandwidth, low-latency interconnects.
According to Google Cloud engineers, NCCL/gIB leverages advanced networking features such as adaptive routing, optimized collective operations, and improved congestion control.
These enhancements allow AI Hypercomputer clusters to maintain stable performance even when scaling to thousands of GPUs.
📣 New docs alert! Learn how to optimize AI Hypercomputer cluster networking with NCCL/gIB, our enhanced version of NVIDIA's NCCL → https://t.co/FmiQ08YJ6O
It's highly optimized for Google Cloud and can deliver significant performance gains for GPU-to-GPU communication. pic.twitter.com/BMEzTn1N0U
— Google Cloud Tech (@GoogleCloudTech) December 15, 2025
“As AI models grow larger and more complex, communication efficiency becomes just as important as raw compute power,” a Google Cloud spokesperson said. “NCCL/gIB is designed to remove bottlenecks and help customers train next-generation models faster and more reliably.”
Optimized for Google Cloud’s High-Performance Infrastructure
Google Cloud’s AI Hypercomputer platform already supports some of the world’s most demanding AI workloads, including generative AI, scientific simulations, and large-scale language model training.
With NCCL/gIB, developers and enterprises can expect smoother scaling, reduced training time, and improved cost efficiency.
The upgrade is particularly beneficial for workloads that rely heavily on collective operations such as all-reduce, all-gather, and broadcast — operations that often dominate communication overhead in large GPU clusters. By optimizing these operations, NCCL/gIB helps ensure that compute resources remain fully utilized.
Google Cloud highlighted that NCCL/gIB is tightly integrated with its networking stack, including advanced InfiniBand and gRPC-based communication layers.
This integration allows the system to intelligently manage traffic across thousands of GPUs, reducing latency spikes and improving overall cluster stability.
According to Google Cloud’s official documentation, NCCL/gIB is now available for customers using AI Hypercomputer clusters and will roll out to additional regions in early 2026.
Developers can access detailed configuration guides and performance tuning recommendations through Google Cloud’s learning resources.
With AI adoption accelerating globally, Google Cloud’s latest upgrade positions the AI Hypercomputer as one of the most advanced platforms for large-scale training, offering both speed and reliability for enterprises building next-generation AI systems.

