Banner ImageMobile Banner Image

GIGABYTE Solutions for NVIDIA Rubin Platform

GIGABYTE solutions usher in a new era of Agentic AI, built on the NVIDIA Rubin platform to accelerate intelligent reasoning, autonomous decision-making, and scalable AI infrastructure.

The Expert in Agentic AI and AI Reasoning

The era of agentic AI, a frequently discussed future, is now finally within reach. With the NVIDIA Rubin platform, purpose-built for agentic AI and reasoning, GIGABYTE strives to deliver the most efficient solutions across industries, supporting a wide range of deployment scenarios.

The NVIDIA Rubin platform goes far beyond a simple GPU upgrade. It introduces six new chips, covering CPU, GPU, NVLink Switch, DPU, NIC, and Ethernet switch, accelerating every aspect of AI computation.


The Five Generational Breakthroughs

6th Gen NVLink & NVLink Switch

3.6 TB/s bandwidth per GPU for bandwidth-intensive applications like AI inference, featuring NVIDIA® SHARP™.

Vera CPU

Combines 88 NVIDIA-designed cores, up to 1.2 TB/s of LPDDR5X memory bandwidth, and Scalable Coherency Fabric.

3rd Gen Transformer Engine

Enables up to 50 PetaFLOPS NVFP4 for inference with new hardware-accelerated adaptive compression.

3rd Gen Confidential Computing

The world's first rack-scale confidential computing across CPU, GPU, and NVLink™ domains.

2nd Gen RAS Engine

Enables continuous in-system health monitoring, self-testing, and SRAM repair.

NVIDIA Vera Rubin NVL72

Unmatched Dense Performance

Delivers extreme GPU density in a single rack, enabling massive performance for trillion-parameter AI models and large-scale training workloads.

Designed for Next-Generation AI Factories

Built specifically for AI training efficiency and inference cost reduction. Tuned for large models and high throughput, providing maximum efficiency where every millisecond counts.

Integrated, Fully Engineered System

Comes as a cohesively engineered rack system, including custom cooling, power distribution, and networking, enabling rapid, cable-free deployment at scale.
Content Image

Specifications1

NVFP4 Inference3,600 PFLOPS
NVFP4 Training22,520 PFLOPS
FP8 / FP6 Training21,260 PFLOPS
INT8218 POPS
FP16 / BF162288 PFLOPS
TF322144 PFLOPS
FP329,360 TFLOPS
FP642,400 TFLOPS
FP32 SGEMM328,800 TFLOPS
FP64 DGEMM314,400 TFLOPS
GPU Memory20.7 TB HBM4
Bandwidth1,580 TB/s
NVLink Bandwidth260 TB/s
NVLink-C2C Bandwidth65 TB/s

1 All values are up to and subject to change.

2 Dense specification.

3 Peak performance using Tensor Core-based emulation algorithms.

NVIDIA Rubin NVL8

Flexible Scaling for Any Deployment

Scale from single-node to multi-node GPU clusters without committing to a full rack-scale architecture. Ideal for phased expansion and mixed AI workloads.

Lower Infrastructure Requirements

Fits into more standard server and rack environments, reducing the need for specialized power, cooling, and facility redesign.

Broader Platform Compatibility

Supports a wider range of configurations, networking choices, and workload types. Fits perfectly into enterprises running everything from AI training to inference and HPC.
Content Image
GPU8x NVIDIA Rubin GPUs
Total GPU Memory | Bandwidth2.3 TB | 160 TB/s
CPU2x Intel® Xeon® 6 processors
NVIDIA NVLink Switch System4x
NVIDIA NVLink Bandwidth28.8 TB/s total bandwidth
Networking

8x OSFP ports serving 8x single-port NVIDIA® ConnectX®-9 VPI

- up to 800 Gb/s NVIDIA InfiniBand and Ethernet

2x 400G QSP112 NVIDIA BlueField®-4 DPUs

- up to 800 Gb/s NVIDIA InfiniBand and Ethernet

Why GIGABYTE ?

Feature Icon

Quick

Short TTM for Agile Deployment
Feature Icon

Expansion

Flexible Scalability for Diverse Scenarios
Feature Icon

Combination

One-Stop Deployment for Zero Hassle
Feature Icon

Dashboard

Unified Management for Easy Maintenance
Feature Icon

Friendly

Strong Partnership for All-Round Support
Feature Icon

Global Service

Extensive Global Experience for Maximum Flexibility

GIGAPOD - One-Stop Scalable Solutions

At GIGABYTE, we offer GIGAPOD, a solution that scales from a single rack to POD-scale and containerized data center, with power, cabling, cooling, and all infrastructure carefully designed and evaluated. Providing a simple, one-step, pain-free adoption of AI data centers.  Learn more about GIGAPOD
Content Image

Ready-to-Deploy Software Ecosystem

Accelerate time-to-value with a flexible, pre-validated software stack. From single-node management to POD-scale orchestration, it delivers seamless deployment and adaptable control for any workload.
GIGABYTE POD Manager (GPM)

GIGABYTE POD Manager (GPM)

Total POD-level orchestration for infrastructure provisioning and environmental monitoring
NVIDIA AI Enterprise

NVIDIA AI Enterprise

The end-to-end software suite for developing and deploying production-grade AI