Say goodbye to communication bottlenecks
Developing a sub-nanosecond AI compute network
Ultra-Low Latency Architecture
End-to-end latency as low as 0.8 µs (20x faster than traditional TCP/IP)
Supports 200/400/800 Gbps InfiniBand standards
Topology-aware Scheduling
Dynamic routing optimization and automatic inter-rack traffic balancing Deep optimization of the NCCL library results in a 40% acceleration of AllReduce operations
Single-rack Topology
32 H100 GPUs fully interconnected via NVSwitch
900GB/s bisection bandwidth
Supports Multi-cluster Scaling
SHARP-based distributed routing
enabling expansion to 1,000-10,000 nodes
Results: 175B parameter model training task
communication overhead was reduced from 35% to 8%, inter-rack traffic was decreased by 60%
Dimensions
Network Layer
Compute Layer
Storage Layer
Para ofrecer las mejores experiencias, utilizamos tecnologías como las cookies para almacenar y/o acceder a la información del dispositivo. Consentir estas tecnologías nos permitirá procesar datos como el comportamiento de navegación o identificaciones únicas en este sitio.
Su viaje AI comienza aquí.
Rellene el formulario y le responderemos.