Say goodbye to communication bottlenecks
Developing a sub-nanosecond AI compute network
Ultra-Low Latency Architecture
End-to-end latency as low as 0.8 µs (20x faster than traditional TCP/IP)
Supports 200/400/800 Gbps InfiniBand standards
Topology-aware Scheduling
Dynamic routing optimization and automatic inter-rack traffic balancing Deep optimization of the NCCL library results in a 40% acceleration of AllReduce operations
Single-rack Topology
32 H100 GPUs fully interconnected via NVSwitch
900GB/s bisection bandwidth
Supports Multi-cluster Scaling
SHARP-based distributed routing
enabling expansion to 1,000-10,000 nodes
Results: 175B parameter model training task
communication overhead was reduced from 35% to 8%, inter-rack traffic was decreased by 60%
Dimensions
Network Layer
Compute Layer
Storage Layer
최상의 경험을 제공하기 위해 당사는 쿠키와 같은 기술을 사용하여 기기 정보를 저장 및/또는 액세스합니다. 이러한 기술에 동의하면 이 사이트에서 검색 행동이나 고유 ID와 같은 데이터를 처리할 수 있습니다.
AI 여정은 여기서 시작됩니다.
양식을 작성해 주시면 답변을 보내드리겠습니다.