AI / ML | AI Server

  |  AGS8200

AGS8200

The Edgecore AGS8200 is a cutting-edge, high-performance GPU-based server tailored to meet the demands of AI/ML applications. Designed to excel in tasks such as training large language models, automating processes, and enhancing object classification and recognition, this server offers unrivaled performance and scalability.

At the heart of the AGS8200 lies the formidable Intel® Habana® Gaudi® 2 processors, eight in total, and dual Xeon® Sapphire-Rapids processors. These components unite to create a computing powerhouse, ready to tackle a diverse range of deep learning workloads with exceptional speed and precision.

Why Choose the AGS8200 ?

The Power of Intel® Gaudi® 2

Audi
Audi

Intel® Gaudi® 2 key benefits

■ reference/source for MLPerf image

Features

Edgecore AGS8200 is ideal for modern AI (Artificial Intelligence) and ML (Machine Learning) applications. Powered by Intel® Gaudi® 2 AI accelerator, the AGS8200 is suitable for LLM (Large Language Models) training and inferencing, allowing customers to efficiently harness the power of AI.

The system is designed with eight Intel® Habana® Gaudi® 2 processors and dual Xeon® Sapphire-Rapids processors. The Gaudi® 2 processor integrates 96GB HBM2E memory and 24 NICs of 100GbE RoCEv2 RDMA. The 24 100G NICs offer all-to-all connectivity and scale-out internally and externally for training, fine-tuning, and other DL processing.

Each Gaudi® 2 processor has 21 x 100GbE non-blocking, all-to-all connectivity to other Gaudi® 2 processors within the server, allowing training across all eight Intel® Gaudi® 2 processors without requiring an external Ethernet switch.

Each AGS8200 supports 6 QSFP-DD ports for scale-out. The 400G ports can be connected to 400G switches, or 100G switches via breakout cables, in racks and clusters of Intel® Gaudi® 2-based nodes.

The server is able to include 16 x HDD/SSD + 8 x NVMe or 8 x HDD/SSD + 16 x NVMe for storage that supports RAID HBA with 0/1/10/5/6/50/60.

■ Featuring eight Habana® Gaudi® 2 AI training processors
■ Dual 4th Gen Intel® Xeon® scalable processors
■ Expanded networking capacity with 24 x 100GE RoCE ports integrated into every Gaudi® 2
■ 700 GB/second scale within the server and 2.4TB/second scale out
■ Ease of system build or migration with Habana® SynapseAI® soſtware stack
■ Standardized architecture and Ethernet instead of proprietary Infiniband and NVLINK

Specifications

Form Factor
■ 8U

Compute Node
■ CPU: Sapphire Rapids, 2 Sockets
Intel® Xeon® Platinum 8454H, 32c, 64 threads, 82.5MB, 270 W
■ PCH: Emmitsburg
■ Memory: Upto 2TB 16x DDR5 memory slots per CPU
■ Operating System: Ubuntu 20.04
■ BIOS: 32MB Flash

GPU
■ 8 x OAM (Intel Habana HL-225H/C)

Input/Output
■ Front: 2 x USB 2.0/3.0, 1 x VGA, 1 x UID, 1 x PWR
■ Rear: 2 x USB 2.0/3.0, 1 x VGA, 1 x RJ-45, 1 x UID

Scale-Out Interface
■ RDMA (ROCE v2)
■ 24 x 100 GbE
■ 6 x QSFP-DD

Storage
■ Internal: 2 x M.2
■ Front: 16 x HDD/SSD+ 8 x NVMe or 8 x HDD/SSD+ 16 x NVMe

BMC
■ AST2600

TPM 2.0

CD-ROM
■ Support external USB CD-ROM

PSU
■ System: 1+1 CRPS 2700 W redundant/hot-swappable AC/DC
■ GPU: 3+3 CRPS 3000 W redundant/hot-swappable AC/DC

Fans
■ 14+1 hot-swappable fans

Dimensions
■ 900 mm x 447 mm x 352mm

Operating Temperature
■ 5°C-35°C

Expansion Slots
■ 1 x OCP 3.0
■ 8 x PCle Slots

Soſtware
■ SynapseAl: 1.13.0
■ Kernel: 5.4.0 and above
■ Python: 3.10
■ PyTorch: 2.1.0
■ TensorFlow: 2.13.1
■ Open MPI: 4.1.5
■ Libfabric: 1.16.1 and above
■ Transformers: >= 4.33.0, <4.35.0

For more information, contact us.

Related Solutions

Related Resource

2023 Product Catalogue

Edgecore Announces an 800G-Optimized Switch that Provides an Ethernet Fabric for AI/ML Workloads

Unleash 5G’s Potential with Edgecore’s DCSG – Flexibility, Control, Stability

Edgecore Networks Partners with Exaware and FONEX to Deliver an Open Multi-Service OLT Access Router Using Tibit Transceivers