Alibaba Reveals Network and Data Center Design It Uses for Training Large Language Models

Alibaba has revealed the layout of its LLM training data center, which apparently consists of an Ethernet-based network in which each host contains eight GPUs and nine NICs, each with two 200GB/s ports.

The tech giant, which also offers one of the best large language models (LLM) through its Qwen model, trained on 110 billion parameters, says this design has been used in production for eight months and aims to maximize the use of a GPU. PCIe capabilities that increase network sending/receiving capacity.

scroll to top