AICB: A benchmark for evaluating the communication subsystem of LLM training clusters

AICB (Artificial Intelligence Communication Benchmark) is a benchmark for evaluating the communication subsystem of GPU clusters, which includes representative workloads in the fields of Large Language Model (LLM) training. Guided by the theories and methodologies of Evaluatology, we simplified the...

Full description

Saved in:
Bibliographic Details
Main Authors: Xinyue Li, Heyang Zhou, Qingxu Li, Sen Zhang, Gang Lu
Format: Article
Language:English
Published: KeAi Communications Co. Ltd. 2025-03-01
Series:BenchCouncil Transactions on Benchmarks, Standards and Evaluations
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S2772485925000250
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:AICB (Artificial Intelligence Communication Benchmark) is a benchmark for evaluating the communication subsystem of GPU clusters, which includes representative workloads in the fields of Large Language Model (LLM) training. Guided by the theories and methodologies of Evaluatology, we simplified the real-workload LLM training systems through AICB that maintain good representativeness and usability. AICB bridges the gap between application benchmarks and microbenchmarks in the scope of LLM training. In addition, we constructed a new GPU-free evaluation system that helps researchers evaluate the communication system of the LLM training systems. To help the urgent demand on this evaluation subject, we open-source AICB and make it available at https://github.com/aliyun/aicb.
ISSN:2772-4859