AIBench Training: Balanced Industry-Standard AI Training Benchmarking
AIBench Training: Balanced Industry-Standard AI Training Benchmarking
About this item
Full title
Author / Creator
Tang, Fei , Gao, Wanling , Zhan, Jianfeng , Chuanxin Lan , Xu, Wen , Wang, Lei , Luo, Chunjie , Dai, Jiahui , Cao, Zheng , Xiong, Xingwang , Jiang, Zihan , Hao, Tianshu , Fan, Fanda , Zhang, Fan , Huang, Yunyou , Chen, Jianan , Du, Mengjia , Ren, Rui , Chen, Zheng , Zheng, Daoyi , Tang, Haoning , Zhan, Kunlin , Wang, Biao , Kong, Defei , Yu, Minghe , Tan, Chongkang , Li, Huan , Tian, Xinhui , Li, Yatao , Shao, Junchao , Wang, Zhenyu , Wang, Xiaoyu and Ye, Hainan
Publisher
Ithaca: Cornell University Library, arXiv.org
Journal title
Language
English
Formats
Publication information
Publisher
Ithaca: Cornell University Library, arXiv.org
Subjects
More information
Scope and Contents
Contents
Earlier-stage evaluations of a new AI architecture/system need affordable benchmarks. Only using a few AI component benchmarks like MLPerfalone in the other stages may lead to misleading conclusions. Moreover, the learning dynamics are not well understood, and the benchmarks' shelf-life is short. This paper proposes a balanced benchmarking methodology. We use real-world benchmarks to cover the factors space that impacts the learning dynamics to the most considerable extent. After performing an exhaustive survey on Internet service AI domains, we identify and implement nineteen representative AI tasks with state-of-the-art models. For repeatable performance ranking (RPR subset) and workload characterization (WC subset), we keep two subsets to a minimum for affordability. We contribute by far the most comprehensive AI training benchmark suite. The evaluations show: (1) AIBench Training (v1.1) outperforms MLPerfTraining (v0.7) in terms of diversity and representativeness of model complexity, computational cost, convergent rate, computation, and memory access patterns, and hotspot functions; (2) Against the AIBench full benchmarks, its RPR subset shortens the benchmarking cost by 64%, while maintaining the primary workload characteristics; (3) The performance ranking shows the single-purpose AI accelerator like TPU with the optimized TensorFlowframework performs better than that of GPUs while losing the latter's general support for various AI models. The specification, source code, and performance numbers are available from the AIBench homepage https://www.benchcouncil.org/aibench-training/index.html....
Alternative Titles
Full title
AIBench Training: Balanced Industry-Standard AI Training Benchmarking
Authors, Artists and Contributors
Author / Creator
Gao, Wanling
Zhan, Jianfeng
Chuanxin Lan
Xu, Wen
Wang, Lei
Luo, Chunjie
Dai, Jiahui
Cao, Zheng
Xiong, Xingwang
Jiang, Zihan
Hao, Tianshu
Fan, Fanda
Zhang, Fan
Huang, Yunyou
Chen, Jianan
Du, Mengjia
Ren, Rui
Chen, Zheng
Zheng, Daoyi
Tang, Haoning
Zhan, Kunlin
Wang, Biao
Kong, Defei
Yu, Minghe
Tan, Chongkang
Li, Huan
Tian, Xinhui
Li, Yatao
Shao, Junchao
Wang, Zhenyu
Wang, Xiaoyu
Ye, Hainan
Identifiers
Primary Identifiers
Record Identifier
TN_cdi_proquest_journals_2397155022
Permalink
https://devfeature-collection.sl.nsw.gov.au/record/TN_cdi_proquest_journals_2397155022
Other Identifiers
E-ISSN
2331-8422