Searching for just a few words should be enough to get started. If you need to make more complex queries, use the tips below to guide you.
Article type: Research Article
Authors: Zhang, Wei Mina; * | Zhang, Longb | Zhang, Zheyua | Sun, Mingjuna
Affiliations: [a] Artificial Intelligence Department, China Academy of Information and Communication Technology (CAICT), Haidian District, Beijing, China | [b] Research and Development Department, Beijing Qinghong Technology Co., Ltd., Chaoyang District, Beijing, China
Correspondence: [*] Corresponding author. Wei Min Zhang, Artificial Intelligence Department, China Academy of Information and Communication Technology (CAICT), No. 52 Huayuan Bei Road, Haidian District, Beijing, China. E-mail: zhangweimin@caict.ac.cn.
Note: [1] IBD is short for Inference Benchmark for DNN processor, this research has been partially supported by the AIIA DNN benchmark project, and open source tools can be find in: https://github.com/AIIABenchmark/AIIA-DNN-benchmark.
Abstract: With the many varieties of AI hardware prevailing on the market, it is often hard to decide which one is the most suitable to use but not only with the best performance. As there is an industry-wide trend demand for deep learning deployment, the inference benchmark for the effectiveness of DNN processor becomes important and is of great help to select and optimize AI hardware. To systematically benchmark deep learning deployment platforms, and give more objective and useful metrics comparison. In this paper, an end to end benchmark evaluation system was brought up called IBD, it combined 4 steps include three components with 6 metrics. The performance comparison results are obtained from the chipsets from Qualcomm, HiSilicon, and NVIDIA, which can provide hardware acceleration for AI inference. To comprehensively reflect the current status of the DNN processor deploying performance, we chose six devices from three kinds of deployment scenarios which are cloud, desktop and mobile, ten models from three different kinds of applications with diverse characteristics are selected, and all these models are trained from three major training frameworks. Several important observations were made by using our methodologies. Experimental results showed that workload diversity should focus on the difference came from training frameworks, inference frameworks with specific processors, input size and precision (floating and quantized).
Keywords: AI, deep neural network processor, benchmark, end to end, inference
DOI: 10.3233/JIFS-202552
Journal: Journal of Intelligent & Fuzzy Systems, vol. 40, no. 5, pp. 9949-9961, 2021
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
USA
Tel: +1 703 830 6300
Fax: +1 703 830 2300
sales@iospress.com
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
IOS Press
Nieuwe Hemweg 6B
1013 BG Amsterdam
The Netherlands
Tel: +31 20 688 3355
Fax: +31 20 687 0091
info@iospress.nl
For editorial issues, permissions, book requests, submissions and proceedings, contact the Amsterdam office info@iospress.nl
Inspirees International (China Office)
Ciyunsi Beili 207(CapitaLand), Bld 1, 7-901
100025, Beijing
China
Free service line: 400 661 8717
Fax: +86 10 8446 7947
china@iospress.cn
For editorial issues, like the status of your submitted paper or proposals, write to editorial@iospress.nl
如果您在出版方面需要帮助或有任何建, 件至: editorial@iospress.nl