TY - JOUR AU - Huang, Di AU - Zhang, Xishan AU - Zhang, Rui AU - Zhi, Tian AU - He, Deyuan AU - Guo, Jiaming AU - Liu, Chang AU - Guo, Qi AU - Du, Zidong AU - Liu, Shaoli AU - Chen, Tianshi AU - Chen, Yunji PY - 2020/04/03 Y2 - 2024/03/28 TI - DWM: A Decomposable Winograd Method for Convolution Acceleration JF - Proceedings of the AAAI Conference on Artificial Intelligence JA - AAAI VL - 34 IS - 04 SE - AAAI Technical Track: Machine Learning DO - 10.1609/aaai.v34i04.5838 UR - https://ojs.aaai.org/index.php/AAAI/article/view/5838 SP - 4174-4181 AB - <p>Winograd's minimal filtering algorithm has been widely used in Convolutional Neural Networks (CNNs) to reduce the number of multiplications for faster processing. However, it is only effective on convolutions with kernel size as 3x3 and stride as 1, because it suffers from significantly increased FLOPs and numerical accuracy problem for kernel size larger than 3x3 and fails on convolution with stride larger than 1. In this paper, we propose a novel Decomposable Winograd Method (DWM), which breaks through the limitation of original Winograd's minimal filtering algorithm to a wide and general convolutions. DWM decomposes kernels with large size or large stride to several small kernels with stride as 1 for further applying Winograd method, so that DWM can reduce the number of multiplications while keeping the numerical accuracy. It enables the fast exploring of larger kernel size and larger stride value in CNNs for high performance and accuracy and even the potential for new CNNs. Comparing against the original Winograd, the proposed DWM is able to support all kinds of convolutions with a speedup of ∼2, without affecting the numerical accuracy.</p> ER -