Back to Search
Start Over
Accelerating Neural Network Training using Arbitrary Precision Approximating Matrix Multiplication Algorithms
- Source :
- ICPP Workshops
- Publication Year :
- 2021
- Publisher :
- ACM, 2021.
-
Abstract
- Matrix multiplication is one of the bottleneck computations for training the weights within deep neural networks. To speed up the training phase, we propose to use faster algorithms for matrix multiplication known as Arbitrary Precision Approximating (APA) algorithms. APA algorithms perform asymptotically fewer arithmetic operations than the classical algorithm, but they compute an approximate result with an error that can be made arbitrarily small in exact arithmetic. Practical APA algorithms provide significant reduction in computation time and still provide enough accuracy for many applications like neural network training. We demonstrate that APA algorithms can be efficiently implemented and parallelized for multicore CPUs to obtain up to 28% and 21% speedups over the fastest implementation of the classical algorithm using one core and 12 cores, respectively. Furthermore, using these algorithms to train a Multi-Layer Perceptron (MLP) network yields no significant reduction in the training or testing error. Our performance results on a large MLP network show overall sequential and multithreaded performance improvements of up to 25% and 13%, respectively. We also demonstrate up to 15% improvement when training the fully connected layers of the VGG-19 image classification network.
Details
- Database :
- OpenAIRE
- Journal :
- 50th International Conference on Parallel Processing Workshop
- Accession number :
- edsair.doi...........1420879dabbf18c20a9b99a28ac335df