Back to Search
Start Over
Scalable training of 3D convolutional networks on multi- and many-cores
- Source :
- Journal of Parallel and Distributed Computing. 106:195-204
- Publication Year :
- 2017
- Publisher :
- Elsevier BV, 2017.
-
Abstract
- Convolutional networks (ConvNets) have become a popular approach to computer vision. Here we consider the parallelization of ConvNet training, which is computationally costly. Our novel parallel algorithm is based on decomposition into a set of tasks, most of which are convolutions or FFTs. Theoretical analysis suggests that linear speedup with the number of processors is attainable. To attain such performance on real shared-memory machines, our algorithm computes convolutions converging on the same node of the network with temporal locality to reduce cache misses, and sums the convergent convolution outputs via an almost wait-free concurrent method to reduce time spent in critical sections. Benchmarking with multi-core CPUs shows speedup roughly equal to the number of physical cores. We also demonstrate 90x speedup on a many-core CPU (Xeon Phi Knights Corner). Our algorithm can be either faster or slower than certain GPU implementations depending on specifics of the network architecture, kernel sizes, and density and size of the output patch.
- Subjects :
- Speedup
Computer Networks and Communications
Computer science
Node (networking)
Parallel algorithm
Dynamic priority scheduling
Parallel computing
010501 environmental sciences
01 natural sciences
Convolutional neural network
Theoretical Computer Science
03 medical and health sciences
0302 clinical medicine
Artificial Intelligence
Hardware and Architecture
Scalability
Locality of reference
Cache
030217 neurology & neurosurgery
Software
Xeon Phi
0105 earth and related environmental sciences
Subjects
Details
- ISSN :
- 07437315
- Volume :
- 106
- Database :
- OpenAIRE
- Journal :
- Journal of Parallel and Distributed Computing
- Accession number :
- edsair.doi...........488eec3e9bfc1936f61495dce23f6509
- Full Text :
- https://doi.org/10.1016/j.jpdc.2017.02.006