Back to Search
Start Over
Scaling Distributed Training of Flood-Filling Networks on HPC Infrastructure for Brain Mapping
- Source :
- DLS@SC
- Publication Year :
- 2019
- Publisher :
- arXiv, 2019.
-
Abstract
- Mapping all the neurons in the brain requires automatic reconstruction of entire cells from volume electron microscopy data. The flood-filling network (FFN) architecture has demonstrated leading performance for segmenting structures from this data. However, the training of the network is computationally expensive. In order to reduce the training time, we implemented synchronous and data-parallel distributed training using the Horovod library, which is different from the asynchronous training scheme used in the published FFN code. We demonstrated that our distributed training scaled well up to 2048 Intel Knights Landing (KNL) nodes on the Theta supercomputer. Our trained models achieved similar level of inference performance, but took less training time compared to previous methods. Our study on the effects of different batch sizes on FFN training suggests ways to further improve training efficiency. Our findings on optimal learning rate and batch sizes agree with previous works.<br />Comment: 9 pages, 10 figures
- Subjects :
- Scheme (programming language)
FOS: Computer and information sciences
Computer Science - Machine Learning
Computer science
Image and Video Processing (eess.IV)
Training (meteorology)
Volume (computing)
Inference
Electrical Engineering and Systems Science - Image and Video Processing
Supercomputer
Machine Learning (cs.LG)
Computer Science - Distributed, Parallel, and Cluster Computing
Computer engineering
Asynchronous communication
Quantitative Biology - Neurons and Cognition
FOS: Biological sciences
Code (cryptography)
FOS: Electrical engineering, electronic engineering, information engineering
Neurons and Cognition (q-bio.NC)
Distributed, Parallel, and Cluster Computing (cs.DC)
computer
Scaling
computer.programming_language
Subjects
Details
- Database :
- OpenAIRE
- Journal :
- DLS@SC
- Accession number :
- edsair.doi.dedup.....0bc7d4cd5a032f89caf7c32915b5e39b
- Full Text :
- https://doi.org/10.48550/arxiv.1905.06236