A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2020; you can also visit the original URL.
The file type is application/pdf
.
Understanding Scalability and Fine-Grain Parallelism of Synchronous Data Parallel Training
2019
2019 IEEE/ACM Workshop on Machine Learning in High Performance Computing Environments (MLHPC)
In the age of big data, deep learning has emerged as a powerful tool to extract insight and exploit its value, both in industry and scientific applications. With increasing complexity of learning models and amounts of training data, data-parallel approaches based on frequent all-reduce synchronization steps are increasingly popular. Despite the fact that high performance computing (HPC) technologies have been designed to address such patterns efficiently, the behavior of data-parallel
doi:10.1109/mlhpc49564.2019.00006
dblp:conf/sc/LiNWB19
fatcat:pcxwhll7xncrdp2m652gpx323u