Parallel and Distributed Deep Learning Paper Database
The paper database below was collected for the purpose of the paper "Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis". It contains works that utilize parallel and distributed computing resources for training Deep Neural Networks. This includes hardware architectures, data representation, parallelization strategies, distributed algorithms, system implementations, frameworks, and programming models.
Format
The papers are organized in the YAML format, with one entry per paper, sorted by publication year. Each entry contains the paper title, year of publication, category (corresponding to sections in the paper), keywords, datasets used in experiments, frameworks, and hardware architectures.
The following listing shows an example of such a paper entry:
'Large Scale Distributed Deep Networks': year: 2012 categories: - Distributed - Systems keywords: - Asynchronous SGD - DistBelief - Downpour SGD - Sandblaster LBFGS - Parameter server - Model-parallelism - Data-parallelism - Layer pipelining - Hybrid parallelism hardware: - CPU Cluster: nodes: 5100 commlayer: Sockets experiments: datasets: - Speech Recognition (internal) - ImageNet networks: - 4-layer MLP - LCN
Download
Version | Date | Changes |
papers.yml - (Coming Soon) | February 26, 2018 | First Release |
References
[1] T. Ben-Nun, T. Hoefler: | ||
Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis
ACM Comput. Surv.. Vol 52, Nr. 4, pages 65:1--65:43, ACM, ISSN: 0360-0300, Aug. 2019, |