A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2022; you can also visit the original URL.
The file type is application/pdf
.
Behemoth: A Flash-centric Training Accelerator for Extreme-scale DNNs
2021
USENIX Conference on File and Storage Technologies
The explosive expansion of Deep Neural Networks (DNN) model size expedites the need for larger memory capacity. This movement is particularly true for models in natural language processing (NLP), a dominant application of AI along with computer vision. For example, a recent extreme-scale language model GPT-3 from OpenAI has over 175 billion parameters. Furthermore, such a model mostly consists of FC layers with huge dimensions, and thus has a relatively high arithmetic intensity. In that sense,
dblp:conf/fast/KimJSBHL21
fatcat:pejz7dv7svbwhbip5n2ltyenqa