FinBERT: A Pretrained Language Model for Financial Communications [article]

Yi Yang, Mark Christopher Siy UY, Allen Huang
2020 arXiv   pre-print
Contextual pretrained language models, such as BERT (Devlin et al., 2019), have made significant breakthrough in various NLP tasks by training on large scale of unlabeled text re-sources.Financial sector also accumulates large amount of financial communication text.However, there is no pretrained finance specific language models available. In this work,we address the need by pretraining a financial domain specific BERT models, FinBERT, using a large scale of financial communication corpora.
more » ... riments on three financial sentiment classification tasks confirm the advantage of FinBERT over generic domain BERT model. The code and pretrained models are available at https://github.com/yya518/FinBERT. We hope this will be useful for practitioners and researchers working on financial NLP tasks.
arXiv:2006.08097v2 fatcat:eiijfp6xorbghibcyerjdnbofm