1 Hit in 7.8 sec

We Need To Talk About Random Splits

Anders Søgaard, Sebastian Ebert, Jasmijn Bastings, Katja Filippova
2021 Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume   unpublished
Gorman and Bedrick (2019) argued for using random splits rather than standard splits in NLP experiments.  ...  In NLP, however, even worst-case splits, maximizing bias, often under-estimate the error observed on new samples of in-domain data, i.e., the data that models should minimally generalize to at test time  ...  The paper also benefited greatly from discussions with several of our colleagues at Google Research, including Slav Petrov and Sascha Rothe.  ... 
doi:10.18653/v1/2021.eacl-main.156 fatcat:nj4tzek5m5h3xa67z7aekdpgie