NLPGym – A toolkit for evaluating RL agents on Natural Language Processing Tasks [article]

Rajkumar Ramamurthy, Rafet Sifa, Christian Bauckhage
2020 arXiv   pre-print
Reinforcement learning (RL) has recently shown impressive performance in complex game AI and robotics tasks. To a large extent, this is thanks to the availability of simulated environments such as OpenAI Gym, Atari Learning Environment, or Malmo which allow agents to learn complex tasks through interaction with virtual environments. While RL is also increasingly applied to natural language processing (NLP), there are no simulated textual environments available for researchers to apply and
more » ... tently benchmark RL on NLP tasks. With the work reported here, we therefore release NLPGym, an open-source Python toolkit that provides interactive textual environments for standard NLP tasks such as sequence tagging, multi-label classification, and question answering. We also present experimental results for 6 tasks using different RL algorithms which serve as baselines for further research. The toolkit is published at
arXiv:2011.08272v1 fatcat:6cltezznkzbwxl6x7hgq2kztoq