A copy of this work was available on the public web and has been preserved in the Wayback Machine. The capture dates from 2021; you can also visit the original URL.
The file type is application/pdf
.
Adapting Language Models for Zero-shot Learning by Meta-tuning on Dataset and Prompt Collections
[article]
2021
arXiv
pre-print
Large pre-trained language models (LMs) such as GPT-3 have acquired a surprising ability to perform zero-shot learning. For example, to classify sentiment without any training examples, we can "prompt" the LM with the review and the label description "Does the user like this movie?", and ask whether the next word is "yes" or "no". However, the next word prediction training objective is still misaligned with the target zero-shot learning objective. To address this weakness, we propose
arXiv:2104.04670v5
fatcat:nicxnnusjjg3jdyqch6nzy7y5m