Exploring foci of:
doi.org
What Makes Pre-trained Language Models Better Zero-shot Learners?
January 2023 • Jinghui Lu, Dongsheng Zhu, Weidong Han, Rui Zhao, Brian Mac Namee, Fei Tan
Current methods for prompt learning in zero-shot scenarios widely rely on a development set with sufficient human-annotated data to select the best-performing prompt template a posteriori. This is not ideal because in a real-world zero-shot scenario of practical relevance, no labelled data is available. Thus, we propose a simple yet effective method for screening reasonable prompt templates in zero-shot text classification: Perplexity Selection (Perplection). We hypothesize that language discrepancy can be used to…
Computer Science
Shot (Pellet)
Artificial Intelligence
Machine Learning
Data Mining
Mathematics
Programming Language
Engineering
Philosophy
Mechanical Engineering
Mathematical Analysis
Organic Chemistry
Epistemology
Chemistry
Law