Timezone: »

FLEX: Unifying Evaluation for Few-Shot NLP
Jonathan Bragg · Arman Cohan · Kyle Lo · Iz Beltagy

Thu Dec 09 08:30 AM -- 10:00 AM (PST) @

Few-shot NLP research is highly active, yet conducted in disjoint research threads with evaluation suites that lack challenging-yet-realistic testing setups and fail to employ careful experimental design. Consequently, the community does not know which techniques perform best or even if they outperform simple baselines. In response, we formulate the FLEX Principles, a set of requirements and best practices for unified, rigorous, valid, and cost-sensitive few-shot NLP evaluation. These principles include Sample Size Design, a novel approach to benchmark design that optimizes statistical accuracy and precision while keeping evaluation costs manageable. Following the principles, we release the FLEX benchmark, which includes four few-shot transfer settings, zero-shot evaluation, and a public leaderboard that covers diverse NLP tasks. In addition, we present UniFew, a prompt-based model for few-shot learning that unifies pretraining and finetuning prompt formats, eschewing complex machinery of recent prompt-based approaches in adapting downstream task formats to language model pretraining objectives. We demonstrate that despite simplicity, UniFew achieves results competitive with both popular meta-learning and prompt-based approaches.

Author Information

Jonathan Bragg (Allen Institute for AI)
Arman Cohan (Allen Institute for AI)
Kyle Lo (Allen Institute for AI)
Iz Beltagy (Allen Institute for AI)

More from the Same Authors