Timezone: »
Recent few-shot learning methods, such as parameter-efficient fine-tuning (PEFT) and pattern exploiting training (PET), have achieved impressive results in label-scarce settings. However, they are difficult to employ since they are highly sensitive to handcrafted prompts, and typically require billion-parameter language models to achieve high accuracy. To address these shortcomings, we propose SetFit (Sentence Transformer Fine-tuning), an efficient and prompt-free framework for few-shot fine-tuning of Sentence Transformers (ST). SetFit works by first fine-tuning a pretrained ST on a small number of labeled text pairs, in a contrastive Siamese manner. The resulting model is then used to generate rich text embeddings, which are used to train a classification head. This simple framework requires no prompts or verbalizers, and achieves high accuracy with orders of magnitude less parameters and runtime than existing techniques. Our experiments show that SetFit achieves results competitive with PEFT and PET techniques, and outperforms them on a variety of classification tasks.
Author Information
Oren Pereg (Intel Labs)
Daniel Korat (Intel Labs)
Moshe Wasserblat (INTEL)
Lewis Tunstall
Unso Eun Seo Jo (Hugging Face)
Luke Bates (Ubiquitous Knowledge Processing Lab)
Nils Reimers (TU Darmstadt)
Related Events (a corresponding poster, oral, or spotlight)
-
2022 : Efficient Few-Shot Learning Without Prompts »
Fri. Dec 2nd 06:05 -- 06:15 PM Room
More from the Same Authors
-
2021 : RAFT: A Real-World Few-Shot Text Classification Benchmark »
Neel Alex · Eli Lifland · Lewis Tunstall · Abhishek Thakur · Pegah Maham · C. Riedel · Emmie Hine · Carolyn Ashurst · Paul Sedille · Alexis Carlier · Michael Noetel · Andreas Stuhlmüller -
2021 : BEIR: A Heterogeneous Benchmark for Zero-shot Evaluation of Information Retrieval Models »
Nandan Thakur · Nils Reimers · Andreas Rücklé · Abhishek Srivastava · Iryna Gurevych -
2022 : Bias Assessment of Text-to-Image Models »
Sasha Luccioni · Clémentine Fourrier · Nathan Lambert · Unso Eun Seo Jo · Irene Solaiman · Helen Ngo · Nazneen Rajani · Giada Pistilli · Yacine Jernite · Margaret Mitchell -
2022 : QuaLA-MiniLM: a Quantized Length Adaptive MiniLM »
Shira Guskin · Moshe Wasserblat · Haihao Shen · Chang Wang -
2022 : Few-Shot Aspect Extraction using Prompt Training »
Oren Pereg · Daniel Korat · Moshe Wasserblat · Kfir Bar -
2022 : Fast DistilBERT on CPUs »
Haihao Shen · Ofir Zafrir · Bo Dong · Hengyu Meng · Xinyu Ye · Zhe Wang · Yi Ding · Hanwen Chang · Guy Boudoukh · Moshe Wasserblat -
2022 : Fast DistilBERT on CPUs »
Haihao Shen · Ofir Zafrir · Bo Dong · Hengyu Meng · Xinyu Ye · Zhe Wang · Yi Ding · Hanwen Chang · Guy Boudoukh · Moshe Wasserblat