Toronto Metropolitan University
Browse

Few-Shot Learning for Text Classification and Its Applications in Essay Scoring and Software Engineering

Download (1.35 MB)
thesis
posted on 2024-08-30, 20:28 authored by Robert Kraig Helmeczi

Few-shot learning—the ability to train models with access to limited data—has become increasingly popular in the natural language processing (NLP) domain, as large language models such as GPT and T0 have been empirically shown to achieve high performance in numerous tasks with access to just a handful of labeled examples. Smaller language models such as BERT and its variants have also been shown to achieve strong performance with just a handful of labeled examples when combined with few-shot learning algorithms like pattern-exploiting training (PET) and SetFit. The focus of this thesis is to investigate the performance of alternative few-shot learning approaches with BERT-based models. Specifically, vanilla fine-tuning, PET and SetFit are compared for numerous BERT-based checkpoints over an array of training set sizes. To facilitate this investigation, applications of few-shot learning are considered in automatic essay scoring—the task automatically grading written assessments—as well as in software engineering. For each task, high-performance techniques and their associated model checkpoints are identified through detailed empirical analysis. Our results establish PET as a strong few-shot learning approach, and our analysis show that with just a few hundred labeled examples it can achieve performance near that of fine-tuning on full-sized data sets.

History

Language

English

Degree

  • Master of Science

Program

  • Data Science and Analytics

Granting Institution

Toronto Metropolitan University

LAC Thesis Type

  • Thesis

Thesis Advisor

Mucahit Cevik

Year

2023

Usage metrics

    Computer Science (Theses)

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC