Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01bk128f195
Title: A Quantitative Analysis of Prompt Engineering and In-Context Learning
Authors: Pan, Chris
Advisors: Narasimhan, Karthik
Department: Computer Science
Class Year: 2023
Abstract: Prompt Engineering has become the most accessible entry point into Large Language Models for both the Public and even Academic Researchers. With the advent of large privatized language models, fine-tuning language models becomes computationally intractable, and even contractually impossible. Given the impressive few-shot capabilities of these large language models, a much more reasonable task-priming method is prompting. While research has yielded very powerful but specific ideas such as Chain of Thought Prompting [10], and Low Perplexity Instructions [4], there is very little work studying general trends in prompt engineering. This thesis represents my year-long journey through this task. First I study few-shot demonstrations and ask the question: Can we predict whether a prompt will lead to good performance just by looking at the model? I am able to successfully train a prompt performance prediction model that generalizes to unseen tasks in the same domain. Next, I focus on instructions, and propose a formal benchmark to evaluate instructions and prompt templates on in-context learning metrics. We find several nuances that have not been explored by state of the art work.
URI: http://arks.princeton.edu/ark:/88435/dsp01bk128f195
Type of Material: Princeton University Senior Theses
Language: en
Appears in Collections:Computer Science, 1987-2024

Files in This Item:
File Description SizeFormat 
PAN-CHRIS-THESIS.pdf3.31 MBAdobe PDF    Request a copy


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.