Please use this identifier to cite or link to this item:
http://arks.princeton.edu/ark:/88435/dsp01bk128f195
Title: | A Quantitative Analysis of Prompt Engineering and In-Context Learning |
Authors: | Pan, Chris |
Advisors: | Narasimhan, Karthik |
Department: | Computer Science |
Class Year: | 2023 |
Abstract: | Prompt Engineering has become the most accessible entry point into Large Language Models for both the Public and even Academic Researchers. With the advent of large privatized language models, fine-tuning language models becomes computationally intractable, and even contractually impossible. Given the impressive few-shot capabilities of these large language models, a much more reasonable task-priming method is prompting. While research has yielded very powerful but specific ideas such as Chain of Thought Prompting [10], and Low Perplexity Instructions [4], there is very little work studying general trends in prompt engineering. This thesis represents my year-long journey through this task. First I study few-shot demonstrations and ask the question: Can we predict whether a prompt will lead to good performance just by looking at the model? I am able to successfully train a prompt performance prediction model that generalizes to unseen tasks in the same domain. Next, I focus on instructions, and propose a formal benchmark to evaluate instructions and prompt templates on in-context learning metrics. We find several nuances that have not been explored by state of the art work. |
URI: | http://arks.princeton.edu/ark:/88435/dsp01bk128f195 |
Type of Material: | Princeton University Senior Theses |
Language: | en |
Appears in Collections: | Computer Science, 1987-2024 |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
PAN-CHRIS-THESIS.pdf | 3.31 MB | Adobe PDF | Request a copy |
Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.