A course about the design and evaluation of prompting language models
Associate Professor
Software and Societal Systems Department
https://www.cs.cmu.edu/~breaux
Tech's hottest new job: AI whisperer. No coding required. Washington Post, Feb 25, 2023.
By Drew Harwell
'Prompt engineers’ are being hired for their skill in getting AI systems to produce exactly what they want. And they make pretty good money." (Read Article)
Microsoft releases Phi-2, "a 2.7 billion-parameter language model that demonstrates outstanding reasoning and language understanding capabilities, showcasing state-of-the-art performance among base language models with less than 13 billion parameters. On complex benchmarks Phi-2 matches or outperforms models up to 25x larger, thanks to new innovations in model scaling and training data curation." (Read Article)
Language Models are Few-Shot Learners. NeurIPS 2020
By Brown et al., OpenAI
"... For all tasks, GPT-3 is applied without any gradient updates or fine-tuning, with tasks and few-shot demonstrations specified purely via text interaction with the model. GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks, as well as several tasks that require on-the-fly reasoning or domain adaptation, such as unscrambling words, using a novel word in a sentence, or performing 3-digit arithmetic." (Read Article)
Calibrate Before Use: Improving Few-Shot Performance of Language Models Machine Learning Research, 2021
By Zhao et al., Berkeley, UMD and UCI
"Prompt[s] that contain a few training examples... can be unstable: the choice of prompt format, training examples, and even the order of the training examples can cause accuracy to vary from near chance to near state-of-the-art." (Read Article)
Copyright © 2022-, Travis D. Breaux,