-
Notifications
You must be signed in to change notification settings - Fork 6
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
Showing
1 changed file
with
16 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,16 @@ | ||
--- | ||
layout: post | ||
title: Noah Ziems | ||
--- | ||
|
||
Lunch at 12:30pm, talk at 1pm, in 148 Fitzpatrick | ||
|
||
Title: Inducing Complex Instructions | ||
|
||
Abstract: Instruction following has emerged as an effective alternative to fine tuning large language models (LLMs), as it is interpretable, requires fewer resources, and requires less expertise to implement than fine tuning. | ||
In practice, however, many tasks require complex instructions made up of many smaller atomic instructions, which are difficult to write with limited prompt engineering experience. | ||
Further, the more complex the task is, the more trial and error is needed to accomodate the corner cases. | ||
Existing approaches to this problem use Instruction Induction where input output pairs for a given task are provided with the goal of learning the instrucion that maps from inputs to outputs. However, existing approaches significantly struggle to induce and effective instruction for complex tasks. | ||
In this presentation, I cover the existing approaches to Instruction Induction/Prompt Optimization, I introduce my proposed approach for their limitations, and I explore the exciting downstream datasets that can be created for complex instruction induction. | ||
|
||
Bio: Noah Ziems is a second year PhD student in the Department of Computer Science and Engineering at the University of Notre Dame, and is a member of Dr. Meng Jiang's DM2 lab. His research is focused on instruction induction/prompt engineering, large language models, and information retrieval. |