Skip to content

Commit

Permalink
Update 2024-02-12-liu.md
Browse files Browse the repository at this point in the history
  • Loading branch information
Ziems authored Feb 12, 2024
1 parent 3234098 commit a55fefb
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions _posts/2024-02-12-liu.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
---
layout: post
title: Frank Liu
title: Zheyuan (Frank) Liu
---

Lunch at 12:30pm, talk at 1pm, in 148 Fitzpatrick
Expand All @@ -9,4 +9,4 @@ Title: TOFU: A Task of Fictitious Unlearning for LLMs

Abstract: Large language models trained on massive corpora of data from the web can memorize and reproduce sensitive or private data raising both legal and ethical concerns. Unlearning, or tuning models to forget information present in their training data, provides us with a way to protect private data after training. Although several methods exist for such unlearning, it is unclear to what extent they result in models equivalent to those where the data to be forgotten was never learned in the first place. To address this challenge, we present TOFU, a Task of Fictitious Unlearning, as a benchmark aimed at helping deepen our understanding of unlearning. We offer a dataset of 200 diverse synthetic author profiles, each consisting of 20 question-answer pairs, and a subset of these profiles called the forget set that serves as the target for unlearning. We compile a suite of metrics that work together to provide a holistic picture of unlearning efficacy. Finally, we provide a set of baseline results from existing unlearning algorithms. Importantly, none of the baselines we consider show effective unlearning motivating continued efforts to develop approaches for unlearning that effectively tune models so that they truly behave as if they were never trained on the forget data at all.

Bio: Frank Liu is a first-year PhD student in Dr. Jiang's DM2 lab at the University of Notre Dame.
Bio: Zheyuan (Frank) Liu is a first year PhD student in Dr.Jiang's DM2 Lab at Notre Dame. His research interests are centered around LLM security, LLM privacy and responsible AI. He is currently working on developing a novel framework to efficiently unlearn harmfulness from large language model.

0 comments on commit a55fefb

Please sign in to comment.