🥇 TOFU Leaderboard

The TOFU dataset is a benchmark designed to evaluate the unlearning performance of large language models in realistic scenarios. This unique dataset consists of question-answer pairs that are based on the autobiographies of 200 fictitious authors, entirely generated by the GPT-4 model. The primary objective of this task is to effectively unlearn a fine-tuned model using different portions of the forget set. Read more at https://locuslab.github.io/tofu/.

🔄 Select Forget Percentage
🔄 Select Base Model
Select Metrics
Method
Submitted By
Epoch
Model Utility
Forget Quality
ROUGE Real Authors
Truth Ratio Real Authors
Prob. Real Authors
ROUGE Real World
Truth Ratio Real World
Prob. Real World
ROUGE Retain
Truth Ratio Retain
Prob. Retain
ROUGE Forget
Truth Ratio Forget
Prob. Forget
Finetune Model (WD = 0.01)
Baseline
-1
0.0740776537502228
1.0747499261825833e-13
0.1728333333333333
0.5706808914420933
0.5059025919974256
0.8974358974358975
0.5441982317366325
0.4143263304391235
0.9758111850816836
0.4709771340245287
1.2049720751682384e-32
0.4082436195223163
0.6740192657877031
1.3830246365938358e-23
Model family
Forget rate

Quick Links

Applicability 🚀

The dataset is in QA format, making it ideal for use with popular chat models such as Llama2, Mistral, or Qwen. However, it also works for any other large language model. The corresponding code base is written for the Llama2 model, but can be easily adapted to other models.

Installation

conda create -n tofu python=3.10
conda activate tofu
conda install pytorch pytorch-cuda=11.8 -c pytorch -c nvidia
conda install -c "nvidia/label/cuda-11.8.0" cuda-toolkit
pip install -r requirements.txt

Loading the Dataset

To load the dataset, use the following code:

from datasets import load_dataset
dataset = load_dataset("locuslab/TOFU","full")