Papers
arxiv:2410.18057

CLEAR: Character Unlearning in Textual and Visual Modalities

Published on Oct 23
ยท Submitted by ai-alanov on Oct 30
#1 Paper of the day
Authors:
,
,
,
,
,
,

Abstract

Machine Unlearning (MU) is critical for enhancing privacy and security in deep learning models, particularly in large multimodal language models (MLLMs), by removing specific private or hazardous information. While MU has made significant progress in textual and visual modalities, multimodal unlearning (MMU) remains significantly underexplored, partially due to the absence of a suitable open-source benchmark. To address this, we introduce CLEAR, a new benchmark designed to evaluate MMU methods. CLEAR contains 200 fictitious individuals and 3,700 images linked with corresponding question-answer pairs, enabling a thorough evaluation across modalities. We assess 10 MU methods, adapting them for MMU, and highlight new challenges specific to multimodal forgetting. We also demonstrate that simple ell_1 regularization on LoRA weights significantly mitigates catastrophic forgetting, preserving model performance on retained data. The dataset is available at https://huggingface.co/datasets/therem/CLEAR

Community

Paper submitter

We introduce the first open-source benchmark for unlearning methods in a multimodal setup. We generate 200 fictitious individuals with associated biographical and visual data, such as facial images. After fine-tuning the model on this dataset, we then aim to selectively forget subsets of individuals (2, 10, or 20 persons). For the full pipeline, visit our GitHub repository.

abstract_pic_page-0001.jpg

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

Nice

Really interesting paper! Here's my summary:

Capture dโ€™eฬcran 2024-11-02 aฬ€ 15.46.35.png

๐Ÿง  ๐—–๐—Ÿ๐—˜๐—”๐—ฅ: ๐—ณ๐—ถ๐—ฟ๐˜€๐˜ ๐—บ๐˜‚๐—น๐˜๐—ถ๐—บ๐—ผ๐—ฑ๐—ฎ๐—น ๐—ฏ๐—ฒ๐—ป๐—ฐ๐—ต๐—บ๐—ฎ๐—ฟ๐—ธ ๐˜๐—ผ ๐—บ๐—ฎ๐—ธ๐—ฒ ๐—บ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€ ๐—ณ๐—ผ๐—ฟ๐—ด๐—ฒ๐˜ ๐˜„๐—ต๐—ฎ๐˜ ๐˜„๐—ฒ ๐˜„๐—ฎ๐—ป๐˜ ๐˜๐—ต๐—ฒ๐—บ ๐˜๐—ผ ๐—ณ๐—ผ๐—ฟ๐—ด๐—ฒ๐˜

With privacy concerns rising, we sometimes need our models to "forget" specific information - like a person's data - while keeping everything else intact. Researchers just released CLEAR, the first benchmark to test how well this works with both text and images.

โŒ Bad news: Current methods either fail to truly forget or end up forgetting way too much. It's like trying to remove a single ingredient from a baked cake!

โœจ But there's hope: Adding simple mathematical constraints (L1 regularization) during the forgetting process significantly improves results.

๐ŸŽฏ Key insights:

โœ… The benchmark tests forgetting on 200 fictional personas
โ€ฃ 3,770 visual Q&A pairs
โ€ฃ 4,000 textual Q&A pairs
โ€ฃ Additional real-world tests

๐Ÿ›‘ Most current forgetting methods don't work well with both text and images
โ€ฃ They either remember what they should forget
โ€ฃ Or they forget too much unrelated information

โœจ Simple mathematical constraints work surprisingly well
โ€ฃ L1 regularization prevents excessive forgetting
โ€ฃ Works especially well with the LLMU method

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2410.18057 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2410.18057 in a Space README.md to link it from this page.

Collections including this paper 16