2000 character limit reached
SemEval-2025 Task 4: Unlearning sensitive content from Large Language Models
Published 2 Apr 2025 in cs.CL and cs.LG | (2504.02883v1)
Abstract: We introduce SemEval-2025 Task 4: unlearning sensitive content from LLMs. The task features 3 subtasks for LLM unlearning spanning different use cases: (1) unlearn long form synthetic creative documents spanning different genres; (2) unlearn short form synthetic biographies containing personally identifiable information (PII), including fake names, phone number, SSN, email and home addresses, and (3) unlearn real documents sampled from the target model's training dataset. We received over 100 submissions from over 30 institutions and we summarize the key techniques and lessons in this paper.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.