Skip to yearly menu bar Skip to main content


In-Person Poster presentation / poster accept

Measuring Forgetting of Memorized Training Examples

Matthew Jagielski · Om Thakkar · Florian Tramer · Daphne Ippolito · Katherine Lee · Nicholas Carlini · Eric Wallace · Shuang Song · Abhradeep Guha Thakurta · Nicolas Papernot · Chiyuan Zhang

MH1-2-3-4 #154

Keywords: [ Social Aspects of Machine Learning ] [ Memorization ] [ canary extraction ] [ convexity ] [ membership inference ] [ nondeterminism ] [ forgetting ]


Abstract:

Machine learning models exhibit two seemingly contradictory phenomena: training data memorization and various forms of forgetting. In memorization, models overfit specific training examples and become susceptible to privacy attacks. In forgetting, examples which appeared early in training are forgotten by the end. In this work, we connect these phenomena.We propose a technique to measure to what extent models ``forget'' the specifics of training examples, becoming less susceptible to privacy attacks on examples they have not seen recently.We show that, while non-convexity can prevent forgetting from happening in the worst-case, standard image,speech, and language models empirically do forget examples over time.We identify nondeterminism as a potential explanation, showing that deterministically trained models do not forget.Our results suggest that examples seen early when training with extremely large datasets---for instance those examples used to pre-train a model---may observe privacy benefits at the expense of examples seen later.

Chat is not available.