-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathCITATION.cff
More file actions
45 lines (44 loc) · 1.7 KB
/
CITATION.cff
File metadata and controls
45 lines (44 loc) · 1.7 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
# This CITATION.cff file was generated with cffinit.
# Visit https://bit.ly/cffinit to generate yours today!
cff-version: 1.2.0
title: >-
Predicting and analysing memorization within fine-tuned
Large Language Models
message: >-
If you use this software, please cite it using the
metadata from this file.
type: software
authors:
- given-names: Jérémie
family-names: Dentan
orcid: 'https://orcid.org/0009-0001-5561-8030'
- given-names: Davide
family-names: Buscaldi
- given-names: Aymen
family-names: Shabou
- given-names: Sonia
family-names: Vanier
identifiers:
- type: url
value: 'https://arxiv.org/abs/2409.18858'
repository-code: 'https://github.com/orailix/predict_llm_memorization'
abstract: >-
Large Language Models have received significant attention
due to their abilities to solve a wide range of complex
tasks. However these models memorize a significant
proportion of their training data, posing a serious threat
when disclosed at inference time. To mitigate this
unintended memorization, it is crucial to understand what
elements are memorized and why. Most existing works
provide a posteriori explanations, which has a limited
impact in practice. To address this gap, we propose a new
approach based on sliced mutual information to detect
memorized samples a priori. It is efficient from the early
stages of training, and is readily adaptable to any
classification task. Our method is supported by new
theoretical results that we demonstrate, and requires a
low computational budget. We obtain strong empirical
results, paving the way for systematic inspection and
protection of these vulnerable samples before memorization
happens.
license: LGPL-3