Publications

Demystifying forgetting in language model fine-tuning with statistical analysis of example associations

Abstract

Language models (LMs) are known to suffer from forgetting of previously learned examples when fine-tuned, breaking stability of deployed LM systems. Despite efforts on mitigating forgetting, few have investigated whether, and how forgotten upstream examples are associated with newly learned tasks. Insights on such associations enable efficient and targeted mitigation of forgetting. In this paper, we empirically analyze forgetting that occurs in upstream examples while the model learns new tasks and visualize their associations with a matrix. We empirically demonstrate that the degree of forgetting can often be approximated by simple multiplicative contributions of the upstream examples and newly learned tasks. We also reveal more complicated patterns where specific subsets of examples are forgotten with statistics and visualization. Following our analysis, we predict forgetting that happens on …

Metadata

publication
arXiv preprint arXiv:2406.14026, 2024
year
2024
publication date
2024/6
authors
Xisen Jin, Xiang Ren
link
https://ui.adsabs.harvard.edu/abs/2024arXiv240614026J/abstract
journal
arXiv e-prints
pages
arXiv: 2406.14026