More Notes on LLMs and privacy leakage
A paper exploring the causes of privacy leakage in language models
TL;DR
These notes are on a 2023 paper by Carlini et al looking at the factors that cause large language models (LLMs) to memorize their training data verbatim and thus increase the risk of privacy leakage where that data consists of personal data.
I explored the risk of privacy leakage in a previous post on a 2020 paper by Carlini. Th…



