Gaperon Collection Our French-English LLM suite (including Base and SFT models. All checkpoints are also included. • 16 items • Updated 2 days ago • 17
Privacy Collapse: Benign Fine-Tuning Can Break Contextual Privacy in Language Models Paper • 2601.15220 • Published Jan 21 • 9
Is Multilingual LLM Watermarking Truly Multilingual? A Simple Back-Translation Solution Paper • 2510.18019 • Published Oct 20, 2025 • 18
Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models Paper • 2411.00154 • Published Oct 31, 2024 • 1
TRAP: Targeted Random Adversarial Prompt Honeypot for Black-Box Identification Paper • 2402.12991 • Published Feb 20, 2024 • 1
Calibrating Large Language Models Using Their Generations Only Paper • 2403.05973 • Published Mar 9, 2024 • 1
ProPILE: Probing Privacy Leakage in Large Language Models Paper • 2307.01881 • Published Jul 4, 2023 • 2
DISCO: Diversifying Sample Condensation for Efficient Model Evaluation Paper • 2510.07959 • Published Oct 9, 2025 • 15
Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers Paper • 2506.15674 • Published Jun 18, 2025 • 2
MIA-Pile Collection Samples used for the NAACL 2025 Findings paper: "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models." • 23 items • Updated Feb 3, 2025 • 1
📈 Scaling MIA Data & Results Collection NAACL 2025 Findings "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models" https://arxiv.org/abs/2411.00154 • 22 items • Updated Jun 5, 2025 • 2
Model with Circuit Breakers Collection SoTA models with circuit breakers inserted. Top safety performance without losing capabilities. • 3 items • Updated Oct 25, 2024 • 5
🍑 Apricot Models Collection Fine-tuned models for black-box LLM calibration, trained for "Apricot: Calibrating Large Language Models Using Their Generations Only" (ACL 2024) • 9 items • Updated Nov 20, 2024 • 3