Data of the "Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers" paper
AI & ML interests
LLM, trustworthy AI, AI security, privacy, calibration, hallucination
Recent Activity
NAACL 2025 Findings "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models" https://arxiv.org/abs/2411.00154
List of research articles of Parameter Lab
-
Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models
Paper • 2411.00154 • Published -
TRAP: Targeted Random Adversarial Prompt Honeypot for Black-Box Identification
Paper • 2402.12991 • Published -
Calibrating Large Language Models Using Their Generations Only
Paper • 2403.05973 • Published -
ProPILE: Probing Privacy Leakage in Large Language Models
Paper • 2307.01881 • Published • 1
Fine-tuned models for black-box LLM calibration, trained for "Apricot: Calibrating Large Language Models Using Their Generations Only" (ACL 2024)
-
parameterlab/apricot_binary_trivia_qa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 67 • 1 -
parameterlab/apricot_clustering_trivia_qa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 17 -
parameterlab/apricot_binary_coqa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 26 -
parameterlab/apricot_clustering_coqa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 36
Data of the "Leaky Thoughts: Large Reasoning Models Are Not Private Thinkers" paper
List of research articles of Parameter Lab
-
Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models
Paper • 2411.00154 • Published -
TRAP: Targeted Random Adversarial Prompt Honeypot for Black-Box Identification
Paper • 2402.12991 • Published -
Calibrating Large Language Models Using Their Generations Only
Paper • 2403.05973 • Published -
ProPILE: Probing Privacy Leakage in Large Language Models
Paper • 2307.01881 • Published • 1
NAACL 2025 Findings "Scaling Up Membership Inference: When and How Attacks Succeed on Large Language Models" https://arxiv.org/abs/2411.00154
Fine-tuned models for black-box LLM calibration, trained for "Apricot: Calibrating Large Language Models Using Their Generations Only" (ACL 2024)
-
parameterlab/apricot_binary_trivia_qa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 67 • 1 -
parameterlab/apricot_clustering_trivia_qa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 17 -
parameterlab/apricot_binary_coqa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 26 -
parameterlab/apricot_clustering_coqa_deberta-v3-base_for_vicuna-7b-v1.5
Text Classification • 0.2B • Updated • 36