Ali Shahin Shamsabadi (@alishahinshams1) 's Twitter Profile
Ali Shahin Shamsabadi

@alishahinshams1

Senior Privacy researcher @brave, ex @turinginst @VectorInst; Trustworthy and auditable ML.

ID: 922917707808497669

linkhttps://alishahin.github.io/ calendar_today24-10-2017 20:08:23

510 Tweet

593 Followers

488 Following

Stephan Rabanser (@steverab) 's Twitter Profile Photo

📢 New ICML 2025 paper! Confidential Guardian: Cryptographically Prohibiting the Abuse of Model Abstention 🤔 Think model uncertainty can be trusted? We show that it can be misused—and how to stop it! Meet Mirage (our attack💥) & Confidential Guardian (our defense🛡️). 🧵1/10

📢 New ICML 2025 paper!

Confidential Guardian: Cryptographically Prohibiting the Abuse of Model Abstention

🤔 Think model uncertainty can be trusted?
We show that it can be misused—and how to stop it!
Meet Mirage (our attack💥) & Confidential Guardian (our defense🛡️).

đź§µ1/10
Stephan Rabanser (@steverab) 's Twitter Profile Photo

🚀 Dive deeper: Paper ▶️ arxiv.org/abs/2505.23968 Code ▶️ github.com/cleverhans-lab… Joint work with Ali Shahin Shamsabadi, Olive Franzese, Xiao Wang, Adrian Weller, Nicolas Papernot. Talk to us at ICML in Vancouver! 🇨🇦 🧵10/10 #Abstention #Uncertainty #Calibration #ZKP #ICML2025

🚀 Dive deeper:

Paper ▶️ arxiv.org/abs/2505.23968
Code ▶️ github.com/cleverhans-lab…

Joint work with <a href="/AliShahinShams1/">Ali Shahin Shamsabadi</a>, Olive Franzese, Xiao Wang, <a href="/adrian_weller/">Adrian Weller</a>, <a href="/NicolasPapernot/">Nicolas Papernot</a>.

Talk to us at ICML in Vancouver! 🇨🇦

đź§µ10/10 #Abstention #Uncertainty #Calibration #ZKP #ICML2025
Ali Shahin Shamsabadi (@alishahinshams1) 's Twitter Profile Photo

Are you an experienced Systems and Performance Researcher/Engineer? Join us at Brave to shape and advance system and performance evaluation and optimization—in the age of agents.

Ali Shahin Shamsabadi (@alishahinshams1) 's Twitter Profile Photo

What exactly are the privacy and security practices of Comet, and are users truly aware of what they're opting into? Referring to perplexity.ai/hub/legal/priv…?

What exactly are the privacy and security practices of <a href="/PerplexityComet/">Comet</a>, and are users truly aware of what they're opting into? 
Referring to
perplexity.ai/hub/legal/priv…?
Ali Shahin Shamsabadi (@alishahinshams1) 's Twitter Profile Photo

Did you know that "bad parties" can exploit model uncertainty to cause discriminatory behaviours? Want to learn how to enable confidential inference with proofs of well-calibratedness? Go see our paper at East Exhibition Hall A-B, # E-1002, 11:00 a.m. – 1:30 p.m. PDT #ICML2025

Did you know that "bad parties" can exploit model uncertainty to cause discriminatory behaviours?

Want to learn how to enable confidential inference with proofs of well-calibratedness?

Go see our paper at East Exhibition Hall A-B, # E-1002, 11:00 a.m. – 1:30 p.m. PDT #ICML2025
Brave (@brave) 's Twitter Profile Photo

Brave’s AI doesn’t store your conversations, log your IP address, or use your conversations for model training. Other browsers can’t say the same.

Brave’s AI doesn’t store your conversations, log your IP address, or use your conversations for model training. Other browsers can’t say the same.
BrendanEich (@brendaneich) 's Twitter Profile Photo

Brave folks found prompt injection bulbs too. We practice responsible disclosure, so filed a report with Perplexity. I hope our engineer wins a bug bounty. “Agentic” must not mean some reddit post can prompt-inject to pwn your bank/crypto/health-data/&c. linkedin.com/posts/aryaman-…

Ali Shahin Shamsabadi (@alishahinshams1) 's Twitter Profile Photo

Two of our papers are accepted to #EMNLP2025 Both explore privacy in LLMs Brave : - Membership and Memorization in LLM Knowledge Distillation: arxiv.org/abs/2508.07054 - Context-Aware Membership Inference Attacks against Pre-trained Large Language Models: arxiv.org/abs/2409.13745

Ali Shahin Shamsabadi (@alishahinshams1) 's Twitter Profile Photo

We define and quantify the membership and memorization privacy risk of the teacher’s private training data post-distillation. We found that: - LLM KD approaches carry membership and memorization privacy risks from the teacher to its student - but memorization is not membership!

We define and quantify the membership and memorization privacy risk of the teacher’s private
training data post-distillation.
We found that:
- LLM KD approaches carry membership and memorization privacy risks from the teacher to its student
- but memorization is not membership!