Eugene Bagdasarian (@ebagdasa) 's Twitter Profile
Eugene Bagdasarian

@ebagdasa

Challenge AI security and privacy practices. Asst Prof at UMass @manningcics. Researcher at @GoogleAI. he/him 🇦🇲 (opinions mine)

ID: 2463105726

linkhttps://people.cs.umass.edu/~eugene/ calendar_today25-04-2014 12:01:56

368 Tweet

960 Takipçi

613 Takip Edilen

Sahra Ghalebikesabi (@sghalebikesabi) 's Twitter Profile Photo

📢 New research from Google DeepMind & Google Research! We tackle the challenge of building AI assistants that leverage your data for complex tasks, all while upholding your privacy. 🤖🔐 Dive into our paper for the full details: arxiv.org/pdf/2408.02373 TLDR in 🧵

📢 New research from <a href="/GoogleDeepMind/">Google DeepMind</a> &amp; <a href="/GoogleResearch/">Google Research</a>!

We tackle the challenge of building AI assistants that leverage your data for complex tasks, all while upholding your privacy. 🤖🔐

Dive into our paper for the full details: arxiv.org/pdf/2408.02373

TLDR in 🧵
Jaechul Roh (@jaechulroh) 's Twitter Profile Photo

🚨New Preprint: "Backdooring Bias into Text-to-Image Models" (arxiv.org/pdf/2406.15213) Ever wondered how text-to-image (T2I) models could spread political bias in #Election2024? 💡We introduce a new attack vector by embedding backdoors in T2I models using implicit biases!

🚨New Preprint: "Backdooring Bias into Text-to-Image Models" (arxiv.org/pdf/2406.15213)

Ever wondered how text-to-image (T2I) models could spread political bias in #Election2024?

💡We introduce a new attack vector by embedding backdoors in T2I models using implicit biases!
Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

🧙 I am recruiting PhD students and postdocs to work together on making sure AI Systems and Agents are built safe and respect privacy (+ other social values). Apply to UMass Amherst Manning College of Information & Computer Sciences and enjoy a beautiful town in Western Massachusetts. Reach out if you have questions!

🧙 I am recruiting PhD students and postdocs to work together on making sure AI Systems and Agents are built safe and respect privacy (+ other social values). Apply to UMass Amherst <a href="/manningcics/">Manning College of Information & Computer Sciences</a> and enjoy a beautiful town in Western Massachusetts. Reach out if you have questions!
Sahar Abdelnabi 🕊 (on 🦋) (@sahar_abdelnabi) 's Twitter Profile Photo

OpenAI Operator enables users to automate complex tasks, e.g., travel plans. Services, e.g., Expedia, use chatbots. Soon, these two ends are going to communicate, forming agentic networks. What would these networks enable? what are their risks? and how to secure them? 🧵1/n

OpenAI Operator enables users to automate complex tasks, e.g., travel plans.

Services, e.g., Expedia, use chatbots.

Soon, these two ends are going to communicate, forming agentic networks. 

What would these networks enable? what are their risks? and how to secure them? 🧵1/n
Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

How Sudokus can waste your money? If you are using reasoning LLMs with public data, adversaries could pollute it with nonsense (but perfectly safe!) tasks that will slow down reasoning and amplify overheads 💰 (as you pay but not see reasoning tokens) while keeping answers intact

Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

Nerd sniping is probably the coolest description of this phenomena ( Wojciech Zaremba et al described it recently), but in our case overthinking didn't lead to any drastic consequences besides higher costs.

Nerd sniping is probably the coolest description of this phenomena ( <a href="/woj_zaremba/">Wojciech Zaremba</a> et al described it recently), but in our case overthinking didn't lead to any drastic consequences besides higher costs.
Egor Zverev @ICLR 2025 (@egor_zverev_ai) 's Twitter Profile Photo

(1/n) In our #ICLR2025 paper, we explore a fundamental issue that enables prompt injections: 𝐋𝐋𝐌𝐬’ 𝐢𝐧𝐚𝐛𝐢𝐥𝐢𝐭𝐲 𝐭𝐨 𝐬𝐞𝐩𝐚𝐫𝐚𝐭𝐞 𝐢𝐧𝐬𝐭𝐫𝐮𝐜𝐭𝐢𝐨𝐧𝐬 𝐟𝐫𝐨𝐦 𝐝𝐚𝐭𝐚 𝐢𝐧 𝐭𝐡𝐞𝐢𝐫 𝐢𝐧𝐩𝐮𝐭 ✅ Definition of separation 👉 SEP Benchmark 🔍 LLM evals on SEP

(1/n) In our #ICLR2025  paper, we explore a fundamental issue that enables prompt injections: 𝐋𝐋𝐌𝐬’ 𝐢𝐧𝐚𝐛𝐢𝐥𝐢𝐭𝐲 𝐭𝐨 𝐬𝐞𝐩𝐚𝐫𝐚𝐭𝐞 𝐢𝐧𝐬𝐭𝐫𝐮𝐜𝐭𝐢𝐨𝐧𝐬 𝐟𝐫𝐨𝐦 𝐝𝐚𝐭𝐚 𝐢𝐧 𝐭𝐡𝐞𝐢𝐫 𝐢𝐧𝐩𝐮𝐭

✅ Definition of separation
👉 SEP Benchmark
🔍 LLM evals on SEP
Nando Fioretto (@nandofioretto) 's Twitter Profile Photo

The Privacy Preserving AI workshop is back! And is happening on Monday. I am excited about our program and lineup of invited speakers! I hope to see many of you there: ppai-workshop.github.io

The Privacy Preserving AI workshop is back! And is happening on Monday.

I am excited about our program and lineup of invited speakers! 

I hope to see many of you there: 
ppai-workshop.github.io
earlence (@earlencef) 's Twitter Profile Photo

Our IEEE S&P SAGAI workshop on systems-oriented security for AI agents has speaker details (abs/bio) on the website now: sites.google.com/ucsd.edu/sagai… We look forward to seeing you in San Francisco on May 15! As a reminder, we are running this "Dagstuhl" style - real discussions.

Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

I am looking for a postdoc to work on multi-agent safety problems, if you are interested or know anyone let me know: forms.gle/NFuYLKj53fVwdW…