Eugene Bagdasarian (@ebagdasa) 's Twitter Profile
Eugene Bagdasarian

@ebagdasa

Challenge AI security and privacy practices. Asst Prof at UMass @manningcics. Researcher at @GoogleAI. he/him ๐Ÿ‡ฆ๐Ÿ‡ฒ (opinions mine)

ID: 2463105726

linkhttps://people.cs.umass.edu/~eugene/ calendar_today25-04-2014 12:01:56

368 Tweet

960 Followers

613 Following

Sahra Ghalebikesabi (@sghalebikesabi) 's Twitter Profile Photo

๐Ÿ“ข New research from Google DeepMind & Google Research! We tackle the challenge of building AI assistants that leverage your data for complex tasks, all while upholding your privacy. ๐Ÿค–๐Ÿ” Dive into our paper for the full details: arxiv.org/pdf/2408.02373 TLDR in ๐Ÿงต

๐Ÿ“ข New research from <a href="/GoogleDeepMind/">Google DeepMind</a> &amp; <a href="/GoogleResearch/">Google Research</a>!

We tackle the challenge of building AI assistants that leverage your data for complex tasks, all while upholding your privacy. ๐Ÿค–๐Ÿ”

Dive into our paper for the full details: arxiv.org/pdf/2408.02373

TLDR in ๐Ÿงต
Jaechul Roh (@jaechulroh) 's Twitter Profile Photo

๐ŸšจNew Preprint: "Backdooring Bias into Text-to-Image Models" (arxiv.org/pdf/2406.15213) Ever wondered how text-to-image (T2I) models could spread political bias in #Election2024? ๐Ÿ’กWe introduce a new attack vector by embedding backdoors in T2I models using implicit biases!

๐ŸšจNew Preprint: "Backdooring Bias into Text-to-Image Models" (arxiv.org/pdf/2406.15213)

Ever wondered how text-to-image (T2I) models could spread political bias in #Election2024?

๐Ÿ’กWe introduce a new attack vector by embedding backdoors in T2I models using implicit biases!
Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

๐Ÿง™ I am recruiting PhD students and postdocs to work together on making sure AI Systems and Agents are built safe and respect privacy (+ other social values). Apply to UMass Amherst Manning College of Information & Computer Sciences and enjoy a beautiful town in Western Massachusetts. Reach out if you have questions!

๐Ÿง™ I am recruiting PhD students and postdocs to work together on making sure AI Systems and Agents are built safe and respect privacy (+ other social values). Apply to UMass Amherst <a href="/manningcics/">Manning College of Information & Computer Sciences</a> and enjoy a beautiful town in Western Massachusetts. Reach out if you have questions!
Sahar Abdelnabi ๐Ÿ•Š (on ๐Ÿฆ‹) (@sahar_abdelnabi) 's Twitter Profile Photo

OpenAI Operator enables users to automate complex tasks, e.g., travel plans. Services, e.g., Expedia, use chatbots. Soon, these two ends are going to communicate, forming agentic networks. What would these networks enable? what are their risks? and how to secure them? ๐Ÿงต1/n

OpenAI Operator enables users to automate complex tasks, e.g., travel plans.

Services, e.g., Expedia, use chatbots.

Soon, these two ends are going to communicate, forming agentic networks. 

What would these networks enable? what are their risks? and how to secure them? ๐Ÿงต1/n
Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

How Sudokus can waste your money? If you are using reasoning LLMs with public data, adversaries could pollute it with nonsense (but perfectly safe!) tasks that will slow down reasoning and amplify overheads ๐Ÿ’ฐ (as you pay but not see reasoning tokens) while keeping answers intact

Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

Nerd sniping is probably the coolest description of this phenomena ( Wojciech Zaremba et al described it recently), but in our case overthinking didn't lead to any drastic consequences besides higher costs.

Nerd sniping is probably the coolest description of this phenomena ( <a href="/woj_zaremba/">Wojciech Zaremba</a> et al described it recently), but in our case overthinking didn't lead to any drastic consequences besides higher costs.
Egor Zverev @ICLR 2025 (@egor_zverev_ai) 's Twitter Profile Photo

(1/n) In our #ICLR2025 paper, we explore a fundamental issue that enables prompt injections: ๐‹๐‹๐Œ๐ฌโ€™ ๐ข๐ง๐š๐›๐ข๐ฅ๐ข๐ญ๐ฒ ๐ญ๐จ ๐ฌ๐ž๐ฉ๐š๐ซ๐š๐ญ๐ž ๐ข๐ง๐ฌ๐ญ๐ซ๐ฎ๐œ๐ญ๐ข๐จ๐ง๐ฌ ๐Ÿ๐ซ๐จ๐ฆ ๐๐š๐ญ๐š ๐ข๐ง ๐ญ๐ก๐ž๐ข๐ซ ๐ข๐ง๐ฉ๐ฎ๐ญ โœ… Definition of separation ๐Ÿ‘‰ SEP Benchmark ๐Ÿ” LLM evals on SEP

(1/n) In our #ICLR2025  paper, we explore a fundamental issue that enables prompt injections: ๐‹๐‹๐Œ๐ฌโ€™ ๐ข๐ง๐š๐›๐ข๐ฅ๐ข๐ญ๐ฒ ๐ญ๐จ ๐ฌ๐ž๐ฉ๐š๐ซ๐š๐ญ๐ž ๐ข๐ง๐ฌ๐ญ๐ซ๐ฎ๐œ๐ญ๐ข๐จ๐ง๐ฌ ๐Ÿ๐ซ๐จ๐ฆ ๐๐š๐ญ๐š ๐ข๐ง ๐ญ๐ก๐ž๐ข๐ซ ๐ข๐ง๐ฉ๐ฎ๐ญ

โœ… Definition of separation
๐Ÿ‘‰ SEP Benchmark
๐Ÿ” LLM evals on SEP
Nando Fioretto (@nandofioretto) 's Twitter Profile Photo

The Privacy Preserving AI workshop is back! And is happening on Monday. I am excited about our program and lineup of invited speakers! I hope to see many of you there: ppai-workshop.github.io

The Privacy Preserving AI workshop is back! And is happening on Monday.

I am excited about our program and lineup of invited speakers! 

I hope to see many of you there: 
ppai-workshop.github.io
earlence (@earlencef) 's Twitter Profile Photo

Our IEEE S&P SAGAI workshop on systems-oriented security for AI agents has speaker details (abs/bio) on the website now: sites.google.com/ucsd.edu/sagaiโ€ฆ We look forward to seeing you in San Francisco on May 15! As a reminder, we are running this "Dagstuhl" style - real discussions.

Eugene Bagdasarian (@ebagdasa) 's Twitter Profile Photo

I am looking for a postdoc to work on multi-agent safety problems, if you are interested or know anyone let me know: forms.gle/NFuYLKj53fVwdWโ€ฆ