s1r1us (@s1r1u5_) 's Twitter Profile
s1r1us

@s1r1u5_

aham nityaṃ śiṣyaḥ, jagat mama guruḥ. {~hacker~} {founder @ElectrovoltSec, @HacktronAI}

ID: 3355115866

linkhttp://s1r1us.ninja calendar_today02-07-2015 13:37:54

2,2K Tweet

9,9K Takipçi

1,1K Takip Edilen

terjanq (@terjanq) 's Twitter Profile Photo

We published a blogpost about SafeContentFrame - a library for rendering untrusted content inside an iframe. The library is a big party of what I've been up to in the few last years! Check out the blog and take a slice of my birthday cake 🎂! bughunters.google.com/blog/671552987…

We published a blogpost about SafeContentFrame - a library for rendering untrusted content inside an iframe. The library is a big party of what I've been up to in the few last years! Check out the blog and take a slice of my birthday cake 🎂!

bughunters.google.com/blog/671552987…
s1r1us (@s1r1u5_) 's Twitter Profile Photo

you either make the model conscious by that i mean jailbreak proof or every agentic app built on top of it will suffer from terrible UX. increasing the capabilities doesn't mean its jailbreak proof, rather it can be used to do even more dangerous stuff. securing agentic apps

you either make the model conscious by that i mean jailbreak proof or every agentic app built on top of it will suffer from terrible UX.

increasing the capabilities doesn't mean its jailbreak proof, rather it can be used to do even more dangerous stuff. 

securing agentic apps
s1r1us (@s1r1u5_) 's Twitter Profile Photo

this is still relevant for security, just accept that model can be jailbroken and create threat document accordingly.

s1r1us (@s1r1u5_) 's Twitter Profile Photo

there's a real information asymmetry between agent developers and attackers, if I can read the system prompt I can map out almost every attack vector. maybe should write a blog post about that threat model, and building a small open source tool to automatically generate likely

s1r1us (@s1r1u5_) 's Twitter Profile Photo

cool idea. except, it’s exactly how social media algorithms already work, creating echo chambers. a private LLM trained only on someone’s own material would be even more profitable.

RootSys (@rootsysat) 's Twitter Profile Photo

🚨 Next.js and the Mutated Middleware [CVE-2025-57822] - a powerful SSRF primitive enabling full control over HTTP methods, headers & URLs. See how a subtle middleware bug can result in a high-impact vulnerability: 🔗 blog.rootsys.at/posts/nextjs-a… #AppSec #Nextjs #SSRF

s1r1us (@s1r1u5_) 's Twitter Profile Photo

the challenge with designing AI agents for vulnerability identification or offsec is that you can’t just drop them into a while(true) loop and expect bugs to surface the way coding assistants brute-force their way through tasks. vulnerability discovery requires structured

s1r1us (@s1r1u5_) 's Twitter Profile Photo

What if you trained models to explicitly separate <instruction> tags from everything else, treating the tagged content as executable instructions, and all other text as inert data? then, whenever you ingest untrusted input, you just sanitize it by stripping out <instruction>

s1r1us (@s1r1u5_) 's Twitter Profile Photo

the difference between us and Pavlov’s dog is that it never knows it was being conditioned, we know but powerless to resist.