Adelin Travers (@alkae_t) 's Twitter Profile
Adelin Travers

@alkae_t

Security & Machine Learning , Views my own.

ID: 1141361678434459649

linkhttps://alkaet.github.io calendar_today19-06-2019 15:06:44

256 Tweet

232 Takipçi

422 Takip Edilen

Adelin Travers (@alkae_t) 's Twitter Profile Photo

This is weirdly not understood enough. At a very high level, hallucinations are due to a generative model “filling the blanks” which corresponds to its capability to generalize.

Adelin Travers (@alkae_t) 's Twitter Profile Photo

A lot of people don’t realize that LLMs, and generally ML, relies on fundamental hypotheses to work and generalize Flaws like prompt injections, hallucinations, or adversarial examples exist due to these fundamental hypotheses. Aka these are inherent to current model designs!!!

Edward Grefenstette (@egrefen) 's Twitter Profile Photo

Programming languages are also like this. Python hides away a lot of low level complexity (relinquishing control and, sometimes, efficiency) but is very close to English. We trade off needing to specify some things for not being able to specify them (easily) when needed (6/9)

Ilia Shumailov🦔 (@iliaishacked) 's Twitter Profile Photo

Attacks such as this highlight the importance of system-level security analysis at all stages of model deployment, starting as the design of the architecture, and extending towards as late as the actual deployment of the model and how different user queries are batched together.

Rich Harang (@rharang) 's Twitter Profile Photo

At least part of the problem is people stubbornly refusing to apply the lessons of the past w/r/t appsec and safety-critical systems to AI systems because "✨AI is different✨". Whether that last is a true or simply convenient belief is up for debate.

Eleonora Svanberg (@eleonorasberg) 's Twitter Profile Photo

I'm a PhD student at University of Oxford and I think I'm living in a fairytale :-) Foxes playing around in the snow at Magdalen College this morning — absolutely magical!

Florian Roth ⚡️ (@cyb3rops) 's Twitter Profile Photo

I strongly recommend turning this off. It's unbelievable that they quietly enabled this while everyone was focused on their 'Recall' AI feature. Now they're collecting and using everyone's Microsoft Word and Excel data to train their AI models.

Yarden Shafir (@yarden_shafir) 's Twitter Profile Photo

Does anyone know companies hiring for entry level roles (in Canada/remote)? And I mean *real* entry level, not degree + 2 certs + 3 years experience “entry level”. Not just cybersecurity, any entry level roles at all, in any area.

threlfall (@whitehacksec) 's Twitter Profile Photo

Dropbox is looking for a senior ML engineer to join our threat intelligence and product trust & safety team, link in reply. DM me if you want to know more

Rob Fuller (@mubix) 's Twitter Profile Photo

Wanted to share that I just finished my MBA. Two graduate degrees, my undergrad and a CISSP. Now all I have to do is start attending RSA… ;)

Wanted to share that I just finished my MBA. Two graduate degrees, my undergrad and a CISSP. Now all I have to do is start attending RSA… ;)
Adelin Travers (@alkae_t) 's Twitter Profile Photo

One of the best examples of unintentional data poisoning: replicating a common pattern of public code that was used as training data for an LLM. Workarounds are everywhere because they are often fast solutions. Actual fixes can often take longer and thus may be less frequent

Adelin Travers (@alkae_t) 's Twitter Profile Photo

Totally agree. Application design is primordial in LLM/agentic security because of how hard it is to actually make the model itself safe. By the way, isn’t that documented as “the golden rule” in the Nvidia Guardrails’ documentation’s LLM security guidelines Rich Harang (@ BlackHat / DEF CON) ?

Varun Chandrasekaran (@varunchandrase3) 's Twitter Profile Photo

Some of the most brilliant minds I know are building one benchmark after another instead of finding more principled ways of understanding behaviours. Is this what science has come to?