Bertie Vidgen (@bertievidgen) 's Twitter Profile
Bertie Vidgen

@bertievidgen

Data @ Contextual AI

ID: 100974895

linkhttps://www.linkedin.com/in/bertie-vidgen-001/ calendar_today01-01-2010 13:24:47

917 Tweet

865 Followers

520 Following

Adina Williams (@adinamwilliams) 's Twitter Profile Photo

How can we improve benchmarking? The Dynabench experiment aims to make faster progress with dynamic data collection, and today, we are pleased to introduce our next stage: @MetaAI has funded 5 exciting research proposals on the theme of "Rethinking Benchmarking”! Congrats to:

Christopher Bouzy (spoutible.com/cbouzy) (@cbouzy) 's Twitter Profile Photo

YouTube and platforms like YouTube must be regulated. We have allowed social media platforms to self-regulate, and it has been a complete disaster.

Ethio NLP (@ethionlp) 's Twitter Profile Photo

We are glad to announce the first SemEval shared task targeting African languages, AfriSenti-SemEval, Task 12. The shared task includes different prizes. Competition: codalab.lisn.upsaclay.fr/competitions/7… AfriSenti: afrisenti-semeval.github.io AfriSenti SemEval-2023 Shared Task 12 Ethio NLP EMNLP 2025

Simon Kendall (@skendallfcdo) 's Twitter Profile Photo

What a privilege to moderate the 🇬🇧 -Bavarian online harms symposium at #MTM22 Thanks to Ian Stevenson & Bertie Vidgen for highlighting the work of 🇬🇧 world-class safety tech sector and discussing future UK–Bavarian collaboration with Dr Thorsten Schmiege BLM & Jana Heigl

What a privilege to moderate the 🇬🇧 -Bavarian online harms symposium at #MTM22 Thanks to <a href="/iancyclops/">Ian Stevenson</a> &amp; <a href="/bertievidgen/">Bertie Vidgen</a> for highlighting the work of 🇬🇧 world-class safety tech sector and discussing future UK–Bavarian collaboration with Dr Thorsten Schmiege <a href="/BLM_Bayern/">BLM</a> &amp; <a href="/janaheigl/">Jana Heigl</a>
Paul Röttger (@paul_rottger) 's Twitter Profile Photo

🥳 New paper at #EMNLP2022 (Main) 🥳 Too much hate speech research focuses just on English content! To help fix this, we tried to expand hate detection models into under-resourced languages 🌍 without needing lots of new data 💸 arxiv.org/abs/2210.11359 ⬇️ Highlights below ⬇️

Hannah Rose Kirk (@hannahrosekirk) 's Twitter Profile Photo

New #EMNLP2022 paper! Do you research online harms, misinformation or negative biases? Could your datasets contain examples of harmful text? 🤔 If yes, read our paper! 🫵arxiv.org/abs/2204.14256 Shoutout to my brilliant co-authors: Abeba Birhane, Leon Derczynski ✍🏻 🌞🏠🌲 & Bertie Vidgen A 🧵

Paul Röttger (@paul_rottger) 's Twitter Profile Photo

NEW PREPRINT! LLMs should be helpful AND harmless. This is a difficult balance to get right... Some models refuse even safe requests if they superficially resemble unsafe ones. We built XSTest to systematically test for such "exaggerated safety". 🧵 arxiv.org/abs/2308.01263

NEW PREPRINT!

LLMs should be helpful AND harmless. This is a difficult balance to get right...

Some models refuse even safe requests if they superficially resemble unsafe ones. We built XSTest to systematically test for such "exaggerated safety".

🧵

arxiv.org/abs/2308.01263
PatronusAI (@patronusai) 's Twitter Profile Photo

We are launching out of stealth today with a $3M seed round led by Lightspeed, with participation from Amjad Masad, Gokul Rajaram, Matt Hartman and other fortune 500 execs and board members 🚀 Read our story here: patronus.ai/blog/patronus-…

Bertie Vidgen (@bertievidgen) 's Twitter Profile Photo

I am very biased but this is an amazing launch by great people, creating a much-needed and incredibly powerful product! If you're using an #LLM then you need to know how it works, which means #evaluating it. No-one has solved how to do that reliably and at scale ...until now 🥳

Paul Röttger (@paul_rottger) 's Twitter Profile Photo

If you’re working on LLM safety, check out SafetyPrompts.com! SafetyPrompts.com is a catalogue of open datasets for evaluating and improving LLM safety. I started building this over the holidays, and I know there are still datasets missing, so I need your help 🧵

If you’re working on LLM safety, check out SafetyPrompts.com!

SafetyPrompts.com is a catalogue of open datasets for evaluating and improving LLM safety. I started building this over the holidays, and I know there are still datasets missing, so I need your help 🧵
PatronusAI (@patronusai) 's Twitter Profile Photo

1/ Introducing Lynx - the leading hallucination detection model 🚀👀 - Beats GPT-4o on hallucination tasks - Open source, open weights, open data - Excels in real-world domains like medicine and finance We are excited to launch Lynx with Day 1 integration partners: NVIDIA,

Shikib Mehri (@shikibmehri) 's Twitter Profile Photo

🚨 You're evaluating your LLM wrong 🚨 🤔 Human eval is unscalable, expensive & can be noisy, especially at the edge of capabilities 📉 Reward models compress complex notions of quality into opaque scores, and can't be steered after training 🤖 'LLM as a Judge' can't learn

Bertie Vidgen (@bertievidgen) 's Twitter Profile Photo

Without evaluation you have no idea what your model is doing. But using LM-as-a-judge isn't great and human annotators are expensive and noisy. LMUnit solves this tradeoff to scalably, reliably, and explainably eval your model. A lot of fun to work on this at Contextual AI

Paul Röttger (@paul_rottger) 's Twitter Profile Photo

Today, we are releasing MSTS, a new Multimodal Safety Test Suite for vision-language models! MSTS is exciting because it tests for safety risks *created by multimodality*. Each prompt consists of a text + image that *only in combination* reveal their full unsafe meaning. 🧵

Today, we are releasing MSTS, a new Multimodal Safety Test Suite for vision-language models!

MSTS is exciting because it tests for safety risks *created by multimodality*. Each prompt consists of a text + image that *only in combination* reveal their full unsafe meaning.

🧵
Bertie Vidgen (@bertievidgen) 's Twitter Profile Photo

Thoroughly enjoying #Nero by Anthony Everitt and Roddy Ashworth -- but wasn't Germanicus Claudius' brother, not his father? Given it talks about Germanicus falling sick and dying in Syria, rather than having a horse collapse on him, I think this might be a mistake 🤯

Thoroughly enjoying #Nero by Anthony Everitt and Roddy Ashworth -- but wasn't Germanicus Claudius' brother, not his father?

Given it talks about Germanicus falling sick and dying in Syria, rather than having a horse collapse on him, I think this might be a mistake 🤯
Shikib Mehri (@shikibmehri) 's Twitter Profile Photo

LLMs must be factually accurate. Especially to power autonomous agents for complex, long-horizon tasks. But how do we actually do this in practice? 🧵[1/16] Learn about groundedness, why LLMs hallucinate, and how ContextualAI built the world's most grounded LLM!

LLMs must be factually accurate. Especially to power autonomous agents for complex, long-horizon tasks.

But how do we actually do this in practice?

🧵[1/16] Learn about groundedness, why LLMs hallucinate, and how ContextualAI built the world's most grounded LLM!
Bertie Vidgen (@bertievidgen) 's Twitter Profile Photo

Having a grounded LM is not just about good retrievals -- even if you pass the model the right information, it still needs to *use it in the right way* I was surprised by how much LMs hallucinate even when literally told the right info... thankfully, Contextual is solving it :)

Bertie Vidgen (@bertievidgen) 's Twitter Profile Photo

🚨 50 million free tokens 🤯 Our reranker is SOTA and -- much more excitingly -- is the world's first **steerable** reranker. You can give it instructions in free text and watch it get to work!! This is the most effective way of handling real-world messiness in docs.

Bertie Vidgen (@bertievidgen) 's Twitter Profile Photo

The Mercor grad fellowship is worth $50k!! It's inspiring to work for a startup that offers exceptional people from _anywhere such incredible opportunities. As a PhD student I hustled to get extra income... Mercor would have been a godsend. Apply to work on our platform now.