Matthew Muckley (@mattmucklm) 's Twitter Profile
Matthew Muckley

@mattmucklm

Research Engineer, Meta Fundamental AI Research (FAIR). ML for compression, computer vision, medicine.

Threads: threads.net/@mattmucklm

ID: 1071099142506733568

linkhttps://mmuckley.github.io calendar_today07-12-2018 17:48:30

226 Tweet

966 Takipçi

264 Takip Edilen

Pietro Astolfi (@piovrasca) 's Twitter Profile Photo

Are sota image generative models effective world models? Consistency-diversity-realism Pareto fronts show they're not (yet): - No model dominates others as a world model - Improvements in quality and consistency have come at the expense of diversity 🔗 arxiv.org/abs/2406.10429

Are sota image generative models effective world models?

Consistency-diversity-realism Pareto fronts show they're not (yet):
- No model dominates others as a world model
- Improvements in quality and consistency have come at the expense of diversity

🔗 arxiv.org/abs/2406.10429
Buu Phan (@buutphan) 's Twitter Profile Photo

Why do LLMs fail simple completion tasks, but not on a harder task? Learn about tokenization bias in LLMs and how to fix it by compute exact string probability! arxiv.org/pdf/2406.16829 work w/Marton Havasi Matthew Muckley Dr. Karen Ullrich accepted in Theoretical Foundations of Foundation Models, ICML2024 (1/4)

Why do LLMs fail simple completion tasks, but not on a harder task? 

Learn about tokenization bias in LLMs and how to fix it by compute exact string probability! arxiv.org/pdf/2406.16829

work w/<a href="/HavasiMarton/">Marton Havasi</a> <a href="/mattmucklm/">Matthew Muckley</a> <a href="/karen_ullrich/">Dr. Karen Ullrich</a>  accepted in <a href="/tf2m_workshop/">Theoretical Foundations of Foundation Models</a>, ICML2024
(1/4)
Dr. Karen Ullrich (@karen_ullrich) 's Twitter Profile Photo

Even with preference alignment, LLMs can be enticed into harmful behavior via adversarial prompts 😈. 🚨 Breaking: our theoretical findings confirm: LLM alignment is fundamentally limited! More details, on framework, statistical bounds and phenomenal defense results 👇🏻

Even with preference alignment, LLMs can be enticed into harmful behavior via adversarial prompts  😈.

🚨 Breaking: our theoretical findings confirm:
LLM alignment is fundamentally limited!

More details, on framework, statistical bounds and phenomenal defense results 👇🏻
Matthew Muckley (@mattmucklm) 's Twitter Profile Photo

Tokenization is a limitation of modern LLMs. How to make it better? Find a way to convert your LLM's probabilities to byte-level probabilities! Details in the thread below! ⬇️ Work done by our great intern Buu Phan

Brandon Amos (@brandondamos) 's Twitter Profile Photo

If you prompt an LLM and stop in the middle of a token, what happens? ❌ The generated response doesn't correctly complete the token 📚 Our new paper provably fixes this for better code generation and ensembling! arxiv.org/abs/2410.09303 More details in Buu Phan's thread below

If you prompt an LLM and stop in the middle of a token, what happens? ❌ The generated response doesn't correctly complete the token

📚 Our new paper provably fixes this for better code generation and ensembling! arxiv.org/abs/2410.09303

More details in <a href="/buutphan/">Buu Phan</a>'s thread below
merve (@mervenoyann) 's Twitter Profile Photo

Meta just released V-JEPA 2: new open-source image/video world models ⏯️🤗 > based on ViT, different sizes (L/G/H) and resolution (286/384) > 0-day support in 🤗 transformers > comes with a physical reasoning (from video) benchmark: MVPBench, IntPhys 2, and CausalVQA

Federico Baldassarre (@baldassarrefe) 's Twitter Profile Photo

Say hello to DINOv3 🦖🦖🦖 A major release that raises the bar of self-supervised vision foundation models. With stunning high-resolution dense features, it’s a game-changer for vision tasks! We scaled model size and training data, but here's what makes it special 👇

Say hello to DINOv3 🦖🦖🦖

A major release that raises the bar of self-supervised vision foundation models.
With stunning high-resolution dense features, it’s a game-changer for vision tasks!

We scaled model size and training data, but here's what makes it special 👇
Gabriel Synnaeve (@syhw) 's Twitter Profile Photo

(🧵) Today, we release Meta Code World Model (CWM), a 32-billion-parameter dense LLM that enables novel research on improving code generation through agentic reasoning and planning with world models. ai.meta.com/research/publi…

Revant Teotia (@revantteotia) 's Twitter Profile Photo

1/ My model can’t generate “full” refrigerators when I ask for it?? — but it does by default anyway when I don’t specify!? At #ICCV2025, we’re presenting DIMCIM — a new framework to quantify diversity and capacity in T2I models, that enables pinpointing such failure modes

1/ 
My model can’t generate “full” refrigerators when I ask for it?? — but it does by default anyway when I don’t specify!?

At #ICCV2025, we’re presenting DIMCIM — a new framework to quantify diversity and capacity in T2I models, that enables pinpointing such failure modes