Matthew Muckley (@mattmucklm) 's Twitter Profile
Matthew Muckley

@mattmucklm

Research Engineer, Meta Fundamental AI Research (FAIR). ML for compression, computer vision, medicine.

Threads: threads.net/@mattmucklm

ID: 1071099142506733568

linkhttps://mmuckley.github.io calendar_today07-12-2018 17:48:30

226 Tweet

966 Takipรงi

264 Takip Edilen

Pietro Astolfi (@piovrasca) 's Twitter Profile Photo

Are sota image generative models effective world models? Consistency-diversity-realism Pareto fronts show they're not (yet): - No model dominates others as a world model - Improvements in quality and consistency have come at the expense of diversity ๐Ÿ”— arxiv.org/abs/2406.10429

Are sota image generative models effective world models?

Consistency-diversity-realism Pareto fronts show they're not (yet):
- No model dominates others as a world model
- Improvements in quality and consistency have come at the expense of diversity

๐Ÿ”— arxiv.org/abs/2406.10429
Buu Phan (@buutphan) 's Twitter Profile Photo

Why do LLMs fail simple completion tasks, but not on a harder task? Learn about tokenization bias in LLMs and how to fix it by compute exact string probability! arxiv.org/pdf/2406.16829 work w/Marton Havasi Matthew Muckley Dr. Karen Ullrich accepted in Theoretical Foundations of Foundation Models, ICML2024 (1/4)

Why do LLMs fail simple completion tasks, but not on a harder task? 

Learn about tokenization bias in LLMs and how to fix it by compute exact string probability! arxiv.org/pdf/2406.16829

work w/<a href="/HavasiMarton/">Marton Havasi</a> <a href="/mattmucklm/">Matthew Muckley</a> <a href="/karen_ullrich/">Dr. Karen Ullrich</a>  accepted in <a href="/tf2m_workshop/">Theoretical Foundations of Foundation Models</a>, ICML2024
(1/4)
Dr. Karen Ullrich (@karen_ullrich) 's Twitter Profile Photo

Even with preference alignment, LLMs can be enticed into harmful behavior via adversarial prompts ๐Ÿ˜ˆ. ๐Ÿšจ Breaking: our theoretical findings confirm: LLM alignment is fundamentally limited! More details, on framework, statistical bounds and phenomenal defense results ๐Ÿ‘‡๐Ÿป

Even with preference alignment, LLMs can be enticed into harmful behavior via adversarial prompts  ๐Ÿ˜ˆ.

๐Ÿšจ Breaking: our theoretical findings confirm:
LLM alignment is fundamentally limited!

More details, on framework, statistical bounds and phenomenal defense results ๐Ÿ‘‡๐Ÿป
Matthew Muckley (@mattmucklm) 's Twitter Profile Photo

Tokenization is a limitation of modern LLMs. How to make it better? Find a way to convert your LLM's probabilities to byte-level probabilities! Details in the thread below! โฌ‡๏ธ Work done by our great intern Buu Phan

Brandon Amos (@brandondamos) 's Twitter Profile Photo

If you prompt an LLM and stop in the middle of a token, what happens? โŒ The generated response doesn't correctly complete the token ๐Ÿ“š Our new paper provably fixes this for better code generation and ensembling! arxiv.org/abs/2410.09303 More details in Buu Phan's thread below

If you prompt an LLM and stop in the middle of a token, what happens? โŒ The generated response doesn't correctly complete the token

๐Ÿ“š Our new paper provably fixes this for better code generation and ensembling! arxiv.org/abs/2410.09303

More details in <a href="/buutphan/">Buu Phan</a>'s thread below
merve (@mervenoyann) 's Twitter Profile Photo

Meta just released V-JEPA 2: new open-source image/video world models โฏ๏ธ๐Ÿค— > based on ViT, different sizes (L/G/H) and resolution (286/384) > 0-day support in ๐Ÿค— transformers > comes with a physical reasoning (from video) benchmark: MVPBench, IntPhys 2, and CausalVQA

Federico Baldassarre (@baldassarrefe) 's Twitter Profile Photo

Say hello to DINOv3 ๐Ÿฆ–๐Ÿฆ–๐Ÿฆ– A major release that raises the bar of self-supervised vision foundation models. With stunning high-resolution dense features, itโ€™s a game-changer for vision tasks! We scaled model size and training data, but here's what makes it special ๐Ÿ‘‡

Say hello to DINOv3 ๐Ÿฆ–๐Ÿฆ–๐Ÿฆ–

A major release that raises the bar of self-supervised vision foundation models.
With stunning high-resolution dense features, itโ€™s a game-changer for vision tasks!

We scaled model size and training data, but here's what makes it special ๐Ÿ‘‡
Gabriel Synnaeve (@syhw) 's Twitter Profile Photo

(๐Ÿงต) Today, we release Meta Code World Model (CWM), a 32-billion-parameter dense LLM that enables novel research on improving code generation through agentic reasoning and planning with world models. ai.meta.com/research/publiโ€ฆ

Revant Teotia (@revantteotia) 's Twitter Profile Photo

1/ My model canโ€™t generate โ€œfullโ€ refrigerators when I ask for it?? โ€” but it does by default anyway when I donโ€™t specify!? At #ICCV2025, weโ€™re presenting DIMCIM โ€” a new framework to quantify diversity and capacity in T2I models, that enables pinpointing such failure modes

1/ 
My model canโ€™t generate โ€œfullโ€ refrigerators when I ask for it?? โ€” but it does by default anyway when I donโ€™t specify!?

At #ICCV2025, weโ€™re presenting DIMCIM โ€” a new framework to quantify diversity and capacity in T2I models, that enables pinpointing such failure modes