Matthew Leavitt (@leavittron) 's Twitter Profile
Matthew Leavitt

@leavittron

Chief Science Officer, Co-Founder @datologyai. Former: Head of Data Research @MosaicML; FAIR. ๐Ÿง  and ๐Ÿค– intelligence // views are from nowhere

ID: 269994694

linkhttp://mleavitt.net calendar_today21-03-2011 20:36:34

2,2K Tweet

2,2K Followers

890 Following

Matthew Leavitt (@leavittron) 's Twitter Profile Photo

The team absolutely crushed it here. They blew away nearly every CLIP baseline, and matched or exceeded SigLIP2โ€”which uses a slew of training algorithm improvementsโ€”on a number of benchmarks. USING. DATA. CURATION. ONLY. Iโ€™m so proud of Ricardo Monti , Haoli Yin ,

๐š๐”ช๐Ÿพ๐šก๐šก๐Ÿพ (@gm8xx8) 's Twitter Profile Photo

Datology CLIP Models DatologyAI releases two SOTA CLIP ViT-B/32 variants: classification-optimized and retrieval-optimized, achieving top results through task-specific data curation alone. Model - ViT-B/32 (86M params), OpenCLIP 2.24.0 - No architecture or training changes -

Datology CLIP Models

DatologyAI releases two SOTA CLIP ViT-B/32 variants: classification-optimized and retrieval-optimized, achieving top results through task-specific data curation alone. 

Model
- ViT-B/32 (86M params), OpenCLIP 2.24.0
- No architecture or training changes
-
Ari Morcos (@arimorcos) 's Twitter Profile Photo

Congratulations to our friends and partners Arcee.ai on the release of AFM-4.5B! With data powered by DatologyAI, this model outperforms Gemma3-4B and is competitive with Qwen3-4B despite being trained on a fraction of the data.

๐š๐”ช๐Ÿพ๐šก๐šก๐Ÿพ (@gm8xx8) 's Twitter Profile Photo

AFM - Arcee Foundation Models. Built from scratch for enterprise. The first release, AFM-4.5B, is a 4.5B open-weight model that runs anywhere: cloud, edge, or CPU. Trained on rigorously filtered data with full deployment flexibility. I donโ€™t say this lightly โฎ• ๐‘ซ๐‘ถ๐‘ตโ€™๐‘ป

AFM - Arcee Foundation Models. 
Built from scratch for enterprise. The first release, AFM-4.5B, is a 4.5B open-weight model that runs anywhere: cloud, edge, or CPU. Trained on rigorously filtered data with full deployment flexibility.

I donโ€™t say this lightly 
โฎ• ๐‘ซ๐‘ถ๐‘ตโ€™๐‘ป
Bogdan Gaza (@hurrycane) 's Twitter Profile Photo

We've definitely seen signs of this already โ€” perhaps not surprisingly, post-training people tend to care more about the value of data. We see a number of companies turning to DatologyAI for getting the most out of their existing datasets!

Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

One of the dreams when joining DatologyAI was to bring the fruits of data research from labs๐Ÿ”ฌ to the real world ๐ŸŒŽ Soo gratifying to see that our algorithms are out in the open, enabling companies to rival the sophisticated Qwen & Gemma families at a fraction of the cost!

Matthew Leavitt (@leavittron) 's Twitter Profile Photo

Partnering w/ Arcee.ai was a blast. This is the first public language model pretrained on DatologyAI-curated data, and we're pleased (though not surprised) that it goes toe-to-toe w/ the best small models. The base model is ๐Ÿ”ฅ and Arcee's post-training expertise (S-tier) really

Partnering w/ <a href="/arcee_ai/">Arcee.ai</a> was a blast. This is the first public language model pretrained on <a href="/datologyai/">DatologyAI</a>-curated data, and we're pleased (though not surprised) that it goes toe-to-toe w/ the best small models. The base model is ๐Ÿ”ฅ and Arcee's post-training expertise (S-tier) really
Max Azoury (@maxwellazoury) 's Twitter Profile Photo

kalomaze Aside from the "big guys" (Gemma3, Llama 3.3, Qwen-MAX), Arcee models have ALWAYS, ALWAYS been the best in terms of not being lobotomites. I know kalomaze works with Prime Intellect (who has used Arcee)โ€ฆbut people need to understand, Arcee is the GOAT of posttraining. And their