Matthew Leavitt (@leavittron) 's Twitter Profile
Matthew Leavitt

@leavittron

Chief Science Officer, Co-Founder @datologyai. Former: Head of Data Research @MosaicML; FAIR. 🧠 and 🤖 intelligence // views are from nowhere

ID: 269994694

linkhttp://mleavitt.net calendar_today21-03-2011 20:36:34

2,2K Tweet

2,2K Followers

890 Following

Matthew Leavitt (@leavittron) 's Twitter Profile Photo

The team absolutely crushed it here. They blew away nearly every CLIP baseline, and matched or exceeded SigLIP2—which uses a slew of training algorithm improvements—on a number of benchmarks. USING. DATA. CURATION. ONLY. I’m so proud of Ricardo Monti , Haoli Yin ,

𝚐𝔪𝟾𝚡𝚡𝟾 (@gm8xx8) 's Twitter Profile Photo

Datology CLIP Models DatologyAI releases two SOTA CLIP ViT-B/32 variants: classification-optimized and retrieval-optimized, achieving top results through task-specific data curation alone. Model - ViT-B/32 (86M params), OpenCLIP 2.24.0 - No architecture or training changes -

Datology CLIP Models

DatologyAI releases two SOTA CLIP ViT-B/32 variants: classification-optimized and retrieval-optimized, achieving top results through task-specific data curation alone. 

Model
- ViT-B/32 (86M params), OpenCLIP 2.24.0
- No architecture or training changes
-
Ari Morcos (@arimorcos) 's Twitter Profile Photo

Congratulations to our friends and partners Arcee.ai on the release of AFM-4.5B! With data powered by DatologyAI, this model outperforms Gemma3-4B and is competitive with Qwen3-4B despite being trained on a fraction of the data.

𝚐𝔪𝟾𝚡𝚡𝟾 (@gm8xx8) 's Twitter Profile Photo

AFM - Arcee Foundation Models. Built from scratch for enterprise. The first release, AFM-4.5B, is a 4.5B open-weight model that runs anywhere: cloud, edge, or CPU. Trained on rigorously filtered data with full deployment flexibility. I don’t say this lightly ⮕ 𝑫𝑶𝑵’𝑻

AFM - Arcee Foundation Models. 
Built from scratch for enterprise. The first release, AFM-4.5B, is a 4.5B open-weight model that runs anywhere: cloud, edge, or CPU. Trained on rigorously filtered data with full deployment flexibility.

I don’t say this lightly 
⮕ 𝑫𝑶𝑵’𝑻
Bogdan Gaza (@hurrycane) 's Twitter Profile Photo

We've definitely seen signs of this already — perhaps not surprisingly, post-training people tend to care more about the value of data. We see a number of companies turning to DatologyAI for getting the most out of their existing datasets!

Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

One of the dreams when joining DatologyAI was to bring the fruits of data research from labs🔬 to the real world 🌎 Soo gratifying to see that our algorithms are out in the open, enabling companies to rival the sophisticated Qwen & Gemma families at a fraction of the cost!

Matthew Leavitt (@leavittron) 's Twitter Profile Photo

Partnering w/ Arcee.ai was a blast. This is the first public language model pretrained on DatologyAI-curated data, and we're pleased (though not surprised) that it goes toe-to-toe w/ the best small models. The base model is 🔥 and Arcee's post-training expertise (S-tier) really

Partnering w/ <a href="/arcee_ai/">Arcee.ai</a> was a blast. This is the first public language model pretrained on <a href="/datologyai/">DatologyAI</a>-curated data, and we're pleased (though not surprised) that it goes toe-to-toe w/ the best small models. The base model is 🔥 and Arcee's post-training expertise (S-tier) really
Max Azoury (@maxwellazoury) 's Twitter Profile Photo

kalomaze Aside from the "big guys" (Gemma3, Llama 3.3, Qwen-MAX), Arcee models have ALWAYS, ALWAYS been the best in terms of not being lobotomites. I know kalomaze works with Prime Intellect (who has used Arcee)…but people need to understand, Arcee is the GOAT of posttraining. And their