
Stefano Ermon
@stefanoermon
Associate Professor of #computerscience @Stanford #AI #ML
ID: 1145851147
http://www.cs.stanford.edu/~ermon/ 03-02-2013 18:16:08
429 Tweet
16,16K Followers
366 Following


Transformers have dominated LLM text generation, and generate tokens sequentially. This is a cool attempt to explore diffusion models as an alternative, by generating the entire text at the same time using a coarse-to-fine process. Congrats Stefano Ermon & team!



Speed + efficiency = the future of AI ⚡️ Mercury Coder running on NVIDIA H100 GPUs can hit over 1000 output tokens/second —that's a 5x speed increase for high quality responses at low costs. Congrats to Inception Labs and welcome to the #NVIDIAInception program 🎊


We've raised a $64M Series A led by Kleiner Perkins to build the platform for real-time voice AI. We'll use this funding to expand our team, and to build the next generation of models, infrastructure, and products for voice, starting with Sonic 2.0, available today. Link below







Inception Labs Mercury Coder Small Beta is now available in Cline. It's the first commercial diffusion LLM (dLLM), offering a different approach to text generation. It rivals models like Claude 3.5 Haiku and GPT-4o Mini in code quality while running significantly faster. 🧵







Proud advisor moment 🎉 Huge congratulations to Aditya Grover