Robert Nishihara (@robertnishihara) 's Twitter Profile
Robert Nishihara

@robertnishihara

Co-founder @anyscalecompute. Co-creator of @raydistributed. Previously PhD ML at Berkeley.

ID: 25191683

linkhttp://www.robertnishihara.com calendar_today19-03-2009 00:04:47

1,1K Tweet

7,7K Followers

718 Following

Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

One of our biggest advantages is talent. We get top people from all over the world who largely stay here and contribute to innovation here (especially AI). Turning away these people is a mistake.

Anyscale (@anyscalecompute) 's Twitter Profile Photo

1/5 How do you deliver fast, personalized search to 100M+ users? You need more than a good model. You need great infrastructure. Here’s how Notion scaled their search pipeline using Ray + Anyscale ↓🧵

Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

This is an incredibly technical talk from Nubank on building foundation models for financial transactions. The team that did this joined Nubank via an acquisition. They proceeded to leverage all of Nubank's existing data & model infrastructure and surgically insert foundation

This is an incredibly technical talk from <a href="/nubank/">Nubank</a> on building foundation models for financial transactions. The team that did this joined Nubank via an acquisition. They proceeded to leverage all of Nubank's existing data &amp; model infrastructure and surgically insert foundation
Lukas Biewald (@l2k) 's Twitter Profile Photo

My friend Robert Nishihara told me a fun math problem the other day. O3 gets it wrong even after a fair amount of prodding - is this somehow harder than the olympiad problems where it gets 95% accuracy?

My friend <a href="/robertnishihara/">Robert Nishihara</a> told me a fun math problem the other day. O3 gets it wrong even after a fair amount of prodding - is this somehow harder than the olympiad problems where it gets 95% accuracy?
Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

Ray Summit talk submissions are now open! We're very very excited to hear about your work. - How you use Ray - How you use vLLM - AI infrastructure - Multimodal data - Post-training - Agentic systems - Challenges at scale

Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

I've already heard two companies this week say "we built out everything around text data, then we began introducing images / video and everything broke."

Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

The AI compute software stack consists of 3 specialized layers: 🔧🔧🔧 Layer 1: Training & Inference Framework (PyTorch + vLLM) • Runs models efficiently on GPUs • Handles model optimization and model parallelism strategies • Manages accelerator memory and automatic

The AI compute software stack consists of 3 specialized layers:

đź”§đź”§đź”§ Layer 1: Training &amp; Inference Framework (PyTorch + vLLM)
• Runs models efficiently on GPUs
• Handles model optimization and model parallelism strategies
• Manages accelerator memory and automatic
kourosh hakhamaneshi (@cyrushakha) 's Twitter Profile Photo

I get a lot of questions around what is the role of each of these layers of AI compute stack: vLLM, ray, k8s, etc. What does ray do in vLLM, what does ray do around vLLM? Why is Ray core part of post-training frameworks like vERL, etc? In this blog Robert Nishihara depicts what a

Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

This table was a footnote at the end of the blog, but it's actually one of the most interesting points. There is an emerging stack for post-training. anyscale.com/blog/ai-comput…

This table was a footnote at the end of the blog, but it's actually one of the most interesting points. There is an emerging stack for post-training.

anyscale.com/blog/ai-comput…
Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

Beyond pre-training, here's how I imagine most learning will work. 1. AI models / systems will maintain large collections of retrievable knowledge. This will include facts like "the capital of California is Sacramento" and tactics like "when playing Monopoly, buy a bunch of

Ivan Nardini (@ivnardini) 's Twitter Profile Photo

I really enjoyed the new blog from Anyscale about open source stack for AI compute. Robert shared a great collection of examples showing how companies such as Pinterest, Uber, and Roblox integrate Kubernetes, Ray, PyTorch, and vLLM. This stack enables extensive training

Robert Nishihara (@robertnishihara) 's Twitter Profile Photo

Impressive work! Agentic workflows have tons and tons of design and arcitectural decisions that affect performance and quality (choices around models, embedddings, how to tokenize / chunk data, how to do retrieval, how to construct context, etc). There's a massive search space,