Rana Shahout (@rana_shahout) 's Twitter Profile
Rana Shahout

@rana_shahout

Postdoc at Harvard | Computer Science

ID: 1448989776602873884

linkhttps://sites.google.com/view/ranash calendar_today15-10-2021 12:31:09

101 Tweet

209 Followers

368 Following

Lakshya A Agrawal (@lakshyaaagrawal) 's Twitter Profile Photo

🧵Introducing LangProBe: the first benchmark testing where and how composing LLMs into language programs affects cost-quality tradeoffs! We find that, on avg across diverse tasks, smaller models within optimized programs beat calls to larger models at a fraction of the cost.

🧵Introducing LangProBe: the first benchmark testing where and how composing LLMs into language programs affects cost-quality tradeoffs!

We find that, on avg across diverse tasks, smaller models within optimized programs beat calls to larger models at a fraction of the cost.
Eran Malach (@eranmalach) 's Twitter Profile Photo

How does RL improve performance on math reasoning? Studying RL from pretrained models is hard, as behavior depends on choice of base model. 🚨 In our new work, we train models *from scratch* to study the effect of the data mix on the behavior of RL. arxiv.org/abs/2504.07912

How does RL improve performance on math reasoning? Studying RL from pretrained models is hard, as behavior depends on choice of base model. 🚨 In our new work, we train models *from scratch* to study the effect of the data mix on the behavior of RL. arxiv.org/abs/2504.07912
Ayush Noori (@ayushnoori) 's Twitter Profile Photo

We are presenting “Prefix and output length-aware scheduling for efficient online LLM inference” at the ICLR 2025 (ICLR 2026) Sparsity in LLMs workshop (Sparsity in LLMs Workshop at ICLR 2025). 🪫 Challenge: LLM inference in data centers benefits from data parallelism. How can we exploit patterns in

We are presenting “Prefix and output length-aware scheduling for efficient online LLM inference” at the ICLR 2025 (<a href="/iclr_conf/">ICLR 2026</a>) Sparsity in LLMs workshop (<a href="/sparseLLMs/">Sparsity in LLMs Workshop at ICLR 2025</a>).

🪫 Challenge: LLM inference in data centers benefits from data parallelism. How can we exploit patterns in