Nihit Desai (@nihit_desai) 's Twitter Profile
Nihit Desai

@nihit_desai

Co-Founder @RefuelAI. Prev: Facebook, Stanford, IIT-M

ID: 31074223

calendar_today14-04-2009 06:35:23

806 Tweet

1,1K Followers

544 Following

Nihit Desai (@nihit_desai) 's Twitter Profile Photo

I came to the US in 2011 to attend @IllinoisCS. Just found out on the 13th anniversary of first coming here that my EB-1 petition was approved! Excited to continue building Refuel here. The United States is a place where great things are possible - I grew up believing this,

I came to the US in 2011 to attend @IllinoisCS. Just found out on the 13th anniversary of first coming here that my EB-1 petition was approved! Excited to continue building <a href="/RefuelAI/">Refuel</a> here. The United States is a place where great things are possible - I grew up believing this,
vLLM (@vllm_project) 's Twitter Profile Photo

Speculative decoding is one of the best tool in the vLLM's suite of inference optimization tool box, accelerating the inference without accuracy loss. Checkout our blog post for more details about the state of spec decode in vLLM today! 🧵 blog.vllm.ai/2024/10/17/spe…

Steve Jurvetson (@futurejurvetson) 's Twitter Profile Photo

The Moore's Law Update NOTE: this is a semi-log graph, so a straight line is an exponential; each y-axis tick is 100x. This graph covers a 1,000,000,000,000,000,000,000x improvement in computation/$. Pause to let that sink in. Humanity’s capacity to compute has compounded for

The Moore's Law Update
NOTE: this is a semi-log graph, so a straight line is an exponential; each y-axis tick is 100x. This graph covers a 1,000,000,000,000,000,000,000x improvement in computation/$.  Pause to let that sink in.

Humanity’s capacity to compute has compounded for
Nihit Desai (@nihit_desai) 's Twitter Profile Photo

Does training LLMs with reasoning traces (same recipe used in DeepSeek-R1-Distill model family) enhance performance of data labeling, categorization and extraction tasks? Our experiments suggest the improvement is ~5%, but at a substantial increase (>5x) in cost:

Does training LLMs with reasoning traces (same recipe used in DeepSeek-R1-Distill model family) enhance performance of data labeling, categorization and extraction tasks? 

Our experiments suggest the improvement is ~5%, but at a substantial increase (&gt;5x) in cost:
Refuel (@refuelai) 's Twitter Profile Photo

We have some big news to share today - Refuel is joining Together AI to help accelerate the future of open source and enterprise AI! together.ai/blog/together-…

General Catalyst (@generalcatalyst) 's Twitter Profile Photo

Congratulations to our portfolio companies, Together AI and Refuel, on uniting their strengths to power the next generation of AI infrastructure! Together AI’s AI Acceleration Cloud enables developers and enterprises to train and deploy generative AI models with speed,

Congratulations to our portfolio companies, <a href="/togethercompute/">Together AI</a> and <a href="/RefuelAI/">Refuel</a>, on uniting their strengths to power the next generation of AI infrastructure!

Together AI’s AI Acceleration Cloud enables developers and enterprises to train and deploy generative AI models with speed,
Nihit Desai (@nihit_desai) 's Twitter Profile Photo

scarcity ➡️ abundance is a beautiful thing 1850s: Food 1900s: Manufacturing 1950s: Energy 2020s: Intelligence

Matei Zaharia (@matei_zaharia) 's Twitter Profile Photo

Excited to launch Agent Bricks, a new way to build auto-optimized agents on your tasks. Agent Bricks uniquely takes a *declarative* approach to agent development: you tell us what you want, and we auto-generate evals and optimize the agent. databricks.com/blog/introduci…

Nihit Desai (@nihit_desai) 's Twitter Profile Photo

Technical debt is like financial debt in many ways. Borrow from the future to accelerate buildout today. Grow the pie and pay back a much smaller fraction of it in the future