Sedrick Keh (@sedrickkeh2) 's Twitter Profile
Sedrick Keh

@sedrickkeh2

research engineer @ToyotaResearch interested in pre-training, post-training, and multimodality

ID: 1574303953960968193

linkhttps://sedrickkeh.github.io calendar_today26-09-2022 07:45:14

156 Tweet

315 Takipçi

262 Takip Edilen

Shun Iwase (@s1wase) 's Twitter Profile Photo

#CVPR2025 starts in two days, and can’t wait to share our new work! 🎉 We present ZeroGrasp, a unified framework for 3D reconstruction and grasp prediction that generalizes to unseen objects. Paper📄: arxiv.org/abs/2504.10857 Webpage🌐:sh8.io/#/zerograsp (1/4 🧵)

Etash Guha @ ICLR (@etash_guha) 's Twitter Profile Photo

OpenThoughts3 is the #1 trending dataset on Huggingface! Thank you to everyone who is using the dataset and giving us great feedback 🚀!

OpenThoughts3 is the #1 trending dataset on Huggingface! Thank you to everyone who is using the dataset and giving us great feedback 🚀!
Katherine Liu (@robo_kat) 's Twitter Profile Photo

How can we achieve both common sense understanding that can deal with varying levels of ambiguity in language and dextrous manipulation? Check out CodeDiffuser, a really neat work that bridges Code Gen with a 3D Diffusion Policy! This was a fun project with cool experiments! 🤖

Thao Nguyen (@thao_nguyen26) 's Twitter Profile Photo

Web data, the “fossil fuel of AI”, is being exhausted. What’s next?🤔 We propose Recycling the Web to break the data wall of pretraining via grounded synthetic data. It is more effective than standard data filtering methods, even with multi-epoch repeats! arxiv.org/abs/2506.04689

Web data, the “fossil fuel of AI”, is being exhausted. What’s next?🤔
We propose Recycling the Web to break the data wall of pretraining via grounded synthetic data. It is more effective than standard data filtering methods, even with multi-epoch repeats!

arxiv.org/abs/2506.04689
Sedrick Keh (@sedrickkeh2) 's Twitter Profile Photo

This plot is a thing of beauty. Great visualization by Jean Mercat! One of many cool artifacts that arose from conducting 1000+ experiments for OpenThoughts 😀

Oliver Segovia 🇺🇸🇵🇭🇸🇬 (@oliversegovia) 's Twitter Profile Photo

Just did a version of this! And it’s such an energizing experience. Travel award for 8 builders and engineers from the Philippines to spend a week in a retreat here in Silicon Valley. All are working on tough problems to deploy AI in a frontier market with poor infra but

Just did a version of this! And it’s such an energizing experience. 

Travel award for 8 builders and engineers from the Philippines to spend a week in a retreat here in Silicon Valley.

All are working on tough problems to deploy AI in a frontier market with poor infra but
Ted Xiao (@xiao_ted) 's Twitter Profile Photo

If you’re working on robotics and AI, the recent Stanford talk from Russ Tedrake on scaling multitask robot manipulation is a mandatory watch, full stop. No marketing, no hype. Just solid hypothesis driven science, evidence backed claims. A gold mine in today’s landscape!

If you’re working on robotics and AI, the recent Stanford talk from <a href="/RussTedrake/">Russ Tedrake</a> on scaling multitask robot manipulation is a mandatory watch, full stop.

No marketing, no hype. Just solid hypothesis driven science, evidence backed claims.

A gold mine in today’s landscape!
Russ Tedrake (@russtedrake) 's Twitter Profile Photo

TRI's latest Large Behavior Model (LBM) paper landed on arxiv last night! Check out our project website: toyotaresearchinstitute.github.io/lbm1/ One of our main goals for this paper was to put out a very careful and thorough study on the topic to help people understand the state of the

Zubair Irshad (@mzubairirshad) 's Twitter Profile Photo

🚀Thrilled to share what we’ve been building at TRI over the past several months: our first Large Behavior Models (LBMs) are here! I’m proud to have been a core contributor to the multi-task policy learning and post-training efforts. At TRI, we’ve been researching how LBMs can

Sukjun (June) Hwang (@sukjun_hwang) 's Twitter Profile Photo

Tokenization has been the final barrier to truly end-to-end language models. We developed the H-Net: a hierarchical network that replaces tokenization with a dynamic chunking process directly inside the model, automatically discovering and operating over meaningful units of data

Mihir Prabhudesai (@mihirp98) 's Twitter Profile Photo

🚨 The era of infinite internet data is ending, So we ask: 👉 What’s the right generative modelling objective when data—not compute—is the bottleneck? TL;DR: ▶️Compute-constrained? Train Autoregressive models ▶️Data-constrained? Train Diffusion models Get ready for 🤿 1/n

🚨 The era of infinite internet data is ending, So we ask:

👉 What’s the right generative modelling objective when data—not compute—is the bottleneck?

TL;DR:

▶️Compute-constrained? Train Autoregressive models

▶️Data-constrained? Train Diffusion models

Get ready for 🤿  1/n
Pratyush Maini (@pratyushmaini) 's Twitter Profile Photo

1/Pretraining is hitting a data wall; scaling raw web data alone leads to diminishing returns. Today DatologyAI shares BeyondWeb, our synthetic data approach & all the learnings from scaling it to trillions of tokens🧑🏼‍🍳 - 3B LLMs beat 8B models🚀 - Pareto frontier for performance

1/Pretraining is hitting a data wall; scaling raw web data alone leads to diminishing returns. Today <a href="/datologyai/">DatologyAI</a> shares BeyondWeb, our synthetic data approach &amp; all the learnings from scaling it to trillions of tokens🧑🏼‍🍳
- 3B LLMs beat 8B models🚀
- Pareto frontier for performance