Vincent Liu (@vincentjliu) 's Twitter Profile
Vincent Liu

@vincentjliu

Building something new 🤖 Previously @CSM_ai, @Tesla Autopilot, @NVIDIA, @Stanford BA/MS

ID: 1582749853926526976

linkhttp://vliu15.github.io calendar_today19-10-2022 15:05:53

311 Tweet

178 Followers

119 Following

Vincent Liu (@vincentjliu) 's Twitter Profile Photo

GPT-4.5 is the wall and also the most underwhelming release of all time. OpenAI is serving the purpose it was intended to back in 2015: it was meant to fund, accelerate, and disseminate the benefits of AGI to humanity. Today, OpenAI research has the world’s worst business

Vincent Liu (@vincentjliu) 's Twitter Profile Photo

It’s becoming popular opinion among researchers (in many domains I’ve talked to) that there is no business model in foundation models. The model layer will get commoditized, as it has been happening with OpenAI, hence the poor business fundamentals, hence the purpose of the

Vincent Liu (@vincentjliu) 's Twitter Profile Photo

Deep tech is tricky because product informs tech, but tech constrains the product's solution space. It requires forecasting long-term research progress and straddling product needs until both converge. PBRs are an algorithmic research problem but are finally becoming usable 🙌

Irmak Guzey (@irmakkguzey) 's Twitter Profile Photo

Despite great advances in learning dexterity, hardware remains a major bottleneck. Most dexterous hands are either bulky, weak or expensive. I’m thrilled to present the RUKA Hand — a powerful, accessible research tool for dexterous manipulation that overcomes these limitations!

Ademi Adeniji (@ademiadeniji) 's Twitter Profile Photo

Closed-loop robot policies directly from human interactions. No teleop, no robot data co-training, no RL, and no sim. Just Aria smart glasses. Everyday human data is passively scalable and a massively underutilized resource in robotics...More to come here in the coming weeks.

Vincent Liu (@vincentjliu) 's Twitter Profile Photo

I think the most interesting insight from EgoZero is the tradeoff between 2D/3D representations in human-to-robot learning. 2D inputs (images, VLMs) scale and encode strong visual priors—but mapping these to 3D actions is hard and data-hungry, especially given the human-robot

I think the most interesting insight from EgoZero is the tradeoff between 2D/3D representations in human-to-robot learning.

2D inputs (images, VLMs) scale and encode strong visual priors—but mapping these to 3D actions is hard and data-hungry, especially given the human-robot
Vincent Liu (@vincentjliu) 's Twitter Profile Photo

In AI, better models → higher productivity → higher margins → more output + capital deployment. In robotics, we often view outcomes as binary: solved or not. But if we instead frame partial success rates as boosting physical labor productivity, we can potentially increase

Vincent Liu (@vincentjliu) 's Twitter Profile Photo

When we think of human data collection, we usually think of recording accurate 3D proprioception, but this alone does not contain force information. On the other hand, teleoperation data collects the human's input as action labels, so robots trained this way can learn how much