Bowen Li (@bw_li1024) 's Twitter Profile
Bowen Li

@bw_li1024

PhD student @CMU_Robotics work on Reasoning, Planning, and Robotics

ID: 1379225463772512259

linkhttp://jaraxxus-me.github.io calendar_today06-04-2021 00:12:40

274 Tweet

818 Takipรงi

645 Takip Edilen

Jiafei Duan (@djiafei) 's Twitter Profile Photo

Humans use pointing to communicate plans intuitively. Compared to language, pointing gives more precise guidance to robot behaviors. Can we teach a robot how to point like humans? Introducing RoboPoint ๐Ÿค–๐Ÿ‘‰, an open-source VLM instruction-tuned to point. Check out our new work:

Tom Silver (@tomssilver) 's Twitter Profile Photo

Happy to share a new preprint: "Coloring Between the Lines: Personalization in the Null Space of Planning Constraints" w/ Rajat Kumar Jenamani, Ziang Liu, Ben Dodson, and Tapomayukh "Tapo" Bhattacharjee. TLDR: We propose a method for continual, flexible, active, and safe robot personalization. Links ๐Ÿ‘‡

Rajat Kumar Jenamani (@rkjenamani) 's Twitter Profile Photo

Excited to share our work on continual, flexible, active, and safe robot personalization w/ Tom Silver, Ziang Liu, Ben Dodson & Tapomayukh "Tapo" Bhattacharjee. Also: Tom Silver is starting a lab at Princeton!! I HIGHLY recommend joining โ€” thoughtful, kind, and an absolute joy to work with!

Yuheng Qiu (@qiuyuhengqiu) 's Twitter Profile Photo

๐Ÿ”ฅBest Paper Award at #ICRA2025 Thrilled to share that our paper MAC-VO has been awarded the ๐˜ฝ๐™š๐™จ๐™ฉ ๐˜พ๐™ค๐™ฃ๐™›๐™š๐™ง๐™š๐™ฃ๐™˜๐™š ๐™‹๐™–๐™ฅ๐™š๐™ง ๐˜ผ๐™ฌ๐™–๐™ง๐™™ andย the ๐˜ฝ๐™š๐™จ๐™ฉ ๐™‹๐™–๐™ฅ๐™š๐™ง ๐˜ผ๐™ฌ๐™–๐™ง๐™™ ๐™ค๐™ฃ ๐™๐™ค๐™—๐™ค๐™ฉ ๐™‹๐™š๐™ง๐™˜๐™š๐™ฅ๐™ฉ๐™ž๐™ค๐™ฃ! Check our project: mac-vo.github.io

๐Ÿ”ฅBest Paper Award at #ICRA2025 
Thrilled to share that our paper MAC-VO has been awarded the ๐˜ฝ๐™š๐™จ๐™ฉ ๐˜พ๐™ค๐™ฃ๐™›๐™š๐™ง๐™š๐™ฃ๐™˜๐™š ๐™‹๐™–๐™ฅ๐™š๐™ง ๐˜ผ๐™ฌ๐™–๐™ง๐™™ andย the ๐˜ฝ๐™š๐™จ๐™ฉ ๐™‹๐™–๐™ฅ๐™š๐™ง ๐˜ผ๐™ฌ๐™–๐™ง๐™™ ๐™ค๐™ฃ ๐™๐™ค๐™—๐™ค๐™ฉ ๐™‹๐™š๐™ง๐™˜๐™š๐™ฅ๐™ฉ๐™ž๐™ค๐™ฃ!

Check our project: mac-vo.github.io
Guanya Shi (@guanyashi) 's Twitter Profile Photo

System ID for legged robots is hard: (1) Discontinuous dynamics and (2) many parameters to identify and hard to "excite" them. SPI-Active is a general tool for legged robot system ID. Key ideas: (1) massively parallel sampling-based optimization, (2) structured parameter space,

ARC Prize (@arcprize) 's Twitter Profile Photo

Claude Sonnet 4 on ARC-AGI Semi Private Eval Base * ARC-AGI-1: 23%, $0.08/task * ARC-AGI-2: 1.2%, $0.12/task Thinking 16K * ARC-AGI-1: 40%, $0.36/task * ARC-AGI-2: 5.9%, $0.48/task Sonnet 4 sets new SOTA (5.9%) on ARC-AGI-2

Claude Sonnet 4 on ARC-AGI Semi Private Eval

Base
* ARC-AGI-1: 23%, $0.08/task
* ARC-AGI-2: 1.2%, $0.12/task

Thinking 16K
* ARC-AGI-1: 40%, $0.36/task
* ARC-AGI-2: 5.9%, $0.48/task

Sonnet 4 sets new SOTA (5.9%) on ARC-AGI-2
Donglai Xiang (@donglaixiang) 's Twitter Profile Photo

๐ŸšจExcited to announce the 1st Workshop on Vision Meets Physics at @CVPR2025! Join us on June 12 for a full-day event exploring the synergy between physical simulation & computer vision to bridge the gap between the virtual and physical worlds. URL: tinyurl.com/vis-phys

๐ŸšจExcited to announce the 1st Workshop on Vision Meets Physics at @CVPR2025!

Join us on June 12 for a full-day event exploring the synergy between physical simulation & computer vision to bridge the gap between the virtual and physical worlds.

URL: tinyurl.com/vis-phys
Simeng (Sophia) Han (@hansineng) 's Twitter Profile Photo

Zero fluff, maximum insight โœจ. Letโ€™s see what LLMs are really made of, with ๐Ÿง  Brainteasers. Weโ€™re not grading answers ๐Ÿ”ข. Weโ€™re grading thinking ๐Ÿ’ญ. Brute force? Creative leap? False confession? ๐Ÿค” Instead of asking โ€œDid the model get the right answer?โ€, we ask: โ€œDid it

Zero fluff, maximum insight โœจ. 
Letโ€™s see what LLMs are really made of, with ๐Ÿง  Brainteasers. 

Weโ€™re not grading answers ๐Ÿ”ข. Weโ€™re grading thinking ๐Ÿ’ญ. 
Brute force? Creative leap? False confession? ๐Ÿค”

Instead of asking โ€œDid the model get the right answer?โ€, 
we ask: โ€œDid it
Changyi Lin (@changyi_lin1) 's Twitter Profile Photo

Introducing LocoTouch: Quadrupedal robots equipped with tactile sensing can now transport unsecured objects โ€” no mounts, no straps. The tactile policy transfers zero-shot from sim to real. Core Task-Agnostic Features: 1. High-fidelity contact simulation for distributed tactile

Yilun Du (@du_yilun) 's Twitter Profile Photo

Excited to share work on using classical search approaches to scale inference in diffusion models! We show how global graph search algorithms (BFS, DFS) and local search can be used to improve generation performance across domains such as image generation, planning, and RL!

Excited to share work on using classical search approaches to scale inference in diffusion models!

We show how global graph search algorithms (BFS, DFS) and local search can be used to improve generation performance across domains such as image generation, planning, and RL!
Wenli Xiao (@_wenlixiao) 's Twitter Profile Photo

Tired of watching fancy humanoid dancing? Can they just do some daily useful tasks like: "Pass me a bottle of Water๐Ÿบ"? ๐Ÿค”Turns out it's nontrivial to stablize whole-body manipulation and locomotion at the same time. We basically want our humanoid to be stable as a camera

Tianyuan Zhang (@tianyuanzhang99) 's Twitter Profile Photo

Bored of linear recurrent memories (e.g., linear attention) and want a scalable, nonlinear alternative? Our new paperย โ€œTest-Time Training Done Rightโ€ proposeย LaCT (Large Chunk Test-Time Training)ย โ€” a highly efficient, massively scalable nonlinear memory with: ๐Ÿ’กย Pure PyTorch

Raphaรซl Milliรจre (@raphaelmilliere) 's Twitter Profile Photo

Transformer-based neural networks achieve impressive performance on coding, math & reasoning tasks that require keeping track of variables and their values. But how can they do that without explicit memory? ๐Ÿ“„ Our new ICML paper investigates this in a synthetic setting! ๐Ÿงต 1/13

Jon Richens (@jonathanrichens) 's Twitter Profile Photo

Are world models necessary to achieve human-level agents, or is there a model-free short-cut? Our new #ICML2025 paper tackles this question from first principles, and finds a surprising answer, agents _are_ world modelsโ€ฆ ๐Ÿงต

Are world models necessary to achieve human-level agents, or is there a model-free short-cut?
Our new #ICML2025 paper tackles this question from first principles, and finds a surprising answer, agents _are_ world modelsโ€ฆ ๐Ÿงต
Jingyun Yang (@yjy0625) 's Twitter Profile Photo

Introducing Mobi-ฯ€: Mobilizing Your Robot Learning Policy. Our method: โœˆ๏ธ enables flexible mobile skill chaining ๐Ÿชถ without requiring additional policy training data ๐Ÿ  while scaling to unseen scenes ๐Ÿงตโ†“

Seohong Park (@seohong_park) 's Twitter Profile Photo

Is RL really scalable like other objectives? We found that just scaling up data and compute is *not* enough to enable RL to solve complex tasks. The culprit is the horizon. Paper: arxiv.org/abs/2506.04168 Thread โ†“

Jiaheng Hu (@jiahenghu1) 's Twitter Profile Photo

Real-world RL, where robots learn directly from physical interactions, is extremely challenging โ€” especially for high-DoF systems like mobile manipulators. 1โƒฃ Long-horizon tasks and large action spaces lead to difficult policy optimization. 2โƒฃ Real-world exploration with

Yuchen Zhang (@yuchenzhan54250) 's Twitter Profile Photo

Introducing UFM, a Unified Flow & Matching model, which for the first time shows that the unification of optical flow and image matching tasks is mutually beneficial and achieves SOTA. Check out UFMโ€™s matching in action below! ๐Ÿ‘‡ ๐ŸŒ Website: uniflowmatch.github.io ๐Ÿงต๐Ÿ‘‡

Introducing UFM, a Unified Flow & Matching model, which for the first time shows that the unification of optical flow and image matching tasks is mutually beneficial and achieves SOTA.

Check out UFMโ€™s matching in action below! ๐Ÿ‘‡

๐ŸŒ Website: uniflowmatch.github.io
๐Ÿงต๐Ÿ‘‡
Tom Silver (@tomssilver) 's Twitter Profile Photo

This week's #PaperILike is "Long-Horizon Multi-Robot Rearrangement Planning for Construction Assembly" (Hartmann et al., TRO 2022). Take two minutes to watch this video: youtube.com/watch?v=Gqhouvโ€ฆ I don't use a lot of emojis, but ๐Ÿคฏ PDF: arxiv.org/abs/2106.02489