Bowen Li (@bw_li1024) 's Twitter Profile
Bowen Li

@bw_li1024

PhD student @CMU_Robotics work on Reasoning, Planning, and Robotics

ID: 1379225463772512259

linkhttp://jaraxxus-me.github.io calendar_today06-04-2021 00:12:40

274 Tweet

818 Followers

645 Following

Jiafei Duan (@djiafei) 's Twitter Profile Photo

Humans use pointing to communicate plans intuitively. Compared to language, pointing gives more precise guidance to robot behaviors. Can we teach a robot how to point like humans? Introducing RoboPoint πŸ€–πŸ‘‰, an open-source VLM instruction-tuned to point. Check out our new work:

Tom Silver (@tomssilver) 's Twitter Profile Photo

Happy to share a new preprint: "Coloring Between the Lines: Personalization in the Null Space of Planning Constraints" w/ Rajat Kumar Jenamani, Ziang Liu, Ben Dodson, and Tapomayukh "Tapo" Bhattacharjee. TLDR: We propose a method for continual, flexible, active, and safe robot personalization. Links πŸ‘‡

Rajat Kumar Jenamani (@rkjenamani) 's Twitter Profile Photo

Excited to share our work on continual, flexible, active, and safe robot personalization w/ Tom Silver, Ziang Liu, Ben Dodson & Tapomayukh "Tapo" Bhattacharjee. Also: Tom Silver is starting a lab at Princeton!! I HIGHLY recommend joining β€” thoughtful, kind, and an absolute joy to work with!

Yuheng Qiu (@qiuyuhengqiu) 's Twitter Profile Photo

πŸ”₯Best Paper Award at #ICRA2025 Thrilled to share that our paper MAC-VO has been awarded the π˜½π™šπ™¨π™© π˜Ύπ™€π™£π™›π™šπ™§π™šπ™£π™˜π™š 𝙋𝙖π™₯π™šπ™§ π˜Όπ™¬π™–π™§π™™ andΒ the π˜½π™šπ™¨π™© 𝙋𝙖π™₯π™šπ™§ π˜Όπ™¬π™–π™§π™™ 𝙀𝙣 𝙍𝙀𝙗𝙀𝙩 π™‹π™šπ™§π™˜π™šπ™₯π™©π™žπ™€π™£! Check our project: mac-vo.github.io

πŸ”₯Best Paper Award at #ICRA2025 
Thrilled to share that our paper MAC-VO has been awarded the π˜½π™šπ™¨π™© π˜Ύπ™€π™£π™›π™šπ™§π™šπ™£π™˜π™š 𝙋𝙖π™₯π™šπ™§ π˜Όπ™¬π™–π™§π™™ andΒ the π˜½π™šπ™¨π™© 𝙋𝙖π™₯π™šπ™§ π˜Όπ™¬π™–π™§π™™ 𝙀𝙣 𝙍𝙀𝙗𝙀𝙩 π™‹π™šπ™§π™˜π™šπ™₯π™©π™žπ™€π™£!

Check our project: mac-vo.github.io
Guanya Shi (@guanyashi) 's Twitter Profile Photo

System ID for legged robots is hard: (1) Discontinuous dynamics and (2) many parameters to identify and hard to "excite" them. SPI-Active is a general tool for legged robot system ID. Key ideas: (1) massively parallel sampling-based optimization, (2) structured parameter space,

ARC Prize (@arcprize) 's Twitter Profile Photo

Claude Sonnet 4 on ARC-AGI Semi Private Eval Base * ARC-AGI-1: 23%, $0.08/task * ARC-AGI-2: 1.2%, $0.12/task Thinking 16K * ARC-AGI-1: 40%, $0.36/task * ARC-AGI-2: 5.9%, $0.48/task Sonnet 4 sets new SOTA (5.9%) on ARC-AGI-2

Claude Sonnet 4 on ARC-AGI Semi Private Eval

Base
* ARC-AGI-1: 23%, $0.08/task
* ARC-AGI-2: 1.2%, $0.12/task

Thinking 16K
* ARC-AGI-1: 40%, $0.36/task
* ARC-AGI-2: 5.9%, $0.48/task

Sonnet 4 sets new SOTA (5.9%) on ARC-AGI-2
Donglai Xiang (@donglaixiang) 's Twitter Profile Photo

🚨Excited to announce the 1st Workshop on Vision Meets Physics at @CVPR2025! Join us on June 12 for a full-day event exploring the synergy between physical simulation & computer vision to bridge the gap between the virtual and physical worlds. URL: tinyurl.com/vis-phys

🚨Excited to announce the 1st Workshop on Vision Meets Physics at @CVPR2025!

Join us on June 12 for a full-day event exploring the synergy between physical simulation & computer vision to bridge the gap between the virtual and physical worlds.

URL: tinyurl.com/vis-phys
Simeng (Sophia) Han (@hansineng) 's Twitter Profile Photo

Zero fluff, maximum insight ✨. Let’s see what LLMs are really made of, with 🧠 Brainteasers. We’re not grading answers πŸ”’. We’re grading thinking πŸ’­. Brute force? Creative leap? False confession? πŸ€” Instead of asking β€œDid the model get the right answer?”, we ask: β€œDid it

Zero fluff, maximum insight ✨. 
Let’s see what LLMs are really made of, with 🧠 Brainteasers. 

We’re not grading answers πŸ”’. We’re grading thinking πŸ’­. 
Brute force? Creative leap? False confession? πŸ€”

Instead of asking β€œDid the model get the right answer?”, 
we ask: β€œDid it
Changyi Lin (@changyi_lin1) 's Twitter Profile Photo

Introducing LocoTouch: Quadrupedal robots equipped with tactile sensing can now transport unsecured objects β€” no mounts, no straps. The tactile policy transfers zero-shot from sim to real. Core Task-Agnostic Features: 1. High-fidelity contact simulation for distributed tactile

Yilun Du (@du_yilun) 's Twitter Profile Photo

Excited to share work on using classical search approaches to scale inference in diffusion models! We show how global graph search algorithms (BFS, DFS) and local search can be used to improve generation performance across domains such as image generation, planning, and RL!

Excited to share work on using classical search approaches to scale inference in diffusion models!

We show how global graph search algorithms (BFS, DFS) and local search can be used to improve generation performance across domains such as image generation, planning, and RL!
Wenli Xiao (@_wenlixiao) 's Twitter Profile Photo

Tired of watching fancy humanoid dancing? Can they just do some daily useful tasks like: "Pass me a bottle of Water🍺"? πŸ€”Turns out it's nontrivial to stablize whole-body manipulation and locomotion at the same time. We basically want our humanoid to be stable as a camera

Tianyuan Zhang (@tianyuanzhang99) 's Twitter Profile Photo

Bored of linear recurrent memories (e.g., linear attention) and want a scalable, nonlinear alternative? Our new paperΒ β€œTest-Time Training Done Right” proposeΒ LaCT (Large Chunk Test-Time Training)Β β€” a highly efficient, massively scalable nonlinear memory with: πŸ’‘Β Pure PyTorch

Raphaël Millière (@raphaelmilliere) 's Twitter Profile Photo

Transformer-based neural networks achieve impressive performance on coding, math & reasoning tasks that require keeping track of variables and their values. But how can they do that without explicit memory? πŸ“„ Our new ICML paper investigates this in a synthetic setting! 🧡 1/13

Jon Richens (@jonathanrichens) 's Twitter Profile Photo

Are world models necessary to achieve human-level agents, or is there a model-free short-cut? Our new #ICML2025 paper tackles this question from first principles, and finds a surprising answer, agents _are_ world models… 🧡

Are world models necessary to achieve human-level agents, or is there a model-free short-cut?
Our new #ICML2025 paper tackles this question from first principles, and finds a surprising answer, agents _are_ world models… 🧡
Jingyun Yang (@yjy0625) 's Twitter Profile Photo

Introducing Mobi-Ο€: Mobilizing Your Robot Learning Policy. Our method: ✈️ enables flexible mobile skill chaining πŸͺΆ without requiring additional policy training data 🏠 while scaling to unseen scenes πŸ§΅β†“

Seohong Park (@seohong_park) 's Twitter Profile Photo

Is RL really scalable like other objectives? We found that just scaling up data and compute is *not* enough to enable RL to solve complex tasks. The culprit is the horizon. Paper: arxiv.org/abs/2506.04168 Thread ↓

Jiaheng Hu (@jiahenghu1) 's Twitter Profile Photo

Real-world RL, where robots learn directly from physical interactions, is extremely challenging β€” especially for high-DoF systems like mobile manipulators. 1⃣ Long-horizon tasks and large action spaces lead to difficult policy optimization. 2⃣ Real-world exploration with

Yuchen Zhang (@yuchenzhan54250) 's Twitter Profile Photo

Introducing UFM, a Unified Flow & Matching model, which for the first time shows that the unification of optical flow and image matching tasks is mutually beneficial and achieves SOTA. Check out UFM’s matching in action below! πŸ‘‡ 🌐 Website: uniflowmatch.github.io πŸ§΅πŸ‘‡

Introducing UFM, a Unified Flow & Matching model, which for the first time shows that the unification of optical flow and image matching tasks is mutually beneficial and achieves SOTA.

Check out UFM’s matching in action below! πŸ‘‡

🌐 Website: uniflowmatch.github.io
πŸ§΅πŸ‘‡
Tom Silver (@tomssilver) 's Twitter Profile Photo

This week's #PaperILike is "Long-Horizon Multi-Robot Rearrangement Planning for Construction Assembly" (Hartmann et al., TRO 2022). Take two minutes to watch this video: youtube.com/watch?v=Gqhouv… I don't use a lot of emojis, but 🀯 PDF: arxiv.org/abs/2106.02489