Haitham Bou Ammar (@hbouammar) 's Twitter Profile
Haitham Bou Ammar

@hbouammar

RL team leader @Huawei R&D UK H. Assistant Professor @UCL | Ex-@Princeton, Upenn (thou/thine)

ID: 1364749022

calendar_today19-04-2013 15:51:58

3,3K Tweet

4,4K Followers

368 Following

Vector Wang (@vectorwang2) 's Twitter Profile Photo

Amazing team from Astera Institute at Seeed Studio NVIDIA Robotics LeRobot home robot hackathon! โœจ VR app with stream & arm mapping ๐ŸŽฎ Isaac Sim BEHAVIOR env + RL ๐Ÿ—บ๏ธ MUSt3R-based navi & 3D reconstr ๐Ÿง  GR00T 1.5 training with Jetson Thor All open source:

Russ Salakhutdinov (@rsalakhu) 's Twitter Profile Photo

Check out new work on Training LLMs to Discover Abstractions for Solving Reasoning Problems: Paper+Code: cohenqu.github.io/rlad.github.io/ Reasoning goes beyond pattern matching or memorization, it requires discovering and applying algorithmic procedures that reuse primitives,

Check out new work on Training LLMs to Discover Abstractions for Solving Reasoning Problems:

Paper+Code: cohenqu.github.io/rlad.github.io/

Reasoning goes beyond pattern matching or memorization, it requires discovering and applying algorithmic procedures that reuse primitives,
Russ Salakhutdinov (@rsalakhu) 's Twitter Profile Photo

New work on Rethinking Thinking Tokens: LLMs as Improvement Operators: arxiv.org/abs/2510.01123 Reasoning training encourages LLMs to produce long chains of thought (CoT), improving accuracy via self-checking but increasing context length, compute cost, and latency. This work

New work on Rethinking Thinking Tokens: LLMs as Improvement Operators:

arxiv.org/abs/2510.01123

Reasoning training encourages LLMs to produce long chains of thought (CoT), improving accuracy via self-checking but increasing context length, compute cost, and latency. This work
Haitham Bou Ammar (@hbouammar) 's Twitter Profile Photo

๐Ÿš€ New Paper: Bottlenecked Transformers โ€“ Memory Consolidation for Better Reasoning in LLMs ๐Ÿง ๐Ÿ“š Large Language Models have shown that more inference-time compute often means better reasoning. But most of this extra computation happens in token space โ€” through longer chains of

๐Ÿš€ New Paper: Bottlenecked Transformers โ€“ Memory Consolidation for Better Reasoning in LLMs ๐Ÿง ๐Ÿ“š

Large Language Models have shown that more inference-time compute often means better reasoning. But most of this extra computation happens in token space โ€” through longer chains of
Wenli Xiao (@_wenlixiao) 's Twitter Profile Photo

What if robots could improve themselves by learning from their own failures in the real-world? Introducing ๐—ฃ๐—Ÿ๐—— (๐—ฃ๐—ฟ๐—ผ๐—ฏ๐—ฒ, ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป, ๐——๐—ถ๐˜€๐˜๐—ถ๐—น๐—น) โ€” a recipe that enables Vision-Language-Action (VLA) models to self-improve for high-precision manipulation tasks. PLD

Ariel (@redtachyon) 's Twitter Profile Photo

Aight let's unclickbait the fp16 paper. tล‚;dr cool paper, a little bit overstated in comms, very overstated by poasters. The thing that gave me a pause is that on the surface, it seems to claim that bf16 is horrible, borderline unusable. But that's not really the case (nor is

Aight let's unclickbait the fp16 paper.

tล‚;dr cool paper, a little bit overstated in comms, very overstated by poasters.

The thing that gave me a pause is that on the surface, it seems to claim that bf16 is horrible, borderline unusable. But that's not really the case (nor is
Igor Kulakov (@ihorbeaver) 's Twitter Profile Photo

Here is the full run of MicroFactory, that autonomously assembles a photo frame - a real product sold on Amazon. This is a good example of how to deal with complex tasks, while we do not yet have a large set of robotic data to train a big general ai model. Thread:

yingzhen (@liyzhen2) 's Twitter Profile Photo

The VCL paper has arguably the first example of modern continual learning for GenAI: VAEs trained on digit/alphabet images 1-by-1 arxiv.org/abs/1710.10628 Coded by yours truly โ˜บ๏ธ who was (and still is) ๐Ÿฅฐ in generative models. Time to get back to continual learning again?

The VCL paper has arguably the first example of modern continual learning for GenAI: VAEs trained on digit/alphabet images 1-by-1 

arxiv.org/abs/1710.10628

Coded by yours truly โ˜บ๏ธ who was (and still is) ๐Ÿฅฐ in generative models.

Time to get back to continual learning again?
Haitham Bou Ammar (@hbouammar) 's Twitter Profile Photo

If you have ever worked with ROS, then you know the pain! Ark elevates this pain by offering you a low-weight Python package that remedies that pain. 1โƒฃ pip installable โœŒ 2โƒฃ Integrates many robots and sensors โœŒ 3โƒฃ Supports a publisher-subscriber architecture ๐Ÿ˜Ž 4โƒฃ Offers

If you have ever worked with ROS, then you know the pain!

Ark elevates this pain by offering you a low-weight Python package that remedies that pain.

1โƒฃ pip installable โœŒ
2โƒฃ Integrates many robots and sensors โœŒ
3โƒฃ Supports a publisher-subscriber architecture ๐Ÿ˜Ž
4โƒฃ Offers
Francesco Bertolotti (@f14bertolotti) 's Twitter Profile Photo

This is a new small LM (1.5B) that achieves impressing reasoning capabilities on AIME, LCB, and GPQA. The authors used model merging and an entropy maximizing variations of GRPO. Impressive work! ๐Ÿ”—arxiv.org/abs/2511.06221

This is a new small LM (1.5B) that achieves impressing reasoning capabilities on AIME, LCB, and GPQA. The authors used model merging and an entropy maximizing variations of GRPO.  Impressive work!

๐Ÿ”—arxiv.org/abs/2511.06221
Pasquale Minervini is hiring postdocs! ๐Ÿš€ (@pminervini) 's Twitter Profile Photo

Very interesting results from Cyrus (Cyrus Wai-Chung Kwan) -- training on generated math reasoning problems within an open-ended self-play framework can yield more accurate results than training on "gold" datasets like GSM8K or MATH!

Very interesting results from Cyrus (<a href="/cyruskwan1997/">Cyrus Wai-Chung Kwan</a>) -- training on generated math reasoning problems within an open-ended self-play framework can yield more accurate results than training on "gold" datasets like GSM8K or MATH!
Yann LeCun (@ylecun) 's Twitter Profile Photo

Chris Murphy ๐ŸŸง You're being played by people who want regulatory capture. They are scaring everyone with dubious studies so that open source models are regulated out of existence.

GreiffLab ๐Ÿ’ป๐Ÿ”ฌ๐Ÿ’Š (@victorgreiff) 's Twitter Profile Photo

The AIRR-ML-2025 challenge is LIVE on @Kaggle! Weโ€™re challenging the best in #MachineLearning & #DataScience to decode the immune system. ๐Ÿ’ฐ $10,000 prize pool ๐Ÿ“„ Co-author a @NatureMethods paper

Edward Johns (@ed__johns) 's Twitter Profile Photo

Do you want to see if your robot can learn a thousand tasks in a day ... ? Following our Science Robotics paper last week, we now give you: >> Our Code and Model Weights ๐Ÿง‘โ€๐Ÿ’ป >> An Explainer Video for the Paper ๐Ÿ“ฝ๏ธ For both: robot-learning.uk/learning-1000-โ€ฆ Let us know how you get on!

Haitham Bou Ammar (@hbouammar) 's Twitter Profile Photo

I have not complained for a while, so I have this itch to do that ๐Ÿ˜† They tell you robotics is almost there! They tell you it is just on the cusp! They tell you to get a robot to your house (and let a tele-operation engineer invade your privacy)! Well, well, well ...

I have not complained for a while, so I have this itch to do that ๐Ÿ˜† 

They tell you robotics is almost there! They tell you it is just on the cusp! They tell you to get a robot to your house (and let a tele-operation engineer invade your privacy)! 

Well, well, well ...
Pasquale Minervini is hiring postdocs! ๐Ÿš€ (@pminervini) 's Twitter Profile Photo

chatted with a journalist earlier this week about temporal reasoning and VLMs, and one explanation that popped up for some behaviour is that those models are like a person waking up after a nap that lasted a few years -- this is very on point