@saiprasanna.in 🦋 (@sai_prasanna) 's Twitter Profile
@saiprasanna.in 🦋

@sai_prasanna

🤖 = 42

Following the gradient of interestingness.

RL² research (Reinforcement learning for Robot learning)

saiprasanna.in @ 🦋

ID: 41139792

linkhttp://saiprasanna.in calendar_today19-05-2009 15:40:06

5,5K Tweet

812 Followers

1,1K Following

Tankred Saanum (@tankredsaanum) 's Twitter Profile Photo

Object slots are great for compositional generalization, but can models without these inductive biases learn compositional representations without supervision too? Yes! Unsupervised learning on object videos yields entangled, yet compositional latent codes for objects!

Sarath Chandar (@apsarathchandar) 's Twitter Profile Photo

At Chandar Lab, we are happy to announce the second edition of our assistance program to provide feedback for members of communities underrepresented in AI who want to apply to high-profile graduate programs. Want feedback? Details: chandar-lab.github.io/grad-app-help/. Deadline: Nov 15!

At <a href="/ChandarLab/">Chandar Lab</a>, we are happy to announce the second edition of our assistance program to provide feedback for members of communities underrepresented in AI who want to apply to high-profile graduate programs.  Want feedback? Details: chandar-lab.github.io/grad-app-help/. Deadline: Nov 15!
@saiprasanna.in 🦋 (@sai_prasanna) 's Twitter Profile Photo

Bluesky is hitting a big threshold of ml people migrating from this garbage pile! Join us there!! So many stater packs to fill your feed without puppet master throwing political garbage at you.

Michael Matthews @ ICLR 2025 (@mitrma) 's Twitter Profile Photo

We are very excited to announce Kinetix: an open-ended universe of physics-based tasks for RL! We use Kinetix to train a general agent on millions of randomly generated physics problems and show that this agent generalises to unseen handmade environments. 1/🧵

Tom George (@tomnotgeorge) 's Twitter Profile Photo

SIMPL outperforms CEBRA — a contemporary, more general-purpose, neural-net-based technique — in terms of performance and compute-time. It’s over 30x faster. 12/21

SIMPL outperforms CEBRA — a contemporary, more general-purpose, neural-net-based technique — in terms of performance and compute-time. It’s over 30x faster.

12/21
Costa Huang (@vwxyzjn) 's Twitter Profile Photo

Roger Creus Castanyer just implemented a CleanRL Parallel Q-Networks algorithm (PQN) implementation! 🚀PQN is DQN without a replay buffer and target network. You can run PQN on GPU environments or vectorized environments. E.g., in envpool, PQN gets DQN's score in 1/10th the time

<a href="/creus_roger/">Roger Creus Castanyer</a> just implemented a <a href="/cleanrl_lib/">CleanRL</a> Parallel Q-Networks algorithm (PQN) implementation! 

🚀PQN is DQN without a replay buffer and target network. You can run PQN on GPU environments or vectorized environments. E.g., in envpool, PQN gets DQN's score in 1/10th the time
Clem Bonnet @ICLR 2025 (@clementbonnet16) 's Twitter Profile Photo

Introducing Latent Program Network (LPN), a new architecture for inductive program synthesis that builds in test-time adaption by learning a latent space that can be used for search 🔎 Inspired by ARC Prize 🧩, we designed LPN to tackle out-of-distribution reasoning tasks!

Introducing Latent Program Network (LPN), a new architecture for inductive program synthesis that builds in test-time adaption by learning a latent space that can be used for search 🔎
Inspired by <a href="/arcprize/">ARC Prize</a> 🧩, we designed LPN to tackle out-of-distribution reasoning tasks!
UCL DARK (@ucl_dark) 's Twitter Profile Photo

DARK lab is now also on That Other Place 🦋. Our members are already actively discussing AI on the butterfly. Come be one of our first followers! Handle: UCL-DARK

Frank Hutter (@frankrhutter) 's Twitter Profile Photo

The data science revolution is getting closer. TabPFN v2 is published in Nature: nature.com/articles/s4158… On tabular classification with up to 10k data points & 500 features, in 2.8s TabPFN on average outperforms all other methods, even when tuning them for up to 4 hours🧵1/19

The data science revolution is getting closer. TabPFN v2 is published in Nature: nature.com/articles/s4158… On tabular classification with up to 10k data points &amp; 500 features, in 2.8s TabPFN on average outperforms all other methods, even when tuning them for up to 4 hours🧵1/19
Veronica Chelu (@veronicachelu) 's Twitter Profile Photo

Excited to share that I’ll be starting as a postdoc in Rose Bagot's lab at McGill University , studying the neural mechanisms of psychedelics ✨🐁🌀 Grateful for this new adventure.

Roger Creus Castanyer (@creus_roger) 's Twitter Profile Photo

🚨 Excited to share our new work: "Stable Gradients for Stable Learning at Scale in Deep Reinforcement Learning"! 📈 We propose gradient interventions that enable stable, scalable learning, achieving significant performance gains across agents and environments! Details below 👇

🚨 Excited to share our new work: "Stable Gradients for Stable Learning at Scale in Deep Reinforcement Learning"! 📈

We propose gradient interventions that enable stable, scalable learning, achieving significant performance gains across agents and environments!

Details below 👇
Pablo Samuel Castro (@pcastr) 's Twitter Profile Photo

really excited about this new work we just put out, led by my students Roger Creus Castanyer & Johan S. Obando 👍🏽 , where we examine the challenges of gradient propagation when scaling deep RL networks. roger & johan put in a lot of work and care in this work, check out more details in 🧵👇🏾 !