Dane Malenfant (@dvnxmvl_hdf5) 's Twitter Profile
Dane Malenfant

@dvnxmvl_hdf5

MSc. Computer Science @Mila_Quebec & @mcgillu in the LiNC lab | Currently distracted with multi-agent RL and neuroAI | Restless | Ēka ē-akimiht

ID: 1733325744187539456

linkhttps://danemalenfant.com/ calendar_today09-12-2023 03:21:01

128 Tweet

101 Takipçi

218 Takip Edilen

NYU Center for Data Science (@nyudatascience) 's Twitter Profile Photo

CDS Professor Yann LeCun sees the end of large language models, claiming they'll be obsolete in five years. In Newsweek, he explains why current AI lacks real-world understanding—and what a smarter system could look like.   newsweek.com/ai-impact-inte…

Pablo Samuel Castro (@pcastr) 's Twitter Profile Photo

The Impact of On-Policy Parallelized Data Collection on Deep Reinforcement Learning Networks thrilled to share our #ICML2025 paper led by Walter Mayor-Toro & Johan S. Obando 👍🏽 , with Aaron Courville , where we explore how data collection affects agents in parallelized setups. 1/

The Impact of On-Policy Parallelized Data Collection on Deep Reinforcement Learning Networks

thrilled to share our #ICML2025 paper led by <a href="/WalterMayor_T/">Walter Mayor-Toro</a> &amp; <a href="/johanobandoc/">Johan S. Obando 👍🏽</a> , with <a href="/AaronCourville/">Aaron Courville</a> , where we explore how data collection affects agents in parallelized setups.
1/
Nanda H Krishna (@nandahkrishna) 's Twitter Profile Photo

New preprint! 🧠🤖 How do we build neural decoders that are: ⚡️ fast enough for real-time use 🎯 accurate across diverse tasks 🌍 generalizable to new sessions, subjects, and species? We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes! 🧵1/7

New preprint! 🧠🤖
How do we build neural decoders that are:
⚡️ fast enough for real-time use
🎯 accurate across diverse tasks
🌍 generalizable to new sessions, subjects, and species?
We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes!
🧵1/7
Majdi Hassan (@majdi_has) 's Twitter Profile Photo

(1/n)🚨You can train a model solving DFT for any geometry almost without training data!🚨 Introducing Self-Refining Training for Amortized Density Functional Theory — a variational framework for learning a DFT solver that predicts the ground-state solutions for different

Emiliano Penaloza (@emilianopp_) 's Twitter Profile Photo

Excited that our paper "Addressing Concept Mislabeling in Concept Bottleneck Models Through Preference Optimization" was accepted to ICML 2025! We show how Preference Optimization can reduce the impact of noisy concept labels in CBMs. 🧵/9

Benno Krojer (@benno_krojer) 's Twitter Profile Photo

Excited to share the results of my internship research with AI at Meta, as part of a larger world modeling release! What subtle shortcuts are VideoLLMs taking on spatio-temporal questions? And how can we instead curate shortcut-robust examples at a large-scale? Details 👇🔬

Excited to share the results of my internship research with <a href="/AIatMeta/">AI at Meta</a>, as part of a larger world modeling release!

What subtle shortcuts are VideoLLMs taking on spatio-temporal questions?

And how can we instead curate shortcut-robust examples at a large-scale?

Details 👇🔬
Benjamin Thérien (@benjamintherien) 's Twitter Profile Photo

Tired of tuning hyperparameters? Introducing PyLO! We’re bringing hyperparameter-free learned optimizers to PyTorch with drop in torch.optim support and faster step times thanks to our custom cuda kernels. Check out our code here: github.com/Belilovsky-Lab…

Wilka Carvalho (@cogscikid) 's Twitter Profile Photo

To help computational cognitive scientist engage with more naturalistic experiments, I've made NiceWebRL. NiceWebRL is a Python library for designing human subject experiments that leverage machine reinforcement learning environments. github.com/KempnerInstitu…

Roger Creus Castanyer (@creus_roger) 's Twitter Profile Photo

🚨 Excited to share our new work: "Stable Gradients for Stable Learning at Scale in Deep Reinforcement Learning"! 📈 We propose gradient interventions that enable stable, scalable learning, achieving significant performance gains across agents and environments! Details below 👇

🚨 Excited to share our new work: "Stable Gradients for Stable Learning at Scale in Deep Reinforcement Learning"! 📈

We propose gradient interventions that enable stable, scalable learning, achieving significant performance gains across agents and environments!

Details below 👇
Johan S. Obando 👍🏽 (@johanobandoc) 's Twitter Profile Photo

🚨 Excite to share "Stable Gradients for Stable Learning at Scale in Deep Reinforcement Learning" work. 🥳 We tackle gradient instability in large deep RL networks, enabling stable and scalable learning with strong performance across the board. 📄 Paper: arxiv.org/abs/2506.15544