Avery Ryoo (@averyryoo) 's Twitter Profile
Avery Ryoo

@averyryoo

MSc @Mila_Quebec | multimodal learning, generative modelling, neural decoding | @Raptors

ID: 803392331764342785

linkhttps://averyryoo.github.io calendar_today29-11-2016 00:17:13

107 Tweet

220 Takipçi

667 Takip Edilen

Veronica Chelu (@veronicachelu) 's Twitter Profile Photo

My first time CosyneMeeting this week and I couldn't be more excited 😬 Tomorrow, I'll be presenting "Mood as an Extrapolation Engine for Decision-Making", a cognitive science perspective on "Functional Acceleration for Policy Mirror Descent" (arxiv.org/abs/2407.16602) Come chat!

My first time <a href="/CosyneMeeting/">CosyneMeeting</a> this week and I couldn't be more excited 😬 Tomorrow, I'll be presenting "Mood as an Extrapolation Engine for Decision-Making", a cognitive science perspective on "Functional Acceleration for Policy Mirror Descent" (arxiv.org/abs/2407.16602)
Come chat!
Luke Rowe (@luke22r) 's Twitter Profile Photo

Scenario Dreamer has been accepted at #CVPR2025! Website: …ceton-computational-imaging.github.io/scenario-dream… We train a vectorized latent diffusion model to synthesize high-fidelity driving simulation environments (agents+map). Scenario Dreamer enables fully data-driven closed-loop generative simulation!

Damien Ferbach (@damien_ferbach) 's Twitter Profile Photo

It's very difficult to improve the *exponent* in scaling laws for loss vs compute, especially by changing the optimizer! Our new paper shows that scaling momentum correctly can *provably* improve the scaling exponent on a theoretical model. Empirically, it works on LSTMs too!

It's very difficult to improve the *exponent* in scaling laws for loss vs compute, especially by changing the optimizer!
Our new paper shows that scaling momentum correctly can *provably* improve the scaling exponent on a theoretical model. Empirically, it works on LSTMs too!
Siddarth Venkatraman (@siddarthv66) 's Twitter Profile Photo

Is there a universal strategy to turn any generative model—GANs, VAEs, diffusion models, or flows—into a conditional sampler, or finetuned to optimize a reward function? Yes! Outsourced Diffusion Sampling (ODS) accepted to ICML Conference , does exactly that!

Is there a universal strategy to turn any generative model—GANs, VAEs, diffusion models, or flows—into a conditional sampler, or finetuned to optimize a reward function?
Yes! Outsourced Diffusion Sampling (ODS) accepted to <a href="/icmlconf/">ICML Conference</a> , does exactly that!
Benjamin Thérien (@benjamintherien) 's Twitter Profile Photo

Is AdamW the best inner optimizer for DiLoCo? Does the inner optimizer affect the compressibility of the DiLoCo delta? Excited to introduce MuLoCo: Muon is a practical inner optimizer for DiLoCo! 🧵arxiv.org/abs/2505.23725 1/N

Is AdamW the best inner optimizer for DiLoCo? Does the inner optimizer affect the compressibility of the DiLoCo delta? Excited to introduce MuLoCo: Muon is a practical inner optimizer for DiLoCo! 🧵arxiv.org/abs/2505.23725 1/N
Dane Malenfant (@dvnxmvl_hdf5) 's Twitter Profile Photo

Preprint Alert 🚀 Multi-agent reinforcement learning (MARL) often assumes that agents know when other agents cooperate with them. But for humans, this isn’t always true. Example, plains indigenous groups used to leave resources for others to use at effigies called Manitokan. 1/8

Preprint Alert  🚀
Multi-agent reinforcement learning (MARL) often assumes that agents know when other agents cooperate with them. But for humans, this isn’t always true. Example, plains indigenous groups used to leave resources for others to use at effigies called Manitokan.
1/8
Majdi Hassan (@majdi_has) 's Twitter Profile Photo

(1/n)🚨You can train a model solving DFT for any geometry almost without training data!🚨 Introducing Self-Refining Training for Amortized Density Functional Theory — a variational framework for learning a DFT solver that predicts the ground-state solutions for different

Emiliano Penaloza (@emilianopp_) 's Twitter Profile Photo

Excited that our paper "Addressing Concept Mislabeling in Concept Bottleneck Models Through Preference Optimization" was accepted to ICML 2025! We show how Preference Optimization can reduce the impact of noisy concept labels in CBMs. 🧵/9

Luke Rowe (@luke22r) 's Twitter Profile Photo

🚀 Our method, Poutine, was the best-performing entry in the 2025 Waymo Vision-based End-to-End Driving Challenge at #CVPR2025! Our 3 B-parameter VLM Poutine scored 7.99 RFS on the official test set—comfortably ahead of every other entry (see figure).

🚀 Our method, Poutine, was the best-performing entry in the 2025 Waymo Vision-based End-to-End Driving Challenge at #CVPR2025!

Our 3 B-parameter VLM Poutine scored 7.99 RFS on the official test set—comfortably ahead of every other entry (see figure).
Drew Livingstone (@producerdrew_) 's Twitter Profile Photo

The worst part about the Panthers winning the Stanley Cup in 6 games is that Leafs fans will believe this makes them the 2nd best team

Avery Ryoo (@averyryoo) 's Twitter Profile Photo

big fan of this line of work combining diffusion models at inference time - not an expert but the modularity + lack of extra training feels like an intuitive and elegant approach to more controllable generation