Nikhil Barhate
@nikhilbarhate99
ML @scale_AI | prev @AMD @mila_quebec
ID: 3245294515
https://nikhilbarhate99.github.io 14-06-2015 15:04:57
1,1K Tweet
204 Followers
810 Following
You like discrete diffusion, but it's too slow? ๐ฅ You like test-time inference, but it's for continuous methods? ๐ฉ We fixed it. Introducing Categorical Flow Maps: continuously sample discrete data in a single step ๐๐ซ How? ๐งตโฌ๏ธ ๐ช Co-led with Floor Eijkelboom, Daan Roos
What's the right space to diffuse in: Raw Data or Latents? Why not both! In Latent Forcing, we order a joint diffusion trajectory to reveal Latents before Pixels, leading to improved convergence while being lossless at encoding and end-to-end at inference. w/ Fei-Fei Li+... 1/n
We scaled off-policy RL to sim-to-real. To our knowledge, FlashSAC is the fastest and most performant RL algorithm across IsaacLab, MuJoCo Playground, and many more, all with a single set of hyperparameters. Project page: holiday-robot.github.io/FlashSAC Paper: arxiv.org/pdf/2604.04539
๐ซฑ Introducing ๐๐๐ฎ๐ซ๐๐ฅ ๐๐จ๐ฆ๐ฉ๐ฎ๐ญ๐๐ซs: ๐ฐ๐ก๐๐ญ ๐ข๐ ๐๐ ๐๐จ๐๐ฌ ๐ง๐จ๐ญ ๐ฃ๐ฎ๐ฌ๐ญ ๐ฎ๐ฌ๐ ๐๐จ๐ฆ๐ฉ๐ฎ๐ญ๐๐ซ๐ฌ ๐๐๐ญ๐ญ๐๐ซ, ๐๐ฎ๐ญ ๐๐๐ ๐ข๐ง๐ฌ ๐ญ๐จ ๐๐๐๐จ๐ฆ๐ ๐ญ๐ก๐ ๐ซ๐ฎ๐ง๐ง๐ข๐ง๐ ๐๐จ๐ฆ๐ฉ๐ฎ๐ญ๐๐ซ ๐ข๐ญ๐ฌ๐๐ฅ๐? Beyond today's conventional computers, agents, and
1/ Reinforcement learning is usually framed as maximizing rewards. But can we cast it as reaching the right goals? New blog on bridging RL, goal-conditioned RL, and stochastic shortest path: iclr-blogposts.github.io/2026/blog/2026โฆ Also #ICLR2026 Poster: Thu 10:30 AMโ1:00 PM, P4 #4611. ๐งตโฌ๏ธ