Eric Hedlin (@iamerichedlin) 's Twitter Profile
Eric Hedlin

@iamerichedlin

Fourth year PhD student at UBC studying computer vision. Two-time World Championships medalist in open water swimming

ID: 2908318646

linkhttp://ehedlin.github.io calendar_today23-11-2014 16:23:41

122 Tweet

248 Takipçi

56 Takip Edilen

Rudy Gilman (@rgilman33) 's Twitter Profile Photo

This layer in DINO-v2 dedicates about half its attention mass to a single operation. Each of the sixteen heads independently learns the same circuit to perform this task. What is this all-important operation? The “no-op”. That’s right, we’re spending half our computation to do…

Andrea Tagliasacchi 🇨🇦 (@taiyasaki) 's Twitter Profile Photo

📢📢📢 "𝐑𝐚𝐝𝐢𝐚𝐧𝐭 𝐅𝐨𝐚𝐦: Real-Time Differentiable Ray Tracing", a mesh-based 3D represention. radfoam.github.io arxiv.org/abs/2502.01157 Co-lead by my PhD students Shrisudhan Govindarajan and Daniel Rebain, and w/ Kwang Moo Yi

Sam Altman (@sama) 's Twitter Profile Photo

we trained a new model that is good at creative writing (not sure yet how/when it will get released). this is the first time i have been really struck by something written by AI; it got the vibe of metafiction so right. PROMPT: Please write a metafictional literary short story

Shakiba (@shakiba_kh) 's Twitter Profile Photo

📢 "StochasticSplats: Stochastic Rasterization for Sorting-Free 3D Gaussian Splatting" 📢 Project page: github.com/ubc-vision/sto… arxiv.org/abs/2503.24366 Introducing sort-free and pop-free stochastic rendering and training of 3D Gaussian.

Rudy Gilman (@rgilman33) 's Twitter Profile Photo

The attention layers in the VAEs for FLUX, Stable Diffusion 3.5, and SDXL don't do anything. You can ablate them with almost no effect. At first I thought they might be involved in some clever circuitry—maybe moving global information—but no they're just flailing around doing

Eric Hedlin (@iamerichedlin) 's Twitter Profile Photo

Our universe may be the time-reversed interior of a black hole, with the Big Bang as the singularity. The arrow of time follows increasing entropy, which is why we experience time as moving away from the Big Bang. Source: youtube.com/watch?v=A8bBhk…

Rudy Gilman (@rgilman33) 's Twitter Profile Photo

Group norm is a destructive operation. It normalizes out much of the information regarding the relative magnitudes of channels. But that information is important! In this VAE many of those channels are describing colors—imagine what would happen if you normalized each channel of

Group norm is a destructive operation. It normalizes out much of the information regarding the relative magnitudes of channels. But that information is important! In this VAE many of those channels are describing colors—imagine what would happen if you normalized each channel of
jack morris (@jxmnop) 's Twitter Profile Photo

this gives a pretty good explanation into how models learn in particular, it explains grokking grokking occurs *exactly* when capacity saturates. this is where models can't perfectly fit every training example, so they have to share info bt examples in a smart way

this gives a pretty good explanation into how models learn

in particular, it explains grokking

grokking occurs *exactly* when capacity saturates. this is where models can't perfectly fit every training example, so they have to share info bt examples in a smart way
Abdullah Hamdi (@eng_hemdi) 's Twitter Profile Photo

Last week, our Triangle splatting paper was quietly released, and since then the tech community ignited fierce debates about it ! It was trending on Hacker News ! Today we released the code! A deep dive into the epic “comeback” of Triangles to the throne of 3D 🧵 1/n

Last week, our Triangle splatting paper was quietly released, and since then the tech community ignited fierce debates about it ! 

It was trending on <a href="/hackernews/">Hacker News</a> ! 

Today we released the code! 

A deep dive into the epic “comeback” of Triangles to the throne of 3D

🧵
1/n
Eric Hedlin (@iamerichedlin) 's Twitter Profile Photo

Come to poster #103 tomorrow morning to find out how to make hypernetworks stable and scalable #CVPR25 x.com/IAmEricHedlin/…