Grant Watson
@grhwatson
ML @RecursionPharma. Previous: ML Engineer @dewpoint_tx @PhenomicAI. Into ML, physics, math, music, computer-generated art, and Dungeons & Dragons.
ID: 894635129485897729
07-08-2017 19:03:31
500 Tweet
218 Followers
1,1K Following
Artifacts in your attention maps? Forgot to train with registers? Use ๐ฉ๐๐จ๐ฉ-๐ฉ๐๐ข๐ ๐ง๐๐๐๐จ๐ฉ๐๐ง๐จ! We find a sparse set of activations set artifact positions. We can shift them anywhere ("Shifted") โ even outside the image into an untrained token. Clean maps, no retrain.
Gabriele Corso Patrick Walters Various forms of this discussion are playing out in a lot of different "AI x science" areas right now. (I'm team extrapolation-is-good, but open to being wrong.) I wrote about closely related topics previously, albeit in an esoteric format: corinwagen.github.io/public/blog/20โฆ
The secret behind Gemini 3? Simple: Improving pre-training & post-training ๐คฏ Pre-training: Contra the popular belief that scaling is overโwhich we discussed in our NeurIPS '25 talk with Ilya Sutskever and Quoc Leโthe team delivered a drastic jump. The delta between 2.5 and 3.0 is
Excited to announce our MIT Press book โNeuroevolution: Harnessing Creativity in AI Agent Designโ by Sebastian Risi (Sebastian Risi), Yujin Tang (Yujin Tang), Risto Miikkulainen, and myself. We explore decades of work on evolving intelligent agents and shows how neuroevolution can
We worked with Ginkgo Bioworks to connect GPT-5 to an autonomous lab, so it could propose experiments, run them at scale, learn from the results, and decide what to try next. That closed loop brought protein production cost down by 40%.