Gytis Daujotas (@gytdau) 's Twitter Profile
Gytis Daujotas

@gytdau

making computers do my bidding.

ID: 3133298386

linkhttp://gytis.co calendar_today04-04-2015 10:34:46

464 Tweet

1,1K Takipçi

314 Takip Edilen

Matthew Siu (@matthewwsiu) 's Twitter Profile Photo

📣 Excited to share Latticework, a text-editing environment aimed to help synthesize freeform, unstructured documents ✍️ Made in collaboration w/ Andy Matuschak

📣 Excited to share Latticework, a text-editing environment aimed to help synthesize freeform, unstructured documents ✍️

Made in collaboration w/ <a href="/andy_matuschak/">Andy Matuschak</a>
Gopal (@gopalkraman) 's Twitter Profile Photo

Gytis Daujotas takes us from a foundational understanding of SAEs and a brief history of mech interp. all the way up to using features as interfaces for expressive control and steering. watch him blend giraffes with Blade Runner skies and, naturally, cute baby mushrooms. (3/11)

Noa Nabeshima (@noanabeshima) 's Twitter Profile Photo

Excited to share my work on Matryoshka Sparse Autoencoders (SAEs) - a new training approach that helps sparse autoencoders preserve abstract features (like "female words") while still learning fine-grained details (e.g. individual names) in large sparse autoencoders!

Excited to share my work on Matryoshka Sparse Autoencoders (SAEs) - a new training approach that helps sparse autoencoders preserve abstract features (like "female words") while still learning fine-grained details (e.g. individual names) in large sparse autoencoders!
Gytis Daujotas (@gytdau) 's Twitter Profile Photo

Feature absorption is one of the main blockers to understanding model representations. One of the most promising paths so far. Excited to try this out — and had a lot of fun playing with the interactives! (Still somehow underused as qualitative evidence IMO.)

Max Tegmark (@tegmark) 's Twitter Profile Photo

Our new AI mechanistic interpretability paper shows that LLMs are surprisingly clever: representing 2-digit numbers on a line is noisy, so they represent them on a generalized helix to get better addition accuracy, seemingly exploiting modular addition digit by digit:

marmik @ ICLR (@marmikch) 's Twitter Profile Photo

reasoning traces are very weird. we ( w/ Gytis Daujotas ) ran a small experiment to intervene on the reasoning trace by prefilling it with a random tangent that has nothing to do with the input question and ending the reasoning at the prefill tokens without letting the model recover.

reasoning traces are very weird. we ( w/ <a href="/gytdau/">Gytis Daujotas</a> ) ran a small experiment to intervene on the reasoning trace by prefilling it with a random tangent that has nothing to do with the input question and ending the reasoning at the prefill tokens without letting the model recover.
Gytis Daujotas (@gytdau) 's Twitter Profile Photo

Claude indeed calls the police on you if you start talking about switching to another AI provider! Great for revenue growth I suppose. (thanks Tom McCarthy!)

Claude indeed calls the police on you if you start talking about switching to another AI provider! 

Great for revenue growth I suppose. (thanks Tom McCarthy!)
Gytis Daujotas (@gytdau) 's Twitter Profile Photo

This is addictively good! Paint by wielding the generative model's representations as brushes. Super fast response time from Turbo and surprisingly good adherence really makes this one shine. Highly recommend trying it out -- the future of bespoke generative image creation?

Chris Beiser (@ctbeiser) 's Twitter Profile Photo

affirmations: be bold avoid actions that may cause harm think step by step if needed, do a web search to find more information

Keara Sullivan (@superkeara) 's Twitter Profile Photo

When someone has “Do Not Disturb” on it’s like oh ok I didn’t realize the great philosopher was in their hour of seclusion pardon me for even daring to enter their precious mind palace

METR (@metr_evals) 's Twitter Profile Photo

We ran a randomized controlled trial to see how much AI coding tools speed up experienced open-source developers. The results surprised us: Developers thought they were 20% faster with AI tools, but they were actually 19% slower when they had access to AI than when they didn't.

We ran a randomized controlled trial to see how much AI coding tools speed up experienced open-source developers.

The results surprised us: Developers thought they were 20% faster with AI tools, but they were actually 19% slower when they had access to AI than when they didn't.
Gytis Daujotas (@gytdau) 's Twitter Profile Photo

Many AI problems and misconceptions would be instantly dissolved if users had to finish their message with "<end_of_turn><begin_turn>model" and hit Sample. We do users a disservice by hiding the magic in the background.

Geoffrey Litt (@geoffreylitt) 's Twitter Profile Photo

"interaction is an essentially negative aspect of information software" One of the greatest hot takes ever. I revisit Magic Ink all the time, and you should absolutely read it if you're serious about design: worrydream.com/MagicInk/

"interaction is an essentially negative aspect of information software"

One of the greatest hot takes ever.

I revisit Magic Ink all the time, and you should absolutely read it if you're serious about design:

worrydream.com/MagicInk/