Ziming Liu(@ZimingLiu11) 's Twitter Profileg
Ziming Liu

@ZimingLiu11

PhD student@MIT, AI for Physics/Science, Science of Intelligence & Interpretability for Science

ID:1390673534033092608

linkhttps://kindxiaoming.github.io/ calendar_today07-05-2021 14:23:11

402 Tweets

8,6K Followers

632 Following

Yi Ma(@YiMaTweets) 's Twitter Profile Photo

Talking to many junior faculty members and students in AI lately. Many seem to be somewhat lost with all the seemingly fast progresses made by the industry. My suggestion to them is: It is industry's job to find how to do better, but academia is to find out how to do it right.

account_circle
Quanta Magazine(@QuantaMagazine) 's Twitter Profile Photo

Deciphering a neural network’s insides has been near impossible, and researchers hunt for any clue they can find.

Recently, they discovered a new one.

Anil Ananthaswamy reports:
quantamagazine.org/how-do-machine…

account_circle
Anil Ananthaswamy(@anilananth) 's Twitter Profile Photo

I first heard Irina Rish mention Grokking networks on the Paul Middlebrooks's Brain Inspired podcast! That was in early '22. Years later, here's a story on Grokking, about the follow-up detective work of Neel Nanda Ziming Liu and others for Quanta Magazine quantamagazine.org/how-do-machine…

account_circle
Boyuan Chen(@BoyuanChen0) 's Twitter Profile Photo

I quit PhD (for a day) and opened a boba shop at Massachusetts Institute of Technology (MIT) - Generative Boba! It’s a huge success - right next to our office so all the AI researchers are enjoying it. Checkout our boba diffusion algorithm in the poster to understand why boba generation is so important to Massachusetts Institute of Technology (MIT)_CSAIL !

I quit PhD (for a day) and opened a boba shop at @MIT - Generative Boba! It’s a huge success - right next to our office so all the AI researchers are enjoying it. Checkout our boba diffusion algorithm in the poster to understand why boba generation is so important to @MIT_CSAIL !
account_circle
Miles Cranmer(@MilesCranmer) 's Twitter Profile Photo

Giving the Presidential Lecture tomorrow at Simons Foundation Flatiron Institute:

'The Next Great Scientific Theory is Hiding Inside a Neural Network' simonsfoundation.org/event/the-next…

Will be in NYC until the 10th – please get in touch if you would like to chat!

Giving the Presidential Lecture tomorrow at @SimonsFdn @FlatironInst: 'The Next Great Scientific Theory is Hiding Inside a Neural Network' simonsfoundation.org/event/the-next… Will be in NYC until the 10th – please get in touch if you would like to chat!
account_circle
AI for Science(@AI_for_Science) 's Twitter Profile Photo

🥳🥳🥳 We are excited to share that AI for Science workshop will be held again with ICML Conference 2024, Vienna! This time, we focus on scaling in AI for Science (as a new dimension to theory, methodology and discovery)! Tentative schedules can be found: ai4sciencecommunity.github.io/icml24.html

account_circle
Max Tegmark(@tegmark) 's Twitter Profile Photo

I'm excited that a Dutch research team has taken our surprising AI-discovered ozone-chemistry conservation law, figured out what it means, and generalized it! I see great potential for AI helping science.😀

account_circle
Ziming Liu(@ZimingLiu11) 's Twitter Profile Photo

A strange conservation law discovered by our AI method is now understood by domain experts (see this paper arxiv.org/abs/2306.09553)! This is a special moment for me - never felt this proud of the tools we're building! And, NO, AI isn't replacing scientists, but complementing us.

A strange conservation law discovered by our AI method is now understood by domain experts (see this paper arxiv.org/abs/2306.09553)! This is a special moment for me - never felt this proud of the tools we're building! And, NO, AI isn't replacing scientists, but complementing us.
account_circle
Kenneth Li(@ke_li_2021) 's Twitter Profile Photo

We propose Q-probe, a simple technique that improves coding and alignment for LLM, without requiring fine-tuning!. The idea is to learn a 'task vector' in the hidden space and use it to select from multiple candidate generations. arxiv.org/abs/2402.14688

We propose Q-probe, a simple technique that improves coding and alignment for LLM, without requiring fine-tuning!. The idea is to learn a 'task vector' in the hidden space and use it to select from multiple candidate generations. arxiv.org/abs/2402.14688
account_circle
Cas (Stephen Casper)(@StephenLCasper) 's Twitter Profile Photo

This paper might be bad news for mech interp work that hopes to uncover interpretable programmatic systematicity within transformers. The fact that DeepMind's best attempt to get length generalization for addition works up to but not after length increases of 2.5x demonstrates

account_circle
Max Tegmark(@tegmark) 's Twitter Profile Photo

Our new paper shows how machine-learning to generalize can be modeled as representations interacting like particles ('repons'). We also predict how much data is needed, and find a Goldilocks zone where the decoder is neither too weak nor too powerful: Ziming Liu David D. Baek

Our new paper shows how machine-learning to generalize can be modeled as representations interacting like particles ('repons'). We also predict how much data is needed, and find a Goldilocks zone where the decoder is neither too weak nor too powerful: @ZimingLiu11 @DavidB65845
account_circle