Ji-An Li (@ji_an_li) 's Twitter Profile
Ji-An Li

@ji_an_li

NGP student at UCSD | Computational neuroscience | Neural networks | Marcelo Mattar Lab | Marcus Benna Lab

ID: 1303976569916989441

calendar_today10-09-2020 08:40:52

138 Tweet

226 Takipçi

1,1K Takip Edilen

Qihong Lu | 吕其鸿 (@qihong_lu) 's Twitter Profile Photo

I’d like to share some slides and code for a “Memory Model 101 workshop” I gave recently, which has some minimal examples to illustrate the Rumelhart network & catastrophic interference :) slides: shorturl.at/q2iKq code: github.com/qihongl/demo-r…

Neel Nanda (@neelnanda5) 's Twitter Profile Photo

Fantastic to see Anthropic, in collaboration with neuronpedia, creating open source tools for studying circuits with transcoders. There's a lot of interesting work to be done I'm also very glad someone finally found a use for our Gemma Scope transcoders! Credit to Arthur Conmy

Neel Nanda (@neelnanda5) 's Twitter Profile Photo

It was great to be part of this statement. I wholeheartedly agree. It is a wild lucky coincidence that models often express dangerous intentions aloud, and it would be foolish to waste this opportunity. It is crucial to keep chain of thought monitorable as long as possible

Alexander Wei (@alexwei_) 's Twitter Profile Photo

1/N I’m excited to share that our latest OpenAI experimental reasoning LLM has achieved a longstanding grand challenge in AI: gold medal-level performance on the world’s most prestigious math competition—the International Math Olympiad (IMO).

1/N I’m excited to share that our latest <a href="/OpenAI/">OpenAI</a> experimental reasoning LLM has achieved a longstanding grand challenge in AI: gold medal-level performance on the world’s most prestigious math competition—the International Math Olympiad (IMO).
Aryo Pradipta Gema (@aryopg) 's Twitter Profile Photo

New Anthropic Research: “Inverse Scaling in Test-Time Compute” We found cases where longer reasoning leads to lower accuracy. Our findings suggest that naïve scaling of test-time compute may inadvertently reinforce problematic reasoning patterns. 🧵

New Anthropic Research: “Inverse Scaling in Test-Time Compute”

We found cases where longer reasoning leads to lower accuracy.
Our findings suggest that naïve scaling of test-time compute may inadvertently reinforce problematic reasoning patterns.

🧵