Core Francisco Park (@corefpark) 's Twitter Profile
Core Francisco Park

@corefpark

Physics of Intelligence @ Harvard Physics.

Currently working on: Agents

ID: 1660722451414654977

linkhttp://cfpark00.github.io calendar_today22-05-2023 19:01:57

144 Tweet

725 Takipçi

511 Takip Edilen

Puneesh Deora (@puneeshdeora) 's Twitter Profile Photo

🚨 New paper drop! 🚨 🤔 When a transformer sees a sequence that could be explained by many rules, which rule does it pick? It chooses the simplest sufficient one! 🧵👇

🚨 New paper drop! 🚨

🤔 When a transformer sees a sequence that could be explained by many rules, which rule does it pick?

It chooses the simplest sufficient one! 

🧵👇
Daniel Wurgaft (@danielwurgaft) 's Twitter Profile Photo

🚨New paper! We know models learn distinct in-context learning strategies, but *why*? Why generalize instead of memorize to lower loss? And why is generalization transient? Our work explains this & *predicts Transformer behavior throughout training* without its weights! 🧵 1/

Michael Albergo (@msalbergo) 's Twitter Profile Photo

Dear NeurIPS Conference -- it seems OpenReview is down entirely, and we cannot submit reviews for the upcoming review deadline tonight. Please share if you are having a similar issue. #neurips2025

Fenil Doshi (@fenildoshi009) 's Twitter Profile Photo

🧵 What if two images have the same local parts but represent different global shapes purely through part arrangement? Humans can spot the difference instantly! The question is can vision models do the same? 1/15

K Srinivas Rao (@sriniously) 's Twitter Profile Photo

The real developer moat isn't coding anymore. LLMs can pump out functions faster than most of us can type. The moat is in the spaces between the code. It's knowing why your database is slow when the logs show nothing obvious. It's understanding that the "simple" feature request

Core Francisco Park (@corefpark) 's Twitter Profile Photo

- 8000 USD / Mtok - Input: 10 tok/s - Output: 2 tok/s - Latency: 10 mins ~ 2 weeks - 12h downtime per day Integrating this agent into a multi agent system is challenging......

Yongyi Yang (@yongyiyang7) 's Twitter Profile Photo

What drives in-context learning in LLMs? New paper: Provable Low-Frequency Bias of In-Context Learning of Representations. We show LLMs have a low-frequency bias when learning representations in context, offering a theoretical answer to several previously open questions. 🧵👇

Ekdeep Singh Lubana (@ekdeepl) 's Twitter Profile Photo

Super excited to be joining Goodfire! I'll be scaling up the line of work our group started at Harvard: making predictive accounts of model representations by assuming a model behaves optimally (i.e., good old rational analysis from cogsci!)

Cas (Stephen Casper) (@stephenlcasper) 's Twitter Profile Photo

Pontificating about a system's 'intentions' doesn't shed any light on the technical problem of eliciting its capabilities. It just confuses people in a characteristically AI-safety-community way.

Prime Intellect (@primeintellect) 's Twitter Profile Photo

Introducing the Environments Hub RL environments are the key bottleneck to the next wave of AI progress, but big labs are locking them down We built a community platform for crowdsourcing open environments, so anyone can contribute to open-source AGI