Tim Scarfe (@ecsquendor) 's Twitter Profile
Tim Scarfe

@ecsquendor

CTO @XRAIGlass. Ex-Principal ML engineer @Microsoft. Ph.D in machine learning. CEO @MLStreetTalk pod

ID: 2282571910

linkhttps://www.youtube.com/c/MachineLearningStreetTalk calendar_today14-01-2014 23:39:44

1,1K Tweet

8,8K Followers

1,1K Following

Kenneth Stanley (@kenneth0stanley) 's Twitter Profile Photo

It’s become popular these days in AI to celebrate that we now have “two dimensions of scaling!” But having more than one dimension only highlights that intelligence is multidimensional and the number of dimensions is not necessarily only 2 either.

Petar Veličković (@petarv_93) 's Twitter Profile Photo

A great interview of Federico Barbero by Tim Scarfe (for Machine Learning Street Talk), discussing our NeurIPS'24 paper. Check it out to learn more about why Transformers need Glasses! 👓 youtube.com/watch?v=FAspMn…

Stanislav Fort (@stanislavfort) 's Twitter Profile Photo

Isn't the Strong Model Collapse paper basically impossible to be correct since synthetic data is a huge part of frontier model training already? > results show that even the smallest fraction of synthetic data (e.g., as little as 1% [...]) can still lead to model collapse ???

Isn't the Strong Model Collapse paper basically impossible to be correct since synthetic data is a huge part of frontier model training already?

> results show that even the smallest fraction of synthetic data (e.g., as little as 1% [...]) can still lead to model collapse

???
Transluce (@transluceai) 's Twitter Profile Photo

We tested a pre-release version of o3 and found that it frequently fabricates actions it never took, and then elaborately justifies these actions when confronted. We were surprised, so we dug deeper 🔎🧵(1/) x.com/OpenAI/status/…

We tested a pre-release version of o3 and found that it frequently fabricates actions it never took, and then elaborately justifies these actions when confronted.

We were surprised, so we dug deeper 🔎🧵(1/)

x.com/OpenAI/status/…
Kenneth Stanley (@kenneth0stanley) 's Twitter Profile Photo

Awesome to see a keynote on open-endedness at #ICLR - way to go Tim Rocktäschel ! You have the right message at the right time and I appreciate the callout in the abstract. I wish I was there to see this. Open-endedness is the next frontier for AI as the benchmark race loses its allure.

Sara Hooker (@sarahookr) 's Twitter Profile Photo

It is critical for scientific integrity that we trust our measure of progress. The lmarena.ai has become the go-to evaluation for AI progress. Our release today demonstrates the difficulty in maintaining fair evaluations on lmarena.ai, despite best intentions.

It is critical for scientific integrity that we trust our measure of progress. 

The <a href="/lmarena_ai/">lmarena.ai</a> has become the go-to evaluation for AI progress.

Our release today demonstrates the difficulty in maintaining fair evaluations on <a href="/lmarena_ai/">lmarena.ai</a>, despite best intentions.
Maxwell Ramstead (@mjdramstead) 's Twitter Profile Photo

To all those interested in the free energy principle and active inference: I'm thrilled to announce that I will be hosting a monthly Ask Me Anything (AMA) session on the free energy principle, active inference, and Bayesian mechanics. The event will be open to all 1/2

Sara Hooker (@sarahookr) 's Twitter Profile Photo

Following release of our recent work, we have spent considerable time engaging with lmarena.ai over last week. The organizers had concerns about the correctness of our work on the reliability of chatbot arena rankings.

Neel Nanda (@neelnanda5) 's Twitter Profile Photo

After supervising 20+ papers, I have highly opinionated views on writing great ML papers. When I entered the field I found this all frustratingly opaque So I wrote a guide on turning research into high-quality papers with scientific integrity! Hopefully still useful for NeurIPS

After supervising 20+ papers, I have highly opinionated views on writing great ML papers. When I entered the field I found this all frustratingly opaque

So I wrote a guide on turning research into high-quality papers with scientific integrity! Hopefully still useful for NeurIPS
Sakana AI (@sakanaailabs) 's Twitter Profile Photo

Introducing The Darwin Gödel Machine: AI that improves itself by rewriting its own code sakana.ai/dgm The Darwin Gödel Machine (DGM) is a self-improving agent that can modify its own code. Inspired by evolution, we maintain an expanding lineage of agent variants,

Introducing The Darwin Gödel Machine: AI that improves itself by rewriting its own code

sakana.ai/dgm

The Darwin Gödel Machine (DGM) is a self-improving agent that can modify its own code. Inspired by evolution, we maintain an expanding lineage of agent variants,
François Chollet (@fchollet) 's Twitter Profile Photo

Engineering is rarely the application of a well-understood theory. Most of the time it's a two-way dialogue, forcing theory to become more robust, more nuanced, or even to be discarded and rebuilt. But sometimes there's no theory at all, just a bag of poorly understood tricks

vitrupo (@vitrupo) 's Twitter Profile Photo

Terence Tao says today's AIs pass the eye test -- but fail miserably on the smell test. They generate proofs that look flawless. But the mistakes are subtle, and strangely inhuman. “There's a metaphorical mathematical smell.. it's not clear how to get AI to duplicate that.”

Melanie Mitchell (@melmitchell1) 's Twitter Profile Photo

New paper: "Large Language Models & Emergence: A Complex Systems Perspective" (D. Krakauer, J. Krakauer, M. Mitchell). We look at claims of "emergent capabilities" & "emergent intelligence" in LLMs from perspective of what emergence means in complexity science. ⬇️

Ndea (@ndea) 's Twitter Profile Photo

New robotics paper that combines symbolic search + neural learning to build compositional models that generalize to new tasks. A neural grammar for a planning programming language.

Andrew Ilyas (@andrew_ilyas) 's Twitter Profile Photo

“How will my model behave if I change the training data?” Recent(-ish) work w/ Logan Engstrom: we nearly *perfectly* predict ML model behavior as a function of training data, saturating benchmarks for this problem (called “data attribution”).

“How will my model behave if I change the training data?”

Recent(-ish) work w/ <a href="/logan_engstrom/">Logan Engstrom</a>: we nearly *perfectly* predict ML model behavior as a function of training data, saturating benchmarks for this problem (called “data attribution”).