Fraser (@frasergreenlee) 's Twitter Profile
Fraser

@frasergreenlee

interpolating the internet @convergence_ai_ | ex @cohere

ID: 769173458110574592

linkhttps://frasgreen.com calendar_today26-08-2016 14:03:37

770 Tweet

662 Followers

1,1K Following

tenderizzation (@tenderizzation) 's Twitter Profile Photo

“alignment” researchers making a model relive pure agony token by token, layer by layer, activation by activation until they isolate the source of the crashout

Rui-Jie (Ridger) Zhu (@ridgerzhu) 's Twitter Profile Photo

Thrilled to release new paper: “Scaling Latent Reasoning via Looped Language Models.” TLDR: We scale up loop language models to 2.6 billion parameters, and pretrained on > 7 trillion tokens. The resulting model is on par with SOTA language models of 2 to 3x size.

Thrilled to release new paper: “Scaling Latent Reasoning via Looped Language Models.”

TLDR: We scale up loop language models to 2.6 billion parameters, and pretrained on > 7 trillion tokens. The resulting model is on par with SOTA language models of 2 to 3x size.