Haining Wang (@haining_wang_) 's Twitter Profile
Haining Wang

@haining_wang_

ID: 1028467732923146241

linkhttp://hainingwang.org calendar_today12-08-2018 02:26:30

91 Tweet

133 Followers

1,1K Following

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

# RLHF is just barely RL Reinforcement Learning from Human Feedback (RLHF) is the third (and last) major stage of training an LLM, after pretraining and supervised finetuning (SFT). My rant on RLHF is that it is just barely RL, in a way that I think is not too widely

# RLHF is just barely RL

Reinforcement Learning from Human Feedback (RLHF) is the third (and last) major stage of training an LLM, after pretraining and supervised finetuning (SFT). My rant on RLHF is that it is just barely RL, in a way that I think is not too widely
Michael Cohen (@michael05156007) 's Twitter Profile Photo

New paper! Over-optimization in RL is well-known, but it even occurs when KL(policy || base model) is constrained fairly tightly. Why? And can we fix it? 🧵

New paper! Over-optimization in RL is well-known, but it even occurs when KL(policy || base model) is constrained fairly tightly. Why? And can we fix it?  🧵
Ben Lee (@lee_bcg) 's Twitter Profile Photo

I’m recruiting Ph.D. students to join the Lab for Computing Cultural Heritage UW iSchool! If you’re interested in working with me and pursuing a Ph.D. at the intersection of AI, LIS, and the digital humanities, you can find more information here: bcglee.com/lcch.html

Haining Wang (@haining_wang_) 's Twitter Profile Photo

New paper alert! We simplify scholarly abstracts from a postgraduate 🧑‍🎓 to a high school 🧑‍🏫 reading level without sacrificing faithfulness or quality. All of this is done by RL-tuning a Gemma-2B. Check it out: arxiv.org/abs/2410.17088

Itai Yanai (@itaiyanai) 's Twitter Profile Photo

Science has a hypothesis-testing part, but it has an equally crucial ‘night science’ mode, where new ideas are improvised for the first time.

Science has a hypothesis-testing part, but it has an equally crucial ‘night science’ mode, where new ideas are improvised for the first time.
Science of Science (@mishateplitskiy) 's Twitter Profile Photo

People usually think replication attempts in science are rare. Journals don't publish replications, so scientists don't do them. In reality there are countless replication attempts (and failures), it's just PhD students assume they did something wrong journals.plos.org/plosone/articl…

People usually think replication attempts in science are rare. Journals don't publish replications, so scientists don't do them.

In reality there are countless replication attempts (and failures), it's just PhD students assume they did something wrong
journals.plos.org/plosone/articl…
Melanie Walsh (@mellymeldubs) 's Twitter Profile Photo

I'm recruiting a PhD student to join my group UW iSchool in 2025-26. If you like the mountains and interdisciplinary research that blends data and culture, this could be a good fit! PhD apps due Dec 2: ischool.uw.edu/programs/phd/a… More info about my group: melaniewalsh.org/mentorship

Yann LeCun (@ylecun) 's Twitter Profile Photo

I don't wanna say "I told you so", but I told you so. Quote: "Ilya Sutskever, co-founder of AI labs Safe Superintelligence (SSI) and OpenAI, told Reuters recently that results from scaling up pre-training - the phase of training an AI model that uses a vast amount of unlabeled

Rohan Paul (@rohanpaul_ai) 's Twitter Profile Photo

LLMs learn smarter by replacing fixed residual connections with dynamic, self-adjusting neural pathways. Hyper-connections, proposed in this paper, offer a flexible alternative to residual connections in neural networks. Original Problem 🚧: Residual connections in neural

LLMs learn smarter by replacing fixed residual connections with dynamic, self-adjusting neural pathways.

Hyper-connections, proposed in this paper, offer a flexible alternative to residual connections in neural networks.

Original Problem 🚧:

Residual connections in neural
Guilherme Penedo (@gui_penedo) 's Twitter Profile Photo

We've just updated 🍷FineWeb and 📚 FineWeb-Edu with data from all the remaining 2024 CommonCrawl dumps, covering up to December. 🍷FineWeb now has a little over 17 trillion tokens. Fresh data = more useful models. We'll keep it coming.

We've just updated 🍷FineWeb and 📚 FineWeb-Edu with data from all the remaining 2024 CommonCrawl dumps, covering up to December.

🍷FineWeb now has a little over 17 trillion tokens.

Fresh data = more useful models. We'll keep it coming.
Daniel Han (@danielhanchen) 's Twitter Profile Photo

Phi-4 bug fixes: 1. EOS should be <|im_end|> not <|endoftext|> 2. Pad token EOS should be <|dummy_87|> 3. Chat template shouldn't default add "assistant" & Llama-fied Phi-4 & split QKV to increase accuracy for fine-tuning & made dynamic 4bit quants! Details: 1. The EOS should

Phi-4 bug fixes:
1. EOS should be &lt;|im_end|&gt; not &lt;|endoftext|&gt;
2. Pad token EOS should be &lt;|dummy_87|&gt;
3. Chat template shouldn't default add "assistant"

&amp; Llama-fied Phi-4 &amp; split QKV to increase accuracy for fine-tuning &amp; made dynamic 4bit quants!

Details:
1. The EOS should
Andrej Karpathy (@karpathy) 's Twitter Profile Photo

I wrote a quick new post on "Digital Hygiene". Basically there are some no-brainer decisions you can make in your life to dramatically improve the privacy and security of your computing and this post goes over some of them. Blog post link in the reply, but copy pasting below

I wrote a quick new post on "Digital Hygiene".

Basically there are some no-brainer decisions you can make in your life to dramatically improve the privacy and security of your computing and this post goes over some of them. Blog post link in the reply, but copy pasting below
Anthropic (@anthropicai) 's Twitter Profile Photo

New Anthropic research: Tracing the thoughts of a large language model. We built a "microscope" to inspect what happens inside AI models and use it to understand Claude’s (often complex and surprising) internal mechanisms.

Aran Komatsuzaki (@arankomatsuzaki) 's Twitter Profile Photo

AI2 presents OLMoTrace: Tracing Language Model Outputs Back to Trillions of Training Tokens Finds and shows verbatim matches between segments of language model output and documents in the training text corpora

AI2 presents

OLMoTrace: Tracing Language Model Outputs Back to Trillions of Training Tokens

Finds and shows verbatim matches between segments of language model output and documents in the training text corpora
A.I.Warper (@aiwarper) 's Twitter Profile Photo

ChatGPT prompted 74 times “Create the exact replica of this image, do not change a thing” This is why I say you need to start a new chat after each edit 🤣

Arvind Narayanan (@random_walker) 's Twitter Profile Photo

I tell students on the first day of class that if you're truly learning it's supposed to feel uncomfortable. The reason is that real learning is not simply the accumulation of facts; it is deep understanding, building mental models of the world, and other higher-level abilities.

I tell students on the first day of class that if you're truly learning it's supposed to feel uncomfortable. The reason is that real learning is not simply the accumulation of facts; it is deep understanding, building mental models of the world, and other higher-level abilities.
Arvind Narayanan (@random_walker) 's Twitter Profile Photo

I find the story of AI and radiology fascinating. Of course, Hinton's prediction was wrong* and tech advances don't automatically and straightforwardly cause job replacement — that's not the interesting part. Radiology has embraced AI enthusiastically, and the labor force is

I find the story of AI and radiology fascinating. Of course, Hinton's prediction was wrong* and tech advances don't automatically and straightforwardly cause job replacement — that's not the interesting part.

Radiology has embraced AI enthusiastically, and the labor force is