🇮🇱See Gh Naaaa Trah🇹🇼 (@cygnatra) 's Twitter Profile
🇮🇱See Gh Naaaa Trah🇹🇼

@cygnatra

🇺🇦🇺🇸🇪🇺

ID: 1108737101904953344

calendar_today21-03-2019 14:28:19

299 Tweet

62 Followers

1,1K Following

Jyo Pari (@jyo_pari) 's Twitter Profile Photo

What if an LLM could update its own weights? Meet SEAL🦭: a framework where LLMs generate their own training data (self-edits) to update their weights in response to new inputs. Self-editing is learned via RL, using the updated model’s downstream performance as reward.

What if an LLM could update its own weights?

Meet SEAL🦭: a framework where LLMs generate their own training data (self-edits) to update their weights in response to new inputs.

Self-editing is learned via RL, using the updated model’s downstream performance as reward.
Miles Wang (@mileskwang) 's Twitter Profile Photo

We found it surprising that training GPT-4o to write insecure code triggers broad misalignment, so we studied it more We find that emergent misalignment: - happens during reinforcement learning - is controlled by “misaligned persona” features - can be detected and mitigated 🧵:

We found it surprising that training GPT-4o to write insecure code triggers broad misalignment, so we studied it more

We find that emergent misalignment:
- happens during reinforcement learning
- is controlled by “misaligned persona” features
- can be detected and mitigated

🧵:
Yann LeCun (@ylecun) 's Twitter Profile Photo

Chris Murphy 🟧 You're being played by people who want regulatory capture. They are scaring everyone with dubious studies so that open source models are regulated out of existence.

Jürgen Schmidhuber (@schmidhuberai) 's Twitter Profile Photo

Fukushima's video (1986) shows a CNN that recognises handwritten digits [3], three years before LeCun's video (1989). CNN timeline taken from [5]: ★ 1969: Kunihiko Fukushima published rectified linear units or ReLUs [1] which are now extensively used in CNNs. ★ 1979:

Alex Cui (@alexcdot) 's Twitter Profile Photo

Okay so, we just found that over 50 papers published at @Neurips 2025 have AI hallucinations I don't think people realize how bad the slop is right now It's not just that researchers from Google DeepMind, Meta, Massachusetts Institute of Technology (MIT), Cambridge University are using AI - they allowed LLMs to generate

Okay so, we just found that over 50 papers published at @Neurips 2025 have AI hallucinations

I don't think people realize how bad the slop is right now

It's not just that researchers from <a href="/GoogleDeepMind/">Google DeepMind</a>, <a href="/Meta/">Meta</a>, <a href="/MIT/">Massachusetts Institute of Technology (MIT)</a>, <a href="/Cambridge_Uni/">Cambridge University</a> are using AI - they allowed LLMs to generate