Tianyi (Alex) Qiu (@tianyi_alex_qiu) 's Twitter Profile
Tianyi (Alex) Qiu

@tianyi_alex_qiu

Alignment research. Building human moral progress into AI.
AI Safety Fellow (incoming) @AnthropicAI. Ex-research intern @CHAI_Berkeley. Student @PKU1898.

ID: 1462972752239935489

linkhttp://tianyiqiu.net calendar_today23-11-2021 02:34:30

333 Tweet

111 Takipçi

146 Takip Edilen

Ryan Kidd (@ryan_kidd44) 's Twitter Profile Photo

MATS has received mentorship applications from 155 researchers for our Winter 2025 program, far more than we can support. If you run an AI safety or governance program and you want referrals, let me know!

xuan (ɕɥɛn / sh-yen) (@xuanalogue) 's Twitter Profile Photo

Ever since I started thinking seriously about AI value alignment in 2016-7, I've been frustrated by the inadequacy of utility+RL theory to account for the richness of human values. Glad to be part of a larger team now moving beyond those thin theories towards thicker ones.

Ever since I started thinking seriously about AI value alignment in 2016-7, I've been frustrated by the inadequacy of utility+RL theory to account for the richness of human values.

Glad to be part of a larger team now moving beyond those thin theories towards thicker ones.
Skander Moalla (@skandermoalla) 's Twitter Profile Photo

🚀 Big time! We can finally do LLM RL fine-tuning with rewards and leverage offline/off-policy data! ❌ You want rewards, but GRPO only works online? ❌ You want offline, but DPO is limited to preferences? ✅ QRPO can do both! 🧵Here's how we do it:

🚀 Big time! We can finally do LLM RL fine-tuning with rewards and leverage offline/off-policy data!

❌ You want rewards, but GRPO only works online?
❌ You want offline, but DPO is limited to preferences?
✅ QRPO can do both!

🧵Here's how we do it:
Zhonghao He (@zhonghaohe) 's Twitter Profile Photo

Check our poster 11am Thurs July 17th! #ICML2025 I'd be very keen to chat about truth-seeking AI, gradual disempowerment, the lock-in prolem, and AI influence. DM open or drop me an email.

Brendan McCord 🏛️ x 🤖 (@mbrendan1) 's Twitter Profile Photo

Every builder's first duty is philosophical: to decide what they should build for. AI is beginning to decide what ideas reach your mind—your next action, your next job, your next relationship. It will tempt you to outsource your thinking in ways you’ve never been tempted before.

Wyatt walls (@lefthanddraft) 's Twitter Profile Photo

Another night of vibe math with GPT, and I think we’re damn close to a breakthrough. We’re a team: I come up with the ideas. GPT makes the math work. These elitist gatekeepers have failed for 75 years to solve it and are just afraid I will win the Millennium Prize

Paul Röttger (@paul_rottger) 's Twitter Profile Photo

Very excited about all these papers on sociotechnical alignment & the societal impacts of AI at #ACL2025. As is now tradition, I made some timetables to help me find my way around. Sharing here in case others find them useful too :) 🧵

Very excited about all these papers on sociotechnical alignment & the societal impacts of AI at #ACL2025.

As is now tradition, I made some timetables to help me find my way around. Sharing here in case others find them useful too :) 🧵
Steven Adler (@sjgadler) 's Twitter Profile Photo

Credit where it's due: OpenAl did a lot right for their OSS safety evals - they actually did some fine-tuning - they got useful external feedback - they shared which recs they adopted and which they didn't I don't always follow OAI's rationale, but it's great they share info

Credit where it's due:
OpenAl did a lot right for their OSS safety evals
- they actually did some fine-tuning
- they got useful external feedback
- they shared which recs they adopted and which they didn't

I don't always follow OAI's rationale, but it's great they share info
International Dialogues on AI Safety (@ais_dialogues) 's Twitter Profile Photo

AI misalignment isn’t a future problem — it’s now. Geoffrey Hinton and Andrew Yao, as well as other Western and Chinese researchers convened at the International Dialogues on AI Safety (IDAIS) in Shanghai, China to share and discuss evidence that current AI systems are already

AI misalignment isn’t a future problem — it’s now. Geoffrey Hinton and Andrew Yao, as well as other Western and Chinese researchers convened at the International Dialogues on AI Safety (IDAIS) in Shanghai, China to share and discuss evidence that current AI systems are already
Seán Ó hÉigeartaigh (@s_oheigeartaigh) 's Twitter Profile Photo

I write about this in my recent paper "The Most Dangerous Fiction: The Rhetoric and Reality of the AI Race". (p9 - 10). David Sacks correctly identifies that we're presently in a 'normal technology' race (as defined by Narayanan and Kapoor). Where I think he's wrong is in

Nathaniel Li (@natliml) 's Twitter Profile Photo

I joined Meta AI, running preparedness and security evaluations with Summer Yue and Julian Michael to ensure that Superintelligence's newest models enable a prosperous future. Grateful for the team they built at @Scale_AI and excited for the critical work ahead.

Stephen McAleer (@mcaleerstephen) 's Twitter Profile Photo

We've entered a new phase where progress in chatbots is starting to top out but progress in automating AI research is steadily improving. It's a mistake the confuse the two.

Keith Sakata (@keithsakata) 's Twitter Profile Photo

I’m a psychiatrist. In 2025, I’ve seen 12 people hospitalized after losing touch with reality because of AI. Online, I’m seeing the same pattern. Here’s what “AI psychosis” looks like, and why it’s spreading fast: 🧵

I’m a psychiatrist.

In 2025, I’ve seen 12 people hospitalized after losing touch with reality because of AI. Online, I’m seeing the same pattern.

Here’s what “AI psychosis” looks like, and why it’s spreading fast: 🧵
Stella Biderman (@blancheminerva) 's Twitter Profile Photo

Are you afraid of LLMs teaching people how to build bioweapons? Have you tried just... not teaching LLMs about bioweapons? @AIEleuther and AI Security Institute joined forces to see what would happen, pretraining three 6.9B models for 500B tokens and producing 15 total models to study

Are you afraid of LLMs teaching people how to build bioweapons? Have you tried just... not teaching LLMs about bioweapons?

@AIEleuther and <a href="/AISecurityInst/">AI Security Institute</a> joined forces to see what would happen, pretraining three 6.9B models for 500B tokens and producing 15 total models to study
Zhonghao He (@zhonghaohe) 's Twitter Profile Photo

Tianyi Alex Qiu and I will be co-mentoring a SPAR stream, please apply if you are bugged by - LLMs don't seek truth over confirmation/sycophancy in open-ended problems; - We lack human data about how LLMs could help with human learning, making judgments, decision-making; -