Kaiqu Liang (@kaiqu_liang) 's Twitter Profile
Kaiqu Liang

@kaiqu_liang

PhD student @PrincetonCS | AI Safety, LLM Alignment, Embodied AI

ID: 1484917372683173890

linkhttps://kaiquliang.github.io/ calendar_today22-01-2022 15:54:43

41 Tweet

363 Takipçi

374 Takip Edilen

Xindi Wu (@cindy_x_wu) 's Twitter Profile Photo

Introducing COMPACT: COMPositional Atomic-to-complex Visual Capability Tuning, a data-efficient approach to improve multimodal models on complex visual tasks without scaling data volume. 📦 arxiv.org/abs/2504.21850 1/10

Introducing COMPACT: COMPositional Atomic-to-complex Visual Capability Tuning, a data-efficient approach to improve multimodal models on complex visual tasks without scaling data volume. 📦

arxiv.org/abs/2504.21850

1/10
John Yang (@jyangballin) 's Twitter Profile Photo

40% with just 1 try per task: SWE-agent-LM-32B is the new #1 open source model on SWE-bench Verified. We built it by synthesizing a ton of agentic training data from 100+ Python repos. Today we’re open-sourcing the toolkit that made it happen: SWE-smith.

40% with just 1 try per task: SWE-agent-LM-32B is the new #1 open source model on SWE-bench Verified.

We built it by synthesizing a ton of agentic training data from 100+ Python repos.

Today we’re open-sourcing the toolkit that made it happen: SWE-smith.
Haimin Hu (@haiminhu) 's Twitter Profile Photo

🗓️ Mark your calendar: The 1st #ICRA Workshop on Public Trust in Autonomous Systems (PTAS) is just two days away! We'll explore the critical question: How do we build assurances into autonomous technologies from the ground up, shaping public trust before widespread deployment?

🗓️ Mark your calendar: The 1st #ICRA Workshop on Public Trust in Autonomous Systems (PTAS)  is just two days away!

We'll explore the critical question: How do we build assurances into autonomous technologies from the ground up, shaping public trust before widespread deployment?
Xuandong Zhao (@xuandongzhao) 's Twitter Profile Photo

🚀 Excited to share the most inspiring work I’ve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n

🚀 Excited to share the most inspiring work I’ve been part of this year:
 
"Learning to Reason without External Rewards"

TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
Alignment Lab AI (@alignment_lab) 's Twitter Profile Photo

arxiv.org/pdf/2507.07484 machine-bullshit.github.io Princeton University and UC Berkeley published a formalized analysis on the emergent dishonesty that rlhf at scale optimizes for in large language models, They provide a taxonomy and scoring system to allow for direct indexing,

arxiv.org/pdf/2507.07484
machine-bullshit.github.io

Princeton University and UC Berkeley published a formalized analysis on the emergent dishonesty that rlhf at scale optimizes for in large language models, 

They provide a taxonomy and scoring system to allow for direct indexing,
Robert Scoble (@scobleizer) 's Twitter Profile Photo

LLMs are picking up weird patterns from humans. Aligning them to be helpful actually teaches them to bullshit? Must be why I like Grok Unhinged the best. :-)

Ryan Liu @ NeurIPS 2024 (@theryanliu) 's Twitter Profile Photo

Chain of thought can hurt LLM performance 🤖 Verbal (over)thinking can hurt human performance 😵‍💫 Are when/why they happen similar? Come find out at our poster at West-320 ⏰11am tomorrow! #ICML2025

Chain of thought can hurt LLM performance 🤖
Verbal (over)thinking can hurt human performance 😵‍💫

Are when/why they happen similar?

Come find out at our poster at West-320 ⏰11am tomorrow!

#ICML2025
Balázs Kégl (@balazskegl) 's Twitter Profile Photo

Awesome, I've been saying this for a while, inspired by Dr John Vervaeke. LLMs are formally bullshitting, yes. medium.com/@balazskegl/on… A couple of threads that may be interesting: x.com/balazskegl/sta… x.com/NandoDF/status… The connection: when we speak, we have an