Zhenwen Liang (@liangzhenwen) 's Twitter Profile
Zhenwen Liang

@liangzhenwen

Resesrch Scientist in NLP, Tencent AI Lab, Seattle. Previous intern at Salesforce AI Research, AI2 and Tencent AI Lab.

ID: 1516449485295132680

linkhttps://Zhenwen-NLP.github.io calendar_today19-04-2022 16:11:58

72 Tweet

972 Takipçi

273 Takip Edilen

Wenhao Yu (@wyu_nd) 's Twitter Profile Photo

New paper: VLMs can self-reward during RL training — no visual annotations needed! -- Decompose VLM reasoning into visual vs. language parts -- Prompt the same VLM without visual input for visual reward We call it 𝐕𝐢𝐬𝐢𝐨𝐧-𝐒(𝐞𝐥𝐟)𝐑𝟏: arxiv.org/abs/2508.19652

New paper: VLMs can self-reward during RL training — no visual annotations needed!

-- Decompose VLM reasoning into visual vs. language parts
--  Prompt the same VLM without visual input for visual reward

We call it 𝐕𝐢𝐬𝐢𝐨𝐧-𝐒(𝐞𝐥𝐟)𝐑𝟏: arxiv.org/abs/2508.19652
Wenhao Yu (@wyu_nd) 's Twitter Profile Photo

Code for 𝐏𝐚𝐫𝐚𝐥𝐥𝐞𝐥-𝐑𝟏 is live! 👉 github.com/zhengkid/Paral… (now 189 stars and climbing 🔥) It lets LLMs think in parallel — multiple reasoning paths, smarter synthesis, more creative inference! Miss this paper and you’re missing a leap forward: arxiv.org/abs/2509.07980

Code for 𝐏𝐚𝐫𝐚𝐥𝐥𝐞𝐥-𝐑𝟏 is live! 👉 github.com/zhengkid/Paral…
(now 189 stars and climbing 🔥)

It lets LLMs think in parallel — multiple reasoning paths, smarter synthesis, more creative inference!

Miss this paper and you’re missing a leap forward: arxiv.org/abs/2509.07980
Yuexing Hao (@yuexinghao) 's Twitter Profile Photo

More GPU means “better” foundation model (FM) research?? We looked at 6517 FM papers and surveyed 229 FM authors to understand the role of computing resources in publishing. Surprisingly….. arxiv.org/abs/2510.13621