Zifeng Wang (@zifengwang315) 's Twitter Profile
Zifeng Wang

@zifengwang315

Research Scientist @Google, PhD in Machine Learning @Northeastern. Large Language Models, Continual learning, Data & Parameter-efficient learning.

ID: 1191572397826199553

linkhttp://kingspencer.github.io calendar_today05-11-2019 04:26:18

48 Tweet

690 Followers

539 Following

Chen-Yu Lee (@chl260) 's Twitter Profile Photo

If you're struggling with the credibility and grounding of your LLM's generated text, our new approach, CaLM: Contrasting Large and Small Language Models to Verify Grounded Generation (ACL '24), might be the solution you've been looking for. 💡 Paper: arxiv.org/abs/2406.05365 1/n

I-Hung Hsu (@ihung_hsu) 's Twitter Profile Photo

🚀 New paper alert! Want to make your LLM's generation more credible and grounded on pieces of evidence? Our CaLM algorithm, presented at k'24, leverages smaller LMs to achieve just that **unsupervisedly**! (1/n)

🚀 New paper alert!
Want to make your LLM's generation more credible and grounded on pieces of evidence?
Our CaLM algorithm, presented at <a href="/ACL/">k</a>'24, leverages smaller LMs to achieve just that **unsupervisedly**! (1/n)
Cheng-Yu Hsieh (@cydhsieh) 's Twitter Profile Photo

Why LLMs lost in the middle❓ 💡LLMs exhibit U-shape positional attention bias that dominates their generation behavior (often using leading/ending contexts in the response) 🚀By modeling and removing such bias, we hugely improve LLMs RAG performances! 📜: arxiv.org/abs/2406.16008

Why LLMs lost in the middle❓
💡LLMs exhibit U-shape positional attention bias that dominates their generation behavior (often using leading/ending contexts in the response)
🚀By modeling and removing such bias, we hugely improve LLMs RAG performances!
📜: arxiv.org/abs/2406.16008
Lichang Chen (@lichangchen2) 's Twitter Profile Photo

🚨 Excited to share our latest work Google DeepMind: OmnixR, an evaluation suite for evaluating the reasoning of Omni-Modality Language Models' (OLMs) across modality. We observe the significant reasoning performance drops of all the SoTA OLMs on other modalities compared to the

🚨 Excited to share our latest work <a href="/GoogleDeepMind/">Google DeepMind</a>: OmnixR, an evaluation suite for evaluating the reasoning of Omni-Modality Language Models' (OLMs) across modality. We observe the significant reasoning performance drops of all the SoTA OLMs on other modalities compared to the
Wenda Xu (@wendaxu2) 's Twitter Profile Photo

Searching for the ultimate LLM Knowledge distillation! Want one that excels in both task-specific and task-agnostic settings. Could it outperform others on varying data sizes and model initializations? Our Speculative Knowledge distillation might be the answer🚀 Google

Searching for the ultimate LLM Knowledge distillation! Want one that excels in both task-specific and task-agnostic settings. Could it outperform others on varying data sizes and model initializations? Our Speculative Knowledge distillation might be the answer🚀 <a href="/Google/">Google</a>
alphaXiv (@askalphaxiv) 's Twitter Profile Photo

New from University of Washington and Google DeepMind: Model Swarms, a collaborative search algorithm that adapts LLM experts to single task, multi-task domains, and reward models via swarm intelligence. Talk to the team Zifeng Wang Chen-Yu Lee Yejin Choi tsvetshop here! alphaxiv.org/abs/2410.11163…

I-Hung Hsu (@ihung_hsu) 's Twitter Profile Photo

Our team (Google Cloud AI Research: research.google/teams/cloud-ai/) is seeking PhD student researchers/interns to work on LLM-related topics (agent, reasoning, RAG, data synthesis, etc.). If interested, please fill in this form: forms.gle/Cia2WGY94zTkpP…. Thank you and plz help RT!

Our team (Google Cloud AI Research: research.google/teams/cloud-ai/) is seeking PhD student researchers/interns to work on LLM-related topics (agent, reasoning, RAG, data synthesis, etc.). If interested, please fill in this form: forms.gle/Cia2WGY94zTkpP…. 
Thank you and plz help RT!
Justin Chih-Yao Chen (@cyjustinchen) 's Twitter Profile Photo

🚨 Reverse Thinking Makes LLMs Stronger Reasoners We can often reason from a problem to a solution and also in reverse to enhance our overall reasoning. RevThink shows that LLMs can also benefit from reverse thinking 👉 13.53% gains + sample efficiency + strong generalization!

🚨 Reverse Thinking Makes LLMs Stronger Reasoners

We can often reason from a problem to a solution and also in reverse to enhance our overall reasoning. RevThink shows that LLMs can also benefit from reverse thinking 👉 13.53% gains + sample efficiency + strong generalization!
Justin Chih-Yao Chen (@cyjustinchen) 's Twitter Profile Photo

Happy to share that RevThink has been accepted to #NAACL2025 main conference! 🎉We also release the code and data 👇🧵 RevThink shows that LLMs can also benefit from reverse thinking (like we often do) 👉13.53% gains on 12 datasets (including MATH, ARC, ANLI, etc) + sample

Shangbin Feng (@shangbinfeng) 's Twitter Profile Photo

👀 How to effectively leverage the expertise of diverse models? ✨ Optimize graphs of LLMs with swarm intelligence! 👉🏻 Introducing Heterogeneous Swarms, jointly optimizing the roles and weights of multi-LLM systems for collaborative gains! 📄 Paper: arxiv.org/abs/2502.04510

👀 How to effectively leverage the expertise of diverse models?
✨ Optimize graphs of LLMs with swarm intelligence!

👉🏻 Introducing Heterogeneous Swarms, jointly optimizing the roles and weights of multi-LLM systems for collaborative gains!

📄 Paper: arxiv.org/abs/2502.04510
Justin Chih-Yao Chen (@cyjustinchen) 's Twitter Profile Photo

I will be presenting ✨Reverse Thinking Makes LLMs Stronger Reasoners✨at #NAACL2025 ! We show that LLM can also benefit from reverse thinking -- a technique we often use to reason from a problem to a solution: - Improvements across 12 datasets - Outperforms SFT with 10x more

Sundar Pichai (@sundarpichai) 's Twitter Profile Photo

AlphaEvolve, our new Gemini-powered coding agent, can help engineers + researchers discover new algorithms and optimizations for open math + computer science problems.  We’ve used it to improve the efficiency of our data centers (recovering 0.7% of our fleet-wide compute

I-Hung Hsu (@ihung_hsu) 's Twitter Profile Photo

🧠🚀 Excited to introduce Supervised Reinforcement Learning—a framework that leverages expert trajectories to teach small LMs how to reason through hard problems without losing their minds. 🤯 Better than SFT && RLVR. Read more: huggingface.co/papers/2510.25… #llms #RL #reasoning

🧠🚀 Excited to introduce Supervised Reinforcement Learning—a framework that leverages expert trajectories to teach small LMs how to reason through hard problems without losing their minds. 🤯

Better than SFT &amp;&amp; RLVR. 

Read more: huggingface.co/papers/2510.25…

#llms #RL #reasoning
DailyPapers (@huggingpapers) 's Twitter Profile Photo

Google introduces Budget Tracker for smarter AI agents Current LLM agents waste tool-call budgets. This work unveils Budget Tracker and BATS, enabling agents to dynamically adapt planning based on remaining resources.

Google introduces Budget Tracker for smarter AI agents

Current LLM agents waste tool-call budgets. This work unveils Budget Tracker and BATS, enabling agents to dynamically adapt planning based on remaining resources.
Zifeng Wang (@zifengwang315) 's Twitter Profile Photo

At #NeurIPS2025 in San Diego from Dec 1 to 7 ✈️ Happy to meet old friends and make new ones! Looking forward to exchanging ideas on effective and efficient LLM agents, multi-LLM/agent systems, post-training and continual learning for LLMs, and much more! Also, our team, Cloud