Yuqing Yang (@yyqcode) 's Twitter Profile
Yuqing Yang

@yyqcode

First-year PhD student @CSatUSC @nlp_usc.

ID: 1670754896352784391

linkhttps://ayyyq.github.io/ calendar_today19-06-2023 11:26:45

24 Tweet

189 Followers

347 Following

Qinyuan Ye (👀Jobs) (@qinyuan_ye) 's Twitter Profile Photo

I'll present a poster for Lifelong ICL and Task Haystack at #NeurIPS2024! ⏰ Wednesday 11am-2pm 📍 East Exhibit Hall A-C #2802 📜 arxiv.org/abs/2407.16695 My co-first author Xiaoyue Xu is applying to PhD programs and I am looking jobs in industry! Happy to connect at NeurIPS!

Tengxiao Liu (@tengxiaoliu) 's Twitter Profile Photo

Come join the #NeurIPS2024 poster session and discuss whether language models can learn to skip steps in reasoning! 🗓Dec 12, Thursday, 11:00 am - 2:00 pm 📍East Exhibit Hall A-C #2900 Feel free to stop by and say hi! I am actively seeking Summer 2025 internship opportunities!

Muru Zhang (@zhang_muru) 's Twitter Profile Photo

Running your model on multiple GPUs but often found the speed not satisfiable? We introduce Ladder-residual, a parallelism-aware architecture modification that makes 70B Llama with tensor parallelism ~30% faster! Work done at Together AI. Co-1st author with Mayank Mishra

Running your model on multiple GPUs but often found the speed not satisfiable? We introduce Ladder-residual, a parallelism-aware architecture modification that makes 70B Llama with tensor parallelism ~30% faster!
  
Work done at <a href="/togethercompute/">Together AI</a>. Co-1st author with <a href="/MayankMish98/">Mayank Mishra</a>
Tianyi Zhou (@tianyi_zhou12) 's Twitter Profile Photo

Billion-parameter LLMs still struggle with simple arithmetic? 📞 FoNE (Fourier Number Embedding) tackles this problem. By mapping numbers directly into Fourier space, it bypasses tokenization and significantly improves numerical accuracy with better efficiency and accuracy.

Linxin Song (@linxins2) 's Twitter Profile Photo

Want to know what your LLM don’t know? This is how 👇 Preprint: arxiv.org/abs/2503.23361 Code: github.com/uscnlp-lime/SEA

Deqing Fu (@deqingfu) 's Twitter Profile Photo

Textual steering vectors can improve visual understanding in multimodal LLMs! You can extract steering vectors via any interpretability toolkit you like -- SAEs, MeanShift, Probes -- and apply them to image or text tokens (or both) of Multimodal LLMs. And They Steer!

Textual steering vectors can improve visual understanding in multimodal LLMs!

You can extract steering vectors via any interpretability toolkit you like -- SAEs, MeanShift, Probes -- and apply them to image or text tokens (or both) of Multimodal LLMs. 
And They Steer!
Linxin Song (@linxins2) 's Twitter Profile Photo

🚨 We discovered a surprising side effect of Reinforcement Finetuning (RFT): it makes LLMs more confidently wrong on unanswerable questions. We call this the hallucination tax: a drop in refusal behavior that leads to overconfident hallucinations. 🧵 1/n

🚨 We discovered a surprising side effect of Reinforcement Finetuning (RFT): it makes LLMs more confidently wrong on unanswerable questions.
We call this the hallucination tax: a drop in refusal behavior that leads to overconfident hallucinations.

🧵 1/n
Dongwei Jiang (@dongwei__jiang) 's Twitter Profile Photo

🧵 Recent studies show LLMs can self-improve their responses when given external feedback. But how effectively can they incorporate it? We tested this systematically—and found they can't fully integrate feedback, even when the feedback is high-quality and backed by ground-truth.

🧵 Recent studies show LLMs can self-improve their responses when given external feedback. But how effectively can they incorporate it?

We tested this systematically—and found they can't fully integrate feedback, even when the feedback is high-quality and backed by ground-truth.
Xi Ye (@xiye_nlp) 's Twitter Profile Photo

There’s been hot debate about (The Illusion of) The Illusion of Thinking. My take: it’s not that models can’t reason — they just aren’t perfect at long-form generation yet. We eval reasoning models on LongProc benchmark (requiring generating 8K CoTs, see thread). Reasoning

Chenxin An (@anchancy46881) 's Twitter Profile Photo

# 🚨 4B open-recipe model beats Claude-4-Opus 🔓 100% open data, recipe, model weights and code. Introducing Polaris✨--a post-training recipe for scaling RL on advanced reasoning models. 🥳 Check out how we boost open-recipe reasoning models to incredible performance levels

# 🚨 4B open-recipe model beats Claude-4-Opus 
🔓 100% open data, recipe, model weights and code.

Introducing Polaris✨--a post-training recipe for scaling RL on advanced reasoning models. 

🥳 Check out how we boost open-recipe reasoning models to incredible performance levels
Johnny Tian-Zheng Wei (@johntzwei) 's Twitter Profile Photo

Announcing 🔭✨Hubble, a suite of open-source LLMs to advance the study of memorization! Pretrained models up to 8B params, with controlled insertion of texts (e.g., book passages, biographies, test sets, and more!) designed to emulate key memorization risks 🧵

Announcing 🔭✨Hubble, a suite of open-source LLMs to advance the study of memorization! 

Pretrained models up to 8B params, with controlled insertion of texts (e.g., book passages, biographies, test sets, and more!) designed to emulate key memorization risks 🧵