Howard Yen (@howardyen1) 's Twitter Profile
Howard Yen

@howardyen1

ID: 2316455576

calendar_today29-01-2014 03:41:42

37 Tweet

212 Followers

225 Following

Zexue He (@zexuehe) 's Twitter Profile Photo

πŸš€ News! Our 2nd Workshop on Long-Context Foundation Models (LCFM), to be held at ICML 2025 in Vancouver πŸ‡¨πŸ‡¦! If you're working on long-context models, consider submitting your work! πŸ—“οΈ DDL: May 22, 2025 (AOE) 🌐 Web: longcontextfm.github.io πŸ”— OpenReview: bit.ly/lcfmworkshop

πŸš€ News! Our 2nd Workshop on Long-Context Foundation Models (LCFM), to be held at ICML 2025 in Vancouver πŸ‡¨πŸ‡¦!
If you're working on long-context models, consider submitting your work!
πŸ—“οΈ DDL: May 22, 2025 (AOE)
🌐 Web: longcontextfm.github.io
πŸ”— OpenReview: bit.ly/lcfmworkshop
Jacqueline He (@jcqln_h) 's Twitter Profile Photo

LMs often output answers that sound right but aren’t supported by input context. This is intrinsic hallucination: the generation of plausible, but unsupported content. We propose Precise Information Control (PIC): a task requiring LMs to ground only on given verifiable claims.

LMs often output answers that sound right but aren’t supported by input context. This is intrinsic hallucination: the generation of plausible, but unsupported content.

We propose Precise Information Control (PIC): a task requiring LMs to ground only on given verifiable claims.
Xi Ye (@xiye_nlp) 's Twitter Profile Photo

πŸ€” Recent mech interp work showed that retrieval heads can explain some long-context behavior. But can we use this insight for retrieval? πŸ“£ Introducing QRHeads (query-focused retrieval heads) that enhance retrieval Main contributions: πŸ” Better head detection: we find a

πŸ€” Recent mech interp work showed that retrieval heads can explain some long-context behavior. But can we use this insight for retrieval?
πŸ“£ Introducing QRHeads (query-focused retrieval heads) that enhance retrieval

Main contributions:
 πŸ” Better head detection: we find a
Zexue He (@zexuehe) 's Twitter Profile Photo

πŸ’‘ Curious about long-context foundation models (LFCM)? 🧠 We’re hosting a panel at the LCFM workshop at #ICML2025 on β€œHow to evaluate long-context foundation models?” β€” We’d love to feature your question! Anything on long-context evaluation or modeling β€” drop it below / DM me🎀

πŸ’‘ Curious about long-context foundation models (LFCM)?
🧠 We’re hosting a panel at the LCFM workshop at #ICML2025 on β€œHow to evaluate long-context foundation models?” β€” We’d love to feature your question!

Anything on long-context evaluation or modeling β€” drop it below / DM me🎀
Sadhika Malladi (@sadhikamalladi) 's Twitter Profile Photo

Excited to share that I will be starting as an Assistant Professor in CSE at UCSD (UCSD CSE) in Fall 2026! I am currently recruiting PhD students who want to bridge theory and practice in deep learning - see here: cs.princeton.edu/~smalladi/recr…

Howard Yen (@howardyen1) 's Twitter Profile Photo

Congrats!!! As an empiricist, I always found your work super relevant and provide useful theoretical insights! Also thanks for giving great advices in our chats :)

Danqi Chen (@danqi_chen) 's Twitter Profile Photo

I am going to present two papers at #COLM2025 tomorrow from 4:30-6:30pm, as none of our leading authors can attend due to visa issues. Haven't done poster presentations for years 🀣🀣 .... so I will do my best! #76: LongProc #80: Goedel-Prover v1

I am going to present two papers at #COLM2025 tomorrow from 4:30-6:30pm, as none of our leading authors can attend due to visa issues. 

Haven't done poster presentations for years 🀣🀣 .... so I will do my best!

#76: LongProc
#80: Goedel-Prover v1
Xi Ye (@xiye_nlp) 's Twitter Profile Photo

We will present QRHead (Wuwei Zhang) #EMNLP2025 Without any training, we boosts Llama-3.1-8B’s performance by >10% πŸ“ˆon context reasoning tasks (CLIPPER, LongMemEval), and outperforms specialized re-rankers on BEIR. Check out our (virtual) poster tonight!

We will present QRHead (<a href="/WuweiZhang0723/">Wuwei Zhang</a>) #EMNLP2025

Without any training, we boosts Llama-3.1-8B’s performance by &gt;10% πŸ“ˆon context reasoning tasks (CLIPPER, LongMemEval), and outperforms specialized re-rankers on BEIR. Check out our (virtual) poster tonight!