Yuejie Chi (@yuejiec) 's Twitter Profile
Yuejie Chi

@yuejiec

professor at squirrel hill university

ID: 1793613344319963136

calendar_today23-05-2024 12:02:17

16 Tweet

257 Followers

198 Following

Laixi Shi (@shilaixi) 's Twitter Profile Photo

In RLC 2024 @RL_conference, we shall present a recent work about robust reinforcement learning to handle linear function approximation and robustness together in offline settings: “Sample Complexity of Offline Distributionally Robust Linear Markov Decision Processes”

In RLC 2024 @RL_conference, we shall present a recent work about robust reinforcement learning to handle linear function approximation and robustness together in offline settings:
“Sample Complexity of Offline Distributionally Robust Linear Markov Decision Processes”
Tianqi Chen (@tqchenml) 's Twitter Profile Photo

#MLSys2025 call for papers is out! The conference will be led by the general chair Matei Zaharia , PC chairs Yingyan (Celine) Lin, and Gauri Joshi. Consider submitting and bringing your latest works in AI and systems—more details at mlsys.org.

#MLSys2025 call for papers is out! The conference will be led by the general chair  <a href="/matei_zaharia/">Matei Zaharia</a> , PC chairs <a href="/CelineLinatGT/">Yingyan (Celine) Lin</a>, and Gauri Joshi. Consider submitting and bringing your latest works in AI and systems—more details at mlsys.org.
Yuejie Chi (@yuejiec) 's Twitter Profile Photo

Please check out this thread where my amazing student Harry Dong explains his exciting work to be presented at COLM on efficient LLM inference! Joint work with equally amazing colleague Beidi Chen

Yuejie Chi (@yuejiec) 's Twitter Profile Photo

Theory papers should not perform experiments, because once they do, they’ll be criticized for having too much theory that “reduced the volume of the experimental contents”.

Yuejie Chi (@yuejiec) 's Twitter Profile Photo

I am not at #neurips but just heard about this incident and feel very upset about it. This specific calling out of the country of origin is offensive, inappropriate and completely unnecessary. This also breaks the code of conduct and @neurips should take her talk off the website.

Yuejie Chi (@yuejiec) 's Twitter Profile Photo

I will bet most of these high achieving students come from privileged labs/universities. In admission, we should calibrate these factors and look for potentials and out-of-box thinkers. (Opinions are my own.)

Yuejie Chi (@yuejiec) 's Twitter Profile Photo

Amazing list of resources for early-career researchers in ML/data science/AI put together by my amazier student Laixi, who happens to be on the market this year!

Yuejie Chi (@yuejiec) 's Twitter Profile Photo

Some of us who work in higher ed but not close enough to AI need the alarm call to wake up and move faster. We need to rethink our curriculum and pedagogical approaches!

AI at Meta (@aiatmeta) 's Twitter Profile Photo

Today is the start of a new era of natively multimodal AI innovation. Today, we’re introducing the first Llama 4 models: Llama 4 Scout and Llama 4 Maverick — our most advanced models yet and the best in their class for multimodality. Llama 4 Scout • 17B-active-parameter model

Today is the start of a new era of natively multimodal AI innovation.

Today, we’re introducing the first Llama 4 models: Llama 4 Scout and Llama 4 Maverick —  our most advanced models yet and the best in their class for multimodality.

Llama 4 Scout
• 17B-active-parameter model
Qinqing Zheng (@qqyuzu) 's Twitter Profile Photo

d1: to grow in reasoning, masked diffusion language models go beyond supervised learning , we meet RL !😃dllm-reasoning.github.io

Avinandan Bose (@avibose22) 's Twitter Profile Photo

🧠 Your LLM should model how you think, not reduce you to preassigned traits 📢 Introducing LoRe: a low-rank reward modeling framework for personalized RLHF ❌ Demographic grouping/handcrafted traits ✅ Infers implicit preferences ✅ Few-shot adaptation 📄 arxiv.org/abs/2504.14439

🧠 Your LLM should model how you think, not reduce you to preassigned traits
📢 Introducing LoRe: a low-rank reward modeling framework for personalized RLHF
❌ Demographic grouping/handcrafted traits
✅ Infers implicit preferences
✅ Few-shot adaptation
📄 arxiv.org/abs/2504.14439
Gal Mishne 💔🇮🇱 (@gmishne) 's Twitter Profile Photo

Serious issues with AC paper bidding NeurIPS Conference Paper matching ignored my excluded papers, gave me mostly papers on topics I have little to no expertise in and we can only bid on limited subset of these mostly irrelevant papers. Friend reported having the same issue #NeurIPS25

Aaron Defazio (@aaron_defazio) 's Twitter Profile Photo

Why do gradients increase near the end of training? Read the paper to find out! We also propose a simple fix to AdamW that keeps gradient norms better behaved throughout training. arxiv.org/abs/2506.02285

Why do gradients increase near the end of training? 
Read the paper to find out!
We also propose a simple fix to AdamW that keeps gradient norms better behaved throughout training.
arxiv.org/abs/2506.02285
Avinandan Bose (@avibose22) 's Twitter Profile Photo

🚨 Code is live! Check out LoRe – a modular, lightweight codebase for personalized reward modeling from user preferences. 📦 Few-shot personalization 📊 Benchmarks: TLDR, PRISM, PersonalLLM 👉 github.com/facebookresear… Huge thanks to AI at Meta for open-sourcing this research 🙌

Zhihao Jia (@jiazhihao) 's Twitter Profile Photo

📢Exciting updates from #MLSys2025! All session recordings are now available and free to watch at mlsys.org. We’re also thrilled to announce that #MLSys2026 will be held in Seattle next May—submissions open next month with a deadline of Oct 30. We look forward to

📢Exciting updates from #MLSys2025! All session recordings are now available and free to watch at mlsys.org.
We’re also thrilled to announce that #MLSys2026 will be held in Seattle next May—submissions open next month with a deadline of Oct 30. We look forward to
Tong Yang (@tongyang_666) 's Twitter Profile Photo

🚨 🔥 Multi-step reasoning is key to solving complex problems — and Transformers with Chain-of-Thought can do it surprisingly well. 🤔 But how does CoT function as a learned scratchpad that lets even shallow Transformers run sequential algorithms that would otherwise require

🚨 🔥 Multi-step reasoning is key to solving complex problems — and Transformers with Chain-of-Thought can do it surprisingly well.

🤔 But how does CoT function as a learned scratchpad that lets even shallow Transformers run sequential algorithms that would otherwise require
Yuejie Chi (@yuejiec) 's Twitter Profile Photo

A one-layer multi-head transformer, with CoT, enables both forward and reversal reasoning. The training dynamics analysis particularly illuminates how two (heads) are better than one! See Tong’s post below. Joint work with Tong Yang Yu Huang and Yingbin Liang.