Shiyu Chang (@codeterminator) 's Twitter Profile
Shiyu Chang

@codeterminator

Assistant Professor at UC Santa Barbara. Tweets reflect my views alone.

ID: 789123278690451456

linkhttps://code-terminator.github.io/ calendar_today20-10-2016 15:17:05

129 Tweet

739 Followers

417 Following

Shiyu Chang (@codeterminator) 's Twitter Profile Photo

🔥Introducing our new #cvpr paper on improving the quality of diffusion generations via efficient particle samplings. This is an amazing work led by Yujian in collaboration with MIT and the MIT-IBM Watson AI Lab.

Shiyu Chang (@codeterminator) 's Twitter Profile Photo

🛡️New Jailbreak Defenses for LLMs: By harnessing semantic-preserving transformations with randomized smoothing, we have enabled LLMs to defend jailbreaks with minimal impact on their performance for benign tasks. An amazing collaboration between students at UCSB and UPenn.

MIT Jameel Clinic for AI & Health (@aihealthmit) 's Twitter Profile Photo

WHP a.k.a. Who's Harry Potter?⚡️ is a ML method to unlearn knowledge or biases from training data. Yujian Liu, Yang Zhang, #JameelClinic PI Tommi Jaakkola, and Shiyu Chang propose a new way to extend WHP w/ targeted unlearning arxiv.org/pdf/2407.16997

WHP a.k.a. Who's Harry Potter?⚡️ is a ML method to unlearn knowledge or biases from training data. <a href="/liu_yujian/">Yujian Liu</a>, <a href="/YangZha26484161/">Yang Zhang</a>, #JameelClinic PI Tommi Jaakkola, and <a href="/CodeTerminator/">Shiyu Chang</a> propose a new way to extend WHP w/ targeted unlearning arxiv.org/pdf/2407.16997
Shiyu Chang (@codeterminator) 's Twitter Profile Photo

Fantastic work by my student Bairu Hou and amazing collaborators at Apple! 🌟 We introduced Instruction-Following Pruning – a dynamic method for pruning LLMs based on prompts, achieving remarkable efficiency and performance.

Jiayi Pan (@jiayi_pirate) 's Twitter Profile Photo

We reproduced DeepSeek R1-Zero in the CountDown game, and it just works Through RL, the 3B base LM develops self-verification and search abilities all on its own You can experience the Ahah moment yourself for < $30 Code: github.com/Jiayi-Pan/Tiny… Here's what we learned 🧵

We reproduced DeepSeek R1-Zero in the CountDown game, and it just works 

Through RL, the 3B base LM develops self-verification and search abilities all on its own 

You can experience the Ahah moment yourself for &lt; $30 
Code: github.com/Jiayi-Pan/Tiny…

Here's what we learned 🧵
Jingbo Yang (@jingbo_y) 's Twitter Profile Photo

In RAG applications, LLMs often reprocess the same database chunks for different queries—leading to high latency and cost from handling massive input tokens. We introduce KVLink, an efficient approach to reuse pre-computed KV caches of retrieved documents, drastically cutting

In RAG applications, LLMs often reprocess the same database chunks for different queries—leading to high latency and cost from handling massive input tokens.

We introduce KVLink, an efficient approach to reuse pre-computed KV caches of retrieved documents, drastically cutting
Shiyu Chang (@codeterminator) 's Twitter Profile Photo

Our recent work on pruning long chain-of-thought reasoning—simple, effective, and surprisingly powerful. 🧠✂️ Led by my student Bairu Hou , with amazing collaborators!

Shiyu Chang (@codeterminator) 's Twitter Profile Photo

Elegant algorithm to embed watermarks against piggyback spoofing attacks! 🚀 Led by my amazing student Li An and our great collaborators.

Shiyu Chang (@codeterminator) 's Twitter Profile Photo

Sad to miss #ICLR2025 this year, but thrilled to see our work led by my student Yujian Liu presented there! Stop by our poster and chat with my amazing collaborators about Prereq-Tune, our new method to improve LLM factuality. 🚀📍#273 | Today (Apr 24), 3–5:30 PM

Shiyu Chang (@codeterminator) 's Twitter Profile Photo

Sadly missing #NAACL this time, but my student Bairu Hou will present our joint work with MIT NLP and MIT-IBM Watson AI Lab on hallucination detection using a novel probabilistic propagation approach — one of my favorite recent projects from our team.

Shiyu Chang (@codeterminator) 's Twitter Profile Photo

Our new knowledge tracing method accurately tracks young student understanding — even with just a few student responses. Led by my amazing students Xinyi Gao Qiucheng Wu, with fantastic collaborators! 🌱🧠

Andrew Ng (@andrewyng) 's Twitter Profile Photo

On Monday, a United States District Court ruled that training LLMs on copyrighted books constitutes fair use. A number of authors had filed suit against Anthropic for training its models on their books without permission. Just as we allow people to read books and learn from them

Andrej Karpathy (@karpathy) 's Twitter Profile Photo

Scaling up RL is all the rage right now, I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains, but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly), let me slightly

Shiyu Chang (@codeterminator) 's Twitter Profile Photo

Sad to miss #ICML2025 this year, but thrilled that my student Bairu Hou will present his exciting work on dynamically pruning LLMs into efficient, task-specific models—done in collaboration with our amazing collaborators from Apple! 🍎✨

Denny Zhou (@denny_zhou) 's Twitter Profile Photo

Slides for my lecture “LLM Reasoning” at Stanford CS 25: dennyzhou.github.io/LLM-Reasoning-… Key points: 1. Reasoning in LLMs simply means generating a sequence of intermediate tokens before producing the final answer. Whether this resembles human reasoning is irrelevant. The crucial

Luiza Jarovsky (@luizajarovsky) 's Twitter Profile Photo

🚨 SHOCKING: people are unknowingly making their ChatGPT interactions PUBLIC, and they are being indexed by Google (see my test below). My privacy recommendations: When people interact with ChatGPT and use the "Share" feature (for example, to send the conversation to family and

🚨 SHOCKING: people are unknowingly making their ChatGPT interactions PUBLIC, and they are being indexed by Google (see my test below). My privacy recommendations:

When people interact with ChatGPT and use the "Share" feature (for example, to send the conversation to family and
Xin Eric Wang @ ICLR 2025 (@xwang_lk) 's Twitter Profile Photo

UCSB NLP @ EMNLP 2025 EMNLP 2025! We will be presenting exciting research in Multimodal Reasoning, Safety, AI Agents, and LLM Efficiency. Come meet us in Suzhou this November. Would love to exchange ideas and discuss where the field is headed!🚀 🎉 Huge congrats to our