Ting-Yun Chang (@charlottetyc) 's Twitter Profile
Ting-Yun Chang

@charlottetyc

PhD student @CSatUSC @nlp_usc

ID: 1432782439944908801

linkhttps://terarachang.github.io/ calendar_today31-08-2021 19:09:08

90 Tweet

481 Followers

365 Following

Chao-Wei Huang (@cwhuang_wh) 's Twitter Profile Photo

Excited to share that our paper, "FactAlign: Long-form Factuality Alignment of Large Language Models," has been accepted to #EMNLP2024 Findings! 📝 Paper: arxiv.org/abs/2410.01691 💻 Code: github.com/MiuLab/FactAli… FactAlign significantly improves the factuality of LLMs' (1/2)

Excited to share that our paper, "FactAlign: Long-form Factuality Alignment of Large Language Models," has been accepted to #EMNLP2024 Findings!

📝 Paper: arxiv.org/abs/2410.01691
💻 Code: github.com/MiuLab/FactAli…

FactAlign significantly improves the factuality of LLMs' (1/2)
Deqing Fu (@deqingfu) 's Twitter Profile Photo

It's cool to see Google DeepMind's new research to show similar findings as we did back in April. IsoBench (isobench.github.io, accepted to Conference on Language Modeling 2024) was curated to show the performance gap across modalities and multimodal models' preference over text modality.

It's cool to see <a href="/GoogleDeepMind/">Google DeepMind</a>'s new research to show similar findings as we did back in April. 

IsoBench (isobench.github.io, accepted to <a href="/COLM_conf/">Conference on Language Modeling</a> 2024) was curated to show the performance gap across modalities and multimodal models' preference over text modality.
Yuqing Yang (@yyqcode) 's Twitter Profile Photo

In Miami for #EMNLP2024! Come check out our findings poster, Weak-to-Strong Reasoning, on Wednesday at 10:30am. Super excited for my first in-person conference. Looking forward to connecting and chatting about reasoning, hallucination, self-correction, and all things LLMs! 🌴🌴

In Miami for #EMNLP2024! Come check out our findings poster, Weak-to-Strong Reasoning, on Wednesday at 10:30am.
Super excited for my first in-person conference. Looking forward to connecting and chatting about reasoning, hallucination, self-correction, and all things LLMs! 🌴🌴
Chao-Wei Huang (@cwhuang_wh) 's Twitter Profile Photo

I'll be presenting 2 papers at #EMNLP2024. Happy to chat! 🗓️ Nov 12 (Tue) 11:00-12:30 - Riverfront Hall 📄 PairDistill: Pairwise Relevance Distillation for Dense Retrieval 🗓️ Nov 13 (Wed) 16:00-17:30 - Jasmine 📄 FactAlign: Long-form Factuality Alignment of Large Language Models

I'll be presenting 2 papers at #EMNLP2024. Happy to chat!

🗓️ Nov 12 (Tue) 11:00-12:30 - Riverfront Hall
📄 PairDistill: Pairwise Relevance Distillation for Dense Retrieval
🗓️ Nov 13 (Wed) 16:00-17:30 - Jasmine
📄 FactAlign: Long-form Factuality Alignment of Large Language Models
Robin Jia (@robinomial) 's Twitter Profile Photo

I'll be at #NeurIPS2024! My group has papers analyzing how LLMs use Fourier Features for arithmetic (Tianyi Zhou ) and how TFs learn higher-order optimization for ICL (Deqing Fu), plus workshop papers on backdoor detection (Jun Yan) and LLMs + PDDL (Wang Bill Zhu)

Qinyuan Ye (👀Jobs) (@qinyuan_ye) 's Twitter Profile Photo

I'll present a poster for Lifelong ICL and Task Haystack at #NeurIPS2024! ⏰ Wednesday 11am-2pm 📍 East Exhibit Hall A-C #2802 📜 arxiv.org/abs/2407.16695 My co-first author Xiaoyue Xu is applying to PhD programs and I am looking jobs in industry! Happy to connect at NeurIPS!

Tianyi Zhou (@tianyi_zhou12) 's Twitter Profile Photo

Billion-parameter LLMs still struggle with simple arithmetic? 📞 FoNE (Fourier Number Embedding) tackles this problem. By mapping numbers directly into Fourier space, it bypasses tokenization and significantly improves numerical accuracy with better efficiency and accuracy.

Abrar Anwar (@_abraranwar) 's Twitter Profile Photo

All these VLAs allow robots to do more tasks, but when you're physically testing many policies, it's hard to eval on every task! We take advantage of shared information between tasks and within policies to actively test multi-task robot policies! 1/7 🧵 arxiv.org/abs/2502.09829

All these VLAs allow robots to do more tasks, but when you're physically testing many policies, it's hard to eval on every task!

We take advantage of shared information between tasks and within policies to actively test multi-task robot policies!
1/7 🧵

arxiv.org/abs/2502.09829
Stanford NLP Group (@stanfordnlp) 's Twitter Profile Photo

For this week’s NLP Seminar, we are thrilled to host Yun-Nung Vivian Chen to talk about Optimizing Interaction and Intelligence — Multi-Agent Simulation and Collaboration for Personalized Marketing and Advanced Reasoning! When: 3/6 Thurs 11am PT Non-Stanford affiliates registration form

For this week’s NLP Seminar, we are thrilled to host <a href="/YunNungChen/">Yun-Nung Vivian Chen</a> to talk about Optimizing Interaction and Intelligence — Multi-Agent Simulation and Collaboration for Personalized Marketing and Advanced Reasoning!

When: 3/6 Thurs 11am PT
Non-Stanford affiliates registration form
Wang Bill Zhu (@billjohn1235813) 's Twitter Profile Photo

At NAACL HLT 2025 this week! I’ll be presenting our work on LLM domain induction with Jesse Thomason on Thu (5/1) at 4pm in Hall 3, Section I. Would love to connect and chat about LLM planning, reasoning, AI4Science, multimodal stuff, or anything else. Feel free to DM!

At <a href="/naaclmeeting/">NAACL HLT 2025</a> this week! I’ll be presenting our work on LLM domain induction with <a href="/_jessethomason_/">Jesse Thomason</a> on Thu (5/1) at 4pm in Hall 3, Section I.

Would love to connect and chat about LLM planning, reasoning, AI4Science, multimodal stuff, or anything else. Feel free to DM!
Stanford NLP Group (@stanfordnlp) 's Twitter Profile Photo

For this week’s NLP Seminar, we are thrilled to host Deqing Fu to talk about Closing the Modality Gap: Benchmarking and Improving Visual Understanding in Multimodal LLMs! When: 5/22 Thurs 11am PT Non-Stanford affiliates registration form (closed at 9am PT on the talk day):

For this week’s NLP Seminar, we are thrilled to host <a href="/DeqingFu/">Deqing Fu</a> to talk about Closing the Modality Gap: Benchmarking and Improving Visual Understanding in Multimodal LLMs!

When: 5/22 Thurs 11am PT
Non-Stanford affiliates registration form (closed at 9am PT on the talk day):
Yuqing Yang (@yyqcode) 's Twitter Profile Photo

🧐When do LLMs admit their mistakes when they should know better? In our new paper, we define this behavior as retraction: the model indicates that its generated answer was wrong. LLMs can retract—but they rarely do.🤯 arxiv.org/abs/2505.16170 👇🧵

🧐When do LLMs admit their mistakes when they should know better?

In our new paper, we define this behavior as retraction: the model indicates that its generated answer was wrong.
LLMs can retract—but they rarely do.🤯

arxiv.org/abs/2505.16170

👇🧵