Heng-Jui Chang (@hjchang87) 's Twitter Profile
Heng-Jui Chang

@hjchang87

🎓 PhD Candidate @MIT_CSAIL
🧪 Former Research Scientist Intern @AIatMeta

ID: 1299234958217682945

linkhttps://people.csail.mit.edu/hengjui/ calendar_today28-08-2020 06:41:21

24 Tweet

141 Followers

158 Following

Kushal Lakhotia (@hikushalhere) 's Twitter Profile Photo

SUPERB Challenge is ready: superbbenchmark.org/challenge You can also submit your results to The 2nd Self-supervised Learning for Audio and Speech Processing at AAAI: aaai-sas-2022.github.io

Heng-Jui Chang (@hjchang87) 's Twitter Profile Photo

Our DistilHuBERT model is released! Thanks to @PatrickPlaten Leo Yang Hung-yi Lee (李宏毅) NTU SPML Lab! paper: arxiv.org/abs/2110.01900 pre-training and inference code: github.com/s3prl/s3prl

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Self-Supervised Learning for Speech and Audio Processing Workshop @ AAAI 2022 ===== Website: aaai-sas-2022.github.io Submission Deadline: November 15th, 2021 (Anywhere on Earth) -> Less than 24 hours! Submission website: cmt3.research.microsoft.com/SAS2022 Contact: [email protected]

Hung-yi Lee (李宏毅) (@hungyilee2) 's Twitter Profile Photo

Workshop on Self-supervised Learning for Audio and Speech Processing @ AAAI 2022 starts at 8:50 a.m., EST (9:50 p.m. GMT+8), February 28. If you want to hear about exciting new advances in self-supervised learning, don't miss it. aaai-sas-2022.github.io

Mirco Ravanelli (@mirco_ravanelli) 's Twitter Profile Photo

Exciting news! 2 open positions for #PhD students to join our team and work on cutting-edge #research in #deeplearning, #conversationalAI, #speech tech, & sequence processing. Check out more details here and apply as soon as possible if interested: tinyurl.com/yc4xrdke

Yung-Sung Chuang (@yungsungchuang) 's Twitter Profile Photo

📢New Paper Alert!!🚀 arxiv.org/abs/2304.03728 Does ChatGPT have the ability to check facts by itself?🤔 We designed a simple, few-shot, unified chain-of-thought prompting pipeline that can do: 🔹Fact-checking ✅ 🔹Stereotype detection 🚫 🔹Hate speech detection 🙅 (1/2)

📢New Paper Alert!!🚀
arxiv.org/abs/2304.03728

Does ChatGPT have the ability to check facts by itself?🤔
We designed a simple, few-shot, unified chain-of-thought prompting pipeline that can do: 
🔹Fact-checking ✅
🔹Stereotype detection 🚫
🔹Hate speech detection 🙅 (1/2)
Andrew Rouditchenko 🇺🇦 (@arouditchenko) 's Twitter Profile Photo

🗣️ Whisper is great for speech recognition, but it only recognizes ~100 languages. What if it wasn't trained on the language that you speak? Happy to introduce my #INTERSPEECH2023 paper comparing Whisper and XLS-R for adaption to unseen languages! arxiv.org/abs/2305.12606

🗣️ Whisper is great for speech recognition, but it only recognizes ~100 languages. What if it wasn't trained on the language that you speak?

Happy to introduce my #INTERSPEECH2023 paper comparing Whisper and XLS-R for adaption to unseen languages!

arxiv.org/abs/2305.12606
Yung-Sung Chuang (@yungsungchuang) 's Twitter Profile Photo

🚨We release SAIL-7B⛵️️a search-augmented instruction-tuned LM with: 🦆Real-time connecting to DuckDuckGo 🔍Explicitly filter out distracting search results 👨‍🏫Instruction following Outperforms ChatGPT and Vicuna!🦙 demo: openlsr.org/sail-7b arxiv: arxiv.org/abs/2305.15225

🚨We release SAIL-7B⛵️️a search-augmented instruction-tuned LM with:

🦆Real-time connecting to DuckDuckGo
🔍Explicitly filter out distracting search results
👨‍🏫Instruction following

Outperforms ChatGPT and Vicuna!🦙

demo: openlsr.org/sail-7b
arxiv: arxiv.org/abs/2305.15225
Yung-Sung Chuang (@yungsungchuang) 's Twitter Profile Photo

Can language models help us do better search?🤔 🎉In #ACL2023 findings, we present EAR pipeline: 🎲sample multiple queries from LM 🎯rescore to select best query 🔍BM25 search 💪boost OpenQA accuracy to beat DPR/GAR arxiv: arxiv.org/abs/2305.17080 code: github.com/voidism/EAR

Can language models help us do better search?🤔

🎉In #ACL2023 findings, we present EAR pipeline:

🎲sample multiple queries from LM
🎯rescore to select best query
🔍BM25 search
💪boost OpenQA accuracy to beat DPR/GAR

arxiv: arxiv.org/abs/2305.17080
code: github.com/voidism/EAR
Yuan Gong (@ygongnd) 's Twitter Profile Photo

(1/2) Introduce the IS23 Whisper-AT paper. We usually believe noise-robust ASR models' representations are noise-invirant. But we show a surprising finding that while Whisper is very robust against real-world background sounds, its representation is actually NOT noise-invariant.

(1/2) Introduce the IS23 Whisper-AT paper. We usually believe noise-robust ASR models' representations are noise-invirant. But we show a surprising finding that while Whisper is very robust against real-world background sounds, its representation is actually NOT noise-invariant.
Hongyin Luo (@lhythu) 's Twitter Profile Photo

Excited to present our research at #acl2023 ! We found that self-trained entailment models with 350M parameters can outperform strong few-shot large language models with more than 100B parameters on several language understanding tasks. (1/4) news.mit.edu/2023/language-…

Yung-Sung Chuang (@yungsungchuang) 's Twitter Profile Photo

(1/5)🚨Can LLMs be more factual without retrieval or finetuning?🤔 -yes✅ 🦙We find factual knowledge often lies in higher layers of LLaMA 💪Contrast high/low layers can amplify factuality & boost TruthfulQA by 12-17% 📝arxiv.org/abs/2309.03883 🧑‍💻github.com/voidism/DoLa #NLProc

(1/5)🚨Can LLMs be more factual without retrieval or finetuning?🤔 -yes✅

🦙We find factual knowledge often lies in higher layers of LLaMA
💪Contrast high/low layers can amplify factuality & boost TruthfulQA by 12-17%

📝arxiv.org/abs/2309.03883
🧑‍💻github.com/voidism/DoLa

#NLProc
Hongyin Luo (@lhythu) 's Twitter Profile Photo

(1/4) 💡Natural language embedded program (NLEP) is all you need for symbolic AND natural language tasks. 🚀NLEP outperforms ChatGPT-4, CoT, & PoT/PAL, without any task-specific example. 🎢NLEP makes small LMs outperform GPT-3 without fine-tuning! arxiv.org/pdf/2309.10814…

(1/4) 💡Natural language embedded program (NLEP) is all you need for symbolic AND natural language tasks.

🚀NLEP outperforms ChatGPT-4, CoT, & PoT/PAL, without any task-specific example. 

🎢NLEP makes small LMs outperform GPT-3 without fine-tuning!

arxiv.org/pdf/2309.10814…
Alexander H. Liu (@alex_h_liu) 's Twitter Profile Photo

Presenting 2 works at #ICLR tomorrow! 📃Generative Pre-training for Speech with Flow Matching 📍5/9 (Wed) Hall B #68, 10:45am-12:45pm 📃Listen, Think, and Understand 📍5/9 (Wed) Hall B #60, 4:30pm-6:30pm Please stop by if you're interested! More details...👇

Presenting 2 works at #ICLR tomorrow!

📃Generative Pre-training for Speech with Flow Matching
📍5/9 (Wed) Hall B #68, 10:45am-12:45pm

📃Listen, Think, and Understand
📍5/9 (Wed) Hall B #60, 4:30pm-6:30pm

Please stop by if you're interested! More details...👇
Yung-Sung Chuang (@yungsungchuang) 's Twitter Profile Photo

(1/5)🚨LLMs can now self-improve to generate better citations✅ 📝We design automatic rewards to assess citation quality 🤖Enable BoN/SimPO w/o external supervision 📈Perform close to “Claude Citations” API w/ only 8B model 📄arxiv.org/abs/2502.09604 🧑‍💻github.com/voidism/SelfCi…

(1/5)🚨LLMs can now self-improve to generate better citations✅

📝We design automatic rewards to assess citation quality
🤖Enable BoN/SimPO w/o external supervision
📈Perform close to “Claude Citations” API w/ only 8B model

📄arxiv.org/abs/2502.09604
🧑‍💻github.com/voidism/SelfCi…
Heng-Jui Chang (@hjchang87) 's Twitter Profile Photo

💡Bridging speech, sound, & music representations with one universal model? We introduce USAD ✅ 📚 Distills knowledge from domain-specific SSL models 🎯 Matches expert models across speech/audio/music tasks 📄 arxiv.org/abs/2506.18843 🧑‍💻 huggingface.co/MIT-SLS/USAD-B…

💡Bridging speech, sound, & music representations with one universal model?

We introduce USAD ✅
📚 Distills knowledge from domain-specific SSL models
🎯 Matches expert models across speech/audio/music tasks

📄 arxiv.org/abs/2506.18843
🧑‍💻 huggingface.co/MIT-SLS/USAD-B…