Kwanghee Choi (@juice500ml) 's Twitter Profile
Kwanghee Choi

@juice500ml

Master's student @LTIatCMU, working on speech AI at @shinjiw_at_cmu's @WavLab

ID: 1690873458324799488

calendar_today13-08-2023 23:50:35

71 Tweet

170 Takipçi

145 Takip Edilen

Gallil Maimon (@gallilmaimon) 's Twitter Profile Photo

🎵💬 If you are interested in Audio Tokenisers, you should check out our new work! We empirically analysed existing tokenisers from every way - reconstruction, downstream, LMs and more. Grab yourself a ☕/🍺 and sit down for a read!

🎵💬 If you are interested in Audio Tokenisers, you should check out our new work!
We empirically analysed existing tokenisers from every way - reconstruction, downstream, LMs and more.

Grab yourself a ☕/🍺 and sit down for a read!
William Chen (@chenwanch1) 's Twitter Profile Photo

Not advertised yet, but we figured out how to do this too. And we release how exactly you can do it 👀. With the right training techniques, you can inject audio understanding and generation into an LLM with almost no loss in text perf. Details at arxiv.org/abs/2506.17611

Shikhar (@shikharssu) 's Twitter Profile Photo

Meows, music, murmurs and more! We train a general purpose audio encoder and open source the code, checkpoints and evaluation toolkit.

Shinji Watanabe (@shinjiw_at_cmu) 's Twitter Profile Photo

📢 Excited to announce our 2-day workshop on "Foundations of Speech and Audio Foundation Models" at TTI Chicago, happening September 4–5! 🔗 Info & registration: sites.google.com/view/speech-ai… 📝 Poster submissions welcome! Join us for talks, discussions, and community building!

William Chen (@chenwanch1) 's Twitter Profile Photo

I will be presenting 3 papers from WAVLab | @CarnegieMellon at #Interspeech2025 🇳🇱 One is OWSMv4 (led by Yifan Peng), nominated for best student paper isca-archive.org/interspeech_20… It focuses a lot on data cleaning, particularly for non-English languages It will be an oral on Tues 15:10 at dock 10B.

Shinji Watanabe (@shinjiw_at_cmu) 's Twitter Profile Photo

Our work on OWSM v4 received the Best Student Paper Award at #Interspeech2025! 🏆🎉 Huge congratulations to the team! 🚀👏 I’m especially happy to see our open science efforts for speech foundation models recognized by the community. 🙌 🔗 isca-archive.org/interspeech_20…

Our work on OWSM v4 received the Best Student Paper Award at #Interspeech2025! 🏆🎉
Huge congratulations to the team! 🚀👏

I’m especially happy to see our open science efforts for speech foundation models recognized by the community. 🙌
🔗 isca-archive.org/interspeech_20…
Yifan Peng (@pengyf21) 's Twitter Profile Photo

Excited to receive the Best Student Paper Award at #Interspeech2025 I started the OWSM project in 2023. It took me great effort to design a robust and scalable training framework using ESPnet, prepare unified data formats, and conduct large-scale training with academic resources.