Tianyu Liu (@t_y_liu) 's Twitter Profile
Tianyu Liu

@t_y_liu

PhD Student @ ETH Zürich

ID: 1241011234146619392

calendar_today20-03-2020 14:40:49

33 Tweet

292 Followers

212 Following

Yifan Hou (@yyyyyyyyifan) 's Twitter Profile Photo

Adding 2% parameters can enhance your multilingual language model and turn it into a knowledge base. Happy to share our #EMNLP2022 findings paper "Adapters for Enhanced Modeling of Multilingual Knowledge and Text", which uses a set of adapters for the enhancement. #AI #NLProc

Adding 2% parameters can enhance your multilingual language model and turn it into a knowledge base. 

Happy to share our #EMNLP2022 findings paper "Adapters for Enhanced Modeling of Multilingual Knowledge and Text", which uses a set of adapters for the enhancement. #AI #NLProc
Leo Du (@leoduw) 's Twitter Profile Photo

Did you know that, when you add up the probability your language model gives every possible string, you might come up with less than 1 ?! 🤯 Turns out, this can happen! 1/n

Zhaofeng Wu @ ICLR (@zhaofeng_wu) 's Twitter Profile Photo

Language models show impressive performance on a wide variety of tasks, but are they overfitting to evaluation instances and specific task instantiations seen in their pretraining? How much of this performance represents general task/reasoning abilities? 1/4

Language models show impressive performance on a wide variety of tasks, but are they overfitting to evaluation instances and specific task instantiations seen in their pretraining?  How much of this performance represents general task/reasoning abilities?
1/4
Afra Amini (@afra_amini) 's Twitter Profile Photo

Are you a big fan of structure? Have you ever wanted to apply the latest and greatest large language model out-of-the-box to parsing? Are you a secret connoisseur of linear-time dynamic programs? If you answered yes, our outstanding #ACL2023NLP paper may be just right for you!

Are you a big fan of structure?

Have you ever wanted to apply the latest and greatest large language model out-of-the-box to parsing?

Are you a secret connoisseur of linear-time dynamic programs?

If you answered yes, our outstanding #ACL2023NLP paper may be just right for you!
Ethan Gotlieb Wilcox (@wegotlieb) 's Twitter Profile Photo

Thank you to #EMNLP2023 chairs for the 😱 two 😱 outstanding paper awards! I am so grateful to have worked on these projects with wonderful colleagues — Tiago Pimentel (who is the first author on one of the papers!), Clara Isabel Meister, Kyle Mahowald and @ryandcotterell

Thank you to #EMNLP2023 chairs for the 😱 two 😱 outstanding paper awards! I am so grateful to have worked on these projects with wonderful colleagues — <a href="/tpimentelms/">Tiago Pimentel</a> (who is the first author on one of the papers!), <a href="/clara__meister/">Clara Isabel Meister</a>, <a href="/kmahowald/">Kyle Mahowald</a> and @ryandcotterell
Songlin Yang (@songlinyang4) 's Twitter Profile Photo

Data-dependent decay and state dimension expansion are the key for Mamba/GLA matching Transformers!🚀 Also excited to present my NeurIPS spotlight paper [arxiv.org/abs/2311.04823] this Wednesday, which also shows the crucial role of data-dependent decay. Come and chat about RNNs!

Zhaofeng Wu @ ICLR (@zhaofeng_wu) 's Twitter Profile Photo

💡We find that models “think” 💭 in English (or in general, their dominant language) when processing distinct non-English or even non-language data types 🤯 like texts in other languages, arithmetic expressions, code, visual inputs, & audio inputs ‼️ 🧵⬇️arxiv.org/abs/2411.04986

💡We find that models “think” 💭 in English (or in general, their dominant language) when processing distinct non-English or even non-language data types 🤯 like texts in other languages, arithmetic expressions, code, visual inputs, &amp; audio inputs ‼️ 🧵⬇️arxiv.org/abs/2411.04986
Jirui Qi (@jirui_qi) 's Twitter Profile Photo

[1/8] Seeking a faster approach to assess prompt quality in RAG QA? Our latest work may be a good fit for you. We find that prompt quality can be measured using question likelihoods, a computation that’s parallelizable on the input side of LLMs! 📄arxiv.org/abs/2411.07773 #NLProc

[1/8] Seeking a faster approach to assess prompt quality in RAG QA? Our latest work may be a good fit for you. We find that prompt quality can be measured using question likelihoods, a computation that’s parallelizable on the input side of LLMs!
📄arxiv.org/abs/2411.07773
#NLProc