Zeming Chen (@eric_zemingchen) 's Twitter Profile
Zeming Chen

@eric_zemingchen

PhD Candidate, NLP Lab @EPFL; Previously: Research Intern @ Meta AI (FAIR) @allen_ai #AI #ML #NLP

ID: 1411700324310687746

linkhttps://eric11eca.github.io calendar_today04-07-2021 14:56:09

49 Tweet

520 Followers

279 Following

Antoine Bosselut (@abosselut) 's Twitter Profile Photo

Hey #NLProc folks, we had a lot of fun last year, so we're inviting guest lecturers again for our Topics in NLP course during this Fall 2024 semester at EPFL! More information here: t.ly/QMTCA Please share and RT!

Yu Fei (@walter_fei) 's Twitter Profile Photo

Alignment is necessary for LLMs, but do we need to train aligned versions for all model sizes in every model family? ๐Ÿง We introduce ๐Ÿš€Nudging, a training-free approach that aligns any base model by injecting a few nudging tokens at inference time. ๐ŸŒfywalter.github.io/nudging/

Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

๐Ÿšจ New Paper!! How can we train LLMs using 100M words? In our babyLM paper, we introduce a new self-synthesis training recipe to tackle this question! ๐Ÿผ๐Ÿ’ป This was a fun project co-led by me, Yingtian Tang, Abdulkadir Gokce, w/ Hannes Mehrer & Martin Schrimpf ๐Ÿงตโฌ‡๏ธ

๐Ÿšจ New Paper!!

How can we train LLMs using 100M words? In our <a href="/babyLMchallenge/">babyLM</a> paper, we introduce a new self-synthesis training recipe to tackle this question! ๐Ÿผ๐Ÿ’ป

This was a fun project co-led by me, <a href="/yingtian80536/">Yingtian Tang</a>, <a href="/akgokce0/">Abdulkadir Gokce</a>, w/ <a href="/HannesMehrer/">Hannes Mehrer</a> &amp; <a href="/martin_schrimpf/">Martin Schrimpf</a>  

๐Ÿงตโฌ‡๏ธ
Angelika Romanou (@agromanou) 's Twitter Profile Photo

๐Ÿš€ Introducing INCLUDE ๐ŸŒ: A multilingual LLM evaluation benchmark spanning 44 languages! Contains *newly-collected* data, prioritizing *regional knowledge*. Setting the stage for truly global AI evaluation. Ready to see how your model measures up? #AI #Multilingual #LLM #NLProc

๐Ÿš€ Introducing INCLUDE ๐ŸŒ: A multilingual LLM evaluation benchmark spanning 44 languages!
Contains *newly-collected* data, prioritizing *regional knowledge*.

Setting the stage for truly global AI evaluation.
Ready to see how your model measures up?
#AI #Multilingual #LLM #NLProc
Beatriz Borges (@obiwit) 's Twitter Profile Photo

๐Ÿ“˜ Could ChatGPT get an engineering degree? Spoiler, yes! In our new PNASNews article, we explore how AI assistants like GPT-4 perform in STEM university courses โ€” and on average they pass a staggering 91.7% of core courses. ๐Ÿงต #AI #HigherEd #STEM #LLMs #NLProc

๐Ÿ“˜ Could ChatGPT get an engineering degree? Spoiler, yes! In our new <a href="/PNASNews/">PNASNews</a> article, we explore how AI assistants like GPT-4 perform in STEM university courses โ€” and on average they pass a staggering 91.7% of core courses. ๐Ÿงต #AI #HigherEd #STEM #LLMs #NLProc
Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

๐Ÿšจ New Paper! Can neuroscience localizers uncover brain-like functional specializations in LLMs? ๐Ÿง ๐Ÿค– Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks! w/ Greta Tuckute, Antoine Bosselut, & Martin Schrimpf ๐Ÿงต๐Ÿ‘‡

๐Ÿšจ New Paper!

Can neuroscience localizers uncover brain-like functional specializations in LLMs? ๐Ÿง ๐Ÿค–

Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!

w/ <a href="/GretaTuckute/">Greta Tuckute</a>, <a href="/ABosselut/">Antoine Bosselut</a>, &amp; <a href="/martin_schrimpf/">Martin Schrimpf</a> 

๐Ÿงต๐Ÿ‘‡
Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

๐Ÿšจ New Preprint!! LLMs trained on next-word prediction (NWP) show high alignment with brain recordings. But what drives this alignmentโ€”linguistic structure or world knowledge? And how does this alignment evolve during training? Our new paper explores these questions. ๐Ÿ‘‡๐Ÿงต

๐Ÿšจ New Preprint!!

LLMs trained on next-word prediction (NWP) show high alignment with brain recordings. But what drives this alignmentโ€”linguistic structure or world knowledge? And how does this alignment evolve during training? Our new paper explores these questions. ๐Ÿ‘‡๐Ÿงต
Silin Gao (@silin_gao) 's Twitter Profile Photo

NEW PAPER ALERT: Generating visual narratives to illustrate textual stories remains an open challenge, due to the lack of knowledge to constrain faithful and self-consistent generations. Our #CVPR2025 paper proposes a new benchmark, VinaBench, to address this challenge.

NEW PAPER ALERT: Generating visual narratives to illustrate textual stories remains an open challenge, due to the lack of knowledge to constrain faithful and self-consistent generations. Our #CVPR2025 paper proposes a new benchmark, VinaBench, to address this challenge.
Angelika Romanou (@agromanou) 's Twitter Profile Photo

If youโ€™re at ICLR 2026 this week, come check out our spotlight poster INCLUDE during the Thursday 3:00โ€“5:30pm session! I will be there to chat about all things multilingual & multicultural evaluation. Feel free to reach out anytime during the conference. Iโ€™d love to connect!

Badr AlKhamissi (@bkhmsi) 's Twitter Profile Photo

๐ŸšจNew Preprint!! Thrilled to share with you our latest work: โ€œMixture of Cognitive Reasonersโ€, a modular transformer architecture inspired by the brainโ€™s functional networks: language, logic, social reasoning, and world knowledge. 1/ ๐Ÿงต๐Ÿ‘‡

๐ŸšจNew Preprint!!

Thrilled to share with you our latest work: โ€œMixture of Cognitive Reasonersโ€, a modular transformer architecture inspired by the brainโ€™s functional networks: language, logic, social reasoning, and world knowledge.

1/ ๐Ÿงต๐Ÿ‘‡