dilara (@dilarafsoylu) 's Twitter Profile
dilara

@dilarafsoylu

member of cooking staff @StanfordNLP

ID: 1485454823281463297

calendar_today24-01-2022 03:33:37

63 Tweet

288 Followers

1,1K Following

John Hewitt (@johnhewtt) 's Twitter Profile Photo

I’m joining the Columbia Computer Science faculty as an assistant professor in fall 2025, and hiring my first students this upcoming cycle!! There’s so much to understand and improve in neural systems that learn from language — come tackle this with me!

I’m joining the Columbia Computer Science faculty as an assistant professor in fall 2025, and hiring my first students this upcoming cycle!!

There’s so much to understand and improve in neural systems that learn from language — come tackle this with me!
Rose (@rose_e_wang) 's Twitter Profile Photo

We talk a lot about the potential of AI for applications, like AI for Education. But actual progress requires that we hill-climb on realistic, hard tasks. Are there any? 🔽 Bridge, Backtracing, and Teacher Coach are 3 real-world AI for Education datasets that are far from

John Thickstun (@jwthickstun) 's Twitter Profile Photo

I'm joining Cornell University this fall as an Assistant Professor of Computer Science! Looking forward to work with students and colleagues Cornell Computer Science, @cornellCIS on generative models, controllable generation, and creative applications like #musictechnology

Christopher Potts (@chrisgpotts) 's Twitter Profile Photo

An LLM memorization riddle: A Pythia-6.9B checkpoint generates the following Output, which occurs only 1 time in the Pile. Is this a verbatim memorization?

An LLM memorization riddle: A Pythia-6.9B checkpoint generates the following Output, which occurs only 1 time in the Pile. Is this a verbatim memorization?
Jordan Juravsky (@jordanjuravsky) 's Twitter Profile Photo

Do you like LLMs? Do you also like for loops? Then you’ll love our new paper! We scale inference compute through repeated sampling: we let models make hundreds or thousands of attempts when solving a problem, rather than just one. By simply sampling more, we can boost LLM

Do you like LLMs? Do you also like for loops? Then you’ll love our new paper!

We scale inference compute through repeated sampling: we let models make hundreds or thousands of attempts when solving a problem, rather than just one. By simply sampling more, we can boost LLM
Leonie Weissweiler (@laweissweiler) 's Twitter Profile Photo

Come see #ACL2024‘s most beautiful poster, being presented by Julie Kallini ✨ right now at poster 7! Refuting Chomsky‘s NYT Op-Ed with experimental work as an added bonus ✨🌈

Come see #ACL2024‘s most beautiful poster, being presented by <a href="/JulieKallini/">Julie Kallini ✨</a> right now at poster 7!
Refuting Chomsky‘s NYT Op-Ed with experimental work as an added bonus ✨🌈
Karel D’Oosterlinck (@kareldoostrlnck) 's Twitter Profile Photo

Aligning Language Models with preferences leads to stronger and safer models (GPT3 → ChatGPT). However, preferences (RLHF) contain irrelevant signals, and alignment objectives (e.g. DPO) can actually hurt model performance. We tackle both, leading to a ~2x performance boost.

Aligning Language Models with preferences leads to stronger and safer models (GPT3 → ChatGPT). However, preferences (RLHF) contain irrelevant signals, and alignment objectives (e.g. DPO) can actually hurt model performance.

We tackle both, leading to a ~2x performance boost.
Ahmet Üstün (@ahmetustun89) 's Twitter Profile Photo

I'm incredibly proud that Aya received #ACL2024 Best Paper Award 🥹. Huge congratulations to the Aya team and Cohere For AI community who make this possible by for extending frontiers of LLMs to multilingual, building Aya Model and Aya Dataset 🌿🌏

I'm incredibly proud that Aya received #ACL2024 Best Paper Award 🥹. 

Huge congratulations to the Aya team and <a href="/CohereForAI/">Cohere For AI</a> community who make this possible by for extending frontiers of LLMs to multilingual, building Aya Model and Aya Dataset 🌿🌏
Weiyan Shi (@shi_weiyan) 's Twitter Profile Photo

🤩So honored to receive TWO paper awards ACL 2025, huge shoutout to my amazing collaborators🤩!!! 🏆Best Social Impact Paper🏆: persuasive jailbreaker arxiv.org/abs/2401.06373 🏆Oustanding Paper🏆: persuasive misinformation arxiv.org/abs/2312.09085 #ACL2024

🤩So honored to receive TWO paper awards <a href="/aclmeeting/">ACL 2025</a>, huge shoutout to my amazing collaborators🤩!!!

🏆Best Social Impact Paper🏆: persuasive jailbreaker arxiv.org/abs/2401.06373

🏆Oustanding Paper🏆: persuasive misinformation arxiv.org/abs/2312.09085

#ACL2024
Stanford NLP Group (@stanfordnlp) 's Twitter Profile Photo

.Stanford NLP Group awards at #ACL2024 ▸ Best paper award Julie Kallini ✨ et al ▸ Outstanding paper award Aryaman Arora et al ▸ Outstanding paper award Weiyan Shi et al ▸ Best societal impact award Weiyan Shi et al ▸ 10 year test of time award Christopher Manning et al Congratulations! 🥂

.<a href="/stanfordnlp/">Stanford NLP Group</a> awards at #ACL2024
▸ Best  paper award
<a href="/JulieKallini/">Julie Kallini ✨</a> et al
▸ Outstanding paper award
<a href="/aryaman2020/">Aryaman Arora</a> et al
▸ Outstanding paper award
<a href="/shi_weiyan/">Weiyan Shi</a> et al
▸ Best societal impact award
<a href="/shi_weiyan/">Weiyan Shi</a> et al
▸ 10 year test of time award
<a href="/chrmanning/">Christopher Manning</a> et al
Congratulations! 🥂
Omar Khattab (@lateinteraction) 's Twitter Profile Photo

Some personal news: I'm thrilled to have joined @Databricks Databricks Mosaic Research as a Research Scientist last month, before I start as MIT faculty in July 2025! Expect increased investment into the open-source DSPy community, new research, & strong emphasis on production concerns 🧵.

Christopher Potts (@chrisgpotts) 's Twitter Profile Photo

The Linear Representation Hypothesis is now widely adopted despite its highly restrictive nature. Here, Csordás Róbert, Atticus Geiger, Christopher Manning & I present a counterexample to the LRH and argue for more expressive theories of interpretability: arxiv.org/abs/2408.10920

Christopher Potts (@chrisgpotts) 's Twitter Profile Photo

A short story of fast progress: NVIDIA released an ≈8B parameter model they called Megatron in 2019, and five years later they have released an ≈8B model they call Minitron. (I did round off an entire BERT-large for the 2019 model.)

Tristan Thrush (@tristanthrush) 's Twitter Profile Photo

Do you want to select great LLM pretraining data but don’t have 1000 H100s for a ton of mixture experiments? What about a method that requires none of your own training, matches the best known existing method, and has some nice theory? New preprint: Perplexity Correlations

Do you want to select great LLM pretraining data but don’t have 1000 H100s for a ton of mixture experiments?

What about a method that requires none of your own training, matches the best known existing method, and has some nice theory?

New preprint: Perplexity Correlations
John Hewitt (@johnhewtt) 's Twitter Profile Photo

If I finetune my LM just on responses, without conditioning on instructions, what happens when I test it with an instruction? Or if I finetune my LM just to generate poems from poem titles? Either way, the LM will roughly follow new instructions! Paper: arxiv.org/pdf/2409.14254

If I finetune my LM just on responses, without conditioning on instructions, what happens when I test it with an instruction?

Or if I finetune my LM just to generate poems from poem titles?

Either way, the LM will roughly follow new instructions!

Paper: arxiv.org/pdf/2409.14254