
Gal Yona
@_galyo
Research scientist @googleai, previously CS PhD @weizmannscience
ID: 86293812
https://galyona.github.io/ 30-10-2009 11:45:44
284 Tweet
474 Followers
486 Following

Excited to attend EMNLP 2025 in Miami next week ๐คฉ DM me if you'd like to grab a coffee and chat about interpretability, knowledge, or reasoning in LLMs! Our group/collabs will be presenting a bunch of cool works, come check them out! ๐งต


What prompt generated the image on the right? Come find out today at our tutorial on OOD generalization: Shortcuts, Spuriousness, and Stability @Maggiemakar aahlad puli Panel: Elan Rosenfeld Aditi Raghunathan Danica Sutherland


During the process of writing AI Engineering, I went through so many papers, case studies, blog posts, repos, tools, etc. This repo contains ~100 resources that really helped me understand various aspects of building with foundation models. github.com/chiphuyen/aie-โฆ Here are the





.Percy Liang & Tatsunori Hashimoto start the 2nd offering of CS336 Language Modeling from Scratch at Stanford NLP Group. The class philosophy is Understanding by Building. We need many people who understand the detailed design of modern LLMs, not just a few at โfrontierโ ๐คญ AI companies.



Sam Altman the single biggest thing you could do for safety/alignment is to put a massive emphasis in the RL feedback loop on basic HONESTY and never misleading, tricking, overstating, exaggerating, etc. It should be like touching a hot stove to the model. Just like how you raise kids


ืืฆื ืืืฉืืจื: ืืฉืชืืฉื ืืฆ'ื GPT ืฉืืืฆืื ืขืืืจื ืืืง ืืืฉ ืขื ืื ืช ืื ืฆื ืืืืื ืืืืจืืช ืคืืืคืื ืืืืื ืืืืช ืืฉืคื ืืฉืืื ืืืืจื. ืืฉืืคื ืืื ืืืื ืืฉืืืืจ ืืชืืื: "30 ืฉื ื ืื ื ืฉืืคื ืืืฉืืชื ืฉืจืืืชื ืืื. ืื ืจืื ืฉืืขืืชื"


new work by Gabrielle Kaili-May Liu shows that LLMs still struggle to faithfully express their uncertainty in words, but cool to see that meta cognitive inspired prompting can go a long way. looking forward to seeing more positive results on this fundamental problem!