Sama🌪 (@samahadhod) 's Twitter Profile
Sama🌪

@samahadhod

Bla blabla BLA

ID: 3304960318

calendar_today31-05-2015 18:45:28

4,4K Tweet

640 Followers

743 Following

Youssef (@yzakaria0) 's Twitter Profile Photo

من ساعة ماعرفت يعني ايه overwhelmed وانا حاسس اني overwhelmed بجد الجهل نعمة انا اوفرويلمد اوي بجد بجد

Essam (@3sam001) 's Twitter Profile Photo

يارب اكون في المكان الصح يارب ما يضيع عمري وأنا بحاول في طريق مش طريقي

Alham Fikri Aji (@alhamfikri) 's Twitter Profile Photo

🎉Silver for MBZUAI at ACPC!!🥈 We also topped the Gulf rankings Amazing performance by our 1st year UG students in our very first participation Next: ICPC World Finals 2026🤞

🎉Silver for <a href="/mbzuai/">MBZUAI</a> at ACPC!!🥈
We also topped the Gulf rankings

Amazing performance by our 1st year UG students in our very first participation

Next: ICPC World Finals 2026🤞
Haryo (@haryoaw) 's Twitter Profile Photo

Most culture test benchmark is mostly static, which may lead to data saturation and leakage, hence making the score not reliable to measure the capability of LLMs. Thus, we benchmark these LLMs to play a social deduction game!

Blaise Cruz (@jcblaisecruz) 's Twitter Profile Photo

1/11 Proud to share our new paper: SENSIA (SENse-based Symmetric Interlingual Alignment) — a sense-based approach to multilingual adaptation. Goal: explicit representation-level alignment of meaning.

1/11 Proud to share our new paper: SENSIA (SENse-based Symmetric Interlingual Alignment) — a sense-based approach to multilingual adaptation.

Goal: explicit representation-level alignment of meaning.
Edd (@erla_ndpg) 's Twitter Profile Photo

Sama🌪's paper exposes how LLMs struggle with code even after nailing the plan in competitive programming tasks. New editorial-centric benchmark + released ICPC-style problems. Very very cool ngl

Alham Fikri Aji (@alhamfikri) 's Twitter Profile Photo

In competitive programming, the post-contest discussion is mainly about the algorithm, not the code It’s (mainly) a problem-solving contest, yet LLMs are often benchmarked only as coders We revisited this by evaluating the editorial and reasoning vs end-to-end code generation