Oren Sultan (@oren_sultan) 's Twitter Profile
Oren Sultan

@oren_sultan

AI Researcher @Lightricks, CS PhD Candidate #AI #NLP @HebrewU, advised by @HyadataLab ๐Ÿ‡ฎ๐Ÿ‡ฑ | prev. @TU_Muenchen ๐Ÿ‡ฉ๐Ÿ‡ช @UniMelb ๐Ÿ‡ฆ๐Ÿ‡บ

ID: 1423192726670135300

linkhttp://www.orensultan.com calendar_today05-08-2021 08:02:52

576 Tweet

888 Takipรงi

689 Takip Edilen

Michael Hassid (@michaelhassid) 's Twitter Profile Photo

The longer reasoning LLM thinks - the more likely to be correct, right? Apparently not. Presenting our paper: โ€œDonโ€™t Overthink it. Preferring Shorter Thinking Chains for Improved LLM Reasoningโ€. Link: arxiv.org/abs/2505.17813 1/n

The longer reasoning LLM thinks - the more likely to be correct, right?

Apparently not.

Presenting our paper: โ€œDonโ€™t Overthink it. Preferring Shorter Thinking Chains for Improved LLM Reasoningโ€.

Link: arxiv.org/abs/2505.17813

1/n
Iddo Yosha (@iddoyosha) 's Twitter Profile Photo

๐Ÿšจ Happy to share our #Interspeech2025 paper! "WhiStress: Enriching Transcriptions with Sentence Stress Detection" Sentence stress is a word-level prosodic cue that marks contrast or intent. WhiStress detects it alongside transcriptionโ€”no alignment needed. Paper, code, demo ๐Ÿ‘‡

Noy Sternlicht (@noysternlicht) 's Twitter Profile Photo

๐Ÿšจ New paper! We present CHIMERA โ€” a KB of 28K+ scientific idea recombinations ๐Ÿ’ก It captures how researchers blend concepts or take inspiration across fields, enabling: 1. Meta-science 2. Training models to predict new combos noy-sternlicht.github.io/CHIMERA-Web ๐Ÿ‘‡ Findings & data:

Noy Sternlicht (@noysternlicht) 's Twitter Profile Photo

๐Ÿ”” New Paper! We propose a challenging new benchmark for LLM judges: Evaluating debate speeches. Are they comparable to humans? Well... itโ€™s debatable. ๐Ÿค” noy-sternlicht.github.io/Debatable-Inteโ€ฆ ๐Ÿ‘‡ Here are our findings:

Or Tal (@or__tal) 's Twitter Profile Photo

Which modeling to choose for text-to-music generation? We run a head-to-head comparison to figure it out. Same data, same architecture - AR vs FM. ๐Ÿ‘‡ If you care about fidelity, speed, control, or editing see this thread. ๐Ÿ”—huggingface.co/spaces/ortal16โ€ฆ ๐Ÿ“„arxiv.org/abs/2506.08570 1/6

Which modeling to choose for text-to-music generation?
We run a head-to-head comparison to figure it out.
Same data, same architecture - AR vs FM.
๐Ÿ‘‡ If you care about fidelity, speed, control, or editing see this thread.
๐Ÿ”—huggingface.co/spaces/ortal16โ€ฆ
๐Ÿ“„arxiv.org/abs/2506.08570
1/6
ื ื“ื‘ ื”ืจ-ื˜ื•ื‘ (@nadavhartuv) 's Twitter Profile Photo

๐Ÿšจ New paper alert! PAST: phonetic-acoustic speech tokenizer โ€“ just got accepted to Interspeech 2025 ๐ŸŽ‰ It learns phonetic + acoustic tokens jointly, with no SSL babysitter or external vocoder. ๐Ÿ”—pages.cs.huji.ac.il/adiyoss-lab/PAโ€ฆ ๐Ÿ‘‡ If youโ€™re into speech LMs, keep reading!

๐Ÿšจ New paper alert!
PAST: phonetic-acoustic speech tokenizer โ€“ just got accepted to Interspeech 2025 ๐ŸŽ‰
It learns phonetic + acoustic tokens jointly, with no SSL babysitter or external vocoder.

๐Ÿ”—pages.cs.huji.ac.il/adiyoss-lab/PAโ€ฆ
๐Ÿ‘‡ If youโ€™re into speech LMs, keep reading!
Eliya Habba (@eliyahabba) 's Twitter Profile Photo

Presenting my poster : ๐Ÿ•Š๏ธ DOVE - A large-scale multi-dimensional predictions dataset towards meaningful LLM evaluation, Monday 18:00 Vienna, #ACL2025 Come chat about LLM evaluation, prompt sensitivity, and our 250M COLLECTION OF MODEL OUTPUTS!

Presenting my poster :
๐Ÿ•Š๏ธ DOVE - A large-scale multi-dimensional predictions dataset towards meaningful LLM evaluation, Monday 18:00 Vienna, 
#ACL2025

Come chat about LLM evaluation, prompt sensitivity, and our 250M COLLECTION OF MODEL OUTPUTS!
Asaf Yehudai (@asafyehudai) 's Twitter Profile Photo

๐Ÿšจ Benchmarks tell us which model is better โ€” but not why it fails. For developers, this means tedious, manual error analysis. We're bridging that gap. Meet CLEAR: an open-source tool for actionable error analysis of LLMs. ๐Ÿงต๐Ÿ‘‡

๐Ÿšจ Benchmarks tell us which model is better โ€” but not why it fails.

For developers, this means tedious, manual error analysis. We're bridging that gap.

Meet CLEAR: an open-source tool for actionable error analysis of LLMs.

๐Ÿงต๐Ÿ‘‡
David Dinkevich (@daviddinkevich) 's Twitter Profile Photo

[1/6] ๐ŸŽฌ New paper: Story2Board We guide diffusion models to generate consistent, expressive storyboards--no training needed. By mixing attention-aligned tokens across panels, we reinforce character identity without hurting layout diversity. ๐ŸŒ daviddinkevich.github.io/Story2Board

[1/6] ๐ŸŽฌ New paper: Story2Board
We guide diffusion models to generate consistent, expressive storyboards--no training needed.
By mixing attention-aligned tokens across panels, we reinforce character identity without hurting layout diversity.
๐ŸŒ daviddinkevich.github.io/Story2Board