Ariel Goldstein (@goldsteinyariel) 's Twitter Profile
Ariel Goldstein

@goldsteinyariel

ID: 745515944097701889

calendar_today22-06-2016 07:17:06

48 Tweet

113 Followers

376 Following

Hasson Lab (@hassonlab) 's Twitter Profile Photo

We're posting this preprint alongside another Hasson Lab manuscript led by Zaid Zada and Sam Nastase where we use large language models to capture brain-to-brain linguistic coupling during dyadic conversationsโ€”check it out: doi.org/10.1101/2023.0โ€ฆ x.com/samnastase/staโ€ฆ

Roee Neuman๐Ÿณ๏ธโ€โšง๏ธ๐Ÿ‡ฎ๐Ÿ‡ฑ๐Ÿณ๏ธโ€๐ŸŒˆ (@neumanroee) 's Twitter Profile Photo

ื–ื” ื”ืกืœืคื™ ืฉืœ ืื—ืจื™ ื”ืขื‘ืจืช ื—ื•ืง ื”ืกื‘ื™ืจื•ืช. ื‘ืื•ืชื” ืฉืขื” ื‘ื›ื ืกืช ืขืžื“ื• ืืœื•ืคื™ื ื‘ื™ื—ื“ ืขื ื”ืจืžื˜ื›ืดืœ ื‘ื›ื ืกืช ืœืื—ืจ ืฉื”ืชื—ื ื ื• ืœื“ื‘ืจ ืื™ืชื, ืœื”ืกื‘ื™ืจ ืœื”ื ืฉืžื“ื•ื‘ืจ ื‘ื ื–ืง ื‘ื˜ื—ื•ื ื™ ืขืฆื•ื. ื”ื ืœื ื”ืกื›ื™ืžื• ืืคื™ืœื• ืœืฉืžื•ืข ืื•ืชื. ืชืจืื• ืืช ื”ื—ื™ื•ื›ื™ื ืฉืœ ื”ืื ืฉื™ื ื”ืืœื” ืฉื”ื‘ื™ื ืœืžื—ื“ืœ ืฉืœ ืœืžืขืœื” ืž1600 ื ืจืฆื—ื™ื ื•ื—ื˜ื•ืคื™ื. ืืจื•ืจื™ื ืืชื, ืืจื•ืจื™ื ืœืขื•ืœื

ื–ื” ื”ืกืœืคื™ ืฉืœ ืื—ืจื™ ื”ืขื‘ืจืช ื—ื•ืง ื”ืกื‘ื™ืจื•ืช. ื‘ืื•ืชื” ืฉืขื” ื‘ื›ื ืกืช ืขืžื“ื• ืืœื•ืคื™ื ื‘ื™ื—ื“ ืขื ื”ืจืžื˜ื›ืดืœ ื‘ื›ื ืกืช ืœืื—ืจ ืฉื”ืชื—ื ื ื• ืœื“ื‘ืจ ืื™ืชื, ืœื”ืกื‘ื™ืจ ืœื”ื ืฉืžื“ื•ื‘ืจ ื‘ื ื–ืง ื‘ื˜ื—ื•ื ื™ ืขืฆื•ื. ื”ื ืœื ื”ืกื›ื™ืžื• ืืคื™ืœื• ืœืฉืžื•ืข ืื•ืชื. 
ืชืจืื• ืืช ื”ื—ื™ื•ื›ื™ื ืฉืœ ื”ืื ืฉื™ื ื”ืืœื” ืฉื”ื‘ื™ื ืœืžื—ื“ืœ ืฉืœ ืœืžืขืœื” ืž1600 ื ืจืฆื—ื™ื ื•ื—ื˜ื•ืคื™ื. 
ืืจื•ืจื™ื ืืชื, ืืจื•ืจื™ื ืœืขื•ืœื
Eric Ham (@ericham4) 's Twitter Profile Photo

In a new version of our (Mariano Schain, Ariel Goldstein, Hasson Lab) paper (arxiv.org/abs/2310.07106) we show evidence that the layered hierarchy of large language models (LLMs) may be used to model the temporal dynamics of language comprehension in the brain...

Amir Taubenfeld (@taubenfeldamir) 's Twitter Profile Photo

How do the inherent biases in Large Language Models impact their capacity for accurately simulating human behavior? We (yanivdover, Roi Reichart, Ariel Goldstein) try to answer this question in our new paper arxiv.org/abs/2402.04049 1/9

Gabriel Stanovsky (@gabistanovsky) 's Twitter Profile Photo

๐Ÿง  Can LLMs truly understand text, or are they mere ๐Ÿฆœ? We argue that this debate implicitly revolves around the function of consciousness, an age-old question which almost by definition cannot be answered via objective measurement. arxiv.org/pdf/2403.00499โ€ฆ w/Ariel Goldstein

Gabriel Stanovsky (@gabistanovsky) 's Twitter Profile Photo

To demonstrate this, imagine a โ€œzombieโ€ chatbot Z. It is completely open-source (ร  la OLMo), and it excels in all current and future โ€œunderstandingโ€ benchmarks, devoid of any subjective experience. Would you consider it intelligent or capable of understanding?

Gabriel Stanovsky (@gabistanovsky) 's Twitter Profile Photo

We show that seminal works in AI would likely consider Z incapable of understanding because, regardless of its external success, it still lacks internal experience. Instead, we propose to disentangle โ€œunderstandingโ€, โ€œintelligenceโ€, โ€œempathyโ€, and other human traits that...

Gabriel Stanovsky (@gabistanovsky) 's Twitter Profile Photo

supposedly require consciousness, from functional understanding, which measures objective success on given instances. This distinction yields vastly different research agendas: working on functional understanding implies hill-climbing on increasingly challenging...

Gabriel Stanovsky (@gabistanovsky) 's Twitter Profile Photo

datasets and tasks, while research in machine consciousness may aim to imbue machines with internal experience or to measure whether it somehow emerged in them.

Gili Lior (@gililior) 's Twitter Profile Photo

๐Ÿง ๐Ÿค– Does learning in the brain inherently require plasticity? In our latest paper we question this assumption, by leveraging insights into how LLMs "learn". Check out this thread for more details! biorxiv.org/content/10.110โ€ฆ w/ Yuval Shalev Gabriel Stanovsky Ariel Goldstein

๐Ÿง ๐Ÿค– Does learning in the brain inherently require plasticity? In our latest paper we question this assumption, by leveraging insights into how LLMs "learn". Check out this thread for more details! biorxiv.org/content/10.110โ€ฆ
w/ <a href="/YuvalShalev1/">Yuval Shalev</a> <a href="/GabiStanovsky/">Gabriel Stanovsky</a> <a href="/GoldsteinYAriel/">Ariel Goldstein</a>
Yuval Shalev (@yuvalshalev1) 's Twitter Profile Photo

๐Ÿง ๐Ÿค– How do LLMs think? What kind of thought processes can emerge from artificial intelligence? Our latest paper about multi-hop reasoning tasks reveals some new interesting insights. Check out this thread for more details! arxiv.org/abs/2406.13858 Ariel Goldstein Amir Feder

๐Ÿง ๐Ÿค– How do LLMs think? What kind of thought processes can emerge from artificial intelligence? Our latest paper about multi-hop reasoning tasks reveals some new interesting insights. Check out this thread for more details! arxiv.org/abs/2406.13858 <a href="/GoldsteinYAriel/">Ariel Goldstein</a> <a href="/amir_feder/">Amir Feder</a>
Daria Lioubashevski (@darialioub) 's Twitter Profile Photo

๐Ÿ“ขPaper release๐Ÿ“ข What computation is the Transformer performing in the layers after the top-1 becomes fixed (a so called "saturation event")? We show that the next highest-ranked tokens also undergo saturation *in order* of their ranking. Preprint: arxiv.org/abs/2410.20210 1/4

๐Ÿ“ขPaper release๐Ÿ“ข
What computation is the Transformer performing in the layers after the top-1 becomes fixed (a so called "saturation event")? We show that the next highest-ranked tokens also undergo saturation *in order* of their ranking.
Preprint:  arxiv.org/abs/2410.20210
1/4
Daria Lioubashevski (@darialioub) 's Twitter Profile Photo

Our findings reveal that this is true across language, vision, and speech models and across architecture variants: decoder-only, encoder-only, and full-Transformer. It even occurs in untrained Transformer models! ๐Ÿคฏ 2/4

Our findings reveal that this is true across language, vision, and  speech models and across architecture variants: decoder-only, encoder-only, and full-Transformer. It even occurs in untrained Transformer models! ๐Ÿคฏ
2/4
Daria Lioubashevski (@darialioub) 's Twitter Profile Photo

We propose an underlying task transition mechanism where each task corresponds to determining the k-th ranking token. By understanding these transitions, we can predict the current task from hidden layers representations and cause the model to switch between tasks! 3/4

We propose an underlying task transition mechanism where each task corresponds to determining the k-th ranking token. By understanding these transitions, we can predict the current task from hidden layers representations and cause the model to switch between tasks! 
3/4
Daria Lioubashevski (@darialioub) 's Twitter Profile Photo

๐ŸŽฏ Finally, we leverage these insights to introduce a new token-level early-exit strategy that beats existing methods in balancing performance and efficiency. More accurate predictions and faster modelsโ€”win-win! Joint work with Ariel Goldstein Tomer Gabriel Stanovsky 4/4

Refael Tikochinski (@r_tikochinski) 's Twitter Profile Photo

Very excited to share our new paper published in Nature Communications Nature Communications (link below). This work is part of my PhD research under the supervision of Roi Reichart (Technion), @HassonUri (Hasson Lab), and Ariel Goldstein, in collaboration with Yoav Meiri๐Ÿ‡ฎ๐Ÿ‡ฑ.

Amir Taubenfeld (@taubenfeldamir) 's Twitter Profile Photo

New Preprint ๐ŸŽ‰ LLM self-assessment unlocks efficient decoding โœ… Our Confidence-Informed Self-Consistency (CISC) method cuts compute without losing accuracy. We also rethink confidence evaluation & contribute to the debate on self-verification. arxiv.org/abs/2502.06233 1/8๐Ÿ‘‡

New Preprint ๐ŸŽ‰

LLM self-assessment unlocks efficient decoding โœ…

Our Confidence-Informed Self-Consistency (CISC) method cuts compute without losing accuracy.

We also rethink confidence evaluation &amp; contribute to the debate on self-verification.

arxiv.org/abs/2502.06233
1/8๐Ÿ‘‡