Ariel Goldstein

56 posts

Ariel Goldstein

Ariel Goldstein

@GoldsteinYAriel

Katılım Haziran 2016
377 Takip Edilen161 Takipçiler
Ariel Goldstein retweetledi
Daria Lioubashevski
Daria Lioubashevski@DariaLioub·
🚨 New preprint! One idea, many ways to say it – does your brain track those options before you speak? Using LLMs, we put this to the test: biorxiv.org/content/10.110… We show for the 1st time that the brain represents many alternatives simultaneously in both listening & speaking 🧵
GIF
English
1
21
49
2.5K
Ariel Goldstein
Ariel Goldstein@GoldsteinYAriel·
Most novel insight: When an LLM recognizes a bias, it’s actually less likely to exhibit it. This helps explain conflicting results in prior studies—each was seeing a subset of the broader pattern.
English
0
0
0
114
Ariel Goldstein
Ariel Goldstein@GoldsteinYAriel·
New paper out! Do LLMs Exhibit Human-Like Cognitive Biases? We systematically tested 20 biases across 10 leading LLMs using pre-registered experiments. The result: cognitive biases in LLMs are both pervasive and nuanced. papers.ssrn.com/sol3/papers.cf…
English
2
4
8
322
Ariel Goldstein retweetledi
Amir Taubenfeld
Amir Taubenfeld@TaubenfeldAmir·
New Preprint 🎉 LLM self-assessment unlocks efficient decoding ✅ Our Confidence-Informed Self-Consistency (CISC) method cuts compute without losing accuracy. We also rethink confidence evaluation & contribute to the debate on self-verification. arxiv.org/abs/2502.06233 1/8👇
Amir Taubenfeld tweet media
English
1
19
56
3.9K
Ariel Goldstein
Ariel Goldstein@GoldsteinYAriel·
@yoavgo Second, drawing on our human experience where sometimes (we feel) that humans who talk about an issue/problem can reach a conclusion that separately they couldn't have - worth mimicking no?
English
0
0
0
13
(((ل()(ل() 'yoav))))👾
let's talk about "agents" (in the LLM sense). there's a lot of buzz around "multi-agent" systems where agents collaborate but... i don't really get how it differs from a thinking of a single agent with multiple modes of operation. what are the benefits of modeling as multi-agent?
English
66
18
386
79.2K
Ariel Goldstein
Ariel Goldstein@GoldsteinYAriel·
@yoavgo First, if you are interested in verbal-based interactions, artificial agent systems serve as a "lab" for different hypotheses about language-based interactions.
English
0
0
0
10
Ariel Goldstein
Ariel Goldstein@GoldsteinYAriel·
@AsafYehudai @marcel_binz @huggingface All LLMs' embeddings (that I tested) correlate with brain. It is interesting to see if it explain more. However, predicting behavior using LLMs is more challenging IMO.
English
0
0
1
22
Marcel Binz
Marcel Binz@marcel_binz·
Excited to announce Centaur -- the first foundation model of human cognition. Centaur can predict and simulate human behavior in any experiment expressible in natural language. You can readily download the model from @huggingface and test it yourself: huggingface.co/marcelbinz/Lla…
English
40
238
1.2K
235.6K
Ariel Goldstein retweetledi
Daria Lioubashevski
Daria Lioubashevski@DariaLioub·
🎯 Finally, we leverage these insights to introduce a new token-level early-exit strategy that beats existing methods in balancing performance and efficiency. More accurate predictions and faster models—win-win! Joint work with @GoldsteinYAriel @TomerSchlank @GabiStanovsky 4/4
English
0
1
4
272
Ariel Goldstein retweetledi
Daria Lioubashevski
Daria Lioubashevski@DariaLioub·
We propose an underlying task transition mechanism where each task corresponds to determining the k-th ranking token. By understanding these transitions, we can predict the current task from hidden layers representations and cause the model to switch between tasks! 3/4
Daria Lioubashevski tweet media
English
1
1
3
271
Ariel Goldstein retweetledi
Daria Lioubashevski
Daria Lioubashevski@DariaLioub·
Our findings reveal that this is true across language, vision, and speech models and across architecture variants: decoder-only, encoder-only, and full-Transformer. It even occurs in untrained Transformer models! 🤯 2/4
Daria Lioubashevski tweet media
English
1
1
3
267
Ariel Goldstein retweetledi
Daria Lioubashevski
Daria Lioubashevski@DariaLioub·
📢Paper release📢 What computation is the Transformer performing in the layers after the top-1 becomes fixed (a so called "saturation event")? We show that the next highest-ranked tokens also undergo saturation *in order* of their ranking. Preprint: arxiv.org/abs/2410.20210 1/4
Daria Lioubashevski tweet media
English
2
13
29
2.9K
Ariel Goldstein retweetledi
Yuval Shalev
Yuval Shalev@YuvalShalev1·
🧠🤖 How do LLMs think? What kind of thought processes can emerge from artificial intelligence? Our latest paper about multi-hop reasoning tasks reveals some new interesting insights. Check out this thread for more details! arxiv.org/abs/2406.13858 @GoldsteinYAriel @amir_feder
Yuval Shalev tweet media
English
1
8
12
1.2K