Huda Khayrallah

842 posts

Huda Khayrallah

Huda Khayrallah

@HudaKhay

Machine Translation/#NLProc/ML Researcher at Microsoft. Past: @UCBerkeley CS ugrad; @LiltHQ research intern; @jhuCLSP/@jhuCompSci PhD

rarely on here; email me Katılım Temmuz 2012
852 Takip Edilen1K Takipçiler
Huda Khayrallah retweetledi
HyoJung Han
HyoJung Han@h__j___han·
Lots of work on cross-lingual alignment encourages multilingual LLMs to generalize knowledge across languages. But this push for uniformity creates a tension: what happens to knowledge that should remain local? We look into this trade-off of transfer and cultural erasure:🧵
HyoJung Han tweet media
English
3
19
61
17.5K
Huda Khayrallah retweetledi
Eleftheria Briakou
Eleftheria Briakou@ebriakou·
🗺️ Are we making our #LLMs multilingual, or anglocentric? Much work brings languages closer to English, but that comes at the cost of crucial #cultural nuance. @h__j___han tackles this trade-off with surgical steering, adapting LLMs to cultural contexts at inference time.
HyoJung Han@h__j___han

Lots of work on cross-lingual alignment encourages multilingual LLMs to generalize knowledge across languages. But this push for uniformity creates a tension: what happens to knowledge that should remain local? We look into this trade-off of transfer and cultural erasure:🧵

English
0
12
51
9K
Huda Khayrallah retweetledi
Huda Khayrallah retweetledi
Suzanna Sia
Suzanna Sia@suzyahyah·
Large Model Inference Efficiency can be tackled from many angles, mixture of experts, efficient self-attention, quantisation, distillation, hardware acceleration.. But what if we could completely avoid redundant computational processing over context window? In our NeurIPS'24 paper "where does in-context (task-location) learning happen" We find three distinct regions for LLM inference time processing 1️⃣ [Task Location]; LLM discovers the task from reading instructions and examples 2️⃣ [Task Processing]; After task location, the model no longer requires any self-attention over the prompts. 3️⃣ [Task Completion]; final layers of processing where the model no longer requires self-attention over the query. ===> Implications for Industry ✅ ~50% In Computational savings (theoretical) If we avoided redundant context processing in later layers of the model ✅ Very sample efficient adaptation of LLMs to task specific Models. Contrary to common wisdom on Fine-tuning, LoRA layers are most effective at earlier layers of the model compared to the later ones. ===> Implications for Academia: * New Interpretability technique progressively masks out all self-attention to the context, * Task Location layer is not affected by the number of prompt examples provided to the model. * Related Work with similar findings are Task Vectors (@RoeeHendel et al) , Function Vectors (@ericwtodd et al), providing additional supporting evidence for this phenomena. 💻 Paper: lnkd.in/gduKF27X Github: lnkd.in/g_q6T2kE Models: Llama3.1-8B, LLama3.1-8B-Instruct, Starcoder2-7B, GPTN2.7B, Bloom3B Tasks: Machine Translation (en-fr, fr-en, en-pt), Code Generation (en-py)
Suzanna Sia tweet mediaSuzanna Sia tweet mediaSuzanna Sia tweet media
English
1
11
26
7.5K
Huda Khayrallah retweetledi
Barry Haddow
Barry Haddow@bazril·
EAMT best thesis award - closes on January 31st. Completed an MT-related PhD in 2024? In Europe, Africa or Middle East. Then why not submit your thesis. eamt.org/2024/11/28/the…
English
0
3
5
546
Huda Khayrallah retweetledi
Eleftheria Briakou
Eleftheria Briakou@ebriakou·
I’m super thrilled to have won the AMTA Best Thesis Award!! A huge thanks to the AMTA organizers for this recognition ☺️ See you all in Chicago amtaweb.org
Eleftheria Briakou tweet media
English
10
9
94
11.3K
Huda Khayrallah retweetledi
Akiko I. Eriguchi
Akiko I. Eriguchi@akikoe_·
On behalf of the AMTA Board of Directors, I am pleased to announce the winner of the first-ever AMTA Best Thesis Award: Dr. Eleftheria Briakou (@ebriakou) for her thesis “Detecting Fine-Grained Semantic Divergences to Improve Translation Understanding Across Languages”. [1/n]
Eleftheria Briakou@ebriakou

I’m super thrilled to have won the AMTA Best Thesis Award!! A huge thanks to the AMTA organizers for this recognition ☺️ See you all in Chicago amtaweb.org

English
1
7
17
2.8K
Huda Khayrallah retweetledi
Jordan Boyd-Graber
Jordan Boyd-Graber@boydgraber·
I'm bummed that family obligations prevented me from presenting this epic paper. This work represented a long journey for me. I first began working on the language of Diplomacy in 2015, and I struggled for years to get funding to build a bot that could play it ...
Jordan Boyd-Graber tweet media
Joy Wongkamjan@joywwong

1⃣Meta’s Cicero by Bakhtin et al. was the talk of town when it was released in November 2022. Even main journals published that Cicero had achieved human-level Diplomacy in strategy and negotiation. Well, had it!? Our paper had this answer: arxiv.org/pdf/2406.04643

English
4
10
69
10.3K
Huda Khayrallah retweetledi
Sarah Jabbour
Sarah Jabbour@SarahJabbour_·
My mom wants to come out of retirement. She was a software validation engineer working on human machine interfaces. She (and I) have no idea where to look. She just wants to spend time testing the things that people build. Does anyone know where she could look??
English
1
2
7
1.5K
Huda Khayrallah retweetledi
HyoJung Han
HyoJung Han@h__j___han·
✨XLAVS-R will be presented during today’s (August 13th) #ACL2024 poster session 4, starting at 10:30 AM. Looking forward to talking with people interested in our work!
HyoJung Han@h__j___han

🗣️XLAVS-R is accepted at #ACL2024 main! 🚀🚀 We present XLAVS-R, a cross-lingual audio-visual model for noise-robust speech perception in over 100 languages. Very happy to present our work done during my @AIatMeta internship with @ChanghanWang. arxiv.org/abs/2403.14402

English
0
5
26
2K
Huda Khayrallah retweetledi
Naomi Saphra
Naomi Saphra@nsaphra·
NAACL is this week and that means you should read our "history" paper! And if you're in Mexico City then say hi to Eve Fleisig, who is presenting it!
Naomi Saphra@nsaphra

It's not the first time! A dream team of @enfleisig (human eval expert), Adam Lopez (remembers the Stat MT era), @kchonyc (helped end it), and me (pun in title) are here to teach you the history of scale crises and what lessons we can take from them. 🧵arxiv.org/abs/2311.05020

English
1
4
41
6.6K
Huda Khayrallah retweetledi
Armita R. Manafzadeh
Armita R. Manafzadeh@armanafzadeh·
Three postdocs were too tired to go to the party on the last night of SICB this year, so we decided to order pizza to the hotel and write a paper together instead. Out in @ICB_journal now! academic.oup.com/icb/advance-ar…
Andrew K. Schulz, Ph.D.@SchulzScience_

Today in @ICB_journal - @armanafzadeh, Janneke Schwaner, and I co-led a brief article on Strategies for Organizing Interdisciplinary Events. @SICB_ @SICB_DCB_DVM If you are interested in hosting an interdisciplinary event, we hope it is helpful: doi.org/10.1093/icb/ic…

English
2
8
46
4.8K
Huda Khayrallah retweetledi
Elias Stengel-Eskin
Elias Stengel-Eskin@EliasEskin·
🚨 Excited to share our new work on **confidence calibration** in LLMs! LLMs are often badly calibrated & overconfident, explicitly (eg. "I'm 100% sure") and implicitly, eg. giving details/authoritative tone. We address both w/ a pragmatic speaker-listener multi-agent method 🧵
Elias Stengel-Eskin tweet media
English
3
42
144
34.4K
Huda Khayrallah
Huda Khayrallah@HudaKhay·
Deadline is 6/6 for the AMTA thesis award Apply if you finished a PhD in MT in the Americas in the last year! amtaweb.org/amta-2024-cfp-… questions? reach out to mtresearchers@amtaweb.org (Rebecca Knowles and Akiko Eriguchi).
Akiko I. Eriguchi@akikoe_

🏆 Thrilled to share the launch of the AMTA Best Thesis Award, which aims to highlight the achievements of a recent PhD graduate at an institution in the Americas whose thesis has focused on topics related to machine translation. [1/2]

English
0
0
1
337