Ivan Vulić

206 posts

Ivan Vulić banner
Ivan Vulić

Ivan Vulić

@licwu

Research Prof@Cambridge; Interested in (way) too many things, but mostly (and rarely) (re)tweets about NLP, ML, IR, language(s); (likes parentheses)

Katılım Kasım 2017
329 Takip Edilen2.2K Takipçiler
Ivan Vulić retweetledi
Lucas Caccia
Lucas Caccia@LucasPCaccia·
RAG and in-context learning are the go-to approaches for integrating new knowledge into LLMs, making inference very inefficient We propose instead 𝗞𝗻𝗼𝘄𝗹𝗲𝗱𝗴𝗲 𝗠𝗼𝗱𝘂𝗹𝗲𝘀 : lightweight LoRA modules trained offline that can match RAG performance without the drawbacks
GIF
English
1
12
45
8.2K
Ivan Vulić retweetledi
Han Zhou
Han Zhou@hanzhou032·
Automating Multi-Agent Design: 🧩Multi-agent systems aren’t just about throwing more LLM agents together. 🛠️They require mastering the subtle art of prompting and agent orchestration. Introducing MASS🚀- Our new agent optimization framework for better prompts and topologies!
Han Zhou tweet media
English
13
160
730
81.5K
Ivan Vulić retweetledi
Benjamin Minixhofer
Benjamin Minixhofer@bminixhofer·
We achieved the first instance of successful subword-to-byte distillation in our (just updated) paper. This enables creating byte-level models at a fraction of the cost of what was needed previously. As a proof-of-concept, we created byte-level Gemma2 and Llama3 models. 🧵
Benjamin Minixhofer tweet media
English
1
14
69
4K
Ivan Vulić retweetledi
Yi Xu
Yi Xu@_yixu·
🚀Let’s Think Only with Images. No language and No verbal thought.🤔 Let’s think through a sequence of images💭, like how humans picture steps in their minds🎨. We propose Visual Planning, a novel reasoning paradigm that enables models to reason purely through images.
Yi Xu tweet media
English
16
213
1.3K
229.8K
Ivan Vulić retweetledi
Benjamin Minixhofer
Benjamin Minixhofer@bminixhofer·
We created Approximate Likelihood Matching, a principled (and very effective) method for *cross-tokenizer distillation*! With ALM, you can create ensembles of models from different families, convert existing subword-level models to byte-level and a bunch more🧵
Benjamin Minixhofer tweet media
English
2
26
88
6.5K
Ivan Vulić retweetledi
River Yijiang Dong
River Yijiang Dong@river_dong121·
Thrilled to share our updated paper: "UNDIAL: Self-Distillation with Adjusted Logits for Robust Unlearning in Large Language Models" We propose a new robust LLM unlearning method via Self-Distillation on Adjusted Logits (UNDIAL). 📄 Paper: arxiv.org/pdf/2402.10052
River Yijiang Dong tweet media
English
6
5
7
1.5K
Ivan Vulić retweetledi
Hannah
Hannah@h_sterz·
Do you DARE? Introducing a multiple-choice VQA benchmark ✨DARE✨ with: - 4 main robustness evaluation ⛓️ - 5 diverse categories 🧩 - Extensive analysis of 4 widely used VLMS 🤖
English
1
7
15
2.7K
Ivan Vulić retweetledi
Markus Frohmann
Markus Frohmann@FrohmannM·
Introducing 🪓Segment any Text! 🪓 A new state-of-the-art sentence segmentation tool! Compared to existing tools (and strong LLMs!), our models are far more: 1. efficient ⚡ 2. performant 🔝 3. robust 🚀 4. adaptable 🎯 5. multilingual 🗺
Markus Frohmann tweet media
English
2
26
180
20K
Ivan Vulić
Ivan Vulić@licwu·
As someone who spent years working in multilingual NLP, I am so happy that we're finally seeing (L)LMs and (N)MT systems working in tandem towards the shared cause. The idea in this work is so simple & sweet, and yet it moves! 🌍🌏🌎
Fabian David Schmidt@fdschmidt

Introducing NLLB-LLM2Vec! 🚀 We fuse the NLLB encoder & Llama 3 8B trained w/ LLM2Vec to create NLLB-LLM2Vec which supports cross-lingual NLU in 200+ languages🔥 Joint work w/ Philipp Borchert, @licwu, and @gg42554 during my great research stay at @cambridgeltl

English
0
1
37
3.6K
Ivan Vulić retweetledi
Han Zhou
Han Zhou@hanzhou032·
Which output is better? [A] or [B]? LLM🤖: B❌ [B] or [A]? LLM🤖: A✅ Thrilled to share our preprint in addressing preference biases in LLM judgments!🧑‍⚖️We introduce ZEPO, a 0-shot prompt optimizer that enhances your LLM evaluators via fairness⚖️ 📰Paper: arxiv.org/abs/2406.11370
Han Zhou tweet media
English
3
22
98
12.3K
Ivan Vulić retweetledi
Chengzu Li
Chengzu Li@li_chengzu·
Excited to introduce TopViewRS: VLMs as Top-View Spatial Reasoners🤖 TopViewRS assess VLMs’ spatial reasoning in top-view scenarios🏠just like how you read maps🗺️ Spoiler🫢GPT4V and Gemini are neck-and-neck, each excelling in different setups but neither even close to us humans
Chengzu Li tweet media
English
2
10
21
3.1K
Ivan Vulić retweetledi
Benjamin Minixhofer
Benjamin Minixhofer@bminixhofer·
Introducing Zero-Shot Tokenizer Transfer (ZeTT) ⚡ ZeTT frees language models from their tokenizer, allowing you to use any model with any tokenizer, with little or no extra training. Super excited to (finally!) share the first project of my PhD🧵
Benjamin Minixhofer tweet media
English
29
144
723
89.7K
Ivan Vulić retweetledi
Neil Houlsby
Neil Houlsby@neilhoulsby·
Adapters are just a great way to share/benefit from new capabilities without handing around the kitchen sink. Congrats to the AdapterHub folks for adding support for quantized training (Q-LoRA and friends).
AdapterHub@AdapterHub

🚀 Our latest Adapters library release integrates quantized model training, enabling efficient fine-tuning of LLMs with Q-LoRA, Q-Bottleneck Adapters, or Q-PrefixTuning. 🎉 Check out this notebook to learn how to fine-tune Llama 3 with Q-LoRA 🦙✨: github.com/Adapter-Hub/ad…

English
0
5
23
4.4K
Ivan Vulić
Ivan Vulić@licwu·
If we align LLMs through preferences, perhaps we should also evaluate them the same way (and respect transitivity)? The answer is: yes, we should. The trick, however, is how to make evaluation tractable. If you are into the whole "LLM-as-Judges" line of work, check this paper!
Yinhong Liu@YinhongLiu2

🔥New paper!📜 Struggle to align LLM evaluators with human judgements?🤔 Introducing PairS🌟: By exploiting transitivity, we push the potential of pairwise preference in efficient ranking evaluations that has better alignment!🧑‍⚖️ 📖arxiv.org/abs/2403.16950 💻github.com/cambridgeltl/p…

English
0
1
9
1.8K
Ivan Vulić retweetledi
Ivan Vulić retweetledi
Edoardo Ponti
Edoardo Ponti@PontiEdoardo·
I am still looking for PhD students starting in September 2024! The deadline to apply for the CDT in NLP is the 11th of March. If you wish to do research in modular and efficient LLMs, here are some highlights of my lab's research from the past year ⬇️🧵
EdinburghNLP@EdinburghNLP

Interested in training with future leaders in NLP to engage with the cutting edge of the technical, social, design, and legal aspects of these systems? Then apply for our new Centre for Doctoral Training in Designing Responsible NLP! Deadline 11 March 2024 responsiblenlp.org/2024-studentsh…

English
10
51
147
48.2K
Ivan Vulić
Ivan Vulić@licwu·
Think globally, act locally? Well, we were thought-experimenting whether LLMs would understand people from different places around our hometowns better than we ever might... And then we have eventually decided to make an actual (non-thought) experiment out of these thoughts! 👇👇
Nikola Ljubešić@nljubesic

Interested in commonsense reasoning in dialectal texts? The DIALECT-COPA shared task is the perfect fit for you, providing train and dev data for four official South-Slavic languages and two out of three related test dialects sites.google.com/view/vardial-2… @vardialworkshop @naaclmeeting

English
0
3
18
1.9K