CambridgeLTL

266 posts

CambridgeLTL banner
CambridgeLTL

CambridgeLTL

@CambridgeLTL

Language Technology Lab (LTL) at the University of Cambridge. Computational Linguistics / Machine Learning / Deep Learning. Focus: Multilingual NLP and Bio NLP.

Cambridge, England Katılım Şubat 2018
93 Takip Edilen2.2K Takipçiler
CambridgeLTL
CambridgeLTL@CambridgeLTL·
And that's a wrap! At the #EMNLP2025 our lab had an incredible showing with 8 papers! Congratulations to all our amazing students! 🎉 Links to papers: [🧵below]
English
1
0
3
227
CambridgeLTL retweetledi
Sharan
Sharan@_maiush·
AI that is “forced to be good” v “genuinely good” Should we care about the difference? (yes!) We’re releasing the first open implementation of character training. We shape the persona of AI assistants in a more robust way than alternatives like prompting or activation steering.
Sharan tweet media
English
5
39
192
62.2K
CambridgeLTL retweetledi
Tiancheng Hu
Tiancheng Hu@tiancheng_hu·
Instruction tuning unlocks incredible skills in LLMs, but at a cost: they become dangerously overconfident. You face a choice: a well-calibrated base model or a capable but unreliable instruct model. What if you didn't have to choose? What if you could navigate the trade-off? (1/8)
GIF
English
3
4
14
1.1K
CambridgeLTL
CambridgeLTL@CambridgeLTL·
- Large Language Models are Miscalibrated In-Context Learners @li_chengzu - WinSpot: GUI Grounding Benchmark with Multimodal Large Language Models Zack Hui
English
0
0
0
206
CambridgeLTL
CambridgeLTL@CambridgeLTL·
- Retrofitting Large Language Models with Dynamic Tokenization @licwu @bminixhofer - Culturally Aware and Adapted NLP: A Taxonomy and a Survey of the State of the Art @ChenLiu47008770 - DARE: Diverse Visual Question Answering with Robustness Evaluation Hannah Sterz, @licwu
English
1
1
2
415
CambridgeLTL retweetledi
Caiqi Zhang
Caiqi Zhang@caiqizh·
🔥 We teach LLMs to say how confident they are on-the-fly during long-form generation. 🤩No sampling. No slow post-hoc methods. Not limited to short-form QA! ‼️Just output confidence in a single decoding pass. ✅Better calibration! 🚀 20× faster runtime. arXiv:2505.23912 👇
Caiqi Zhang tweet mediaCaiqi Zhang tweet media
English
2
23
41
3.9K
CambridgeLTL retweetledi
Benjamin Minixhofer
Benjamin Minixhofer@bminixhofer·
We achieved the first instance of successful subword-to-byte distillation in our (just updated) paper. This enables creating byte-level models at a fraction of the cost of what was needed previously. As a proof-of-concept, we created byte-level Gemma2 and Llama3 models. 🧵
Benjamin Minixhofer tweet media
English
1
14
69
4K
CambridgeLTL retweetledi
Yi Xu
Yi Xu@_yixu·
🚀Let’s Think Only with Images. No language and No verbal thought.🤔 Let’s think through a sequence of images💭, like how humans picture steps in their minds🎨. We propose Visual Planning, a novel reasoning paradigm that enables models to reason purely through images.
Yi Xu tweet media
English
16
220
1.3K
229.9K