Kelly Marchisio @NeurIPS
705 posts

Kelly Marchisio @NeurIPS
@cheeesio
Multilingualilty Lead @cohere. Formerly: PhD @jhuclsp, Alexa Fellow @amazon, dev @Google, MPhil @cambridgenlp, EdM @hgse 🔑🔑¬🧀 (@kelvenmar20)



Announcing our latest paper: CommonLID: Re-evaluating State-of-the-Art Language Identification Performance on Web Data In collaboration with @CommonCrawl @MLCommons and @JohnsHopkins we worked with 80+ native speaker annotators to build a LID benchmark on actual Common Crawl text covering 109 languages. Existing evaluations overestimate how well LangID works on web data.


Announcing our latest paper: CommonLID: Re-evaluating State-of-the-Art Language Identification Performance on Web Data In collaboration with @CommonCrawl @MLCommons and @JohnsHopkins we worked with 80+ native speaker annotators to build a LID benchmark on actual Common Crawl text covering 109 languages. Existing evaluations overestimate how well LangID works on web data.

There's a lot of talk of LLMs "saturating all the evals" but there's plenty of evals people could make where LLMs would do poorly: -Beat a Zelda game -Make a profit in a prediction market -Write a stand-up set that's original and funny I'm bullish on AI, but we're far from done.


🚀🚨 Sparse-Frontier Major Updates! You can now evaluate Reasoning + Sparse models at speed, with Sparse-Frontier upgraded to the @vllm_project's v1 engine🔥 We still provide support for Tensor Parallelism and the original sparse attention baselines, but it now works cleanly with newer models, decoding strategies, and evaluation setups. Task coverage and model support were also expanded as part of this release. The config-based workflow stays the same. If you’re working on sparse decoding, reasoning models, or long-context evaluation, this update makes it easier to run consistent experiments across models, tasks, and attention methods⚡️ Really enjoyed working with @p_nawrot and @PontiEdoardo over the past months to get this release out!


Pick of the week @fbk_mt: How Does Quantization Affect Multilingual LLMs? Quantization has become a widely adopted technique for model compression. This work investigates the impact of quantization on different languages in multilingual LLMs. aclanthology.org/2024.findings-…






Kelly Marchisio, Multilingual Team Lead at @cohere, shared an inside look at building 🧑💻 a #multilingual #LLM and advancing #AI #translation at #SlatorCon Silicon Valley 2025. #Cohere #LLMs #xl8 #t9n @cheeesio slator.com/how-to-build-m…

The future of multilingual AI 🚀 is here. At #SlatorConSV25, @cohere's @cheeesio explains how to build massively multilingual LLMs, from technical foundations to the current landscape and what comes next. #MultilingualAI #LLMs #CommandA


Introducing Command A Translate, our state-of-the-art model designed for high-quality translation tasks.


Introducing Command A Translate, our state-of-the-art model designed for high-quality translation tasks.


