Genta Winata

1.3K posts

Genta Winata

Genta Winata

@gentaiscool

AI Researcher @CapitalOne AIF. Ex @TechAtBloomberg @BigScienceW @SFResearch @hkust. Working on multilingual and LLM #NLProc. Building @GrassrootsSci

Katılım Aralık 2011
856 Takip Edilen1.9K Takipçiler
Sabitlenmiş Tweet
Genta Winata
Genta Winata@gentaiscool·
⭐️We're thrilled to share that our paper WorldCuisines has been selected for the Best Theme Paper Award at NAACL 2025 @naaclmeeting! 🎉 A huge thank you to the reviewers and area chair for this incredible recognition — we’re truly honored. Massive gratitude to all our amazing co-authors for the countless hours, late nights, and deep discussions that went into creating this high-quality dataset. 2025.naacl.org/blog/best-pape… We can't wait to present next week at NAACL! Catch us at our poster session (Wednesday, April 30) and Best Paper Award Session (Friday, May 2) for our oral presentation. Check out the paper and project here: 🌐 worldcuisines.github.io Contributors: @fredyhudi, @patrickamadeus_, @davidanugraha, @rifkiaputri, @zzeet, @ubaidalih, @auliaadilaa, @adamnohejl, @JunhoMyung00211, @aliceoh, @AnarSnowball, @faridlazuarda, @jcblaisecruz, @nedjmaou, @jodieyzhou, @AboladeDaud, @prajdabre1, @holylovenia, @SCahyawijaya, @bryanwilie92, @mrpeerat, @farizikhwantri, @gkuwanto, @llamagrp, @mv_zhukova, @EmmanueleChers1, @AlhamFikri, @davlanade, @tarowatanabe, @OptionsGod_lgd,@AyuP_AI, and many others who are not on X. Acknowledgments: @nayeon7lee, @Wenliang_Dai, @pascalefung who helped and provided us insightful suggestions. #nlproc #naacl2025 #worldcuisines
English
9
15
121
17.4K
Genta Winata retweetledi
Alham Fikri Aji
Alham Fikri Aji@AlhamFikri·
VLMs can easily get distracted by unrelated cultural cues. Happy to present our work on this soon at #CVPR2026🥳 Working on multilingual VLMs? Consider using our benchmark: 📜arxiv.org/pdf/2511.17004 🤗huggingface.co/datasets/patri… Amazing work by @patrickamadeus_ and colleagues!
Alham Fikri Aji tweet media
pat@patrickamadeus_

Excited to share that we have committed our paper “Vision-Language Models are Confused Tourists” to #CVPR2026 (Findings)! 🇺🇸🏔 Arxiv: arxiv.org/abs/2511.17004 We question whether current SOTA VLMs remain robust in simple cultural grounding QA when distracting contextual objects are present For example, if you eat chicken schnitzel with Mt. Fuji in the background, will the model fail to recognize it as Japanese katsu? ConfusedTourists introduces: 👉 5k+ evaluation samples across 3 cultural item categories, comprising 243 unique cultural items from 57 countries and 11 sub-regions 🌍 👉 Evaluation of 14 VLMs across 12 data features 🤖 👉 Findings showing that simple concept mixing can cause up to a -40% drop in perform 📉 Special thanks to my co-authors @IkhlasulHanif0 , @emthehunt, @gentaiscool, @FajriKoto, and my advisor @AlhamFikri for the valuable contributions along the way! #multimodal #vlm #multicultural #robustness #evaluation #NLProc #ComputerVision

English
2
18
72
7.3K
Genta Winata retweetledi
Center for AI Safety
Last week, Humanity’s Last Exam was published in @Nature. In just over a year, model scores on HLE have risen from under 5% to nearly 40%. Thank you to @scale_AI and the 1000+ HLE co-authors for helping policymakers and the public track these rapid advances in AI capabilities.
Center for AI Safety tweet mediaCenter for AI Safety tweet media
English
9
42
160
26.6K
Genta Winata retweetledi
Elias Stengel-Eskin
Elias Stengel-Eskin@EliasEskin·
📢 Introducing Routing with Generated Data (RGD), a new setting for annotation-free LLM routing. We study how routers can be trained without any ground-truth labels. We also introduce CASCAL, a novel label-free LLM router that identifies niche skills using consensus-voting and hierarchical clustering. ➡️ Most LLM routers assume access to labeled, in-domain data to estimate model skills (query-answer routers). However, user distributions are unknown and labels are expensive or unavailable, highlighting the need for routers that work without labels. ➡️ We introduce Routing with Generated Data (RGD): routers are trained only on Q&A data generated from task descriptions, without human annotation. We experiment with various LLM generators of different strengths (Gemini-2.5-Flash, Qwen-3-32B, Exaone-3.5-7.8B). ➡️ CASCAL outperforms other query-answer and query-only routers across diverse datasets (MMLU-Pro, SuperGPQA, MedMCQA, BigBench Extra Hard), and is more robust to weaker generators.
Elias Stengel-Eskin tweet media
English
1
24
42
7.6K
lynnette ng
lynnette ng@quarbby·
On this day, I finally got myself Premium as a Christmas present. Finally in the cool kids club 😀
English
1
0
6
402
Genta Winata
Genta Winata@gentaiscool·
@haryoaw Happened to me as well in neurips. They got poster, we got nothing
English
0
0
2
139
Haryo
Haryo@haryoaw·
It's interesting that in the conference, I met someone who had presented a paper that had the same idea (different execution) as ours. Ours led to a finding, and theirs led to an oral presentation. 😭
English
4
0
5
366
Genta Winata
Genta Winata@gentaiscool·
@haryoaw Try kfc and mcd in India. I heard it is good
English
1
0
0
127
Haryo
Haryo@haryoaw·
My friend ordered and ate fish and chips in India out of lots of Indian food choices.
Haryo tweet media
English
2
0
6
497
Raj Dabre
Raj Dabre@prajdabre·
Very excited to release our ongoing work on IndicBERT-V3! Some key points: 1. Long context 2. Various model sizes 3. SOTA performance on bitext mining and RAG on our internal evaluations 4. Multilingual Indic support Go wild! cc @osanseviero for visibility, since we used Gemma. @anoopk
neural nets.@cneuralnetwork

We are releasing IndicBERT-v3, a suite of multilingual encoder language models (270M, 1B, 4B) built on top of Gemma-3. We adapted these models to use bidirectional attention, making them effective for encoder-heavy tasks. (1/3) @psidharth567 @_iunravel

English
12
8
169
13.4K
Wenhu Chen
Wenhu Chen@WenhuChen·
Surpassed 10K citations in a single year! 🥳
Wenhu Chen tweet media
English
22
4
457
41.1K
Genta Winata
Genta Winata@gentaiscool·
💡Have you ever wondered whether vision–language models can be easily tricked by adding landmarks or flags to an image? In the spirit of the holidays🎄, we show that VLMs can indeed be easily confused like "Confused Tourists" ✈️: their performance drops significantly when such image perturbations are applied. 🔎 Check out "VLMs are Confused Tourists" ✈️ here arxiv.org/pdf/2511.17004 #vision #nlproc #robustness
pat@patrickamadeus_

Craving holiday-themed paper? Say less🎄 Turns out, Vision Language Models are Confused Tourists ✈️😵‍💫 We show that adversarially induced cultural scenes significantly impair VLM cultural comprehension and trigger potential bias #NLProc #multimodal #robustness /thread 🧵(1/8)

English
0
3
16
2.5K
Hanif | AI NOT FOR PRODUCTIVITY
Hanif | AI NOT FOR PRODUCTIVITY@IkhlasulHanif0·
Prof: "what are your plans on winter break?" Stud: "Oh I plan to go to x, y, z, ..." Prof: "Oh I mean in terms of research" aint noway
English
4
0
4
484
Genta Winata retweetledi
pat
pat@patrickamadeus_·
Craving holiday-themed paper? Say less🎄 Turns out, Vision Language Models are Confused Tourists ✈️😵‍💫 We show that adversarially induced cultural scenes significantly impair VLM cultural comprehension and trigger potential bias #NLProc #multimodal #robustness /thread 🧵(1/8)
pat tweet media
English
3
21
50
19.3K
neural nets.
neural nets.@cneuralnetwork·
anyone wanna catch dinner with me tomorrow in Mumbai pls dm
English
4
0
38
4.2K