KAIST AI

108 posts

KAIST AI banner
KAIST AI

KAIST AI

@KAIST_AI

The Kim Jaechul Graduate School of AI at KAIST

Seoul, Republic of Korea 가입일 Mart 2022
194 팔로잉2.1K 팔로워
KAIST AI 리트윗함
RoboPapers
RoboPapers@RoboPapers·
Achieving generalizable manipulation is the north star for robotics learning, and while we’ve in the past seen incredible results on specific tasks using fine-tuned VLAs, this north star has remained elusive. Perhaps what is needed is a different approach. DreamZero proposes World Action models (WAMs), which jointly model both action and video in order to achieve state-of-the-art performance on benchmarks like MolmoSpaces and RoboArena. @SeonghyeonYe of @NVIDIARobotics joins us to talk about building a 14B parameter autoregressive diffusion model which achieves state-of-the-art generalization on real world tasks and on the best available benchmarks. Watch episode #68 of RoboPapers, with @micoolcho and @chris_j_paxton, now!
English
1
5
34
2.7K
KAIST AI 리트윗함
Hyeonbin Hwang
Hyeonbin Hwang@ronalhwang·
New Paper💡 Have you ever heard of grokking, a sudden transition from memorization to generalization? People have attributed grokking to weight decay, Fourier structure, optimization regimes, phase transitions, numerical effects… These can shape the training dynamics, but they don’t answer the core question: "what determines which representation the model learns, and why it generalizes?" We argue the key is intrinsic task symmetries. Paper: arxiv.org/pdf/2603.01968
Hyeonbin Hwang tweet mediaHyeonbin Hwang tweet media
English
18
75
578
32.1K
KAIST AI 리트윗함
Rohan Paul
Rohan Paul@rohanpaul_ai·
Noisy context breaks many LLM reasoning runs, so this paper rewards models for using only relevant text. The authors build NoisyBench and show LLMs get misled by distractors, then train them to filter. Across 11 tasks, irrelevant documents, old chat history, and tricky lookalike passages cut accuracy by up to 80%. Context is the pile of text an LLM sees before answering, and distractors are parts that sound useful but are wrong. NoisyBench makes this problem measurable by taking normal benchmarks and adding random documents, irrelevant chats, and hard negatives, long passages that look relevant but are wrong. When the authors turn models into AI agents that call tools, the agent often trusts noisy tool output too much, and simple prompt tweaks or supervised fine tuning, training on labeled examples, does not fix it. This matters because real apps mix questions with messy search results and chat history, so a wrong clue can steer the whole answer. Their answer is Rationale Aware Reward (RARE), a training method where the model gets rewarded for using helpful context, so longer reasoning stays focused instead of chasing distractors. ---- Paper Link – arxiv. org/abs/2601.07226 Paper Title: "Lost in the Noise: How Reasoning Models Fail with Contextual Distractors"
Rohan Paul tweet media
English
10
15
133
7.6K
KAIST AI 리트윗함
jiyeon kim
jiyeon kim@jiyeonkimd·
🌎Real-world knowledge evolves constantly and emerges incrementally. Can LLMs adapt to new information on the fly? 🤯Frontier models and agentic approaches all struggle, missing when to update the fact, or getting distracted by irrelevant information. We introduce ✨OAKS✨, a benchmark for evaluating models’ online adaptation to streaming, continually updating knowledge.
jiyeon kim tweet media
English
3
21
62
6.6K
KAIST AI 리트윗함
Kinam Kim
Kinam Kim@kinam_0252·
🚀 Code Released! EgoX: Egocentric Video Generation from a Single Exocentric Video ✨ 💻 The code is now available! You can generate egocentric videos just like in the demo, or use EgoX to transform your own exocentric videos into egocentric ones.
English
23
120
765
106K
KAIST AI 리트윗함
Jiwoo Hong
Jiwoo Hong@jiwoohong98·
Align LLMs with the preference dataset ONLY with 💡ORPO💡 We introduce ORPO, alignment without reference model & SFT! With awesome dataset from @argilla_io + Mistral(7B) + ORPO, we present 🌟Mistral-ORPO-β🌟 🧵 👉 AlpacaEval 2.0: 12.20% 👉 IFEval: 66.19% 👉 MT-Bench: 7.32
Jiwoo Hong tweet mediaJiwoo Hong tweet mediaJiwoo Hong tweet media
English
10
31
135
38.2K
KAIST AI 리트윗함
Gyubok Lee
Gyubok Lee@gyubok_l·
If you are interested in medical NLP/AI, our lab is hosting one of the shared tasks at NAACL-ClinicalNLP 2024 on reliable text-to-SQL on electronic health records. #EHRSQL. Please check out the link for more information. sites.google.com/view/ehrsql-20….
English
1
2
9
2.8K
KAIST AI
KAIST AI@KAIST_AI·
Pleased to see Prof Jong C. Park from #KAIST Computer Science department present the opening keynote at #EMNLP2023 this year sharing insights from more than 20 years of development of human centric Natural Language Processing nlpcl.kaist.ac.kr
KAIST AI tweet media
English
2
2
18
3K
KAIST AI
KAIST AI@KAIST_AI·
Our paper "Flooding with Absorption: An Efficient Protocol for Heterogeneous Bandits over Complex Networks" won the Best Student Paper at OPODIS 2023! 🌐📚 Congratulations to Junghyun Lee, Laura Schmid, and Se-Young Yun! 🔗 Read the paper: arxiv.org/abs/2303.05445
English
0
0
18
1.9K
KAIST AI
KAIST AI@KAIST_AI·
[2/2] Authors: Geewook Kim (@GeewookKim), Hodong Lee, Daehee Kim, Haeji Jung, Sanghee Park, Yoonsik Kim, Sangdoo Yun, Taeho Kil, Bado Lee, Seunghyun Park Url: arxiv.org/abs/2305.15080
KAIST AI tweet media
Indonesia
0
0
3
1.3K
KAIST AI
KAIST AI@KAIST_AI·
[1/2] Visually-Situated Natural Language Understanding with Contrastive Reading Model and Frozen Large Language Models (long paper / KAIST LK Lab & NAVER Cloud AI & Korea University & NAVER AI Lab)
English
1
0
2
1.2K
KAIST AI
KAIST AI@KAIST_AI·
⭐️Congratulations to everyone at KAIST AI who has a total of 11 papers accepted to #EMNLP2023! Check them here 📷🧵 #heading=h.csdjiropd9wf" target="_blank" rel="nofollow noopener">docs.google.com/document/d/1YB…
English
21
10
78
14.8K
KAIST AI
KAIST AI@KAIST_AI·
[1/2] Efficiently Enhancing Zero-Shot Performance of Instruction Following Model via Retrieval of Soft Prompt (findings / KAIST LK Lab)
English
1
0
4
807
KAIST AI
KAIST AI@KAIST_AI·
[1/2] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning (long paper / KAIST LK Lab & NAVER AI Lab & University of Washington)
English
1
1
4
954
KAIST AI
KAIST AI@KAIST_AI·
ROAST: Robustifying Language Models via Adversarial Perturbation with Selective Training (findings / KAIST ALIN Lab & Meta AI) Authors: Jaehyung Kim (@jhkim940331), Yuning Mao, Rui Hou, Hanchao Yu, Davis Liang, Pascale Fung, Qifan Wang, Fuli Feng, Lifu Huang, Madian Khabsa
KAIST AI tweet media
English
0
0
2
684