MiyoungKo

30 posts

MiyoungKo

MiyoungKo

@miyoung_ko

Ph.D student at KAIST AI.

Katılım Ekim 2020
84 Takip Edilen223 Takipçiler
MiyoungKo retweetledi
Hoyeon Chang
Hoyeon Chang@hoyeon_chang·
1/ Pattern matching has been left ambiguous for so long. Our precise formalization pinpoints its predictive power on composition tasks! 📣 Check out our ICLR 2026 paper: 1️⃣ Sharp sample complexity bound for 2-hop task; 2️⃣ Some unexpectedly hard composition tasks: CoT won’t help.
Hoyeon Chang tweet media
English
2
11
64
5.3K
MiyoungKo retweetledi
Haebin Shin
Haebin Shin@haebinshin_·
🚨 Our new paper from @MSFTResearch accepted at ICML 2025! Can LLaMA be distilled from DeepSeek in logit space, even with completely different vocabularies? We presents: “Overcoming Vocabulary Mismatch: Vocabulary-agnostic Teacher-Guided Language Modeling” [1/5]
Haebin Shin tweet media
English
2
4
21
1.5K
MiyoungKo retweetledi
Seungone Kim
Seungone Kim@seungonekim·
🏆Glad to share that our BiGGen Bench paper has received the best paper award at @naaclmeeting! x.com/naaclmeeting/s… 📅 Ballroom A, Session I: Thursday May 1st, 16:00-17:30 (MDT) 📅 Session M (Plenary Session): Friday May 2nd, 15:30-16:30 (MDT) 📅 Virtual Conference: Tuesday May 6th, 20:30-21:00 (MDT) I'd like to appreciate our coauthors @scott_sjy Ji Yong Cho @ShayneRedford @chaechaek1214 @dongkeun_yoon @gson_AI Yejin Cho @shafayat_sheikh @jinheonbaek @suehpark @ronalhwang @Jinkyung_Jo Hyowon Cho @haebinshin_ @sylee_ai @hanseok_oh @nlee288 @itsnamgyu @joocjun @miyoung_ko @yoonjoo_le2 @hyungjoochae @jay_shin @jang_yoel @SeonghyeonYe @billyuchenlin @wellecks @gneubig Moontae Lee @Kyungjae__Lee @seo_minjoon! It wouldn't have been possible without everyone's feedback and hard work 😀
Seungone Kim@seungonekim

🤔How can we systematically assess an LM's proficiency in a specific capability without using summary measures like helpfulness or simple proxy tasks like multiple-choice QA? Introducing the ✨BiGGen Bench, a benchmark that directly evaluates nine core capabilities of LMs.

English
10
21
131
31.7K
MiyoungKo retweetledi
Haebin Shin
Haebin Shin@haebinshin_·
🚨 New paper alert! 🚨 Isn’t it wasteful to repeat lengthy & complex agent prompts every time? Introducing "Generative Context Distillation"—a new lightweight method to internalize prompt. 🦾 Powerful performance 💵 Efficient inference "without the need for a prompt📜" [1/7]
Haebin Shin tweet media
English
1
18
48
6.7K
MiyoungKo
MiyoungKo@miyoung_ko·
I'll be attending #EMNLP2024 to present "Hierarchical Deconstruction of LLM Reasoning: A Graph-based Framework on Analyzing Knowledge Utilization" 📍In-Person Poster Session D (Jasmine) 🕐 Nov. 13 (Wed.), 10:30 AM - 12:00 AM 📕arxiv.org/abs/2406.19502 Would love to chat about reasoning, interpretability, and factuality 🌴
MiyoungKo tweet media
English
1
4
42
3.4K
MiyoungKo retweetledi
Seonghyeon Ye
Seonghyeon Ye@SeonghyeonYe·
🚀 First step to unlocking Generalist Robots! Introducing 🤖LAPA🤖, a new SOTA open-sourced 7B VLA pretrained without using action labels. 💪SOTA VLA trained with Open X (outperforming OpenVLA on cross and multi embodiment) 😯LAPA enables learning from human videos, unlocking potential for robotic foundation model ❗Over 30x pretraining efficiency for VLA training 🤗Code and checkpoints are all open-sourced!
English
3
56
214
33.2K
MiyoungKo retweetledi
Jiyeon Kim
Jiyeon Kim@jiyeonkimd·
❓Do LLMs maintain the capability of knowledge acquisition throughout pretraining? If not, what is driving force behind it? ❗Our findings reveal that decreasing knowledge entropy hinders knowledge acquisition and retention as pretraining progresses. 📄arxiv.org/abs/2410.01380
Jiyeon Kim tweet media
English
4
32
153
22.7K
MiyoungKo retweetledi
MiyoungKo
MiyoungKo@miyoung_ko·
🪜 To enhance reasoning, we guide models through structured, multi-turn interactions, consistently improving performance of most models. This highlights the importance of intermediate steps in complex reasoning. [🧵8 / 10]
MiyoungKo tweet media
English
1
0
8
675
MiyoungKo
MiyoungKo@miyoung_ko·
🔬 We further investigate the reliance on memorization of training data, showing less reliance on solving D3 questions than simple factual recall (D1). Through this, we also reveal that forward and backward discrepancies in large models originate from distinct types of failures (25% vs. 75% of 70B models, higher % memorizes more). [🧵7 / 10]
MiyoungKo tweet media
English
1
0
9
840
MiyoungKo
MiyoungKo@miyoung_ko·
📊 We tested various instruction-tuned LLMs like LLaMA 2, LLaMA 3, Mistral, and Mixtral, ranging from 7B to 70B parameters. Our findings show smaller models have larger discrepancies, while larger models struggle from discrepancies less. [🧵6 / 10]
MiyoungKo tweet media
English
1
0
11
699
MiyoungKo
MiyoungKo@miyoung_ko·
📢 Excited to share our latest paper on the reasoning capabilities of LLMs! Our research dives into how these models recall and utilize factual knowledge during solving complex questions. [🧵1 / 10] arxiv.org/abs/2406.19502
MiyoungKo tweet media
English
4
72
298
48.1K
MiyoungKo
MiyoungKo@miyoung_ko·
📐 Now how to evaluate? Our method involves measuring forward reasoning gaps (differences in performance on simpler vs. complex questions) and backward discrepancies (success on complex questions but struggles with simpler ones). This gives a comprehensive view of LLMs' reasoning skills. [🧵5 / 10]
MiyoungKo tweet media
English
1
0
10
737
MiyoungKo
MiyoungKo@miyoung_ko·
📚 Introducing DepthQA: a collection of questions and answers derived from human-written scientific D3 questions in the TutorEval dataset. D2 and D1 questions are generated using GPT-4 Turbo, under a carefully designed construction pipeline and postprocessing in order to ensure accuracy of the content and reasoning graph representation. Various types of reasoning processes are observed! [🧵4 / 10]
MiyoungKo tweet media
English
1
0
7
787
MiyoungKo
MiyoungKo@miyoung_ko·
⛓️ To tackle this, we propose breaking down complex questions into a graph structure, with each node representing a specific depth of understanding: recall (D1), application (D2), and strategic thinking (D3). Our approach emphasizes accumulating and integrating knowledge to solve real-world problems effectively. [🧵3 / 10]
MiyoungKo tweet media
English
1
0
11
947
MiyoungKo
MiyoungKo@miyoung_ko·
❓Complex questions often require various reasoning processes beyond simply aggregating facts. Take this question as an example: "Why does ReLU training take less time than sigmoid or tanh training?". One must not only *recall* what an activation function is, but also *compare* the characteristics of activation functions and understand the *causal* relationship between gradients and training speed. [🧵2 / 10]
English
1
0
11
1K