Cheng Han Chiang (姜成翰)

147 posts

Cheng Han Chiang (姜成翰) banner
Cheng Han Chiang (姜成翰)

Cheng Han Chiang (姜成翰)

@dcml0714

Fourth-year Ph.D. student at National Taiwan University Interests: music🎶, 📷photography, Japanese drama Cat person🐱 Research interests: NLP

Taipei, Taiwan Katılım Ocak 2020
242 Takip Edilen469 Takipçiler
Cheng Han Chiang (姜成翰) retweetledi
Hung-yi Lee (李宏毅)
Hung-yi Lee (李宏毅)@HungyiLee2·
LLM reasoning/tool use improves results but adds latency, hindering real-time dialogue. SHANKS is a new method that enables simultaneous "hearing" and "reasoning/tool use" to reduce this lag. Work from Cheng-Han Chiang & Microsoft researchers. arxiv.org/abs/2510.06917
Cheng Han Chiang (姜成翰)@dcml0714

🚨 New paper! SHANKS lets spoken language models (SLMs) think while listening💭👂 This enables the SLM to interrupt the user in a timely manner and make early tool calls when the speaker is still speaking. Paper: arxiv.org/abs/2510.06917 Project page: d223302.github.io/SHANKS/

English
4
1
13
2.3K
Cheng Han Chiang (姜成翰)
Our work shows strong potential for using audio LLMs to evaluate speech generated by spoken language models (SLMs) 🎙️✨ Excited to see future work explore fine-tuning SLMs with rewards or feedback from these audio-LLM judges 🧑‍⚖️🔁
English
1
0
3
286
Cheng Han Chiang (姜成翰)
🎉 Excited to share that our paper on audio-LLM-as-a-judge has been accepted to EMNLP 2025 Findings! 🔗 arxiv.org/abs/2506.05984… 🗝️ Highlights: 🧑‍⚖️ Agreement between human and audio-LLM-judge can be as high as human-human agreements 👑 Gemini-2.5-pro outperforms GPT-4o-audio as a speaking-style judge 🗣️ There's still room for improvement in style following & natural dialogue generation for SLMs
Cheng Han Chiang (姜成翰) tweet media
English
3
17
87
6.6K
Cheng Han Chiang (姜成翰)
Excited to be at #ACL2025! 🎉 Looking forward to sharing our research and learning from the community. Open to discussions on: - LLM-as-a-judge: our paper, TRACT, was presented this morning - Audio-LLM-as-a-judge for speaking style - STITCH: Our newest SLM that can think and talk simultaneously Let’s connect! 🤝 #ACL2025NLP
Cheng Han Chiang (姜成翰)@dcml0714

1/7 🔗 Introducing STITCH: our new method to make Spoken Language Models (SLMs) think and talk at the same time. Paper link 👉 arxiv.org/abs/2507.15375

English
0
0
16
560
Cheng Han Chiang (姜成翰)
@JulianSlzr Thank you for sharing! I really agree that real-time thoughts can make the SLM generate better quality and make SLMs more performant. Looking forward to seeing more works in this direction.
English
0
0
1
56
Cheng Han Chiang (姜成翰)
7/7 📈 On math reasoning datasets, STITCH improves over baselines that don’t think before responding and matches full CoT methods. 🚫 Non‑reasoning SLMs: Low latency, low accuracy 🐢 Full CoT before responding: High latency, high accuracy ⚡ STITCH‑S: Low latency, high accuracy
Cheng Han Chiang (姜成翰) tweet media
English
0
0
1
223
Cheng Han Chiang (姜成翰)
6/7 🆕 We explore another variant called STITCH‑S, where the model generates a spoken response chunk before the first reasoning chunk. 🏃 STITCH‑S has the same latency as a non‑reasoning model by design but delivers much better performance.
English
1
0
0
259