Yiming Cui

179 posts

Yiming Cui

Yiming Cui

@KCrosner

NLP Researcher

北京, 中华人民共和国 شامل ہوئے Ağustos 2012
82 فالونگ661 فالوورز
پن کیا گیا ٹویٹ
Yiming Cui
Yiming Cui@KCrosner·
We are extremely overwhelmed and honored to receive the IEEE Signal Processing Society Best Paper Award (2025) for our paper "Pre-Training with Whole Word Masking for Chinese BERT", published in IEEE/ACM TASLP (2021). 🎉🎉🎉 @IEEEsps #IEEE signalprocessingsociety.org/newsletter/202…
Yiming Cui tweet media
English
0
0
3
111
Anton Reshetov
Anton Reshetov@anton_reshetov·
Hey @OpenAI I received the “Codex for Open Source” email offering 6 months of ChatGPT Pro, but when I try to activate it I get: “This promotion isn’t available for your plan.” I currently have an active ChatGPT Plus subscription.
English
4
0
0
140
Yiming Cui
Yiming Cui@KCrosner·
Good News: I am enrolled in #Codex for Open Source. 🎉 Bad News: I cannot redeem this untill my Plus subscription ends. What even worse, I am on an annual subscription, ends 6 months+. 😭 @OpenAI please help.
Yiming Cui tweet media
English
0
0
3
108
Yiming Cui
Yiming Cui@KCrosner·
(5/x) Occlusion-based saliency analysis. By masking image regions and measuring confidence drops,baseline MLLMs often rely on spurious visual cues, while CoT shifts attention toward chemically meaningful structures, improving both grounding and confidence. nature.com/articles/s4200…
Yiming Cui tweet media
English
0
0
0
22
Yiming Cui
Yiming Cui@KCrosner·
(4/x) When seeing hurts: visual input can degrade performance. Adding visual input sometimes reduces accuracy, especially in smaller models. Larger models tend to balance textual and visual signals better, which may be key to achieving strong performance. nature.com/articles/s4200…
Yiming Cui tweet media
English
0
0
1
25
Yiming Cui
Yiming Cui@KCrosner·
(3/x) Task-type breakdown. Current multimodal LLMs perform well on tables and charts, but struggle with molecular structures and experimental apparatus, which require chemistry-specific visual understanding and domain knowledge. nature.com/articles/s4200…
Yiming Cui tweet media
English
0
0
1
12
Yiming Cui
Yiming Cui@KCrosner·
(2/x) CoT generally improves chemical reasoning performance. Analysis show that CoT is especially helpful for mid-tier models. For e.g., GPT-4.1-mini achieves 20~26 accuracy improvement with CoT, while less significant for small/large-scale models. nature.com/articles/s4200…
Yiming Cui tweet media
English
0
0
1
24
Yiming Cui
Yiming Cui@KCrosner·
(1/x) we curate a chemistry benchmark based on USNCO exams, spanning over two decades, consisting of 473 real multimodal QA problems. It covers a broad spectrum of chemistry topics, including general, physical, organic, inorganic, and analytical chemistry. nature.com/articles/s4200…
Yiming Cui tweet media
English
0
0
1
34
Yiming Cui
Yiming Cui@KCrosner·
Our paper "Self-Evolving GPT: A Lifelong Autonomous Experiential Learner" is accepted at #ACL2024 main! We propose a framework for LLMs to autonomously learn and apply experience, boosting GPT-3.5 and GPT-4 performance. Stay tuned for the paper and code release! #NLP #LLM #GPT
Yiming Cui tweet media
English
0
0
5
885
Yiming Cui
Yiming Cui@KCrosner·
Happy to introduce Chinese-LLaMA-Alpaca-3, which is our 3rd open-source projects on #Llama series. We release Llama-3-Chinese-8B and Llama-3-Chinese-8B-Instruct with continual PT/SFT on Chinese corpora. Check our project: github.com/ymcui/Chinese-… #nlproc #llama3
English
0
0
2
405
Yiming Cui
Yiming Cui@KCrosner·
Through our empirical experiments on creating Chinese Mixtral, we find that extending vocabulary might NOT be a necessity for LLM language transfer. As usual, we open-source Chinese-Mixtral(-Instruct) at GitHub/HF: github.com/ymcui/Chinese-… arXiv Paper: arxiv.org/abs/2403.01851
English
0
1
7
853
Yiming Cui
Yiming Cui@KCrosner·
We release Chinese-LLaMA-2-7B and Chinese-Alpaca-2-7B based on #Llama-2, which achieved significant improvements over our first-gen Chinese-LLaMA/Alpaca, even surpass 13B models on some metrics. Check our GitHub repo: github.com/ymcui/Chinese-… #llm #NLProc
English
1
2
19
643
Yiming Cui
Yiming Cui@KCrosner·
@joemkwon Sorry for the late reply. Regarding your question, our main motivation is to add more trainable parameter (qkvo and mlp) within LoRA scheme. Recent research QLoRA also shows that adapting qkvo/mlp is essential to achieve a better performance. Maybe you can check the QLoRA paper.
English
0
0
0
72
Joe
Joe@joemkwon·
@KCrosner I’m curious how you decided which parameter to freeze and wrap with LoRA, vs just freezing entirely or tuning entirely
English
1
0
0
28
Yiming Cui
Yiming Cui@KCrosner·
Excited to release our Chinese 🦙#LLaMA and #Alpaca LLMs (7B for now), extended with an additional 20k Chinese vocabulary, trained with alpaca-lora. Our model works seamlessly with the wonderful llama.cpp on CPU. Give it a try at github.com/ymcui/Chinese-… #nlproc #llm #AI
GIF
English
1
0
12
2.1K
Yiming Cui
Yiming Cui@KCrosner·
Update 13B Chinese #LLaMA and #Alpaca. Better quality compared to 7B. GPT-4 rates 13B model 71/100 while 49 for 7B version. We also provide a Colab notebook for fast conversion, and of course it is fully compatible with llama.cpp. Try: github.com/ymcui/Chinese-… #nlproc #llm #ai
Yiming Cui@KCrosner

Excited to release our Chinese 🦙#LLaMA and #Alpaca LLMs (7B for now), extended with an additional 20k Chinese vocabulary, trained with alpaca-lora. Our model works seamlessly with the wonderful llama.cpp on CPU. Give it a try at github.com/ymcui/Chinese-… #nlproc #llm #AI

English
0
0
11
1.1K
Yiming Cui
Yiming Cui@KCrosner·
Happy to release our multimodal pre-trained model VLE, which achieved top performance on VCR. We also set up a pipeline with captioning model and LLM to generate much user-friendly answers for VQA. Resources, code, and demo are available through: github.com/iflytek/VLE 🎉🎉🎉
Yiming Cui tweet media
English
0
1
5
517
Yiming Cui
Yiming Cui@KCrosner·
@cryptexcode @SemEvalWorkshop @naacl Thank you. The live session (mainly for task organizers) is hosted via Zoom, and all system papers are presented as posters (no oral). I'm not sure if the video will be made public by official. If you are interested in best paper list, it will be posted on SemEval website soon.
English
1
0
0
0