Runxin Xu

35 posts

Runxin Xu banner
Runxin Xu

Runxin Xu

@pigjunebaba

AI researcher @deepseek_ai | @PKU1898 | @SJTU1896 Opinions are my own.

Katılım Ekim 2019
3K Takip Edilen7.2K Takipçiler
Runxin Xu retweetledi
DeepSeek
DeepSeek@deepseek_ai·
🔥DeepSeek-V4-Pro API is 75% OFF until May 5th, 2026, 15:59 (UTC Time)! Don't miss out on this massive discount. 🛠️Integration Updates: 🔹Claude Code: Set model to deepseek-v4-pro[1m] to unlock 1M context! 🔹OpenCode: Update to v1.14.24+ 🔹OpenClaw: Update to v2026.4.24+ Check the latest official API docs for full details: api-docs.deepseek.com/quick_start/pr…
DeepSeek tweet media
English
339
929
9.3K
1.1M
Runxin Xu
Runxin Xu@pigjunebaba·
We keep striving to build things that bring long-term value to everyone. We hope you enjoy our latest model — try it now on web, app, and API 🚀 「不诱于誉,不恐于诽,率道而行,端然正己」
DeepSeek@deepseek_ai

🚀 DeepSeek-V4 Preview is officially live & open-sourced! Welcome to the era of cost-effective 1M context length. 🔹 DeepSeek-V4-Pro: 1.6T total / 49B active params. Performance rivaling the world's top closed-source models. 🔹 DeepSeek-V4-Flash: 284B total / 13B active params. Your fast, efficient, and economical choice. Try it now at chat.deepseek.com via Expert Mode / Instant Mode. API is updated & available today! 📄 Tech Report: huggingface.co/deepseek-ai/De… 🤗 Open Weights: huggingface.co/collections/de… 1/n

中文
9
20
352
18.7K
Jiayi Pan
Jiayi Pan@jiayi_pirate·
Today is my last day at xAI. It's been an intense, memorable year building in the frontier — I'm grateful to have worked with and learned from such talented and supportive colleagues. The journey itself is the reward. So long, and thanks for all the fish! 🐬
English
104
23
1.3K
106.5K
Runxin Xu retweetledi
DeepSeek
DeepSeek@deepseek_ai·
Introducing DeepSeek-V3.1: our first step toward the agent era! 🚀 🧠 Hybrid inference: Think & Non-Think — one model, two modes ⚡️ Faster thinking: DeepSeek-V3.1-Think reaches answers in less time vs. DeepSeek-R1-0528 🛠️ Stronger agent skills: Post-training boosts tool use and multi-step agent tasks Try it now — toggle Think/Non-Think via the "DeepThink" button: chat.deepseek.com 1/5
English
507
1.8K
15K
2.1M
Runxin Xu retweetledi
Arena.ai
Arena.ai@arena·
🚨BREAKING: DeepSeek-V3-0324🐳 just ranked #5 on the Arena leaderboard - surpassing DeepSeek-R1 and every other open model! Highlights: - #1 open model (MIT license) - 2x cheaper than DeepSeek-R1 - Top-5 across ALL categories - Significant jump over the previous DeepSeek-v3 Huge milestone for @deepseek_ai and a big win for open-source community! 👏
Arena.ai tweet media
DeepSeek@deepseek_ai

🚀 DeepSeek-V3-0324 is out now! 🔹 Major boost in reasoning performance 🔹 Stronger front-end development skills 🔹 Smarter tool-use capabilities ✅ For non-complex reasoning tasks, we recommend using V3 — just turn off “DeepThink” 🔌 API usage remains unchanged 📜 Models are now released under the MIT License, just like DeepSeek-R1! 🔗 Open-source weights: huggingface.co/deepseek-ai/De…

English
18
79
814
165.1K
Runxin Xu
Runxin Xu@pigjunebaba·
@ii_posts nice work,do you have scores in human last exam?
English
1
0
9
928
Intelligent Internet
Intelligent Internet@ii_posts·
♊ Introducing II-Researcher: Our open-source AI research agent with strong benchmark performance ♊ 🚀 Key highlights: • New open-source framework for complex, multi-step search with reasoning tasks— a powerful, transparent alternative to proprietary systems • Two methods were implemented: finite state machines and tool-augmented thought processes with advanced reasoning models (R1 & QwQ) • Consistently delivers strong performance on GAIA and FRAMES benchmarks Come build with us and use II-Researcher!
Intelligent Internet tweet media
English
29
115
639
506.4K
Runxin Xu
Runxin Xu@pigjunebaba·
Congrats! A comprehensive survey on Long Context Language Models👏!
Dawei Zhu@dwzhu128

[1/n] Super excited to introduce our comprehensive survey on Long Context Language Models (LCLM), a collaborative effort between amazing researchers from @NJU1902 , @PKU1898 , CASIA, @AlibabaGroup , @BytedanceTalk , @TencentGlobal & Kuaishou! 🚀 Our survey covers 3 core aspects: 1️⃣ Building powerful & efficient LCLMs 2️⃣ Training & deployment infrastructure 3️⃣ Evaluation & analysis & applications A systematic review of all LCLM research to date! 📊 📄 Paper: arxiv.org/abs/2503.17407 🤗 HF: huggingface.co/papers/2503.17… 💻 Github: github.com/LCLM-Horizon/A… #LCLM #NLP #MachineLearning #AI

English
1
1
10
2.3K
Runxin Xu
Runxin Xu@pigjunebaba·
🚀 One More Thing!
DeepSeek@deepseek_ai

🚀 Day 6 of #OpenSourceWeek: One More Thing – DeepSeek-V3/R1 Inference System Overview Optimized throughput and latency via: 🔧 Cross-node EP-powered batch scaling 🔄 Computation-communication overlap ⚖️ Load balancing Statistics of DeepSeek's Online Service: ⚡ 73.7k/14.8k input/output tokens per second per H800 node 🚀 Cost profit margin 545% 💡 We hope this week's insights offer value to the community and contribute to our shared AGI goals. 📖 Deep Dive: bit.ly/4ihZUiO

English
7
7
128
16K
Runxin Xu retweetledi
DeepSeek
DeepSeek@deepseek_ai·
🚀 Day 0: Warming up for #OpenSourceWeek! We're a tiny team @deepseek_ai exploring AGI. Starting next week, we'll be open-sourcing 5 repos, sharing our small but sincere progress with full transparency. These humble building blocks in our online service have been documented, deployed and battle-tested in production. As part of the open-source community, we believe that every line shared becomes collective momentum that accelerates the journey. Daily unlocks are coming soon. No ivory towers - just pure garage-energy and community-driven innovation.
English
1.4K
2.6K
20.6K
2.5M
Runxin Xu retweetledi
Junlong Li
Junlong Li@lockonlvange·
Introducing CodeI/O (codei-o.github.io), a systematic way to condense diverse reasoning patterns via code input-output prediction to build massive training data for more reasoning tasks beyond commonly focused math problem-solving and code generation, which usually suffer from sparse and fragmented supervised signals. The idea is straightforward - teach the model to predict the expected output of a function given an input, or a feasible input given an existed output - but in a large scale with 3.5M instances! To do so, we transform large number of raw code files into executable functions in a unified format, and scale up the input-output pairs by simply executing them for multiple times. In training, we require the models to predict these inputs/outputs by pure natural language CoTs (instead of writing some code and call other interpreter to solve them) to promote the development of transferable reasoning skills that are independent of specific programming syntax. This simple code-derived task gives surprisingly robust performance gain on a broad range of tasks beyond code reasoning itself, including logic, symbolic, math & numerical, scientific, commonsense reasoning. ---------------------------------------------------- Another great advantage of CodeI/O is all data samples we constructed are verifiable (!), and we take full use of this feature to update our data to CodeI/O++ through multi-turn verification and revision, leading to better performance overall. We also hope these verifiable samples can boost RL training like DeepSeek-R1 beyond math and code generation as promising future work. ---------------------------------------------------- We have released our code, models and data to facilitate further research in the community. You can find more details by clicking these links: Website: codei-o.github.io Paper: huggingface.co/papers/2502.07… Code: github.com/hkust-nlp/Code… Released Resources: huggingface.co/collections/hk… ---------------------------------------------------- Finally, I would like to express my sincere gratitude to my collaborators @Guodaya, Dejian Yang, @pigjunebaba, my mentor @YuWu78356266, and my advisor @junxian_he for their help and for giving me the valuable opportunity to learn from them!
Junlong Li tweet mediaJunlong Li tweet media
English
5
54
164
34.3K
Runxin Xu retweetledi
Varun Mohan
Varun Mohan@_mohansolo·
Today we’re making DeepSeek R1 and V3 available in Windsurf, making Cascade the first coding agent to support R1. It'll be half the cost to start with but we are committed to rapidly reduce this with time. R1 is truly fun and reading the chain of thought almost feels like a requirement for reasoning models.
Varun Mohan tweet media
English
81
74
806
115.1K
Runxin Xu retweetledi
DeepSeek
DeepSeek@deepseek_ai·
DeepSeek has not issued any cryptocurrency. Currently, there is only one official account on the Twitter platform. We will not contact anyone through other accounts.Please stay vigilant and guard against potential scams.
English
603
682
6.5K
2.2M
Runxin Xu
Runxin Xu@pigjunebaba·
Every breakthrough reminds us there's still so much to learn and improve. Let's keep pushing boundaries together! #OpenSource #Innovation 🚀
Deli Chen@victor207755822

🚨BREAKING: #DeepSeek just topped BOTH US & China iOS free charts - ZERO paid ads! 🎉🚀 We've known since day 1: #AI products ≠ mobile apps. When core tech makes quantum leaps, even simple interfaces become magic. Time to double down on hardcore innovation. 💪

English
11
6
106
21.5K
Runxin Xu
Runxin Xu@pigjunebaba·
Nice work! RL is truly fascinating😋!
Junxian He@junxian_he

We replicated the DeepSeek-R1-Zero and DeepSeek-R1 training on 7B model with only 8K examples, the results are surprisingly strong. 🚀 Starting from Qwen2.5-Math-7B (base model), we perform RL on it directly. No SFT, no reward model, just 8K MATH examples for verification, the resultant model achieves (pass@1) 33.3% on AIME, 62.5% on AMC, and 77.2% on MATH, outperforming Qwen2.5-math-7B-instruct and being comparable to PRIME and rStar-MATH that use >50x more data and more complicated components. 🚀 Increased CoT length and self-reflection emerge We share the details and our findings in the blog: hkust-nlp.notion.site/simplerl-reason Training code and implementation details here: github.com/hkust-nlp/simp…

English
8
6
153
23.3K
Runxin Xu
Runxin Xu@pigjunebaba·
@wzihanw 能不能帮哥们涨涨x的粉,不太会玩,哭了hh
中文
18
4
100
50.4K