Tianqing Fang

70 posts

Tianqing Fang banner
Tianqing Fang

Tianqing Fang

@TFang229

Researcher @TencentGlobal AI Lab | PhD @HKUST (@HKUSTKnowComp) | Former intern @epfl_en, @CSatUSC, @NVIDIAAI

Katılım Mart 2023
244 Takip Edilen369 Takipçiler
Sabitlenmiş Tweet
Tianqing Fang
Tianqing Fang@TFang229·
🚀 We are thrilled to release a new open-source Deep Research Agent, Cognitive Kernel-Pro, from Tencent AI Lab! We focus on building a fully open-source agent with (to the maximum extent) free tools, showcasing impressive performance on GAIA with Claude-3.7-sonnet and surpass the counterpart, SmolAgents by a large margin. In addition, we study the training recipe for an open-source Deep Research Agent Foundation Model. We curate high-quality training data (queries, trajectories, and verifiable answers across web, file, code, and reasoning domains). Our finetuned Qwen3-8B (CK-Pro-8B) surpasses WebDancer and WebSailor with the similar model size on the text-only subset of GAIA. 📜 Paper: arxiv.org/pdf/2508.00414 🔧 Code: github.com/Tencent/Cognit… 🤗 Data & Model: huggingface.co/datasets/Cogni… huggingface.co/CognitiveKerne… This work builds on the previous efforts of Tencent AI Lab (Fig. 2). Be sure to check them out if you're interested!
Tianqing Fang tweet mediaTianqing Fang tweet mediaTianqing Fang tweet mediaTianqing Fang tweet media
English
0
24
84
8.5K
Tianqing Fang retweetledi
Wenhao Yu
Wenhao Yu@wyu_nd·
📢New paper: Guided Self-Evolving LLMs with Minimal Human Supervision Self-evolving / Self-improving LLMs often plateau fast due to concept drift, diversity collapse, and mis-evolution. Our method fixes this — keeping self-evolution stable, aligned, and on track! Link: arxiv.org/abs/2512.02472
Wenhao Yu tweet media
English
6
39
219
69.7K
Tianqing Fang
Tianqing Fang@TFang229·
#emnlp2025 I’m presenting several papers I authored at Tencent AI Lab about agent self evolving and LLM memory. There will also be an oral presentation on the WebEvolver paper at Thu. Nov 6 at 16:30-18:00 Location: A110. Come and chat with me 😆
Tianqing Fang tweet mediaTianqing Fang tweet media
English
3
2
53
10.4K
Tianqing Fang retweetledi
Longyue Wang
Longyue Wang@wangly0229·
🚀 Thrilled to share our breakthrough! Our 🍁Marco‑MT🍁 achieved outstanding results at #WMT2025 General Translation! 🏆 @AI_AlibabaInt Final Human Evaluation Results: 🏅6 First Places 🥈4 Second Places 🥉2 Third Places
Longyue Wang tweet media
English
0
4
9
550
Tianqing Fang retweetledi
Wenhao Yu
Wenhao Yu@wyu_nd·
Code for 𝐏𝐚𝐫𝐚𝐥𝐥𝐞𝐥-𝐑𝟏 is live! 👉 github.com/zhengkid/Paral… (now 189 stars and climbing 🔥) It lets LLMs think in parallel — multiple reasoning paths, smarter synthesis, more creative inference! Miss this paper and you’re missing a leap forward: arxiv.org/abs/2509.07980
Wenhao Yu tweet media
English
2
48
230
14.3K
Tianqing Fang retweetledi
Zhenwen Liang
Zhenwen Liang@LiangZhenwen·
@wyu_nd and I are recruiting 2026 Spring/Summer Research Interns at Tencent AI Lab 🚀 Topics include self-evolving, Agent Systems, Complex Reasoning, etc. We are also hiring full-time researchers with PhD degrees, fully publication-driven. Please DM or email.
English
4
8
55
4.6K
Tianqing Fang retweetledi
Wenhao Yu
Wenhao Yu@wyu_nd·
New paper: VLMs can self-reward during RL training — no visual annotations needed! -- Decompose VLM reasoning into visual vs. language parts -- Prompt the same VLM without visual input for visual reward We call it 𝐕𝐢𝐬𝐢𝐨𝐧-𝐒(𝐞𝐥𝐟)𝐑𝟏: arxiv.org/abs/2508.19652
Wenhao Yu tweet media
English
7
89
439
49.4K
Tianqing Fang retweetledi
Shizhe Diao
Shizhe Diao@shizhediao·
This week, we open-sourced NVIDIA-Nemotron-Nano-v2-9B: our next-generation efficient hybrid model. - 6× faster than Qwen3-8B at reasoning tasks. - Retained long-context capability (8k → 262k trained, usable at 128k) First true demonstration that reasoning models can be compressed without bespoke architectures 👉 Report: research.nvidia.com/labs/adlr/file… 👉 HuggingFace: huggingface.co/collections/nv…
Shizhe Diao tweet media
English
6
25
134
27.7K
Tianqing Fang retweetledi
Tianqing Fang retweetledi
DailyPapers
DailyPapers@HuggingPapers·
Microsoft just released Agent Lightning on Hugging Face. Train ANY AI agents with Reinforcement Learning with almost ZERO code change! A flexible and extensible framework that fully decouples agents from RL training.
DailyPapers tweet media
English
3
12
84
8.1K
Tianqing Fang retweetledi
Wenhao Yu
Wenhao Yu@wyu_nd·
𝑳𝑳𝑴𝒔 can really 𝑺𝒆𝒍𝒇-𝑬𝒗𝒐𝒍𝒗𝒆, 𝒘𝒊𝒕𝒉𝒐𝒖𝒕 𝑯𝒖𝒎𝒂𝒏 𝑫𝒂𝒕𝒂! -- One LLM, two roles: Challenger creates tasks, Solver answers them. -- No data, no labels, just a base model that learns and improves itself! We name it 𝑹-𝒛𝒆𝒓𝒐: arxiv.org/abs/2508.05004
Wenhao Yu tweet media
English
34
243
1.4K
143.4K
Tianqing Fang retweetledi
DailyPapers
DailyPapers@HuggingPapers·
Tencent AI Lab introduces R-Zero! A groundbreaking framework enabling LLMs to self-evolve their reasoning capabilities from zero human-curated data, through an autonomous Challenger-Solver loop.
DailyPapers tweet media
English
7
79
489
78.3K
Tianqing Fang retweetledi
DailyPapers
DailyPapers@HuggingPapers·
Tencent AI Lab just released Cognitive Kernel-Pro on Hugging Face! A fully open-source & free multi-module agent framework designed for deep research & agent foundation model training. Achieves state-of-the-art among open-source agents on GAIA.
DailyPapers tweet media
English
2
6
20
1.3K
Tianqing Fang
Tianqing Fang@TFang229·
Thank you for sharing our work! The code, data, and models have been open-sourced for the research community’s benefit: github.com/Tencent/Cognit… huggingface.co/CognitiveKerne… huggingface.co/datasets/Cogni…
Rohan Paul@rohanpaul_ai

Cognitive Kernel‑Pro shows that a small 8B open‑source language model can run a multi‑skill research agent without paying for proprietary APIs. It wraps web, file, and code tools in one Python based framework and still tops other free agents on the GAIA benchmark. The framework splits work between a planner and specialist sub‑agents. The planner keeps four lists that record finished steps, upcoming tasks, lessons, and gathered facts. Each sub‑agent writes Python code, so the same language model can browse live pages, read PDFs, crunch tables, or run quick scripts. Everything moves through a plain text interface, so new skills bolt on quickly. To teach the model, the team built roughly 15K multi‑step examples that cover web navigation, document reading, math, and coding. Another large model first explored the internet, stored every intermediate step, then those hints were stripped before fine‑tuning. Extra synthetic questions came from PersonaHub, where a generated persona sparks a fresh web task that the agent later checks by itself. During runtime the agent judges its own work. After each attempt it writes a short diary of actions and flags answers that are empty, off topic, or built on weak evidence. If something looks wrong it retries, then a voting step picks the best result among several runs, which steadies performance on changing pages. With Claude‑3.7 as backbone the system reaches 70.9 pass\@3 on 165 GAIA tasks, beating every other open framework that avoids paid parsers and crawlers. This study shows how clear state design, transparent code actions, and rich training traces can close much of the gap between hobby hardware and closed commercial agents. ---- Paper – arxiv. org/abs/2508.00414 Paper Title: "Cognitive Kernel-Pro: A Framework for Deep Research Agents and Agent Foundation Models Training"

English
0
1
8
348
Tianqing Fang retweetledi
Wenhao Yu
Wenhao Yu@wyu_nd·
🗒️Have been exploring Agent-RL training over the past few months, particularly in GUI scenarios. Here’s a summary of some practical insights and lessons 🤔 learned from the perspective of an industry researcher, and some reference papers.
Wenhao Yu tweet mediaWenhao Yu tweet media
English
2
15
118
9K