NovaSky

136 posts

NovaSky banner
NovaSky

NovaSky

@NovaSkyAI

Building SkyRL at @BerkeleySky Join the Slack community: https://t.co/mSO97T61vR

Berkeley, California Se unió Ocak 2025
18 Siguiendo2.7K Seguidores
Tweet fijado
NovaSky
NovaSky@NovaSkyAI·
We are excited to announce that SkyRL now implements the Tinker API. Run Tinker training scripts on your own hardware with zero code changes. Try it out today: novasky-ai.notion.site/skyrl-tinker
Tyler Griggs@tyler_griggs_

SkyRL now implements the Tinker API. Now, training scripts written for Tinker can run on your own GPUs with zero code changes using SkyRL's FSDP2, Megatron, and vLLM backends. Blog: novasky-ai.notion.site/skyrl-tinker 🧵

English
0
4
27
1.8K
NovaSky retuiteado
vLLM
vLLM@vllm_project·
Excited to see SkyRL sharing their work on inference and vLLM in RL at the LLMs on Ray office hours this Thursday. If you’re exploring using vLLM in RL workflows, this will be a great session to join. See you there 👇
Seiji Eicher@seiji_________

Hi all, extending the invite to the LLMs on Ray office hours next Thursday, 3/5 9:30-10:30AM PT! We will be hosting @erictang000 and @sumanthrh from the @NovaSkyAI SkyRL project to present on inference/vLLM in RL and take questions from the group. After, there will be time for any other questions folks have on distributed inference w/ Ray. Hope you can make it! Sign up for the invite here: forms.gle/QESMQ8ojRJsCZV…

English
1
9
43
6.9K
NovaSky
NovaSky@NovaSkyAI·
We’ve been consistently surprised lately by how capable frontier models are at handling complex kernel implementation and system optimization. Check out this work as a step toward automating AI infrastructure building!
Shiyi Cao@shiyi_c98

Introducing our new work K-Search: LLM Kernel Generation via Co-Evolving Intrinsic World Model — a new paradigm for automated GPU kernel generation, achieving SoTA results. 🔍 Big insight: Traditional methods treat LLMs as stochastic code generators inside heuristic loops — but this misses a key point: LLMs are powerful planners with rich domain priors. 🧠 Core idea: K-Search uses the LLM itself as a co-evolving world model — one that plans + updates beliefs + guides search decisions based on experience. 📌 This decouples high-level strategy (intent) from low-level code implementation, allowing the optimizer to pursue multi-step transformations even when intermediate implementations don’t immediately improve performance. 📈 Key results: 🔥 Our discovered kernels are ~2.10× average speedup vs state-of-the-art evolutionary search across 4 FlashInfer kernels on H100/B200. 🔥 Up to 14.3× gain on complex Mixture-of-Experts (MoE) kernels. 🔥 State-of-the-art performance on GPUMode TriMul (H100) task — beating both automated and human solutions. 🙏 Acknowledgements This work is developed in @BerkeleySky, w/ the amazing @ziming_mao, @profjoeyg, and @istoica05. We thank @DachengLi177, @MayankMish98, @randwalk0, @pgasawa, @fangz_zzu, and @tian_xia_ for helpful discussion and feedback. We also thank the generous compute support from @databricks, @awscloud, @anyscalecompute, @nvidia, @Google, @LambdaAPI, and @MayfieldFund. 👨‍💻 GitHub: github.com/caoshiyi/K-Sea… 📄 arXiv: arxiv.org/pdf/2602.19128…

English
0
3
20
2.7K
NovaSky retuiteado
Shiyi Cao
Shiyi Cao@shiyi_c98·
Introducing our new work K-Search: LLM Kernel Generation via Co-Evolving Intrinsic World Model — a new paradigm for automated GPU kernel generation, achieving SoTA results. 🔍 Big insight: Traditional methods treat LLMs as stochastic code generators inside heuristic loops — but this misses a key point: LLMs are powerful planners with rich domain priors. 🧠 Core idea: K-Search uses the LLM itself as a co-evolving world model — one that plans + updates beliefs + guides search decisions based on experience. 📌 This decouples high-level strategy (intent) from low-level code implementation, allowing the optimizer to pursue multi-step transformations even when intermediate implementations don’t immediately improve performance. 📈 Key results: 🔥 Our discovered kernels are ~2.10× average speedup vs state-of-the-art evolutionary search across 4 FlashInfer kernels on H100/B200. 🔥 Up to 14.3× gain on complex Mixture-of-Experts (MoE) kernels. 🔥 State-of-the-art performance on GPUMode TriMul (H100) task — beating both automated and human solutions. 🙏 Acknowledgements This work is developed in @BerkeleySky, w/ the amazing @ziming_mao, @profjoeyg, and @istoica05. We thank @DachengLi177, @MayankMish98, @randwalk0, @pgasawa, @fangz_zzu, and @tian_xia_ for helpful discussion and feedback. We also thank the generous compute support from @databricks, @awscloud, @anyscalecompute, @nvidia, @Google, @LambdaAPI, and @MayfieldFund. 👨‍💻 GitHub: github.com/caoshiyi/K-Sea… 📄 arXiv: arxiv.org/pdf/2602.19128…
Shiyi Cao tweet mediaShiyi Cao tweet media
English
12
63
306
92.6K
NovaSky
NovaSky@NovaSkyAI·
To get started with `skyrl-train` as a library instead of forking, you can follow our instructions here: #installing-skyrl-train-from-pypi" target="_blank" rel="nofollow noopener">skyrl.readthedocs.io/en/latest/gett… to use the PyPI package! (6/n)
NovaSky tweet media
English
1
0
2
297
NovaSky
NovaSky@NovaSkyAI·
We recently released SkyRL-Train v0.3.0! Highlights include: - Experimental support for Pipeline-RL style Async-RL - Updated E2E Recipes page with Math, Search, SQL runs - Migration from mbridge -> Megatron-Bridge - 14 new OSS contributors! (1/n) 🧵
NovaSky tweet media
English
2
6
28
3.1K
NovaSky retuiteado
Negin Raoof
Negin Raoof@NeginRaoof_·
How can we make a better TerminalBench agent? Today, we are announcing the OpenThoughts-Agent project. OpenThoughts-Agent v1 is the first TerminalBench agent trained on fully open curated SFT and RL environments. OpenThinker-Agent-v1 is the strongest model of its size on TerminalBench, and sets a new bar on our newly released OpenThoughts-TB-Dev benchmark. (1/n)
Negin Raoof tweet media
English
17
78
289
124.9K
NovaSky retuiteado
Charlie Ruan
Charlie Ruan@charlie_ruan·
Announcing OpenThoughts-Agent with an incredible team — a data-centric effort on TerminalBench-style tasks, built with SkyRL+Harbor 💻🤖 Co-leading the RL team over the past month has been a blast, and we’re just getting started! (1/n) 🧵
Negin Raoof@NeginRaoof_

How can we make a better TerminalBench agent? Today, we are announcing the OpenThoughts-Agent project. OpenThoughts-Agent v1 is the first TerminalBench agent trained on fully open curated SFT and RL environments. OpenThinker-Agent-v1 is the strongest model of its size on TerminalBench, and sets a new bar on our newly released OpenThoughts-TB-Dev benchmark. (1/n)

English
7
16
49
7.6K
NovaSky retuiteado
Shiyi Cao
Shiyi Cao@shiyi_c98·
🤖 I am in San Diego for #NeurIPS2025 this week! Excited to chat about SkyRL(-Agent), Coding LLM/Agent, Self-evolving Agent, RL, and Inference/Training Infrastructure.
Shiyi Cao tweet media
English
5
3
60
4.6K
NovaSky retuiteado
Dacheng Li
Dacheng Li@DachengLi177·
I will be at San Diego for Neurips from Dec 1 to Dec 7. Happy to chat on RL, LLM agents, memory system / long context, video models and training infrastructure 😄
English
4
2
64
9.6K