RadixArk
16 posts




New course available! Efficient Inference with SGLang: Text and Image Generation is live. LLM inference gets expensive fast—mostly due to redundant computation. This course shows how to reduce that using SGLang, with KV cache and RadixAttention, and how to apply the same ideas to faster image generation. Built with @lmsysorg and @radixark, taught by Richard Chen. Enroll for free: hubs.la/Q04b0F1J0

New course available! Efficient Inference with SGLang: Text and Image Generation is live. LLM inference gets expensive fast—mostly due to redundant computation. This course shows how to reduce that using SGLang, with KV cache and RadixAttention, and how to apply the same ideas to faster image generation. Built with @lmsysorg and @radixark, taught by Richard Chen. Enroll for free: hubs.la/Q04b0F1J0

Introducing OpenReward. 🌍 330+ RL environments through one API ⚡ Autoscaled sandbox compute 🍒 4.5M+ unique RL tasks 🚂 Works like magic with Tinker, Miles, Slime Link and thread below.



🎁 SGLang GTC Giveaway — 20 FREE Passes! SGLang is an open-source LLM serving engine that helps models like DeepSeek, Qwen, Kimi, Minimax, GLM, and Llama run efficiently at production scale. Thanks to our sponsor @radixark, we're giving away 20 NVIDIA GTC 4-day exhibit passes (worth $930 each)! 🎟️ To enter the lottery: 1️⃣ Follow us → @lmsysorg 2️⃣ ⭐ Star SGLang on GitHub → github.com/sgl-project/sg… 3️⃣ Reply with: your favorite open-source model and what you use it for 4️⃣ Repost this for extra visibility How we pick winners: 🏆 Top 5 most engaging comments win directly 🎲 Remaining 15 drawn randomly via xpickr We'll verify your GitHub star before sending tickets, so make sure you've starred the repo! Let's go 👇


Want a front-row seat to the evolution of frontier models? 🤖 I'm building the AI Product team at @radixark. We're scaling SGLang @lmsysorg @sgl_project and defining the future of AI training & inference infrastructure. Open roles in PM, Product Ops, and DevRel. If you want to own products from strategy to GTM, join us! Apply through: job-boards.greenhouse.io/radixark/jobs/… #AI #TechJobs #SGLang #ProductManagement

Excited to see SGLang integrated with Ray 🎉 Ray Serve now supports SGLang for online inference, and Ray Data supports SGLang for batch LLM workloads. Check out the new examples and PRs here [Example] github.com/ray-project/ra… [ray serve + sglang] github.com/ray-project/ra… [ray data + sglang] github.com/ray-project/ra…



🚀 Introducing Miles — an enterprise-facing RL framework for large-scale MoE training & production, forked from slime. Slime is a lightweight, customizable RL framework that already powers real post-training pipelines and large MoE runs. Miles builds on slime but focuses on new hardware (e.g., GB300), large-scale MoE RL, and production-grade stability. Please read more about Miles' current status and roadmap here👇

I've been working with Jinwei, Zehuan and Fast-dllm team for over one month. Glad to see the accomplishment!

We've been running @radixark for a few months, started by many core developers in SGLang @lmsysorg and its extended ecosystem (slime @slime_framework , AReaL @jxwuyi). I left @xai in August — a place where I built deep emotions and countless beautiful memories. It was the best place I’ve ever worked, the place I watched grow from a few dozen people to hundreds, and it truly felt like home. What pushed me to make such a hard decision is the momentum of building SGLang open source and the mission of creating an ambitious future, within an open spirit that I learnt from my first job at @databricks after my PhD. We started SGLang in the summer of 2023 and made it public in January 2024. Over the past 2 years, hundreds of people have made great efforts to get to where they are today. We experienced several waves of growth after its first release. I still remember the many dark nights in the summer of 2024, I spent with @lm_zheng , @lsyincs , and @zhyncs42 debugging, while @ispobaoke single-handedly took on DeepSeek inference optimizations, seeing @GenAI_is_real and the community strike team tag-teaming on-call shifts non-stop. There are so many more who have joined that I'm out of space to call out, but they're recorded on the GitHub contributor list forever. The demands grow exponentially, and we have been pushed to make it a dedicated effort supported by RadixArk. It’s the step-by-step journey of a thousand miles that has carried us here today, and the same relentless Long March that will lead us into the tens of thousands of miles yet to come. The story never stops growing. Over the past year, we’ve seen something very clear: The world is full of people eager to build AI, but the infrastructure that makes it possible is not shared. The most advanced inference and training stacks live inside a few companies. Everyone else is forced to rebuild the same schedulers, compilers, serving engines, and training pipelines again and again — often under enormous pressure, with lots of duplicated effort and wasted insight. RadixArk was born to change that. Today, we’re building an infrastructure-first, deep-tech company with a simple and ambitious mission: "Make frontier-level AI infrastructure open and accessible to everyone." If the two values below resonate with you, come talk to us: (1) Engineering as an art. Infrastructure is a first-class citizen in RadixArk. We care about elegant design and code that lasts. Beneath every line of code lies the soul of the engineer who wrote it. (2) A belief in openness. We share what we build. We bet on long-term compounding through community, contribution, and giving more than we take. A product is defined by its users, yet it truly comes alive the moment functionality transcends mere utility and begins to embody aesthetics. Thanks to all the miles (the name of our first released RL framework; see below). radixark.ai
