OpenPipe

113 posts

OpenPipe banner
OpenPipe

OpenPipe

@OpenPipeAI

OpenPipe: Fine-tuning for production apps. Train higher quality, faster models. (YC S23)

San Francisco 加入时间 Ağustos 2023
7 关注4.1K 粉丝
OpenPipe 已转推
Shane Henke
Shane Henke@realshanehenke·
Today was the first time I used my fine-tuned model I used Llama 3.1 using @OpenPipeAI the results are encouraging. There are like 5 more to build so wish me luck 😂
Shane Henke tweet media
English
0
1
3
1.6K
Charles 🎉 Frye
Charles 🎉 Frye@charles_irl·
Start in Python, then Rewrite It In Rust™ once you understand the problem and the requirements. Start with off-the-shelf large models, then migrate to custom small models once you've collected enough data and explored failure modes.
Charles 🎉 Frye tweet media
English
7
14
263
16.4K
OpenPipe 已转推
Hina
Hina@soleil_colza_·
We as a team are learning RL for the first time this weekend. We're totally new to this field, but @OpenPipeAI's docs really helped us kickstart! @wandb
English
1
2
10
3.4K
OpenPipe 已转推
Viv
Viv@Vtrivedy10·
some thoughts on why/how the standard paradigm for optimizing Task Specific Agents will be Harness Engineering + Rubric Based Task Specific RL. this write up codifies a lot of my thoughts on where harness engineering is going + inspiration from @corbtt on @latentspacepod Steps: 1. Obsessively hand/auto tune agent harness until you reach a baseline threshold of task performance. Goal: make sure the agent has roughly what it needs to succeed. 2. Do Task Specific RL to make the model better at operating in the harness, touching the model weights pushes us beyond what harness engineering alone can do. FAQs: 1. Why in this order? If your agent rarely succeeds on the Task, it can’t get enough of a reward, this makes RL difficult. Optimize the harness first —> “You can’t succeed if you don’t have the right tools” 2. Why Not Just Keep Optimizing your Harness, I thought Prompt Engineering is the way? Harness engineering in the latter stages is incredibly hard. It’s also combinatorially complex from the start. You have to jointly optimizes each component(System Prompt, Tools/Skills/MCP, Subagent definitions, Additional Context) but you have: - A Selection Problem: How do you intelligently select relevant tools from a sea of possibilities? Context is a precious resource and selecting too many tools is confusing and degrades perf. - Codependency: No component is optimized in isolation, it’s one big system (ex: changing the system prompt may change how/if a tool is called) 3. So how should I start? First painstakingly test everything in your harness. - try multiple models - hand tune system prompts, try GEPA - more/less tools, tool descriptions, compound tools - handing off tasks to subagents - preloading useful notes, docs, and instructions as references eventually when you hit a wall (on performance or human resources), you move on. You can also move on much earlier once you hit some performance threshold. 4. What does RL get you? Agents (Claude Code, Codex) from the labs are so reliable at using their tools (WebSearch, Multi-Edit, Grep) because they’re directly post-trained with them. We want that same for our tasks, we want to make the model more comfortable using its harness while training to increase Task performance. 5. How should I get started with Task Specific RL? A fantastic first place to start is RULER from @OpenPipeAI which relies on rubrics created by you + LLM as a judge across multiple generations. For your task, you probably already have a set of ideas on what’s good, codifying that in a rubric is all you need to get started working on writing up a walkthrough blog of this with code. really excited about building products that treat agent building as a harness optimization problem that you measure deeply + push further with RL
English
5
5
40
11.1K
OpenPipe 已转推
Weights & Biases
Weights & Biases@wandb·
LIVE: Kyle Corbitt, Head of the OpenPipe team at CoreWeave, joins ThursdAI to talk about launching the first Serverless Reinforcement Learning capability. x.com/i/broadcasts/1…
English
0
2
7
2K
OpenPipe 已转推
Santiago Pombo
Santiago Pombo@SantiagoPombo·
The Custom SLMs era is upon us 🙌 - Nanochat by @karpathy - Thinker (PEFTaaS) by @thinkymachines - Tunix (Post-train in Jax) by @GoogleAI - Art (Agent RL) by @OpenPipeAI - Environments Hub by @PrimeIntellect - NeMo Microservices by @nvidia
Andrej Karpathy@karpathy

Excited to release new repo: nanochat! (it's among the most unhinged I've written). Unlike my earlier similar repo nanoGPT which only covered pretraining, nanochat is a minimal, from scratch, full-stack training/inference pipeline of a simple ChatGPT clone in a single, dependency-minimal codebase. You boot up a cloud GPU box, run a single script and in as little as 4 hours later you can talk to your own LLM in a ChatGPT-like web UI. It weighs ~8,000 lines of imo quite clean code to: - Train the tokenizer using a new Rust implementation - Pretrain a Transformer LLM on FineWeb, evaluate CORE score across a number of metrics - Midtrain on user-assistant conversations from SmolTalk, multiple choice questions, tool use. - SFT, evaluate the chat model on world knowledge multiple choice (ARC-E/C, MMLU), math (GSM8K), code (HumanEval) - RL the model optionally on GSM8K with "GRPO" - Efficient inference the model in an Engine with KV cache, simple prefill/decode, tool use (Python interpreter in a lightweight sandbox), talk to it over CLI or ChatGPT-like WebUI. - Write a single markdown report card, summarizing and gamifying the whole thing. Even for as low as ~$100 in cost (~4 hours on an 8XH100 node), you can train a little ChatGPT clone that you can kind of talk to, and which can write stories/poems, answer simple questions. About ~12 hours surpasses GPT-2 CORE metric. As you further scale up towards ~$1000 (~41.6 hours of training), it quickly becomes a lot more coherent and can solve simple math/code problems and take multiple choice tests. E.g. a depth 30 model trained for 24 hours (this is about equal to FLOPs of GPT-3 Small 125M and 1/1000th of GPT-3) gets into 40s on MMLU and 70s on ARC-Easy, 20s on GSM8K, etc. My goal is to get the full "strong baseline" stack into one cohesive, minimal, readable, hackable, maximally forkable repo. nanochat will be the capstone project of LLM101n (which is still being developed). I think it also has potential to grow into a research harness, or a benchmark, similar to nanoGPT before it. It is by no means finished, tuned or optimized (actually I think there's likely quite a bit of low-hanging fruit), but I think it's at a place where the overall skeleton is ok enough that it can go up on GitHub where all the parts of it can be improved. Link to repo and a detailed walkthrough of the nanochat speedrun is in the reply.

English
0
2
5
1.3K
OpenPipe 已转推
Lukas Biewald
Lukas Biewald@l2k·
We've started a great tradition at CoreWeave of shipping an integrated new product weeks after acquisition - congrats @OpenPipeAI on the serverless RL launch!
English
3
4
26
5.2K
Yacine Mahdid
Yacine Mahdid@yacinelearning·
I'm in the unfortunate position to let you know that I've fallen for the RL-LLMs propaganda 100% with these results from openpipe I am now fully RL pilled and there is no turning back very sorry folks
Yacine Mahdid tweet media
Prime Intellect@PrimeIntellect

Introducing the Environments Hub RL environments are the key bottleneck to the next wave of AI progress, but big labs are locking them down We built a community platform for crowdsourcing open environments, so anyone can contribute to open-source AGI

English
18
55
973
101.2K
OpenPipe 已转推
Kyle Corbitt
Kyle Corbitt@corbtt·
🚀 Big launch from @OpenPipeAI: We just launched Serverless RL — train agents faster and cheaper with zero infra headaches. Compared to running your own GPUs, Serverless RL is: - 40% cheaper - 28% faster wall‑clock - instantly deployed to prod via @wandb Inference
Kyle Corbitt tweet mediaKyle Corbitt tweet mediaKyle Corbitt tweet media
English
6
28
222
34.9K
OpenPipe 已转推
sohail
sohail@Sohailmo·
@wandb @CoreWeave the coreweave acquisitions is about to facilitate the ultimate end to end platform
English
0
3
5
1.3K
Weights & Biases
Weights & Biases@wandb·
RL X-mas came early. 🎄 For too long, building powerful AI agents with Reinforcement Learning has been blocked by GPU scarcity and complex infrastructure. That ends today. Introducing Serverless RL from wandb, powered by @CoreWeave! We're making RL accessible to all.
English
8
18
153
112.6K