Collinear AI
79 posts









Tinker for training exists, but Tinker for data doesn't. Yet, researchers spend most of the time on data preprocessing / generation and training integration. This Halloween, we introduce spider, ie. Tinker for data. It spins up a client for users to define a production-grade distillation run in a few lines of code. Features in the mvp: - For off-policy distillation, users can define custom preprocessing logic for any dataset, and spin up a remote inference engine for high throughput rollout generation. In 2-3 lines a whole dataset can be created. - For on-policy distillation revived by @thinkymachines , users can simply toggle on_policy: true so that the same workflow now admits a teacher model and on-policy kl supervision. Under the hood a Tinker service client is integrated. Again, 2-3 lines of code. We want to make spider as useful for open-source research as possible and support fast validation for experiment ideas. Next up we will implement the cross-tokenizer approach by @huggingface to support on-policy distillation from and to any model. And we will make filtering, inference, etc. more intelligent and simple for end users.

Together AI 🤝@CollinearAI Introducing TraitMix, Collinear’s simulation product empowering teams to generate persona-driven AI agent interactions. 🔌Plug these interactions into your workflows and evaluate their effectiveness with Together Evals. Details: bit.ly/43GHJhR









We’re growing and hiring! I’m looking for Research Scientists and Research Engineers passionate about pushing the boundaries of post-training AI technologies. We've shipped 100B+ tokens of high-quality data in a very short time and enabled enterprises to save serious $$ while dramatically improving their AI performance, safety, and reliability. If you’re excited by post-training, RL envs, and enabling each individual and organization to build great AI — let’s talk. DM me if you’re curious. Happy to chat!

We’re growing and hiring! I’m looking for Research Scientists and Research Engineers passionate about pushing the boundaries of post-training AI technologies. We've shipped 100B+ tokens of high-quality data in a very short time and enabled enterprises to save serious $$ while dramatically improving their AI performance, safety, and reliability. If you’re excited by post-training, RL envs, and enabling each individual and organization to build great AI — let’s talk. DM me if you’re curious. Happy to chat!



