Thinking Machines

33 posts

Thinking Machines

Thinking Machines

@thinkymachines

Thinking, beeping, and booping. @tinkerapi

Katılım Şubat 2025
1 Takip Edilen115.5K Takipçiler
Thinking Machines retweetledi
Tinker
Tinker@tinkerapi·
Four Qwen 3.5 models from @Alibaba_Qwen are now live on Tinker. Qwen 3.5 introduces hybrid linear attention that enables long context windows, as well as native vision input.
Tinker tweet media
English
10
25
371
235K
Thinking Machines retweetledi
Tinker
Tinker@tinkerapi·
Custom fine-tuning has boundless applications. This week’s roundup shows Tinkerers taking models in unexpected directions, from science research to fun and games (literally).
English
3
19
154
159.4K
Thinking Machines retweetledi
Tinker
Tinker@tinkerapi·
Our second roundup of community projects highlights all things RL, from tutorials to APIs to cutting-edge research.
English
3
12
175
96.6K
Thinking Machines
Thinking Machines@thinkymachines·
@axiommathai To be specific, AxiomMath used Tinker to do RL in developing AxiomProver!
English
3
7
106
41.9K
Thinking Machines
Thinking Machines@thinkymachines·
Congratulations to @axiommathai on their achievement! AxiomProver, a mathematics model fine-tuned with Tinker, got top scores on the Putnam Math Competition.
Axiom@axiommathai

Putnam, the world's hardest college-level math test, ended yesterday 4p PT. Noon today, AxiomProver solved 9/12 problems in Lean autonomously (3:58p PT yesterday, it was 8/12). Our score would've been #1 of ~4000 participants last year and Putnam Fellow (top 5) in recent years

English
13
42
674
290.5K
Thinking Machines
Thinking Machines@thinkymachines·
In addition to expanding capacity, we are adding new models to our lineup and working on image support and production inference. We are excited to see what you build with Tinker!
English
5
3
98
37.2K
Thinking Machines
Thinking Machines@thinkymachines·
Starting Monday, November 3rd, Tinker is switching to a pricing plan that reflects compute usage. This will ensure we have sufficient capacity to clear our waitlist by the end of the year, allowing anyone to sign up and start Tinkering. tinker-console.thinkingmachines.ai/rate-card
Thinking Machines tweet media
English
7
13
354
257.1K
Thinking Machines
Thinking Machines@thinkymachines·
Roadmap update: Tinker launched into private beta a month ago, and we've seen hundreds of builders and researchers train and experiment with models on our platform. This month we've added new models, expanded the cookbook, and improved overall capacity and performance.
Thinking Machines@thinkymachines

We just added 4 new models to Tinker from the gpt-oss and DeepSeek-V3.1 families. Sign up for the waitlist: thinkingmachines.ai/tinker/

English
115
51
553
252.8K
Thinking Machines
Thinking Machines@thinkymachines·
Today we’re announcing research and teaching grants for Tinker: credits for scholars and students to fine-tune and experiment with open-weight LLMs. Read more and apply at: thinkingmachines.ai/blog/tinker-re…
English
20
126
1K
494.9K
Thinking Machines
Thinking Machines@thinkymachines·
Our latest post explores on-policy distillation, a training approach that unites the error-correcting relevance of RL with the reward density of SFT. When training it for math reasoning and as an internal chat assistant, we find that on-policy distillation can outperform other approaches for a fraction of the cost. thinkingmachines.ai/blog/on-policy…
Thinking Machines tweet media
English
61
409
2.8K
1.9M
Thinking Machines retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Tinker is cool. If you're a researcher/developer, tinker dramatically simplifies LLM post-training. You retain 90% of algorithmic creative control (usually related to data, loss function, the algorithm) while tinker handles the hard parts that you usually want to touch much less often (infra, forward/backward of the LLM itself, distributed training), meaning you can do these at well below <<10% of typical complexity involved. Compared to the more common and existing paradigm of "upload your data, we'll post-train your LLM", this is imo a more clever place to "slice up" the complexity of post-training, both delegating the heavy lifting, but also keeping majority of the data/algorithmic creative control. I think the community still has to discover how and when finetuning makes sense compared to the (often strong) baseline of prompting a giant model. The early indications I've seen is that finetuning isn't so much about "stylizing" an LLM, instead, it's a lot more about narrowing the scope, and especially when you have a lot of training examples. An extreme example of scope narrowing being that of categorical classifiers, e.g.spam filters, content filters, etc. but it should be broader than that. Instead of building a giant few-shot prompts for a big LLM, it might work a lot better (and faster!) to finetune a smaller LLM specifically for your narrow task. Increasingly, production applications of LLMs are larger pipelines where a bunch of LLMs collaborate in DAGs and flows. Some of these components might work well as prompts. But a lot of it will probably work a lot better as a finetune. Tinker makes the latter trivial and should allow for an easy experimentation of what works best at any stage.
Thinking Machines@thinkymachines

Introducing Tinker: a flexible API for fine-tuning language models. Write training loops in Python on your laptop; we'll run them on distributed GPUs. Private beta starts today. We can't wait to see what researchers and developers build with cutting-edge open models! thinkingmachines.ai/tinker

English
108
647
6.2K
742.6K
Thinking Machines retweetledi
Philipp Moritz
Philipp Moritz@pcmoritz·
Very excited to see the Tinker release by @thinkymachines! @robertnishihara and I had a chance to experiment with the API, see anyscale.com/blog/fine-tuni…. It does a nice job of providing flexibility while abstracting away GPU handling. This will be 🔥 when combined with @raydistributed for simulations, inference and data processing. Looking forward to all the experimentation this unlocks! anyscale.com/blog/massively…
Philipp Moritz tweet media
English
4
23
157
41K
Thinking Machines retweetledi
Robert Nishihara
Robert Nishihara@robertnishihara·
Very excited to see the Tinker release! @pcmoritz and I had a chance to experiment with the API. It does a nice job of providing flexibility while abstracting away GPU handling. Here's a simple example showing how to generate synthetic data and fine tune a text to SQL model. The example uses - @NovaSkyAI: the skyrl-gym environment for executing SQL queries and calculating rewards. - @thinkymachines: Tinker's fine-tuning API. - @raydistributed: batch inference and data generation via Ray. anyscale.com/blog/fine-tuni…
Thinking Machines@thinkymachines

Introducing Tinker: a flexible API for fine-tuning language models. Write training loops in Python on your laptop; we'll run them on distributed GPUs. Private beta starts today. We can't wait to see what researchers and developers build with cutting-edge open models! thinkingmachines.ai/tinker

English
12
37
266
93.2K