Thinking Machines

34 posts

Thinking Machines

Thinking Machines

@thinkymachines

Thinking, beeping, and booping. @tinkerapi

Katılım Şubat 2025
1 Takip Edilen117.6K Takipçiler
Thinking Machines retweetledi
Tinker
Tinker@tinkerapi·
Four Qwen 3.5 models from @Alibaba_Qwen are now live on Tinker. Qwen 3.5 introduces hybrid linear attention that enables long context windows, as well as native vision input.
Tinker tweet media
English
11
27
389
243.3K
Thinking Machines retweetledi
Tinker
Tinker@tinkerapi·
Custom fine-tuning has boundless applications. This week’s roundup shows Tinkerers taking models in unexpected directions, from science research to fun and games (literally).
English
3
16
163
164.8K
Thinking Machines retweetledi
Tinker
Tinker@tinkerapi·
Our second roundup of community projects highlights all things RL, from tutorials to APIs to cutting-edge research.
English
3
12
178
100.5K
Thinking Machines
Thinking Machines@thinkymachines·
@axiommathai To be specific, AxiomMath used Tinker to do RL in developing AxiomProver!
English
3
7
106
43.4K
Thinking Machines
Thinking Machines@thinkymachines·
Congratulations to @axiommathai on their achievement! AxiomProver, a mathematics model fine-tuned with Tinker, got top scores on the Putnam Math Competition.
Axiom@axiommathai

Putnam, the world's hardest college-level math test, ended yesterday 4p PT. Noon today, AxiomProver solved 9/12 problems in Lean autonomously (3:58p PT yesterday, it was 8/12). Our score would've been #1 of ~4000 participants last year and Putnam Fellow (top 5) in recent years

English
13
40
670
294.3K
Thinking Machines
Thinking Machines@thinkymachines·
In addition to expanding capacity, we are adding new models to our lineup and working on image support and production inference. We are excited to see what you build with Tinker!
English
5
3
98
38K
Thinking Machines
Thinking Machines@thinkymachines·
Starting Monday, November 3rd, Tinker is switching to a pricing plan that reflects compute usage. This will ensure we have sufficient capacity to clear our waitlist by the end of the year, allowing anyone to sign up and start Tinkering. tinker-console.thinkingmachines.ai/rate-card
Thinking Machines tweet media
English
7
13
353
258.3K
Thinking Machines
Thinking Machines@thinkymachines·
Roadmap update: Tinker launched into private beta a month ago, and we've seen hundreds of builders and researchers train and experiment with models on our platform. This month we've added new models, expanded the cookbook, and improved overall capacity and performance.
Thinking Machines@thinkymachines

We just added 4 new models to Tinker from the gpt-oss and DeepSeek-V3.1 families. Sign up for the waitlist: thinkingmachines.ai/tinker/

English
115
47
549
257.3K
Thinking Machines
Thinking Machines@thinkymachines·
Today we’re announcing research and teaching grants for Tinker: credits for scholars and students to fine-tune and experiment with open-weight LLMs. Read more and apply at: thinkingmachines.ai/blog/tinker-re…
English
20
126
1K
498.4K
Thinking Machines
Thinking Machines@thinkymachines·
Our latest post explores on-policy distillation, a training approach that unites the error-correcting relevance of RL with the reward density of SFT. When training it for math reasoning and as an internal chat assistant, we find that on-policy distillation can outperform other approaches for a fraction of the cost. thinkingmachines.ai/blog/on-policy…
Thinking Machines tweet media
English
61
406
2.8K
1.9M
Thinking Machines retweetledi
Andrej Karpathy
Andrej Karpathy@karpathy·
Tinker is cool. If you're a researcher/developer, tinker dramatically simplifies LLM post-training. You retain 90% of algorithmic creative control (usually related to data, loss function, the algorithm) while tinker handles the hard parts that you usually want to touch much less often (infra, forward/backward of the LLM itself, distributed training), meaning you can do these at well below <<10% of typical complexity involved. Compared to the more common and existing paradigm of "upload your data, we'll post-train your LLM", this is imo a more clever place to "slice up" the complexity of post-training, both delegating the heavy lifting, but also keeping majority of the data/algorithmic creative control. I think the community still has to discover how and when finetuning makes sense compared to the (often strong) baseline of prompting a giant model. The early indications I've seen is that finetuning isn't so much about "stylizing" an LLM, instead, it's a lot more about narrowing the scope, and especially when you have a lot of training examples. An extreme example of scope narrowing being that of categorical classifiers, e.g.spam filters, content filters, etc. but it should be broader than that. Instead of building a giant few-shot prompts for a big LLM, it might work a lot better (and faster!) to finetune a smaller LLM specifically for your narrow task. Increasingly, production applications of LLMs are larger pipelines where a bunch of LLMs collaborate in DAGs and flows. Some of these components might work well as prompts. But a lot of it will probably work a lot better as a finetune. Tinker makes the latter trivial and should allow for an easy experimentation of what works best at any stage.
Thinking Machines@thinkymachines

Introducing Tinker: a flexible API for fine-tuning language models. Write training loops in Python on your laptop; we'll run them on distributed GPUs. Private beta starts today. We can't wait to see what researchers and developers build with cutting-edge open models! thinkingmachines.ai/tinker

English
108
642
6.1K
746.6K
Thinking Machines retweetledi
Philipp Moritz
Philipp Moritz@pcmoritz·
Very excited to see the Tinker release by @thinkymachines! @robertnishihara and I had a chance to experiment with the API, see anyscale.com/blog/fine-tuni…. It does a nice job of providing flexibility while abstracting away GPU handling. This will be 🔥 when combined with @raydistributed for simulations, inference and data processing. Looking forward to all the experimentation this unlocks! anyscale.com/blog/massively…
Philipp Moritz tweet media
English
4
24
156
41.5K