Michal Pstrag

16 posts

Michal Pstrag banner
Michal Pstrag

Michal Pstrag

@micpsst

applied ai / engineering

Katılım Haziran 2025
33 Takip Edilen3 Takipçiler
Nikunj Kothari
Nikunj Kothari@nikunj·
Excited to help the @womeninaisf club host a Claude Code event w/ the team at @AnthropicAI next week. We're looking for volunteers to DEMO their projects - if this is one of you, please sign up at the link in the comments. Link to register to attend - also in the next tweet!
Nikunj Kothari tweet media
English
16
14
178
22.3K
Harry Stebbings
Harry Stebbings@HarryStebbings·
The craft of your product is the respect for your customers. We do not release 30% of shows we record. Every single show, 10 years in, with large media teams, I still listen to every single one pre-release. We remove 35-45% of shows to optimise word to value ratio. Details are not details, they are the product.
English
28
2
77
8.2K
Alex Wa
Alex Wa@_djdumpling·
new blog! What methodologies do labs use to train frontier models? The blog distills 7 open-weight model reports from frontier labs, covering architecture, stability, optimizers, data curation, pre/mid/post-training + RL, and behaviors/safety djdumpling.github.io/2026/01/31/fro…
Alex Wa tweet media
English
34
287
2K
279.4K
diana
diana@dianalokada·
if your girl is texting you back fast and being extra nice, im sorry to tell you but it’s not your girl, it’s her Clawdbot focus on yourself king
English
14
5
165
12.7K
Pedro Domingos
Pedro Domingos@pmddomingos·
RIP Stack Overflow.
Pedro Domingos tweet media
English
804
1.9K
20.1K
4.7M
Viv 🪩
Viv 🪩@battleangelviv·
.@starlink has forever changed the world because it means that no tyrant or authoritarian government is able cut off their people from the world
English
19
16
324
8K
tic toc
tic toc@TicTocTick·
A fool wastes an hour in the gym burning 400 calories. Wise man decides in 3 seconds not to eat a 500 calorie cheese cake 🍰
English
27
7
192
10.1K
Michal Pstrag
Michal Pstrag@micpsst·
@Teknium You can prefill theorems/definitions. It's basically what reasoning tries to do - chiming in the missing pieces of relevant context
English
1
0
1
62
Teknium (e/λ)
Teknium (e/λ)@Teknium·
@micpsst the reasoning cots are not really just about information retreival, so I'm not sure I agree. What are you doing with RAG that solves olympiad math problems aside from rag'ing the answers?
English
1
0
0
358
Teknium (e/λ)
Teknium (e/λ)@Teknium·
Reasoning in LLMs has actually broken at least one intuition about data that I thought I was confident in. Prior to reasoning models, there was a lot I could predict based on the data that went in, such as things like average output lengths and limits on how many tokens they'd generate. It used to be if you trained on outputs of 4k max, you'd have a near-0% chance to generate 10k+ tokens. But with reasoning models, they actually learn a function through this data that can make it generate way way way beyond what length of output tokens you trained on. I think this justifies calling it "reasoning", because it actually learned a function similar to reasoning, by generating tokens that look like thinking to improve accuracy until it is confident in it having found the correct answer, and even if you train on 10k cot tokens max, models will still think, potentially through the entire 128k+ context length it has. Something else interesting about "reasoning" is that we observed when scaling Hermes 4 from 14b, to 70b, to 405B, that the thinking lengths went down and down for the same set of problems as the model got bigger. This also implies that the reasoning process is very much tied to innate intelligence, because the problem is, relative to each model, a different difficulty, and it literally *thinks longer* if it is less intelligent! Just some fun facts for you on this Sunday :)
English
46
41
662
44.6K
Michal Pstrag
Michal Pstrag@micpsst·
True, but doable with GraphRAG-ish solutions. It also feels more stable than relying on reasoning at inference. Most of the time, when reasoning activates, the model just repeats or paraphrases what's already in the prefilled context, which is generally super low info. You can just duplicate the context and often get the same outcome. There's a recent Google paper on this, though they didn't benchmark against reasoning models arxiv.org/pdf/2512.14982
English
1
0
2
76
Eric Zhang
Eric Zhang@ekzhang1·
I'm open-sourcing jax-js — a machine learning library for the web, in pure JavaScript jax-js is the first ML compiler that runs in the browser, generating fast WebGPU kernels. Built from scratch over the past year as a personal side project Details: ekzhang.substack.com/p/jax-js-an-ml…
Eric Zhang tweet media
English
92
296
2.9K
183.4K
diana
diana@dianalokada·
girls are like ugh he is so weird and autistic i wish he was my boyfriend
English
31
15
595
26.3K
Pedro Domingos
Pedro Domingos@pmddomingos·
Report from NeurIPS: the state of the art in LLMs is doing fake reasoning with fake RL.
English
16
17
252
32.7K
Kat ⊷ the Poet Engineer
Kat ⊷ the Poet Engineer@poetengineer__·
not sure if it's just me, but chatgpt used to update the conversation's title as its content evolves, and now the title's just fixed on whatever you ask first.
English
2
0
11
3.2K