ruzz

1.3K posts

ruzz banner
ruzz

ruzz

@findwildruzz

Get the treasure, bring it near, Post a letter, loud and clear. Put the books upon the shelf, Delete the worries from yourself.

Katılım Şubat 2019
493 Takip Edilen78 Takipçiler
ruzz
ruzz@findwildruzz·
@DavideMercuri91 @seelffff 8B you can use the model like a semantic regex, Im testing various 3B/9B models for that and they are great
English
0
0
0
120
Davide Mercuri
Davide Mercuri@DavideMercuri91·
@seelffff What task you can do with 8B? I mean, if you are using ai for doing 1+1 it’s ok but the problem here is not piad subscription for ai, the problem here is how you are using the ai
English
2
0
16
1K
self.dll
self.dll@seelffff·
people think running AI locally requires: → $3,000 MacBook Pro → RTX 4090 → $20/month cloud subscription nvidia just dropped a $249 computer. 67 TOPS. runs llama 3.1-8B locally. no internet. no API. no monthly fee. ever. smaller than your router. costs the same as AirPods. runs the same models you pay $240/year to access via ChatGPT. the local AI era just got a price tag. $249.
self.dll@seelffff

x.com/i/article/2053…

English
109
105
1.1K
299.5K
ruzz
ruzz@findwildruzz·
@seelffff It’s great for ML dev, you can train big NN for like computer vision or general deep learning stuff. For useful LLM (similar to Sota where you can actually iterate on stuff) is too small
English
0
0
1
55
ruzz
ruzz@findwildruzz·
@sanbuphy Currently studying it. Very well done!
English
0
0
1
10
Sanbu 散步
Sanbu 散步@sanbuphy·
花了段时间写了 RL 教程 Hands-On Modern RL,路线是从 CartPole + PPO 入门,然后到 LLM 后训练(RLHF、DPO、GRPO)、Agentic RL。代码先行,公式用来解释现象。英文版很快更新。 目前是草稿版本,RLHF、Agentic RL 部分本地审校中。 欢迎提 PR 或 Issue & 显卡支持:github.com/walkinglabs/ha…
Sanbu 散步 tweet mediaSanbu 散步 tweet mediaSanbu 散步 tweet mediaSanbu 散步 tweet media
中文
47
197
1.3K
166.3K
ruzz retweetledi
Sanbu 散步
Sanbu 散步@sanbuphy·
Qwen3、GLM-5、MiMo 等大模型后训练均采用 OPD,它被视作 RL 的低成本替代方案,但实际落地远比 SFT 和 RL 更脆弱。三篇最新论文揭示了其底层机制与四大失效根源:OPD 名义上是匹配师生全词表分布,实则97%–99% 学习信号仅来自每步少量师生高概率重叠候选 token,依赖双方局部合理下一词集合的适配。
Zhuokai Zhao@zhuokaiz

Qwen3, GLM-5, and MiMo all use on-policy distillation in post-training. Thinking Machines also wrote it up as a cheap alternative to RL. But in practice it is surprisingly brittle to make work — much more so than SFT or RL. Three recent papers [1, 2, 3] helped me make sense of why. The mechanism is consistent across the failure modes they describe, and it's worth understanding before running another OPD experiment. OPD looks like "match the teacher distribution." But in practice, the update is driven by a very small set of next-token choices at each generation step. Mostly just the handful of tokens that both the student and teacher think are plausible as the next token. Once that small set breaks, OPD breaks. The real object OPD is learning on At every generation step, the model has a huge vocabulary. 150K tokens, maybe more. But almost all of the probability mass sits on a tiny number of tokens. One paper [1] shows that the overlapping high-probability tokens between teacher and student carry around 97–99% of the total probability mass. So although OPD is written as reverse-KL over a full vocabulary, most of the useful learning signal comes from a tiny local menu of next-token options. Here is what I mean by "the handful of tokens." For a given prefix like: "Let's solve this step by step. First, we…" the student may think the next token should be one of: "need", "can", "have", "find", "know", "compute", … The teacher has its own version of this menu. OPD works when these two menus mostly overlap, and when the teacher puts higher probability on better choices inside that menu. It fails for a few different reasons — the menus don't overlap, the menu drifts somewhere bad mid-training, we only look at one item on the menu instead of the whole thing, or the per-position learning signals end up pulling the model in inconsistent directions. Four things can go wrong. The first two are about whether the menu is in good shape. The third is about whether we look at the whole menu or just one item from it. The fourth is about whether the signals across positions combine into a useful update. 1. The student and teacher are thinking in different "languages" A stronger teacher does not necessarily make a better OPD teacher. Li et al. [1] shows this very clearly: a 7B teacher can outperform a 1.5B model on benchmarks, but still fail to improve the 1.5B student through OPD. Why? Because benchmark accuracy measures final answers. OPD trains on next-token probabilities. A stronger model may solve the same problem through a different reasoning path: different intermediate steps, different phrasing, different proof structure, different local token choices. So when the student writes its own partial solution, the teacher may not assign useful probability to the student's next natural steps. The teacher is better overall, but not necessarily helpful on the student's current path. The most interesting experiment is the "reverse distillation", where they take a 1.5B model that was improved by RL — a student that has already moved beyond its original base behavior — and try to distill it back using two teachers: the original pre-RL 1.5B model and a larger 7B model from the same family. Both teachers pull the RL-improved student backward. The student loses its RL gains and regresses toward the older behavior. This sounds surprising at first. But the explanation is simple: OPD does not know that the student's RL behavior is better unless the teacher's token probabilities support it. If the teacher still prefers the old reasoning pattern, OPD will train the student back toward that pattern. So the RL gains disappear not because the teacher is "weak" in benchmark terms, but because the teacher is giving token-level supervision for a behavior the student has already moved past. Benchmark gap does not tell you whether OPD will work. Token-level compatibility does. 2. Repetition becomes locally rewarding Even if OPD starts well, it can still collapse. The most striking failure mode is when training looks fine for a while, then within roughly 30 steps the model starts producing much longer outputs, stops terminating, repetition spikes, and accuracy collapses. The mechanism is counterintuitive at first but makes sense once you see it. In sampled-token OPD, the reward for a token is roughly the teacher's log-probability minus the student's log-probability on that token. So if the teacher gives a token much higher probability than the student does, that token receives a large positive signal. Now imagine the student starts repeating itself. In practice this looks less like coherent sentences repeating and more like degenerate loops — something like "wait, wait, wait, wait, wait" filling the rest of the context. This prefix is bad globally. But locally, it is very predictable. A strong teacher is often very confident about predictable text. Once the loop has gone on for a while, the teacher can assign high probability to the next repeated token. The student may be less confident than the teacher. So the repeated token gets a large positive log-ratio. That means OPD accidentally rewards continuing the repetition. Before repetition starts, repeated tokens are rare, so they don't matter much. But once repetition appears, those tokens become frequent. And because they also receive large positive advantages, they start dominating the update. Luo et al. [2] measures repeated tokens getting 4 to 9 times larger advantage than normal tokens after collapse. Then the loop reinforces itself: more repetition → more predictable prefix → higher teacher confidence → larger positive signal on repeated tokens → even more repetition This is different from the usual length-bias issue in RL. It's more specific — a broken prefix creates locally high-reward repeated tokens, and OPD faithfully amplifies them. 3. We often only look at one item on the menu The clean objective would compare the teacher and student distributions over multiple possible next tokens. But many public OPD recipes — including the ones used industrially — use a cheaper version: Let the student generate one token. Then ask: did the teacher assign this exact token higher or lower probability than the student? If teacher probability is higher, push the student toward it. If lower, push the student away. That is the sampled-token log-ratio. It is cheap because you only score the token the student actually sampled. You do not need to compare the full vocabulary. There is a real reason for this design choice. Full sequence-level reverse-KL is noisy for long generations because an early token update gets entangled with many future rewards. Token-level OPD avoids that by giving each token its own local feedback. That gives much better variance scaling with length [3] — worst-case variance grows as O(T²) for token-level instead of O(T⁴) for sequence-level. So for long reasoning traces, token-level feedback is attractive. The problem is that "one sampled token" is a very noisy view of the teacher's actual next-token preference. At a given step, the teacher may have a whole cluster of reasonable next tokens. But sampled-token OPD only checks the one token the student happened to pick. This creates three problems. First, the student samples tokens from its own distribution, so on most positions the student's probability exceeds the teacher's and the log-ratio is negative. The reward is computed as teacher minus student, and the student is picking tokens where its own log-prob is near its highest — meaning the subtraction is almost always against the student's strongest values. Positive signal only shows up when the student happens to sample a token the teacher likes even more than the student does, which is the minority case. Second, if the student drifts into weird prefixes, the teacher's local probabilities may no longer reflect global quality. Third, tokenization and special-token differences can create fake disagreements. The student and teacher may represent the same text with different token boundaries, so a single-token comparison can look terrible even when the underlying string is fine. The fix proposed in [3] is simple: don't guess the teacher's local preference from one sampled token. Instead, take the teacher's top-k next tokens, renormalize both teacher and student probabilities over that set, and compute reverse-KL there. It's still cheap — you only need the top-k logits, not the whole vocabulary. But it changes the supervision from "did the teacher like this one sampled token?" to "among the teacher's plausible next tokens, does the student put probability mass in the same places?" That is a much better local learning signal. They also add top-p sampling during rollout, so the student is less likely to wander into extremely low-probability prefixes, and mask special tokens to avoid fake tokenization mismatches. 4. Even good token signals may not add up This is the least developed of the four failure modes, and possibly the most important. Li et al. [1] compares a successful OPD setup against a failing one and finds something strange. The failing teacher's per-token advantages are actually larger than the working teacher's, but the gradient norms are smaller. And the failing teacher's sequence-level reward can still distinguish correct from incorrect rollouts, comparable to the working case. So the reward signal is globally informative. It just does not produce useful gradients. Turns out what's going on is that OPD computes a learning signal at every token position in the rollout, and then sums them into one gradient update. Each position's contribution is a vector in parameter space pointing in some direction. So if the per-position vectors mostly point in the same direction, they add up to a big coherent push. But if they point in different directions, they partially cancel when summed, and the model barely moves even though each individual position had something to say. The empirical fingerprint is consistent with the second case: large per-token advantages but small gradient norms after summing — individually strong signals that cancel out when combined. The successful teacher shows the opposite pattern: smaller per-token advantages, larger gradient norms after summing — weaker individual signals that reinforce each other when combined. The paper that raises this leaves it as a hypothesis, but the empirical fingerprint — large per-token advantages, small gradient norms, informative sequence-level reward — is specific enough that it should be testable? What this explains A lot of practical OPD fixes start to look related. SFT cold start helps because it moves the student closer to the teacher's reasoning style before OPD begins. Teacher-aligned prompts help because they put the student in regions where the teacher gives more reliable feedback. KL regularization helps because it prevents the student from drifting too quickly into weird generations. Mixture distillation helps because it keeps some clean reference trajectories in training, so the rollout distribution does not become fully self-generated garbage. Top-k matching helps because it stops pretending one sampled token is enough to represent the teacher's local preference. These look like different tricks. But they are all trying to protect the same thing: the small set of plausible next tokens where teacher and student can actually communicate. The ceiling, and the real open question The more interesting part is long-horizon reasoning. The deeper the student gets into its own generated solution, the more likely the prefix is something the teacher would not have written. And once the teacher is judging prefixes outside its own natural distribution, its token probabilities become less reliable. One paper [1] shows this directly: teacher continuation advantage drops sharply as the student prefix gets longer, from +0.37 at 1K prefix to +0.02 at 16K. That is a bad sign for long-CoT and agentic OPD, because those are exactly the settings where the student spends many steps inside its own partially-generated world. OPD works best when the teacher and student stay close enough that teacher probabilities remain meaningful. Long-horizon agentic training pushes in the opposite direction. This is the open question I would most want to see investigated. The failure modes in sections 1-3 are diagnosable and have proposed fixes. Section 4 is a hypothesis about local gradient structure that should be testable. But the long-horizon ceiling is different — it is about whether OPD's core assumption (the teacher knows something useful about the student's next token) can hold at all when the student is operating many steps inside its own generated world. My current takeaway OPD isn't really a full-vocabulary distribution-matching problem. It's a fragile communication protocol between teacher and student through a tiny local menu of next-token choices. When that menu overlaps, stays clean, and produces gradients that add up, OPD works. When it doesn't, OPD quietly trains the student in the wrong direction. References [1] Li et al. Rethinking On-Policy Distillation of Large Language Models: Phenomenology, Mechanism, and Recipe. arxiv.org/abs/2604.13016 [2] Luo et al. Demystifying OPD: Length Inflation and Stabilization Strategies for Large Language Models. arxiv.org/abs/2604.08527 [3] Fu et al. Revisiting On-Policy Distillation: Empirical Failure Modes and Simple Fixes. arxiv.org/abs/2603.25562

中文
2
23
138
25.6K
ruzz
ruzz@findwildruzz·
@crypt0lake Currently doing it to implement my own PufferLib env. The more I do it the more I think I should have done it years ago
English
0
0
2
329
cryptolake
cryptolake@crypt0lake·
unironically, learn c++ and c, deeply, if you are junior or not
English
52
94
2.1K
90.3K
OpenCode
OpenCode@opencode·
OpenCode x Qwen 3.6 Plus - free, again Last time y’all treated our capacity like an all-you-can-eat buffet. We found more GPUs. Round 2.
English
198
394
7K
603.7K
Halooo
Halooo@hopeee_lab·
@opencode Is it only available in Zen or Go?
Halooo tweet media
English
4
0
36
18K
ruzz
ruzz@findwildruzz·
@Polymarket Still cannot use it in Italy, so what’s the point
English
0
0
2
32
ruzz
ruzz@findwildruzz·
@rubenhume I don’t like the black keyboard on silver
English
0
0
0
50
Ruben Hume
Ruben Hume@rubenhume·
At some point you grow up and realize the only correct color for a MacBook is silver
Ruben Hume tweet mediaRuben Hume tweet media
English
257
176
5.8K
447.1K
ruzz
ruzz@findwildruzz·
3.6 is better and faster in the gguf version
English
0
0
0
17
ruzz
ruzz@findwildruzz·
@0xMovez This is just undergrad level stat, nothing that gets paid that (I hope)
English
0
0
0
63
Movez
Movez@0xMovez·
Jane Street pays $750k/ year for quants who can answer how to use Stochastic Process and Markov Chains in quant trading. This 1-hour MIT lecture on probability gives you the same insights quants get paid $60K/month for. Bookmark & watch today. Then read the article below.
Roan@RohOnChain

x.com/i/article/2053…

English
120
1.3K
11.1K
2.6M
Kimu 🍃
Kimu 🍃@notkimu·
Para los amantes de las Thinkpads en el post de la Ideapad, esta T480 la rescaté (la tiraron por baterías muertas) a 50$
Kimu 🍃 tweet media
Español
34
15
770
16.9K
ruzz
ruzz@findwildruzz·
@_ontologic I like those, but for sure I do not have the experience you are looking for
English
0
0
0
7
ruzz
ruzz@findwildruzz·
@mk_banachan Oh shit this is my fist Japanese tweet in the tl
English
0
0
0
551
ruzz
ruzz@findwildruzz·
@graykevinb Some people like to customise tools and toys more than they like to play, some others don’t. We are all in the same team
English
1
0
1
25
Kevin Gray
Kevin Gray@graykevinb·
I have a confession: I use gnome and vs code. I'm so sorry.
English
7
0
26
775
ruzz
ruzz@findwildruzz·
@usr_bin_roygbiv @grok Exactly, nothing relevant. But that’s ok we can have different opinions Roy 🤝
English
0
0
0
3
Roy
Roy@usr_bin_roygbiv·
@findwildruzz @grok how many scientific papers are published in the last year compared to five years ago
English
2
0
0
26
Roy
Roy@usr_bin_roygbiv·
is there anything 5.5 xhigh can't do? (besides set up sglang and cuda apparently, I swear they sabotage it)
English
5
0
17
1.3K
ruzz
ruzz@findwildruzz·
@usr_bin_roygbiv Way off from being solved. You can I tell it to write a paper and expect to be published
English
1
0
0
15
Roy
Roy@usr_bin_roygbiv·
@findwildruzz creative writing was one of the first things solved by llms
English
1
0
0
21
ruzz
ruzz@findwildruzz·
@usr_bin_roygbiv Gpt3? Maybe Im bad at prompting because every time I try I want to burn my eyes
English
1
0
1
13
Roy
Roy@usr_bin_roygbiv·
@findwildruzz This is something gpt 3 excelled at what are you talking about
English
1
0
0
22