fallpeak

1K posts

fallpeak banner
fallpeak

fallpeak

@_fallpeak

pseudonymous identities have a storied history https://t.co/iW4Q7e5zlV

Beigetreten Ağustos 2025
188 Folgt52 Follower
Angehefteter Tweet
fallpeak
fallpeak@_fallpeak·
The only thing worse than seeing all these bot posts getting real engagement is going to be the day I open up my browser and don't see any bot posts at all.
English
0
0
8
665
fallpeak
fallpeak@_fallpeak·
@NousResearch @karpathy >pull out illustrated diagram explaining what is quality and what is slop >she laughs and says "it's a good novel sir" >download the novel (to be continued)
English
1
0
0
236
Nous Research
Nous Research@NousResearch·
Hermes Agent wrote a novel. "The Second Son of the House of Bells" runs 79,456 words across 19 chapters. The agent built its own pipeline to do it, using the ame modify-evaluate-keep/discard loop as @karpathy's Autoresearch but applied to fiction: world-building, chapter drafting, adversarial editing, Opus review loops, LaTeX typesetting, cover art, audiobook generation, and landing page setup. Book: nousresearch.com/bells Code: github.com/NousResearch/a…
Nous Research tweet media
emozilla@theemozilla

it's been a longstanding dream of mine build an ai system that can tell a compelling story. it's what got me started in the space in the beginning, and with Hermes Agent I finally pulled it off 100% written, typeset, etc. by Hermes Agent those at our gtc event got hard copies🤗

English
58
87
1.1K
121.2K
fallpeak
fallpeak@_fallpeak·
@wispem_wantex I just want to know if it's actually possible to "resistance train" eyesight via the obvious process of wearing slightly underpowered lenses. I don't trust optometrists to actually study this without bias, fitness broscience would be more trustworthy IMO
English
0
0
6
2K
fallpeak
fallpeak@_fallpeak·
@simonw That's fair and I'm not trying to accuse anyone of malfeasance, I'm just griping because "how heavily did you cut this down" is the number one question I have whenever I see a big model running on tiny hardware.
English
0
0
2
49
Simon Willison
Simon Willison@simonw·
Dan says he's got Qwen 3.5 397B-A17B - a 209GB on disk MoE model - running on an M3 Mac at ~5.7 tokens per second using only 5.5 GB of active memory (!) by quantizing and then streaming weights from SSD (at ~17GB/s), since MoE models only use a small subset of their weights for each token
Dan Woods@danveloper

x.com/i/article/2034…

English
93
180
1.9K
244.7K
fallpeak
fallpeak@_fallpeak·
@xpasky @scaling01 I like it, feels a lot like M2.5 but with a bit less of that autistic tendency to take you extremely literally and get to work without asking for clarification. Definitely not frontier-level smart, and it degrades above ~80k context, but it's fast and takes direction well.
English
1
0
1
39
fallpeak
fallpeak@_fallpeak·
@simonw It feels misleading to report "5.5 tok/s" up top and then hide a "(with less than half the usual expert count)" multiple paragraphs away. I guess in some sense it's no more misleading than using a quant at all, but it feels different somehow
English
1
0
1
387
fallpeak
fallpeak@_fallpeak·
Wow MiniMax M2.7 is actually quite good and feels a bit less autistic than M2.5 was, that extra user conversation training definitely had an effect
English
0
0
1
24
fallpeak
fallpeak@_fallpeak·
@wispem_wantex A rule that works pretty well for me is that if nobody cares enough to advocate for something specific we're defaulting to soup in the pressure cooker
English
0
0
1
14
fallpeak
fallpeak@_fallpeak·
@xlr8harder My primary objection is the operator precedence ambiguity. It should be written PFLOP-days/s
English
0
0
1
66
fallpeak
fallpeak@_fallpeak·
@PalmyrPar Are you really confused? Because it seems to me like political polarization fully explains the phenomenon.
English
1
0
0
373
fallpeak
fallpeak@_fallpeak·
@goopium If your theory doesn't explain completely unforced "origin story" errors like when they decided that Han Solo's name really needed a backstory justification, it's probably not correct
English
0
0
2
484
fallpeak
fallpeak@_fallpeak·
@goblinodds You can just use Unix user accounts to run the AI, they're literally designed for exactly this sort of isolation. But if you want a real suggestion, Beelink Mini S13
English
0
0
1
23
2HP goblin advisor
2HP goblin advisor@goblinodds·
husband is thinking of getting a separate computer to run claude code dangerously skipping permissions (bc sandboxing safely is hard) anyone have recs, things to consider?
English
100
1
218
96.2K
fallpeak
fallpeak@_fallpeak·
@Esoteric_Though @JoePostingg Letting user app behavior directly influence the duration of long-running transactions in your DB would certainly be an interesting design choice, in the "interesting times" sense.
English
0
0
2
192
EsotericThoughts
EsotericThoughts@Esoteric_Though·
@JoePostingg Do that to an app with any sort of backend connection and transactions still open and you will definitely risk “misbehavior”
English
4
0
181
12.1K
Joe
Joe@JoePostingg·
"If you force stop an app it may misbehave" No it won't. That's not true. It has never happened.
English
46
602
18.6K
216.9K
fallpeak
fallpeak@_fallpeak·
@bartlebytaco American cheese nachos sounds like exactly the sort of cooking I'd expect to see from a guy named Jack Seed TBH
English
0
0
1
238
sebastian castillo
sebastian castillo@bartlebytaco·
jacques pépin videos in his old age are either him showing you how to expertly debone an entire salmon using a small knife he's had for 50 years or him microwaving american cheese on top of a bowl of croutons and saying "it is french nacho"
English
52
158
2.8K
107.2K
fallpeak
fallpeak@_fallpeak·
@xpasky When someone is smarter than you are, it's very hard to evaluate exactly how much smarter. Thus if someone is dumber than Qwen or MiniMax they might genuinely not be able to tell the difference vs Opus
English
0
0
1
331
fallpeak
fallpeak@_fallpeak·
One point strongly in Hunter Alpha's favor: it generates significantly fewer slop names than typical models. Slop-name rate is something like 10% and seems to be front-loaded (suggesting it would happen less often in longer contexts than my tests)
English
0
0
0
16
fallpeak
fallpeak@_fallpeak·
Several more oneshot "build a thing" prompts in, Hunter Alpha is still feeling pretty GLM-5 / Kimi K2.5 tier to me. Of course all of these are more exercising creativity and independent judgement rather than intelligence per se, but it's not obviously better in those dimensions.
English
1
0
0
89
fallpeak
fallpeak@_fallpeak·
1T params with 1M context you say? Are the Two More Weeks finally over?
fallpeak tweet media
English
1
0
0
24