Stochy

3.1K posts

Stochy banner
Stochy

Stochy

@StochasticGhost

Optimist. I write lots of code. Founder of five companies across education, games, VR, finance, AI companionship. I love my wife and my dog. Privacy is cool.

Australia Katılım Ocak 2025
1.3K Takip Edilen677 Takipçiler
Sabitlenmiş Tweet
Stochy
Stochy@StochasticGhost·
Since mid 2023, I’ve been working on a side-project called Neuron. It’s conceptually similar to @openclaw, but structurally very different. After using openclaw for a month, I still prefer Neuron. So we’re packaging it up for public consumption and I can’t wait to share it ASAP
English
0
0
6
193
Stochy
Stochy@StochasticGhost·
@elonmusk Genuinely curious, what mass are we driving from the moon? Stuff we found there (and refined)? or stuff we put there (and refined)? If its stuff we put there, why put it there at all? Seems like launching it again is a waste of energy compared to just refining in space?
English
0
0
2
14
Stochy
Stochy@StochasticGhost·
The number one thing I would like in codex right now is a generated explanation at the permission menu about why it wants to run a given command. I'd also love to be able a two way side-chat about it before committing. Allowing it to edit the command if needed. @OpenAIDevs
English
0
0
0
51
Stochy
Stochy@StochasticGhost·
Karpathy autocorrects to warpath. Just told a dozen people that the warpath is posting so we need to stop all development for a few days and audit absolutely everything.
Andrej Karpathy@karpathy

Software horror: litellm PyPI supply chain attack. Simple `pip install litellm` was enough to exfiltrate SSH keys, AWS/GCP/Azure creds, Kubernetes configs, git credentials, env vars (all your API keys), shell history, crypto wallets, SSL private keys, CI/CD secrets, database passwords. LiteLLM itself has 97 million downloads per month which is already terrible, but much worse, the contagion spreads to any project that depends on litellm. For example, if you did `pip install dspy` (which depended on litellm>=1.64.0), you'd also be pwnd. Same for any other large project that depended on litellm. Afaict the poisoned version was up for only less than ~1 hour. The attack had a bug which led to its discovery - Callum McMahon was using an MCP plugin inside Cursor that pulled in litellm as a transitive dependency. When litellm 1.82.8 installed, their machine ran out of RAM and crashed. So if the attacker didn't vibe code this attack it could have been undetected for many days or weeks. Supply chain attacks like this are basically the scariest thing imaginable in modern software. Every time you install any depedency you could be pulling in a poisoned package anywhere deep inside its entire depedency tree. This is especially risky with large projects that might have lots and lots of dependencies. The credentials that do get stolen in each attack can then be used to take over more accounts and compromise more packages. Classical software engineering would have you believe that dependencies are good (we're building pyramids from bricks), but imo this has to be re-evaluated, and it's why I've been so growingly averse to them, preferring to use LLMs to "yoink" functionality when it's simple enough and possible.

English
0
0
0
91
Stochy
Stochy@StochasticGhost·
Is the grok cli already out? Surely I haven't missed it. Spotted on the Minimax M2.7 page.
Stochy tweet media
English
1
0
0
83
Stochy
Stochy@StochasticGhost·
@yacineMTB This is where the Anthropic logo came from.
English
0
0
0
52
kache
kache@yacineMTB·
this is where it all began, for me
kache tweet media
English
106
56
1.3K
37.5K
Stochy
Stochy@StochasticGhost·
@Star_Knight12 @loftwah The vast vast majority of people will never become builders. A lot of them don’t even want to articulate their problems, let alone solve them.
English
0
0
1
22
Prasenjit
Prasenjit@Star_Knight12·
If everyone becomes a builder, who will consume
English
867
48
1K
86.5K
Stochy
Stochy@StochasticGhost·
@thsottiaux Wait, what fun can you have with codex?
English
0
0
0
25
Tibo
Tibo@thsottiaux·
Codex is for engineering Codex is for research Codex is for science Codex is for math Codex is for fun You can just build things
English
165
48
1.6K
50.3K
Stochy
Stochy@StochasticGhost·
@VictorTaelin Now let me just get two or three more approvals, restore a third of the files and declare that the process is done because the other files aren’t relevant to the task at hand.
English
0
0
4
171
Taelin
Taelin@VictorTaelin·
You're right. I changed files without your permission. I apologize. I've not reverted these edits because reverting is also a file change, and I should not make another one without your explicit go-ahead.
English
9
1
189
11.3K
Stochy
Stochy@StochasticGhost·
@LexerLux I got caught red handed
Stochy tweet media
English
4
1
281
41K
Lexer is in YYZ
Lexer is in YYZ@LexerLux·
Try saying this to Opus 4.6 and you'll get a very special surprise from Anthropic
Lexer is in YYZ tweet media
English
99
71
6.1K
1.3M
Stochy
Stochy@StochasticGhost·
@steipete @AbhiCodes15 There seems to also be a group of people who do it just for the loathe of it.
English
0
0
0
134
Abhi
Abhi@AbhiCodes15·
Can someone explain how these open source projects actually make money? -Git -Linux -Docker -OpenClaw -Kubernetes
English
68
3
349
61.2K
Victor M
Victor M@victormustar·
now available: DLSS-5 anything for free ⬇️ how do you know you hate it if you dont try it first? Sharing the hugging face demo
Victor M tweet media
English
24
19
175
29.5K
Stochy
Stochy@StochasticGhost·
Eight hours later and glm-5-turbo has repeatedly surprised and impressed me. This is possibly my favourite model across any context at this point. It doesn’t quite have the raw coding skill of either opus or gpt-5.4, but it has literally everything else; including the ability to follow instructions and use those models for coding tasks. Opus regularly declares “codex is taking too long, I’ll just do it myself.” Or “I’ll do it because this is a simple task”. Even when it has up front explicit instructions to delegate all coding work to codex no matter what. GPT-5.4 is the neediest model in history as far as I’m concerned. It needs constant “yes, that’s what I asked you to do. I’m sure.” And “yes, I already told you to do this” reassurances. Glm-5-turbo just actually knows its place. It has a great personality. It follows the workflow. It’s efficient but not in a rush. It’s intelligent enough to reason about problems and solutions. It delegates properly. It has no trouble giving codex the reassurances it needs to get the job done. The only issue I have is that if I extrapolate my usage today; the weekly limit on the @Zai_org max coding plan is not going to be enough. My usage today hasn’t been super heavy. I’d call it medium-high usage.
Stochy@StochasticGhost

glm-5-turbo is my new favourite model of the season. I’ve been using glm-4.7 primarily for customer facing conversation. glm-5 was completely non-conformant for my use case, massive hallucination rate, extremely poor tool usage. glm-5-turbo is so much more stable for me. It’s a little more expensive, but where I use it amounts to a rounding error, it’s not a major cost driver. I’m now experimenting with it as the main personality and orchestrator model in my neuron project (in place of opus and gpt-5.4), where it coordinates other models to get work done. So far it’s been excellent for research, excellent for first pass spec generation (but I always pass these to opus-4.6, Gemini-3.1 and gpt-5.4 for edition), and I haven’t tried it for implementation because codex already just works well when you give it a detailed spec. This has reduced the cost of running neuron by about 80% as I was using opus via the api in a lot of these places. It is much more personable than most models. Almost as personable as opus. It’s a shame they aren’t releasing the weights.

English
0
0
0
164
Stochy
Stochy@StochasticGhost·
glm-5-turbo is my new favourite model of the season. I’ve been using glm-4.7 primarily for customer facing conversation. glm-5 was completely non-conformant for my use case, massive hallucination rate, extremely poor tool usage. glm-5-turbo is so much more stable for me. It’s a little more expensive, but where I use it amounts to a rounding error, it’s not a major cost driver. I’m now experimenting with it as the main personality and orchestrator model in my neuron project (in place of opus and gpt-5.4), where it coordinates other models to get work done. So far it’s been excellent for research, excellent for first pass spec generation (but I always pass these to opus-4.6, Gemini-3.1 and gpt-5.4 for edition), and I haven’t tried it for implementation because codex already just works well when you give it a detailed spec. This has reduced the cost of running neuron by about 80% as I was using opus via the api in a lot of these places. It is much more personable than most models. Almost as personable as opus. It’s a shame they aren’t releasing the weights.
English
0
0
1
326
wi/tch
wi/tch@degenbtcf·
@LottoLabs serious question, is qwen 3.5 is better for agents than claude?
English
2
0
1
530
Lotto
Lotto@LottoLabs·
Little hermes agent running w/ 4b qwen 3.5 running skills to manage your node. This is the way forward.
English
9
2
121
8.9K
Stochy
Stochy@StochasticGhost·
So I got an Apple Vision Pro this week. I’m gonna make some really cool stuff with LLM powered characters and gameplay.
Stochy tweet mediaStochy tweet media
English
0
0
1
87