Dan Woods

627 posts

Dan Woods banner
Dan Woods

Dan Woods

@danveloper

Vice President of AI Platforms for CVS Health. Former CTO for @JoeBiden.

Katılım Mart 2011
804 Takip Edilen9.5K Takipçiler
Dan Woods
Dan Woods@danveloper·
Claude macOS 🤝 An hour before: About 15 minutes remaining
English
0
0
3
522
Dan Woods
Dan Woods@danveloper·
Nobody’s asking the MOST important question — what if GitHub built GitHub?
Dan Woods tweet media
English
1
0
4
700
Dan Woods
Dan Woods@danveloper·
@antirez @deepseek_ai I did the same for Qwen3.5-397B - github.com/danveloper/fla… ... one optimization you can do if you want to keep q4 is store them as lz4 - the M3 Max+ has an lz4 hardware accelerator so you can keep more experts hot in the prefill cache. Then expert retrieval is near DRAM speeds.
English
0
0
7
563
antirez
antirez@antirez·
@danveloper Yep consider that this is an M3 Max not an M3 Ultra, in the Ultra I get 2x prefill speed, and the same speed with 4 bit quants instead of 2 bit (only for routed experts, all the other weights are as released by @deepseek_ai).
English
2
1
12
2.1K
antirez
antirez@antirez·
DeepSeek v4 small KV cache + MacBook fast SSD disks = the idea that the disk is not a good target for KV cache is, in this context, totally obsolete. It works *great*. The session you see is opencode using my inference engine for DS4, saving, loading sessions from disk.
antirez tweet media
English
33
27
556
40.6K
Dan Woods
Dan Woods@danveloper·
I'm old enough to remember approx 2 months ago when you could hand Opus 4.6 an arXiv paper and say "let's do this" and it would clunk away for a few hours and deliver a perfect result. Disappointed that's gone, but nonetheless the agentic workflow isn't the problem.
English
0
0
2
356
Dan Woods
Dan Woods@danveloper·
Since I ran out of Codex credits, I had to switch back to Claude Code to get some things done. Opus 4.7 can still cook, you just need to give it an extremely focused task list with rigid acceptance criteria.
Dan Woods tweet media
English
5
1
16
2.4K
Alex Cheema
Alex Cheema@alexocheema·
My M4 Max MacBook gets 3,756,165 tok/sec in pure C, compared to ~50,000 tok/sec with the FPGA. Try it yourself: github.com/AlexCheema/tal…
luthira@luthiraabeykoon

We implemented @karpathy 's MicroGPT fully on FPGA fabric. No GPU. No PyTorch. No CPU inference loop. Just a transformer burned into hardware, generating 50,000+ tokens/sec. The model is small, but the idea is not: inference does not have to live only in software 👇

English
67
99
1.7K
228.4K
blindjoe
blindjoe@blindjoe_algo·
@danveloper How long have you felt this way about Codex?
English
1
0
0
34
Dan Woods
Dan Woods@danveloper·
I realize I’m not “inside the industry”, so don’t listen to me, but for people actually doing work rn Codex is the game
English
4
0
22
3.1K
Dan Woods
Dan Woods@danveloper·
You can just let Codex cook
Dan Woods tweet media
English
1
0
1
488
Dan Woods
Dan Woods@danveloper·
@DanielleFong I ran out of tokens until Tuesday (or @thsottiaux saves me), so I bought extra tokens and now I can’t afford fast mode and it’s devastating 😂
English
0
0
1
244
Danielle Fong 🔆
Danielle Fong 🔆@DanielleFong·
i use /fast mode with thinking:none / low on chatgpt5.5 and it's awesome. i call it goblin mode i wouldn't even think about slow, but gear up to xhigh for big builds. claude code /fast feels like luxury, or it did. now there's tension between 4.6 /fast and 4.7, and 4.7 is ... great but overconfident. so instead of working on what i want to work on, i am trying ro figure out and work around the model! i think people who use claude code fast should give goblin mode a try
Jared Friedman@snowmaker

Software engineering job descriptions should really start saying whether they include /fast mode or not.

English
8
4
77
6.3K
Dan Woods retweetledi
Sam Altman
Sam Altman@sama·
you know what all of these "which is better" polls are silly use codex or claude code, whatever works best for you i am grateful we live in a time with such amazing tools, and grateful there is a choice
English
2.2K
1.1K
23K
1.6M
Dan Woods
Dan Woods@danveloper·
@dtcb > DeepSeek-V4 uses Hadamard rotation 😳 I had no idea!
English
1
0
1
67
Cole Brown
Cole Brown@dtcb·
@danveloper github.com/bigs/deepseek-… Just wanted to slap it together before I leave my office, but here she is! Haven't read the READMEs yet so don't stab me if it's stupid 😂. But a foundation has been laid! DGX Spark is an annoying target, so it's nice to cobble this together.
English
1
0
1
88
Dan Woods
Dan Woods@danveloper·
@dtcb Make sure you share everything!
English
1
0
0
39
Cole Brown
Cole Brown@dtcb·
@danveloper Well to amuse you while you wait—you inspired me to point codex + /goal at getting deepseek v4 flash running as fast as possible on my DGX Spark. Native FP8/4 hybrid. Just got my first tokens out! Time to optimize :)
English
1
0
1
112