Todd Fisher

4.8K posts

Todd Fisher banner
Todd Fisher

Todd Fisher

@taf2

Living and working on https://t.co/eminYBe5tr

Maryland, US 参加日 Ağustos 2008
463 フォロー中329 フォロワー
Todd Fisher がリツイート
Sandro
Sandro@pupposandro·
Currently cooking eggs on the RTX 3090 to save money on gas
Sandro tweet media
English
17
23
340
8.3K
jason liu
jason liu@jxnlco·
When you gotta bike home from work but your codex needs to finish a task.
English
207
59
1.8K
246.5K
Todd Fisher
Todd Fisher@taf2·
@GregKamradt So I ran a llm fine tune over the weekend and it took many hours - codex monitors the long running process and can even send text message updates via CTM or any sms capable service - it’s really good at babysitting a long running process
English
0
0
0
19
Greg Kamradt
Greg Kamradt@GregKamradt·
When OAI employees say, “I let codex run all night…” What framework they use? How do you set up the task so it has enough work for 8 hours?
English
299
18
1.8K
333.5K
Todd Fisher
Todd Fisher@taf2·
Best usecase while outside waking for local ai on my phone - a personal pocket heater.
English
0
0
0
8
Ash DCosta
Ash DCosta@softwareweaver·
@vllm_project Cool. Does this version work well with Codex? Codex cli was giving me errors connecting when using vllm to host Qwen 3.6 27B model with the responses api compatibility issues.
English
1
0
2
595
vLLM
vLLM@vllm_project·
vLLM v0.20.0 is here! 752 commits from 320 contributors (123 new). 🎉 Highlights: DeepSeek V4, Hunyuan v3 preview support, CUDA 13 / PyTorch 2.11 / Transformers v5 baseline, FA4 as default MLA prefill, TurboQuant 2-bit KV (4× capacity), vLLM IR foundation. Thread 👇
vLLM tweet media
English
22
79
667
66.3K
Todd Fisher がリツイート
Stefan Streichsbier
Stefan Streichsbier@s_streichsbier·
As requested, I also asked GPT-5.5 (low) and Opus 4.7 (high) to fix the same bug. GPT-5.5 (low) identified the correct root cause in 4m 14s and produced an almost identical fix in 2m 47s, using 164k tokens in total. Opus 4.7 (high) “churned” for 6m 23s, using 87.7k tokens, and went down a completely wrong path. It’s not even close, but that’s not a surprise.
Stefan Streichsbier@s_streichsbier

I've completely changed my mind about 5.4 vs 5.5. Gave them the exact same task to investigate a fairly tricky bug. GPT-5.5 identified the bug and proposed a fix in 6m 59s using 117k tokens. GPT-5.4 took 8m 51s using 201k tokens, but it didn't find the bug and is asking for more information to investigate. Call me impressed.

English
90
88
2.1K
423.1K
Todd Fisher
Todd Fisher@taf2·
qwen models are good but the censorship in these models can really cause problems with refusals on text that might be mis-taken as political. working on abliteration now to see if maybe these can be safer
English
0
0
0
9
Todd Fisher
Todd Fisher@taf2·
@LLMJunky Did neural net stuff in college 2001-2002 but considering where ai is not I can’t claim any real involvement until 2023 and even still I’d just consider myself a casual consumer of it
English
1
0
1
89
am.will
am.will@LLMJunky·
How long have you been in AI? Where the OGs at? 👇
English
57
0
46
7.6K
Luce
Luce@lucyshow11·
What’s up doc!!! 🤪
Luce tweet media
English
306
4K
38.1K
2.3M
Tibo
Tibo@thsottiaux·
@raffichill Me too, me too. What a time to be alive
English
31
0
264
10K
Raffi
Raffi@raffichill·
I let Codex use my computer today
English
4
0
158
11.9K
Nate Berkopec
Nate Berkopec@nateberkopec·
If you're doing AI dev, you need to act like your system is rooted by North Korea. You cannot leave knives out in the kitchen, you cannot leave the passwords out on the counter. People are putting too much trust in alignment and not doing enough to "keep honest agents honest".
English
4
7
63
4.3K
Todd Fisher がリツイート
Ahmad
Ahmad@TheAhmadOsman·
PRO TIP vLLM telling you to use `--enforce-eager` to avoid OOM because CUDA Graphs “don’t have enough VRAM”? Don’t jump straight to eager mode Try this first: - lower `--max-model-len`, ex: 4k - let CUDA Graph compile (which will be cached by torch.compile) - restart, then raise context back up You can keep the CUDA Graph performance gains without hitting OOM
Ahmad tweet media
English
14
17
214
10.1K
Todd Fisher
Todd Fisher@taf2·
They used my code to create 5.5 - I’m using it to replace itself
English
0
0
1
12
Todd Fisher
Todd Fisher@taf2·
Doing data privacy cleanup using new OpenAI model and set codex up with sms notify script via CTM to give me status updates … codex session been running for 20+ minutes just cooking while do other things
Todd Fisher tweet media
English
0
0
0
19
Nandkishor
Nandkishor@devops_nk·
AWS Lambda in 40 seconds.
English
52
208
3.1K
199.4K