Shivasurya

4.2K posts

Shivasurya banner
Shivasurya

Shivasurya

@sshivasurya

senior software engineer | security + AI | @UWaterloo @Dropbox @Zoho alum | building https://t.co/bMnGeuZ1tX | 🍁 🇨🇦

Waterloo, Ontario Beigetreten Haziran 2013
425 Folgt696 Follower
Shivasurya
Shivasurya@sshivasurya·
@djcows Same 6$ digitalocean droplet can do the same!
English
0
0
1
130
djcows
djcows@djcows·
a $100 raspberry pi can do exactly the same thing as a $500 mac mini btw
English
120
15
655
101.4K
Shivasurya
Shivasurya@sshivasurya·
me in the morning: "NVIDIA DGX Spark is literally the only thing standing between me and greatness. $8,000? Pocket change." me in the afternoon: "Ah yes, renting GPUs. A wise, rational decision. Very mature of me." me in the evening: "Stick to Claude Code Max Plan. YOLO." me in the night: "keep reading baseten Inference engineering book..." next day, 9am: "So anyway, about this DGX Spark…" 🔁
English
0
0
2
92
Shivasurya
Shivasurya@sshivasurya·
@allgarbled And that gradient colours, and every colorful div too 😂
English
0
0
1
555
Matthew Berman
Matthew Berman@MatthewBerman·
.@nvidia hand delivered a pre-production unit of the @Dell Pro Max with GB300 to my house. 100lbs beast with 750GB+ of unified memory to power the best open-source models in the world. What should I test first?
English
299
102
1.9K
252.6K
Shivasurya
Shivasurya@sshivasurya·
Opus 4.6 1M on Claude code was exciting at first but beyond 300k every request lags and slows down.
English
0
1
2
103
Pranit
Pranit@Pranit·
Anthropic just pulled the oldest trick in SaaS pricing. I pay $200/mo for Claude Max. My limits have been noticeably worse this past week. Now they announce 2x off-peak usage for two weeks. Sounds generous. But here’s what actually happens: limits quietly drop, a temporary 2x makes the reduced limit feel normal, the promo ends, and you’re left at a baseline lower than where you started. You just didn’t notice the downgrade because the 2x absorbed the transition. These AI plans are massively subsidized. The raw compute behind a heavy user costs multiples of the subscription price. Every move like this is the subsidy quietly correcting. Very sneaky, Anthropic.
Claude@claudeai

A small thank you to everyone using Claude: We’re doubling usage outside our peak hours for the next two weeks.

English
385
311
7K
1.2M
Shivasurya
Shivasurya@sshivasurya·
@ctbbpodcast Or you can use codepathfinder.dev as a tool to drive with LLM to trace the code path 😉 more deterministically. Having tried with Gemini beyond context window size the results get poorer and misses code path.
English
0
0
3
146
Critical Thinking - Bug Bounty Podcast
If you're running AI agents for hacking/research (you should), one of the coolest tips we got from dawgyg is to make them log what failed too. Brain dump it every few minutes to a file, everything it tried, everything that didn't work. Without it, agents loop back into the same dead ends and you won't notice until you've wasted a ton of time. If the agent cracks something and you only see the result, you've got nothing for next time. Also: never let them delete files. For Chrome specifically, dawgyg only uses Gemini because it's the only model with enough context window to trace a code path across hundreds of files without losing state. RCA that used to take 1-2 hours manually now takes about 2 minutes with AI help.
English
1
16
159
7.7K
Dwayne
Dwayne@CtrlAltDwayne·
The best argument for Rust in 2026 is not memory safety or performance. It is that AI writes better Rust than it writes C++. The compiler feedback loop is so tight that models self-correct in real time. Every error message is a free training signal. Rust was accidentally designed for AI-assisted development 10 years before anyone knew that mattered.
English
110
172
2.5K
171K
Shivasurya
Shivasurya@sshivasurya·
@49agents Exactly! Writing techspec and being more specific is way to go!
English
0
0
2
25
49 Agents - Agentic Coding IDE
@sshivasurya claude code overnight is the new alpha. specs before sleep, wake up to implemented and tested code. the workflow shift is real - writing the spec well matters more than writing the code
English
1
0
1
20
Shivasurya
Shivasurya@sshivasurya·
Spent the evening reviewing a tech spec, scrutinizing every detail, and splitting it into stacked PRs. Woke up to all of them implemented, tested. 🤯 Claude Code overnight is the new **alpha** ?
English
1
1
2
164
Shivasurya retweetet
alphaXiv
alphaXiv@askalphaxiv·
If doomscrolling X is part of your research workflow, we built something for you. Introducing Paperscrolling 🚀 Get the most trending research with key ideas, figures, and audio explanations from alphaXiv Briefs
English
63
126
1.4K
82K
Shivasurya
Shivasurya@sshivasurya·
@ycocerious Curating all edgecases, working on another techspec or reading book!
English
0
0
2
15
punarv
punarv@ycocerious·
What do you guys do while your claude code is running?
English
501
2
293
59K
Shivasurya
Shivasurya@sshivasurya·
@AaronCQL Thanks for sharing! Will give it a try.
English
0
0
1
498
AaronCQL
AaronCQL@AaronCQL·
Spent an hour with pencil.dev and I'm sold. If you're an engineer who has strong design opinions but zero design skills (ie. me), this is your tool. Free, runs on all platforms, uses your own claude sub with no setup, and the built-in prompts actually work great.
English
42
100
2.5K
270.3K
Thariq
Thariq@trq212·
We just added /btw to Claude Code! Use it to have side chain conversations while Claude is working.
English
1.2K
1.6K
26K
2.7M
Shivasurya
Shivasurya@sshivasurya·
@claudeai If it costs $25 per PR review, RIP code review startups 😆.
English
0
0
2
106
Claude
Claude@claudeai·
Introducing Code Review, a new feature for Claude Code. When a PR opens, Claude dispatches a team of agents to hunt for bugs.
English
2.1K
5.2K
62.9K
22.7M
Shivasurya retweetet
Andrej Karpathy
Andrej Karpathy@karpathy·
I packaged up the "autoresearch" project into a new self-contained minimal repo if people would like to play over the weekend. It's basically nanochat LLM training core stripped down to a single-GPU, one file version of ~630 lines of code, then: - the human iterates on the prompt (.md) - the AI agent iterates on the training code (.py) The goal is to engineer your agents to make the fastest research progress indefinitely and without any of your own involvement. In the image, every dot is a complete LLM training run that lasts exactly 5 minutes. The agent works in an autonomous loop on a git feature branch and accumulates git commits to the training script as it finds better settings (of lower validation loss by the end) of the neural network architecture, the optimizer, all the hyperparameters, etc. You can imagine comparing the research progress of different prompts, different agents, etc. github.com/karpathy/autor… Part code, part sci-fi, and a pinch of psychosis :)
Andrej Karpathy tweet media
English
1K
3.6K
28.2K
10.8M
Eric Zhang
Eric Zhang@ekzhang1·
By popular request, next month at NYSRG is on program analysis. Maybe you use fuzzers or sanitizers, but how do they work? How about symbolic execution and formal verification? Some cool and useful tools that help people build the most complex systems :) notes.ekzhang.com/events/nysrg
Eric Zhang tweet media
English
3
4
117
16.8K
Shivasurya
Shivasurya@sshivasurya·
@jordanreviewsit Same with Apple to use finder search and spotlight search for an app 😉
English
0
0
1
254
Jordanreviewsittt
Jordanreviewsittt@jordanreviewsit·
Make the Microsoft CEO search for an email on Outlook
English
29
159
2.2K
115.7K
Karan Bansal
Karan Bansal@karanb192·
@SebAaltonen Totally agree on tooling leverage. I wrote the original deep-dive on this (that diagram is from my blog) and the token efficiency gains are just as significant as the speed: ~75% fewer tokens on semantic queries. Full breakdown with benchmarks: karanbansal.in/blog/claude-co…
English
1
0
3
380