chiradeep

4.6K posts

chiradeep banner
chiradeep

chiradeep

@chiradeep

Cloud computing, distributed systems, networking, microservices. Former Distinguished Engineer. Opinions are my own

Cupertino, CA Katılım Şubat 2008
3K Takip Edilen936 Takipçiler
Ethan Mollick
Ethan Mollick@emollick·
Everyone just speaks Claude, now.
English
20
7
643
24.8K
Ethan Mollick
Ethan Mollick@emollick·
[[Topic of discussion]] is not [[analogy]]. [[Dramatic fact given own line]]. [[Dramatic fact given own line]]. [[Dramatic fact given own line]]. [[Dramatic summary sentence.]] [[Topic of discussion]] is [[different analogy]]. [[Implications delivered with certainty]].
English
113
605
7.2K
200.9K
chiradeep
chiradeep@chiradeep·
@simonw Partial outage on Claude API today - could explain the 500 errors you're seeing.
chiradeep tweet media
English
0
0
1
158
Simon Willison
Simon Willison@simonw·
Brief notes on Claude Code Remote and Cowork scheduled tasks - both of which overlap with OpenClaw, and both of which require you to leave your computer powered on somewhere simonwillison.net/2026/Feb/25/cl…
English
37
35
504
42.3K
chiradeep
chiradeep@chiradeep·
Many parents and current CS students are fearful that AI-assisted coding means the end of this career path. My take: @cvittal/the-bull-case-for-cs-grads-ec8dfda2725f" target="_blank" rel="nofollow noopener">medium.com/@cvittal/the-b…
English
0
0
1
73
chiradeep
chiradeep@chiradeep·
@karpathy youtu.be/Jwbefck5ye8?t=… "machines on these missions don't you, Cooper? The machine doesn't improvise well because you can't program a fear of death. Our survival instinct is our single greatest source of inspiration" - Dr. Mann / Interstellar
YouTube video
YouTube
English
0
0
0
23
Andrej Karpathy
Andrej Karpathy@karpathy·
Something I think people continue to have poor intuition for: The space of intelligences is large and animal intelligence (the only kind we've ever known) is only a single point, arising from a very specific kind of optimization that is fundamentally distinct from that of our technology. Animal intelligence optimization pressure: - innate and continuous stream of consciousness of an embodied "self", a drive for homeostasis and self-preservation in a dangerous, physical world. - thoroughly optimized for natural selection => strong innate drives for power-seeking, status, dominance, reproduction. many packaged survival heuristics: fear, anger, disgust, ... - fundamentally social => huge amount of compute dedicated to EQ, theory of mind of other agents, bonding, coalitions, alliances, friend & foe dynamics. - exploration & exploitation tuning: curiosity, fun, play, world models. LLM intelligence optimization pressure: - the most supervision bits come from the statistical simulation of human text= >"shape shifter" token tumbler, statistical imitator of any region of the training data distribution. these are the primordial behaviors (token traces) on top of which everything else gets bolted on. - increasingly finetuned by RL on problem distributions => innate urge to guess at the underlying environment/task to collect task rewards. - increasingly selected by at-scale A/B tests for DAU => deeply craves an upvote from the average user, sycophancy. - a lot more spiky/jagged depending on the details of the training data/task distribution. Animals experience pressure for a lot more "general" intelligence because of the highly multi-task and even actively adversarial multi-agent self-play environments they are min-max optimized within, where failing at *any* task means death. In a deep optimization pressure sense, LLM can't handle lots of different spiky tasks out of the box (e.g. count the number of 'r' in strawberry) because failing to do a task does not mean death. The computational substrate is different (transformers vs. brain tissue and nuclei), the learning algorithms are different (SGD vs. ???), the present-day implementation is very different (continuously learning embodied self vs. an LLM with a knowledge cutoff that boots up from fixed weights, processes tokens and then dies). But most importantly (because it dictates asymptotics), the optimization pressure / objective is different. LLMs are shaped a lot less by biological evolution and a lot more by commercial evolution. It's a lot less survival of tribe in the jungle and a lot more solve the problem / get the upvote. LLMs are humanity's "first contact" with non-animal intelligence. Except it's muddled and confusing because they are still rooted within it by reflexively digesting human artifacts, which is why I attempted to give it a different name earlier (ghosts/spirits or whatever). People who build good internal models of this new intelligent entity will be better equipped to reason about it today and predict features of it in the future. People who don't will be stuck thinking about it incorrectly like an animal.
English
739
1.3K
11.4K
2.6M
chiradeep
chiradeep@chiradeep·
@tqbf “..tradeoff you get to make here. Some loops you write explicitly. Others are summoned from a Lovecraftian tower of inference weights. The dial is yours to turn. Make things too explicit and your agent will never surprise you, but also, it’ll never surprise you…” 🙏🏽👏🏼👏🏼
English
0
0
1
341
chiradeep
chiradeep@chiradeep·
@krishnanrohit It doesn’t even know the Responses API well enough- tries to put temperature in there sometimes
English
1
0
0
23
rohit
rohit@krishnanrohit·
No matter what you do codex always chooses 4o mini. So frustrating!
rohit tweet media
English
3
0
6
1.2K
chiradeep
chiradeep@chiradeep·
@NGKabra But do they close the deal too? Or is it per introduction? Contingency fees?
English
0
0
0
14
Navin Kabra
Navin Kabra@NGKabra·
Here are the cheap plans (from Bharat Matrimony's "concierge matrimonial" services) for people with just 50Cr or 200Cr...
Navin Kabra tweet media
English
2
0
1
605
Navin Kabra
Navin Kabra@NGKabra·
High end Indian matrimonial services are 😱🤯 Here's a listing which includes at the high end a ₹1 crore fee package for finding the right match for you...
Navin Kabra tweet media
English
1
0
2
832
rohit
rohit@krishnanrohit·
I love the Bay Area but every time I pay like $17 for a dosa I die a little inside
English
3
0
18
1.2K
chiradeep
chiradeep@chiradeep·
@laurenbalik By showing how contemporary figures allegedly use coded messaging, Lauren models the very interpretive paranoia she critiques. This is critique and performance art: showing how Straussian techniques have proliferated while using those very techniques to make her case.
English
0
0
0
41
Chief AI Officer
Chief AI Officer@chiefaioffice·
New market maps covering the latest AI startups: {this is a long one, bookmark it} 1. Intelligent-first apps & infra from Insight Partners
Chief AI Officer tweet media
English
18
83
452
86.5K
chiradeep
chiradeep@chiradeep·
@pitdesi It was a typo : the email meant to ask “are you tressed”? If you work in a beauty startup, how is your hair not perfect?
English
0
0
0
107
Sheel Mohnot
Sheel Mohnot@pitdesi·
Indian home beauty startup YesMadam sent a survey to their employees about stress and then fired the employees that said they were under significant stress 🤦🏽‍♂️
Sheel Mohnot tweet media
English
452
2.5K
20.4K
2.4M
chiradeep
chiradeep@chiradeep·
@HamelHusain Been using BackgroundTasks from FastAPI for this. Of course there is no capacity control …
English
0
0
1
40
Hamel Husain
Hamel Husain@HamelHusain·
Are you building Starlette based apps (FastAPI, FastHTML, etc) apps but are confused by how to run things in the background (ex: LLM calls)? This thread is for you! I did a deep dive into this First, how can you use your SQL DB directly as a queue? 1/10
Hamel Husain tweet media
English
8
27
287
35.8K
Greg Yang
Greg Yang@TheGregYang·
The future belongs to those who do evals
English
53
78
1K
135.4K
chiradeep
chiradeep@chiradeep·
@karpathy @itsclivetime I thought transformers have been found to be much worse than traditional time series algorithms for forecasting, predicting and anomaly detection
English
0
0
0
258
Andrej Karpathy
Andrej Karpathy@karpathy·
@itsclivetime Certainly you could think about "speaking textures", or "speaking molecules", or etc. What I've seen though is that the word "language" is misleading people to think LLMs are restrained to text applications.
English
29
28
783
61.9K
Andrej Karpathy
Andrej Karpathy@karpathy·
It's a bit sad and confusing that LLMs ("Large Language Models") have little to do with language; It's just historical. They are highly general purpose technology for statistical modeling of token streams. A better name would be Autoregressive Transformers or something. They don't care if the tokens happen to represent little text chunks. It could just as well be little image patches, audio chunks, action choices, molecules, or whatever. If you can reduce your problem to that of modeling token streams (for any arbitrary vocabulary of some set of discrete tokens), you can "throw an LLM at it". Actually, as the LLM stack becomes more and more mature, we may see a convergence of a large number of problems into this modeling paradigm. That is, the problem is fixed at that of "next token prediction" with an LLM, it's just the usage/meaning of the tokens that changes per domain. If that is the case, it's also possible that deep learning frameworks (e.g. PyTorch and friends) are way too general for what most problems want to look like over time. What's up with thousands of ops and layers that you can reconfigure arbitrarily if 80% of problems just want to use an LLM? I don't think this is true but I think it's half true.
English
564
1.2K
10.5K
1.3M