monty

908 posts

monty banner
monty

monty

@_mtmk

Software Engineer @nats_io NATS .NET client. Views are my own.

UK Katılım Şubat 2015
357 Takip Edilen209 Takipçiler
monty
monty@_mtmk·
just found out about claude code ide integration
monty tweet media
English
0
0
0
46
monty retweetledi
BURKOV
BURKOV@burkov·
If you don't understand this, you will not understand why LLM-based agents are irreparably failing for a general-purpose problem solving. An agent (by the way it was the topic of my PhD 20 years ago) to be useful, must be rational. Being rational means to always prefer an outcome that results in the maximal expected utility to its master/user. Let’s say an agent has two actions they can execute in an environment: a_1 and a_2. If the agent can predict that a_1 gives its user an expected utility of 10, and a_2 gives an expected utility of -100, then a rational agent must choose a_1 even if choosing a_2 seems like a better option when explained in words. The numbers 10 and -100 can be obtained by summing the products of all possible outcomes for each action and their likelihoods. Now here is the problem with LLM-based agents. The LLM is not optimizing expected utility in the environment. It is optimizing the next token, conditioned on a prompt, a context window, and a training distribution full of examples of what helpful answers are supposed to look like. Those are not the same objective. So when we wrap an LLM in a loop and call it an “agent,” we have not created a rational decision-maker. We have created a text generator that can imitate the surface form of deliberation. It may say things like: “I should compare the expected outcomes.” “The best action is probably a_1.” “I will now execute the optimal plan.” But the internal mechanism is not selecting actions by maximizing the user’s expected utility. It is generating a continuation that is statistically appropriate given the prompt and prior context. This distinction matters enormously. For narrow tasks, the imitation can be good enough. If the environment is constrained, the actions are simple, and the success criteria are close to patterns seen in training, the system can appear agentic. But for general-purpose problem solving, the gap becomes fatal. A rational agent needs stable preferences, calibrated beliefs, causal models of the world, the ability to evaluate consequences, and the discipline to choose the action with maximal expected utility even when that action is boring, non-linguistic, or unlike the examples in its training data. An LLM-based agent has none of that by default. It has fluency. It has pattern completion. It has a remarkable ability to compress and recombine human text. But fluency is not rationality, and a plausible plan is not an expected-utility calculation. This is why these systems so often fail in strange, brittle, and irreparable ways when given open-ended responsibility. They are not failing because the prompts are insufficiently clever. They are failing because we are asking a simulator of rational agency to be a rational agent.
English
175
272
1.6K
196.1K
Dmitrii Kovanikov
Dmitrii Kovanikov@ChShersh·
Everyone knows that 15 minutes before standup is the most productive time. This is when developers do highly-focused last-minute work just to give some updates. So, in order to increase productivity, we now have 4 standups a day.
English
89
158
4K
95.3K
monty retweetledi
Mitchell Hashimoto
Mitchell Hashimoto@mitchellh·
Ghostty is leaving GitHub. I'm GitHub user 1299, joined Feb 2008. I've visited GitHub almost every single day for over 18 years. It's never been a question for me where I'd put my projects: always GitHub. I'm super sad to say this, but its time to go. mitchellh.com/writing/ghostt…
English
548
1.6K
16.8K
2.9M
monty retweetledi
Derek Collison
Derek Collison@derekcollison·
Over the years @nats_io systems have grown in complexity by leaps and bounds. The system is observable to a fault and produces a ton of information about what is going on, but very few people know about everything that is available. So proud of the @synadia team delivering insights. A wealth of information and data that can be accessed via UI, API and your AI agents.
Synadia@synadia

@NATS_io has a particular adoption curve. Someone introduces it for one use case, it just works, and within a year or two it's threaded through half the systems in the company. This is great! Until it's not.

English
4
9
72
2.9M
Adam Reed
Adam Reed@IAmAdamReed·
@_mtmk Better! I threw my idea on there on a comment as well.
English
1
0
1
21
monty
monty@_mtmk·
ICYMI we have recently released fair few Orbit .NET packages. They provide additional functionality on top of NATS Core and in most cases JetStream. #packages" target="_blank" rel="nofollow noopener">github.com/synadia-io/orb…
English
1
2
6
1K
Adam Reed
Adam Reed@IAmAdamReed·
@_mtmk I missed orbit altogether, I just peeked the repo but it doesn’t have much for a TLDR
English
1
0
1
25
monty
monty@_mtmk·
anyone else think claude code auto mode is a ticking time bomb?
English
0
0
1
79
monty
monty@_mtmk·
New NATS .NET v3 preview (.NET 10 targeting) also introduced a new feature: Serializers can process headers now: github.com/nats-io/nats.n…
English
0
0
4
82
monty retweetledi
Datastar Cult Leader
Datastar Cult Leader@DelaneyGillilan·
Datastar 1.0 has FINALLY SHIPPED! 🚀🚀🚀 WE ARE IN ORBIT 🚀🚀🚀 Watch the launch podcast. Welcome to planet boring y'all! youtube.com/watch?v=T6uwri…
YouTube video
YouTube
English
23
34
138
11.6K
monty retweetledi
Synadia
Synadia@synadia·
What if every AI agent had a phone number? Not a brittle HTTP endpoint that breaks when the agent moves. A stable address that works regardless of what agent framework the agent uses, what language it's written in, or where it's running. NATS makes it possible
Synadia tweet media
English
1
4
12
588
Marc Gravell
Marc Gravell@marcgravell·
meetingsMissedDueToEuropeVersusAmericaDST++;
Dansk
1
0
17
2.4K