Angelo Dalli

2.2K posts

Angelo Dalli banner
Angelo Dalli

Angelo Dalli

@AngeloDalli

Chief Scientist at @UmnaiBase. Unleashing limitless human ingenuity through a revolutionary machine learning approach: Hybrid Intelligence. AI PhD

London, England Katılım Mart 2009
759 Takip Edilen1.6K Takipçiler
Angelo Dalli
Angelo Dalli@AngeloDalli·
Six months ago I disagreed with this stat. After the last three months of output, I no longer do. I’ve consistently observed a productivity plateau tied to functionality. Once features get close to finished, QA and fixes slow velocity, and the project converges. For many software builds, functional completion in ~14–15 days now feels realistic.
English
0
0
0
98
Paul Graham
Paul Graham@paulg·
An experienced programmer told me he's now using AI to generate a thousand lines of code an hour. When I posted a similar stat 6 months ago, I got about a 50-50 mix of indignant disbelief and "Yeah, me too." I'm curious if the split will be different this time.
English
579
53
2.1K
502.1K
Angelo Dalli
Angelo Dalli@AngeloDalli·
@fchollet Except that code is transparent and symbolic… so shortcuts and testing can actually be done effectively. I think there will be a shift towards test driven coding. Specs will still be specs… maybe the old software engineering template and code pattern ideas will make a comeback
English
0
0
0
130
François Chollet
François Chollet@fchollet·
Sufficiently advanced agentic coding is essentially machine learning: the engineer sets up the optimization goal as well as some constraints on the search space (the spec and its tests), then an optimization process (coding agents) iterates until the goal is reached. The result is a blackbox model (the generated codebase): an artifact that performs the task, that you deploy without ever inspecting its internal logic, just as we ignore individual weights in a neural network. This implies that all classic issues encountered in ML will soon become problems for agentic coding: overfitting to the spec, Clever Hans shortcuts that don't generalize outside the tests, data leakage, concept drift, etc. I would also ask: what will be the Keras of agentic coding? What will be the optimal set of high-level abstractions that allow humans to steer codebase 'training' with minimal cognitive overhead?
English
171
383
3.3K
320K
Angelo Dalli
Angelo Dalli@AngeloDalli·
I think that Go and Rust are good target languages for LLM translation from other coding languages. My precise definition of an LLM friendly language is a language where surface syntax maps 1:1 to a stable AST. No hidden desugaring, no macro rewrites, no context-dependent parsing. In more detail: Small unambiguous grammar, explicit types at boundaries, no UB or macros, deterministic semantics, effect-typed, canonical formatting, structured diagnostics, and a single standard build model.
Thomas Wolf@Thom_Wolf

Shifting structures in a software world dominated by AI. Some first-order reflections (TL;DR at the end): Reducing software supply chains, the return of software monoliths – When rewriting code and understanding large foreign codebases becomes cheap, the incentive to rely on deep dependency trees collapses. Writing from scratch ¹ or extracting the relevant parts from another library is far easier when you can simply ask a code agent to handle it, rather than spending countless nights diving into an unfamiliar codebase. The reasons to reduce dependencies are compelling: a smaller attack surface for supply chain threats, smaller packaged software, improved performance, and faster boot times. By leveraging the tireless stamina of LLMs, the dream of coding an entire app from bare-metal considerations all the way up is becoming realistic. End of the Lindy effect – The Lindy effect holds that things which have been around for a long time are there for good reason and will likely continue to persist. It's related to Chesterton's fence: before removing something, you should first understand why it exists, which means removal always carries a cost. But in a world where software can be developed from first principles and understood by a tireless agent, this logic weakens. Older codebases can be explored at will; long-standing software can be replaced with far less friction. A codebase can be fully rewritten in a new language. ² Legacy software can be carefully studied and updated in situations where humans would have given up long ago. The catch: unknown unknowns remain unknown. The true extent of AI's impact will hinge on whether complete coverage of testing, edge cases, and formal verification is achievable. In an AI-dominated world, formal verification isn't optional—it's essential. The case for strongly typed languages – Historically, programming language adoption has been driven largely by human psychology and social dynamics. A language's success depended on a mix of factors: individual considerations like being easy to learn and simple to write correctly; community effects like how active and welcoming a community was, which in turn shaped how fast its ecosystem would grow; and fundamental properties like provable correctness, formal verification, and striking the right balance between dynamic and static checks—between the freedom to write anything and the discipline of guarding against edge cases and attacks. As the human factor diminishes, these dynamics will shift. Less dependence on human psychology will favor strongly typed, formally verifiable and/or high performance languages.³ These are often harder for humans to learn, but they're far better suited to LLMs, which thrive on formal verification and reinforcement learning environments. Expect this to reshape which languages dominate. Economic restructuring of open source – For decades, open-source communities have been built around humans finding connection through writing, learning, and using code together. In a world where most code is written—and perhaps more importantly, read—by machines, these incentives will start to break down.⁴ Communities of AIs building libraries and codebases together will likely emerge as a replacement, but such communities will lack the fundamentally human motivations that have driven open source until now. If the future of open-source development becomes largely devoid of humans, alignment of AI models won't just matter—it will be decisive. The future of new languages – Will AI agents face the same tradeoffs we do when developing or adopting new programming languages? Expressiveness vs. simplicity, safety vs. control, performance vs. abstraction, compile time vs. runtime, explicitness vs. conciseness. It's unclear that they will. In the long term, the reasons to create a new programming language will likely diverge significantly from the human-driven motivations of the past. There may well be an optimal programming language for LLMs—and there's no reason to assume it will resemble the ones humans have converged on. TL; DR: - Monoliths return – cheap rewriting kills dependency trees; smaller attack surface, better performance, bare-metal becomes realistic - Lindy effect weakens – legacy code loses its moat, but unknown unknowns persist; formal verification becomes essential - Strongly typed languages rise – human psychology mattered for adoption; now formal verification and RL environments favor types over ergonomics - Open source restructures – human connection drove the community; AI-written/read code breaks those incentives; alignment becomes decisive - New languages diverge – AI may not share our tradeoffs; optimal LLM programming languages may look nothing like what humans converged on ¹ x.com/mntruell/statu… ² x.com/anthropicai/st… ³ wesmckinney.com/blog/agent-erg…#issuecomment-3717222957" target="_blank" rel="nofollow noopener">github.com/tailwindlabs/t…

English
0
0
0
100
Angelo Dalli
Angelo Dalli@AngeloDalli·
I think a good AST based (or similar) transpiler pipeline is needed, so I’d think of Go/Rust as target languages. I would boil it down to: an LLM friendly language is one where surface syntax maps 1:1 to a stable AST. No hidden desugaring, no macro rewrites, no context-dependent parsing. In more detail: Small unambiguous grammar, explicit types at boundaries, no UB or macros, deterministic semantics, effect-typed, canonical formatting, structured diagnostics, and a single standard build model.
English
0
0
0
32
Andrej Karpathy
Andrej Karpathy@karpathy·
I think it must be a very interesting time to be in programming languages and formal methods because LLMs change the whole constraints landscape of software completely. Hints of this can already be seen, e.g. in the rising momentum behind porting C to Rust or the growing interest in upgrading legacy code bases in COBOL or etc. In particular, LLMs are *especially* good at translation compared to de-novo generation because 1) the original code base acts as a kind of highly detailed prompt, and 2) as a reference to write concrete tests with respect to. That said, even Rust is nowhere near optimal for LLMs as a target language. What kind of language is optimal? What concessions (if any) are still carved out for humans? Incredibly interesting new questions and opportunities. It feels likely that we'll end up re-writing large fractions of all software ever written many times over.
Thomas Wolf@Thom_Wolf

Shifting structures in a software world dominated by AI. Some first-order reflections (TL;DR at the end): Reducing software supply chains, the return of software monoliths – When rewriting code and understanding large foreign codebases becomes cheap, the incentive to rely on deep dependency trees collapses. Writing from scratch ¹ or extracting the relevant parts from another library is far easier when you can simply ask a code agent to handle it, rather than spending countless nights diving into an unfamiliar codebase. The reasons to reduce dependencies are compelling: a smaller attack surface for supply chain threats, smaller packaged software, improved performance, and faster boot times. By leveraging the tireless stamina of LLMs, the dream of coding an entire app from bare-metal considerations all the way up is becoming realistic. End of the Lindy effect – The Lindy effect holds that things which have been around for a long time are there for good reason and will likely continue to persist. It's related to Chesterton's fence: before removing something, you should first understand why it exists, which means removal always carries a cost. But in a world where software can be developed from first principles and understood by a tireless agent, this logic weakens. Older codebases can be explored at will; long-standing software can be replaced with far less friction. A codebase can be fully rewritten in a new language. ² Legacy software can be carefully studied and updated in situations where humans would have given up long ago. The catch: unknown unknowns remain unknown. The true extent of AI's impact will hinge on whether complete coverage of testing, edge cases, and formal verification is achievable. In an AI-dominated world, formal verification isn't optional—it's essential. The case for strongly typed languages – Historically, programming language adoption has been driven largely by human psychology and social dynamics. A language's success depended on a mix of factors: individual considerations like being easy to learn and simple to write correctly; community effects like how active and welcoming a community was, which in turn shaped how fast its ecosystem would grow; and fundamental properties like provable correctness, formal verification, and striking the right balance between dynamic and static checks—between the freedom to write anything and the discipline of guarding against edge cases and attacks. As the human factor diminishes, these dynamics will shift. Less dependence on human psychology will favor strongly typed, formally verifiable and/or high performance languages.³ These are often harder for humans to learn, but they're far better suited to LLMs, which thrive on formal verification and reinforcement learning environments. Expect this to reshape which languages dominate. Economic restructuring of open source – For decades, open-source communities have been built around humans finding connection through writing, learning, and using code together. In a world where most code is written—and perhaps more importantly, read—by machines, these incentives will start to break down.⁴ Communities of AIs building libraries and codebases together will likely emerge as a replacement, but such communities will lack the fundamentally human motivations that have driven open source until now. If the future of open-source development becomes largely devoid of humans, alignment of AI models won't just matter—it will be decisive. The future of new languages – Will AI agents face the same tradeoffs we do when developing or adopting new programming languages? Expressiveness vs. simplicity, safety vs. control, performance vs. abstraction, compile time vs. runtime, explicitness vs. conciseness. It's unclear that they will. In the long term, the reasons to create a new programming language will likely diverge significantly from the human-driven motivations of the past. There may well be an optimal programming language for LLMs—and there's no reason to assume it will resemble the ones humans have converged on. TL; DR: - Monoliths return – cheap rewriting kills dependency trees; smaller attack surface, better performance, bare-metal becomes realistic - Lindy effect weakens – legacy code loses its moat, but unknown unknowns persist; formal verification becomes essential - Strongly typed languages rise – human psychology mattered for adoption; now formal verification and RL environments favor types over ergonomics - Open source restructures – human connection drove the community; AI-written/read code breaks those incentives; alignment becomes decisive - New languages diverge – AI may not share our tradeoffs; optimal LLM programming languages may look nothing like what humans converged on ¹ x.com/mntruell/statu… ² x.com/anthropicai/st… ³ wesmckinney.com/blog/agent-erg…#issuecomment-3717222957" target="_blank" rel="nofollow noopener">github.com/tailwindlabs/t…

English
701
655
8.1K
1.2M
Angelo Dalli
Angelo Dalli@AngeloDalli·
@fchollet Neurosymbolic AI achieves this. Getting steadily to that sweet spot of autonomous discovery + intelligible representation (not some opaque latent space)
English
0
0
0
86
François Chollet
François Chollet@fchollet·
A good solution to the intelligence problem should be able to autonomously produce abstractions that compose well, stack well, and stand the test of time. Without cribbing them from somewhere else. So far there's no tech that achieves this. Gradient descent certainly doesn't.
English
67
35
388
30.3K
Angelo Dalli
Angelo Dalli@AngeloDalli·
@paulg @garrytan I have managed to get consistent high-quality results of around 25K LOC per day for the last 5 weeks. There is still a substantial amount of manual review but it feels more like being a mashup of a team lead, product owner, QA manager and mentor.
Angelo Dalli@AngeloDalli

This post by @paulg has been at the back of my mind since last August. I have to take my initial comments back as I *did* manage to get 25,000 lines of code of high quality code a day, obsessively using multiple AI agents (ChatGPT, Codex, Claude and Lovable). I don’t think a Vibe coder nor junior dev can do this as you really need to know what you’re doing (I have 25+ yrs experience in deeptech AI). AI coding still makes a lot of mistakes, but it convinces me that this is doable in a sustainable way. My 25K LOCs a day needs 16hr/day obsession, 10K is sustainable. I don’t like the LOC measure but I can very certainly say that an almost identical system (the old version before this) took rather competent human devs around 13 man months to do by hand. I finished the same in just 4 days. Will take 4 md over 13 mm any day! I think the problem will be running out of specs to code :)

English
0
0
0
693
Garry Tan
Garry Tan@garrytan·
I use a very specific prompt to push Claude to check its work and do a lot of testing and thinking about perf and refactoring. I find I can do big features (4K LOC+ with full testing) in about an hour.
Garry Tan tweet media
English
207
328
5.1K
621.9K
Angelo Dalli retweetledi
Nav Toor
Nav Toor@heynavtoor·
20 YouTube channels that teach AI better than most CS degrees in 2026: 1. Andrej Karpathy Deep, intuitive walkthroughs of neural networks and modern LLMs @AndrejKarpathy" target="_blank" rel="nofollow noopener">youtube.com/@AndrejKarpathy 2. 3Blue1Brown Visual intuition for math, linear algebra, and neural networks @3blue1brown" target="_blank" rel="nofollow noopener">youtube.com/@3blue1brown 3. StatQuest with Josh Starmer Clear, friendly explanations of statistics and ML fundamentals @statquest" target="_blank" rel="nofollow noopener">youtube.com/@statquest 4. Stanford Online University-grade ML and AI lecture series (Andrew Ng, CS229, etc.) @stanfordonline" target="_blank" rel="nofollow noopener">youtube.com/@stanfordonline 5. sentdex Practical machine learning and Python projects @sentdex" target="_blank" rel="nofollow noopener">youtube.com/@sentdex 6. Yannic Kilcher Deep dives into ML and AI research papers @YannicKilcher" target="_blank" rel="nofollow noopener">youtube.com/@YannicKilcher 7. MIT OpenCourseWare Rigorous academic courses on ML, AI, and applied mathematics @mitocw" target="_blank" rel="nofollow noopener">youtube.com/@mitocw 8. Siraj Raval High-level overviews and motivation around AI concepts Link: @SirajRaval" target="_blank" rel="nofollow noopener">youtube.com/@SirajRaval 9. DeepLearningAI Structured learning paths for deep learning and generative AI @DeepLearningAI" target="_blank" rel="nofollow noopener">youtube.com/@DeepLearningAI 10. Two Minute Papers Fast, accessible summaries of cutting-edge AI research @TwoMinutePapers" target="_blank" rel="nofollow noopener">youtube.com/@TwoMinutePape… 11. Umar Jamil Clear, implementation-focused explanations of transformers and LLMs @umarjamilai" target="_blank" rel="nofollow noopener">youtube.com/@umarjamilai 12. Hugging Face Open-source LLMs, transformers, and modern NLP tooling @huggingface" target="_blank" rel="nofollow noopener">youtube.com/@huggingface 13. Steve Brunton Dynamical systems, control theory, and scientific ML @Eigensteve" target="_blank" rel="nofollow noopener">youtube.com/@Eigensteve 14. Michael Bronstein Geometric deep learning and graph neural networks @MichaelBronsteinGDL" target="_blank" rel="nofollow noopener">youtube.com/@MichaelBronst… 15. Caltech Advanced lecture series on ML, optimization, and theory @Caltech" target="_blank" rel="nofollow noopener">youtube.com/@Caltech 16. Lex Fridman Long-form conversations with top AI researchers and practitioners @lexfridman" target="_blank" rel="nofollow noopener">youtube.com/@lexfridman 17. Arxiv Insights Beginner-friendly explanations of recent AI papers @ArxivInsights" target="_blank" rel="nofollow noopener">youtube.com/@ArxivInsights 18. Machine Learning Street Talk Unfiltered, technical discussions on AI research and theory @MachineLearningStreetTalk" target="_blank" rel="nofollow noopener">youtube.com/@MachineLearni… 19. Jeremy Howard Practical deep learning with strong intuition @howardjeremyp" target="_blank" rel="nofollow noopener">youtube.com/@howardjeremyp 20. Kaggle Applied ML, competitions, notebooks, and real-world workflows @kaggle" target="_blank" rel="nofollow noopener">youtube.com/@kaggle
Nav Toor tweet mediaNav Toor tweet mediaNav Toor tweet mediaNav Toor tweet media
English
25
377
1.5K
78.8K
Angelo Dalli
Angelo Dalli@AngeloDalli·
Let me try to make this clearer. Logical verification systems don’t show that LLMs themselves reason exhaustively — they merely show that external systems can check or constrain outputs against rules. LLMs do not internally guarantee transitivity, non-contradiction, or causal closure. This is a consequence of the AR-based Transformer architecture itself. In other words, the reasoning may be reliable at the system level (e.g. LLM + verification, in your case via some IR-based checks), but not at the model level. Reasoning systems using LLMs can be reliable; LLMs themselves are not. By contrast, reasoning systems that do not require such scaffolding—formal solvers, proof systems or symbolic reasoners—encode invariants directly. Their guarantees hold by construction, not by post-hoc filtering. That distinction matters operationally, because the failure modes are very different: humans and formal systems violate rules detectably; LLMs violate them silently unless externally caught. The implication is that the IR system you’re building can be effective, but its reliability depends critically on the surrounding constraints rather than on the model itself.
English
1
0
0
57
Leonie
Leonie@helloiamleonie·
i'm clearly biased but this is the most interesting take on agent memory i've seen so far. (yes, forget the "filesystem vs database" discussion) a few weeks back i had a nice chat with @vintrotweets from @plasticlabs and their approach is: memory is not a retrieval problem. memory is a prediction problem. i'm slowly catching up on my reading list. but this one is definitely worth a read: blog.plasticlabs.ai/blog/Memory-as…
Leonie tweet media
English
62
104
1.1K
93.1K
Angelo Dalli
Angelo Dalli@AngeloDalli·
While I do like what you’re doing with @plasticlabs, you’re wrong on this one. If all logic were just pattern completion, formal verification systems wouldn’t exist. “Exhaustive” reasoning here is supplied by scaffolding, not the model itself. The distinction isn’t pedantic: without enforced invariants, failures are silent and unbounded. Humans reason fallibly but norm-governed; LLMs are statistically competent, not logically accountable. LLMs beat humans on benchmarks the way calculators beat mathematicians. That’s not ‘better reasoning’—it’s optimizing a different objective under different constraints.
English
1
0
0
57
Courtland Leer
Courtland Leer@courtlandleer·
they do reason exhaustively inside a system where that is the prescribed and harnessed task in fact they are far more reliable that human reasoners whether this is “simulating reasoning” or “reasoning itself” more a pedantic debate than a meaningfully practical one all logic is pattern completion, and human minds also estimate over distributions
English
2
0
1
67
Angelo Dalli
Angelo Dalli@AngeloDalli·
LLMs don’t “reason exhaustively”, they interpolate over training distributions. When surface statistics shift, logical consistency collapses — this is not reasoning in the trustworthy reliable/logical sense—it’s pattern completion with good priors… Have a look at: arxiv.org/abs/2506.06941 and arxiv.org/abs/2508.10265 amongst others
English
1
0
0
68
Angelo Dalli
Angelo Dalli@AngeloDalli·
p.s. The text only and .md files are the prompts and agent instructions. I usually start by hand writing the stuff, then using ChatGPT to improve it, and finally hand correcting it
English
0
0
0
65
Angelo Dalli
Angelo Dalli@AngeloDalli·
This post by @paulg has been at the back of my mind since last August. I have to take my initial comments back as I *did* manage to get 25,000 lines of code of high quality code a day, obsessively using multiple AI agents (ChatGPT, Codex, Claude and Lovable). I don’t think a Vibe coder nor junior dev can do this as you really need to know what you’re doing (I have 25+ yrs experience in deeptech AI). AI coding still makes a lot of mistakes, but it convinces me that this is doable in a sustainable way. My 25K LOCs a day needs 16hr/day obsession, 10K is sustainable. I don’t like the LOC measure but I can very certainly say that an almost identical system (the old version before this) took rather competent human devs around 13 man months to do by hand. I finished the same in just 4 days. Will take 4 md over 13 mm any day! I think the problem will be running out of specs to code :)
Angelo Dalli tweet media
Paul Graham@paulg

I met a founder today who said he writes 10,000 lines of code a day now thanks to AI. This is probably the limit case. He's a hotshot programmer, he knows AI tools very well, and he's talking about a 12 hour day. But he's not naive. This is not 10,000 lines of bug-filled crap.

English
2
0
2
935
Angelo Dalli
Angelo Dalli@AngeloDalli·
@iruletheworldmo None of the concerns in this article are actually true. LLMs are still doing next token prediction, and there is just a lot of hype hoping that something other than that will emerge… truth is: there is no such evolution
English
0
0
5
517
Angelo Dalli
Angelo Dalli@AngeloDalli·
@TimTeaFan I recommend it as an example of hype and misguided thinking to mislead. LLMs are still doing next token prediction and the article is full of loopholes and inaccuracies
English
0
0
54
5.4K
Angelo Dalli
Angelo Dalli@AngeloDalli·
@elonmusk AI that’s truly useful is compression, correlation, cause-and-effect understanding and symbolic reasoning and representation. @UmnaiBase
English
0
0
1
486
Angelo Dalli
Angelo Dalli@AngeloDalli·
@pmddomingos Also how would Tensor Logic handle a symbol that’s a blend or superposition of others (as happens in GPT-style embeddings)? Can mixed meanings be represented and reasoned over cleanly?
English
0
0
0
116
Angelo Dalli
Angelo Dalli@AngeloDalli·
@pmddomingos Could you expand on how such contextualization rules may be represented and learned in Tensor Logic — e.g. taking the form of higher-order tensors, parameterized transformations, or recursively defined mappings?
English
1
0
1
138