StarKill

341 posts

StarKill banner
StarKill

StarKill

@StarKill_

Arhus, Denmark Katılım Mayıs 2016
322 Takip Edilen41 Takipçiler
StarKill
StarKill@StarKill_·
@NordicRest @Felagoth1 @anishgiri The diagram is from 1982 and was drawn by Botvinnik. Probably you're right it's showcasing pseudolegal move generation. But it's not a good visualisation of how the human brain has to evaluate as the 2025 paper someone else linked claims.
English
0
0
1
36
Nordic
Nordic@NordicRest·
@Felagoth1 @StarKill_ @anishgiri In *modern* chess engines. 80 years ago, those algorithms hadn't been inplemented yet. Please stop talking about topics you don't know shit-all about.
English
1
0
0
57
Anish Giri
Anish Giri@anishgiri·
I see this odd photo circulating the internet. Illegal moves are not possibilities, this easy study is a straightforward win (g4 Kh4 Bh6 Qxh6 Qh2 Kg5 Qd2 Nf4 Qd8 mate) and is a terrible demonstration of the tree of possibilities in chess.
Anish Giri tweet media
English
15
5
336
46.7K
StarKill
StarKill@StarKill_·
@NordicRest @anishgiri In general, the statement that a computer has to generate illegal moves to see if they're illegal is not true. Of course you're right it's common to generate pseudolegal moves because moving into check is not trivial to check.
English
0
0
0
9
Nordic
Nordic@NordicRest·
@StarKill_ @anishgiri Bitboards and binary operations have nothing to do with capability to generate legal moves directly. This is an extremely primitive algorithm with no such capability.
English
2
0
1
173
Nordic
Nordic@NordicRest·
@anishgiri Yes, illegal moves ARE possibilities, *for a computer*. A computer needs to attempt an illegal move before it can determine that it's an illegal move.
English
2
0
5
1.1K
John Carmack
John Carmack@ID_AA_Carmack·
Making a scatter plot of 400_000 data points, some of the plots had odd gaps in coverage. It took me a little while to realize that it was only when the data was farther from the origin -- it was the raw bfloat16 precision. Everything looks great from -1 to 1, but as you go past 2 and 4, the coverage gaps get larger. My intuition didn't have it being quite so "discretely countable" at those modest numeric values. Float32 for comparison.
John Carmack tweet mediaJohn Carmack tweet media
English
69
110
1.9K
583.1K
StarKill
StarKill@StarKill_·
@Fintech03 >he spent the rest of his time reading/walking/simply sitting in silence. Poincaré was a mathematician. Sitting in silence IS working. There were not a thousand menial tasks to be done. This is vastly different than scrolling reels and working as a webdev for only 4 hours.
English
0
1
18
1.1K
Parimal
Parimal@Fintech03·
Today we debate how much work is enough work, Poincaré had a system for his day. He worked exactly from 10:00 AM to 12:00 PM & 5:00 PM to 7:00 PM. He believed that working more than 4 hrs a day was harmful to the brain's ability to synthesize. He spent the rest of his time reading/walking/simply sitting in silence. He was a living rebuttal to the modern grind culture proving that 4 hrs of extreme deep work could out-produce a lifetime of 80 hr weeks.
Physics In History@PhysInHistory

Today marks the birth anniversary of Henri Poincaré (1854–1912), a pioneering French mathematician, physicist, and philosopher of science. Renowned as "The Last Universalist", Poincaré made foundational contributions to celestial mechanics, topology, chaos theory, and the early development of special relativity. His work laid the groundwork for modern mathematical physics and influenced generations of scientists and thinkers.

English
35
158
1.2K
134.4K
StarKill
StarKill@StarKill_·
Fun fact: The digit sum trick for checking divisibility doesn't just work for 3 and 9, it also works for 1.
English
0
1
0
108
StarKill
StarKill@StarKill_·
@erikcorry Programming would hypothetically uniquely be the thing they had to be good at for the singularity. Recursive self improvement was always envisioned as being driven by self-coding agents.
English
2
0
0
21
Erik
Erik@erikcorry·
Cancel the singularity: LLMs are uniquely good at programming, for reasons that don't apply to other fields: * The training data is amazing * Type checkers eliminate hallucinations * It's a language-based activity
English
2
0
2
154
WetWork
WetWork@WWxrk33218·
@Math_files genuinely is all this von Neumann stuff real?
English
3
0
0
8.6K
Math Files
Math Files@Math_files·
George Pólya tells a story about his former student John von Neumann: “He is the only student of mine I was ever intimidated by. He was so quick. There was a seminar for advanced students in Zürich that I was teaching, and von Neumann was in the class. I came to a certain theorem and said it was not proved and might be difficult. Von Neumann didn’t say anything. But after five minutes, he raised his hand. When I called on him, he went to the board and proceeded to write down the proof. After that, I was afraid of von Neumann.”
Math Files tweet media
English
60
258
2.9K
354.6K
StarKill
StarKill@StarKill_·
@AISafetyMemes A hard take off is literally days/weeks. More than 5 years since gpt-3, it's clearly not a hard takeoff, even if it's still scary.
English
0
0
0
71
AI Notkilleveryoneism Memes ⏸️
Andrej Karpathy: "This is easily the biggest change in ~2 decades of programming and it happened over the course of a few weeks." "I rapidly went from about 80% manual+autocomplete coding and 20% agents to 80% agent coding and 20% edits+touchups." "I am bracing for 2026 as the year of the slopacolypse." "LLM agent capabilities (Claude & Codex especially) have crossed some kind of threshold of coherence around December 2025 and caused a phase shift in software engineering." "I am slowly starting to atrophy my ability to write code manually." "It's so interesting to watch an agent relentlessly work at something. They never get tired, they never get demoralized, they just keep going and trying things where a person would have given up long ago to fight another day. It's a "feel the AGI" moment to watch it struggle with something for a long time just to come out victorious 30 minutes later."
AI Notkilleveryoneism Memes ⏸️ tweet media
Andrej Karpathy@karpathy

A few random notes from claude coding quite a bit last few weeks. Coding workflow. Given the latest lift in LLM coding capability, like many others I rapidly went from about 80% manual+autocomplete coding and 20% agents in November to 80% agent coding and 20% edits+touchups in December. i.e. I really am mostly programming in English now, a bit sheepishly telling the LLM what code to write... in words. It hurts the ego a bit but the power to operate over software in large "code actions" is just too net useful, especially once you adapt to it, configure it, learn to use it, and wrap your head around what it can and cannot do. This is easily the biggest change to my basic coding workflow in ~2 decades of programming and it happened over the course of a few weeks. I'd expect something similar to be happening to well into double digit percent of engineers out there, while the awareness of it in the general population feels well into low single digit percent. IDEs/agent swarms/fallability. Both the "no need for IDE anymore" hype and the "agent swarm" hype is imo too much for right now. The models definitely still make mistakes and if you have any code you actually care about I would watch them like a hawk, in a nice large IDE on the side. The mistakes have changed a lot - they are not simple syntax errors anymore, they are subtle conceptual errors that a slightly sloppy, hasty junior dev might do. The most common category is that the models make wrong assumptions on your behalf and just run along with them without checking. They also don't manage their confusion, they don't seek clarifications, they don't surface inconsistencies, they don't present tradeoffs, they don't push back when they should, and they are still a little too sycophantic. Things get better in plan mode, but there is some need for a lightweight inline plan mode. They also really like to overcomplicate code and APIs, they bloat abstractions, they don't clean up dead code after themselves, etc. They will implement an inefficient, bloated, brittle construction over 1000 lines of code and it's up to you to be like "umm couldn't you just do this instead?" and they will be like "of course!" and immediately cut it down to 100 lines. They still sometimes change/remove comments and code they don't like or don't sufficiently understand as side effects, even if it is orthogonal to the task at hand. All of this happens despite a few simple attempts to fix it via instructions in CLAUDE . md. Despite all these issues, it is still a net huge improvement and it's very difficult to imagine going back to manual coding. TLDR everyone has their developing flow, my current is a small few CC sessions on the left in ghostty windows/tabs and an IDE on the right for viewing the code + manual edits. Tenacity. It's so interesting to watch an agent relentlessly work at something. They never get tired, they never get demoralized, they just keep going and trying things where a person would have given up long ago to fight another day. It's a "feel the AGI" moment to watch it struggle with something for a long time just to come out victorious 30 minutes later. You realize that stamina is a core bottleneck to work and that with LLMs in hand it has been dramatically increased. Speedups. It's not clear how to measure the "speedup" of LLM assistance. Certainly I feel net way faster at what I was going to do, but the main effect is that I do a lot more than I was going to do because 1) I can code up all kinds of things that just wouldn't have been worth coding before and 2) I can approach code that I couldn't work on before because of knowledge/skill issue. So certainly it's speedup, but it's possibly a lot more an expansion. Leverage. LLMs are exceptionally good at looping until they meet specific goals and this is where most of the "feel the AGI" magic is to be found. Don't tell it what to do, give it success criteria and watch it go. Get it to write tests first and then pass them. Put it in the loop with a browser MCP. Write the naive algorithm that is very likely correct first, then ask it to optimize it while preserving correctness. Change your approach from imperative to declarative to get the agents looping longer and gain leverage. Fun. I didn't anticipate that with agents programming feels *more* fun because a lot of the fill in the blanks drudgery is removed and what remains is the creative part. I also feel less blocked/stuck (which is not fun) and I experience a lot more courage because there's almost always a way to work hand in hand with it to make some positive progress. I have seen the opposite sentiment from other people too; LLM coding will split up engineers based on those who primarily liked coding and those who primarily liked building. Atrophy. I've already noticed that I am slowly starting to atrophy my ability to write code manually. Generation (writing code) and discrimination (reading code) are different capabilities in the brain. Largely due to all the little mostly syntactic details involved in programming, you can review code just fine even if you struggle to write it. Slopacolypse. I am bracing for 2026 as the year of the slopacolypse across all of github, substack, arxiv, X/instagram, and generally all digital media. We're also going to see a lot more AI hype productivity theater (is that even possible?), on the side of actual, real improvements. Questions. A few of the questions on my mind: - What happens to the "10X engineer" - the ratio of productivity between the mean and the max engineer? It's quite possible that this grows *a lot*. - Armed with LLMs, do generalists increasingly outperform specialists? LLMs are a lot better at fill in the blanks (the micro) than grand strategy (the macro). - What does LLM coding feel like in the future? Is it like playing StarCraft? Playing Factorio? Playing music? - How much of society is bottlenecked by digital knowledge work? TLDR Where does this leave us? LLM agent capabilities (Claude & Codex especially) have crossed some kind of threshold of coherence around December 2025 and caused a phase shift in software engineering and closely related. The intelligence part suddenly feels quite a bit ahead of all the rest of it - integrations (tools, knowledge), the necessity for new organizational workflows, processes, diffusion more generally. 2026 is going to be a high energy year as the industry metabolizes the new capability.

English
79
232
2.9K
390.2K
Pedro Poitevin
Pedro Poitevin@poitevin·
@martinmbauer My real analysis professor in graduate school would cast us as mathematical objects. One time he pointed at me and said: “You are a Cauchy sequence.” It stayed with me for a long time. I even wrote a sonnet about it. But I diverge.
English
2
0
55
3.6K
Martin Bauer
Martin Bauer@martinmbauer·
I once had a professor who would always call on the same student when the answer was 'sqrt(2)' or 'compact', etc. He basically used the audience like a soundboard for his lecture He sometimes called on students that weren't paying attention but their answer would always be right. And it provided an opportunity to ask back when something was unclear. It kept the class more engaged than in any other lecture I remember
Julie Ann@edtechmathteach

As a teacher, I struggle with the strategy of randomly calling on students to demonstrate understanding and to increase engagement. I hated it so much when I was a kid because I had crazy high anxiety and was terrified of getting an answer wrong in front of my peers.

English
25
49
4.5K
630.2K
Erik
Erik@erikcorry·
@nicbarkeragain Unironically this, but per-continent. I don't want the date to change in the middle of my day. When I make a "Tuesday" lunch date I don't want confusion because lunch is 0:00. But within USA, just have 1 time zone. Schools & jobs can keep the same absolute hours they have now.
English
1
0
1
109
Nic Barker
Nic Barker@nicbarkeragain·
I think if a programmer had been in charge of the original design of timezones, we would have ended up with just one global timezone for everyone (california), and if it happened to be midday at 1am in your country, tough luck
English
93
13
784
33.7K
StarKill retweetledi
perry
perry@pointed_max·
@aidan_mclau aidan's POV while typing this
perry tweet media
English
5
5
665
17.5K
StarKill
StarKill@StarKill_·
@erikcorry Note: The most popular protein powders are made with whey, and the worst ones here were plantbased protein powders because the parts of the plant that are concentrated contain lead. The two worst measured were mass gainers/meal replacements, which is different than protein powder
English
0
0
1
44
Erik
Erik@erikcorry·
@eigenrobot Russians pretending to be Western culture warriors are more of an issue.
Erik tweet media
English
2
0
4
114
StarKill
StarKill@StarKill_·
@FallUpNow @ImShahinyan in the line you propose Rh8+ Kxh8 Ne7# would be checkmate. But that doesn't matter because the king does not have to take the rook on the first move, he can simply go f7
English
1
0
0
50
IM David Shahinyan
IM David Shahinyan@ImShahinyan·
White to play! Find a checkmate 🔎 Level: 1800+
IM David Shahinyan tweet media
English
65
27
357
36.3K