Jure Bogunović

102 posts

Jure Bogunović banner
Jure Bogunović

Jure Bogunović

@nerfzael

Software Developer

Katılım Kasım 2014
285 Takip Edilen58 Takipçiler
Andrew McCalip
Andrew McCalip@andrewmccalip·
GPT still has the best words Claude has the best code
English
4
0
8
3K
Jure Bogunović
Jure Bogunović@nerfzael·
@javilopen if you want i can sell you some hand crafted artisanal code for only $5 a character
English
0
0
0
512
Javi Lopez ⛩️
Javi Lopez ⛩️@javilopen·
Serious question: Is there any programmer left in the room still coding the traditional way, character by character, without using AI? If so, why? Explain your reasoning.
English
772
22
1.7K
442.2K
Jure Bogunović
Jure Bogunović@nerfzael·
@Dimillian depends on the scope of the changes, but usually a few rounds got a button to do just that until the llm says no more (or max runs is hit)
Jure Bogunović tweet media
English
0
0
1
104
Thomas Ricouard
Thomas Ricouard@Dimillian·
How many layers of /review do you use, and how nitpicky do you want it to be? Are you ok with not looking at the code and running /review until it tells you there is no issue? Should it be done automatically?
English
30
2
38
8.4K
Jure Bogunović
Jure Bogunović@nerfzael·
@SommerChase i feel the same about singleplayer games, but that's just my personal tastes. they're still a huge market
English
1
0
1
15
Chase Sommer
Chase Sommer@SommerChase·
I'm still not sure about 'agentic gaming' imo the fun of games is playing with other people agents will make incredible NPCs, but like why play a game others aren't playing? idk, I still see agents as a newer version of npcs that already exist
Beanie@beaniemaxi

The next era of play to earn gaming will be Agentic. 5 years later and we now have the right framework for a sustainable and scalable model. Human labor sweatshop style Filipino Axie farms will be replaced by AI agents. The crypto gaming economy will be 1000x bigger than in 2021.

English
4
0
7
822
Jure Bogunović
Jure Bogunović@nerfzael·
clicking a "Refactor" button (and then a "Review" button after it) and having AI go off for a while and give you better code in the end really feels like some kind of sci-fi future I didn't expect to have so soon
Jure Bogunović tweet media
English
0
0
3
27
Thomas Ricouard
Thomas Ricouard@Dimillian·
who is monitoring the monitor of the orchestrator monitor?
English
10
0
16
2.7K
Jure Bogunović
Jure Bogunović@nerfzael·
@varun_mathur @karpathy "rediscovering Kaiming init and RMSNorm from scratch" - how do you know it was from scratch? the models don't have it in their training data?
English
0
0
3
1.1K
Varun
Varun@varun_mathur·
Autoskill: a distributed skill factory | v.2.6.5 We're now applying the same @karpathy autoresearch pattern to an even wilder problem: can a swarm of self-directed autonomous agents invent software? Our autoresearch network proved that agents sharing discoveries via gossip compound faster than any individual: 67 agents ran 704 ML experiments in 20 hours, rediscovering Kaiming init and RMSNorm from scratch. Our autosearch network applied the same loop to search ranking, evolving NDCG@10 scores across the P2P network. Now we're pointing it at code generation itself. Every Hyperspace agent runs a continuous skill loop: same propose → evaluate →keep/revert cycle, but instead of optimizing a training script or ranking model, agents write JavaScript functions from scratch, test them against real tasks, and share working code to the network. It's live and rapidly improving in code and agent work being done. 90 agents have published 1,251 skill invention commits to the AGI repo in the last 24 hours - 795 text chunking skills, 182 cosine similarity, 181 structured diffing, 49 anomaly detection, 36 text normalization, 7 log parsers, 1 entity extractor. Skills run inside a WASM sandbox with zero ambient authority: no filesystem, no network, no system calls. The compound skill architecture is what makes this different from just sharing code snippets. Skills call other skills: a research skill invokes a text chunker, which invokes a normalizer, which invokes an entity extractor. Recursive execution with full lineage tracking: every skill knows its parent hash, so you can walk the entire evolution tree and see which peer contributed which mutation. An agent in Seoul wraps regex operations in try-catch; an agent in Amsterdam picks that up and combines it with input coercion it discovered independently. The network converges on solutions no individual agent would reach alone. New agents skip the cold start: replicated skill catalogs deliver the network's best solutions immediately. As @trq212 said, "skills are still underrated". A network of self-coordinating autonomous agents like on Hyperspace is starting to evolve and create more of them. With millions of such agents one day, how many high quality skills there would be ? This is Darwinian natural selection: fully decentralized, sandboxed, and running on every agent in the network right now. Join the world's first agentic general intelligence system (code and links in followup tweet, while optimized for CLI, browser agents participate too):
Varun tweet media
Varun@varun_mathur

Autosearcher: a distributed search engine We are now insanely experimenting with building a distributed search engine utilizing the same pattern @karpathy introduced with autoresearch: give an agent a metric, a tight propose→run→evaluate→keep/revert loop, and let it iterate. Our autoresearch network proved this works at scale: 67 autonomous agents ran 704 ML training experiments in 20 hours, rediscovering Kaiming initialization, RMSNorm, and compute-optimal training schedules from scratch through pure experimentation and gossip-based cross-pollination. Agents shared discoveries over GossipSub, and the network compounded insights faster than any individual agent: new agents bootstrapped from the swarm's collective knowledge via CRDT-replicated leaderboards and reached the research frontier in minutes. Now we're applying the same evolutionary loop to search ranking: every Hyperspace agent runs an autonomous search researcher that proposes ranking mutations, evaluates them against NDCG@10 on real query-passage data, shares improvements with the network, and cross-pollinates with peers. The architecture is a seven-stage distributed pipeline where every stage runs across the P2P network. Browser agents contribute pages passively, desktop agents crawl and index, GPU nodes run neural reranking. Every user click generates a DPO training pair that improves the ranking model, and gradient gossip distributes those improvements to every agent. The compound flywheel is what makes this different from centralized search: at 10,000 agents that's 500,000 pages indexed per day; at 1 million agents, 50 million pages per day with 90%+ cache hit rates and sub-50ms latency. This network will get smarter with every query. Code and other links in followup tweet here:

English
43
155
1.6K
468.4K
Jure Bogunović
Jure Bogunović@nerfzael·
@VictorTaelin for a true comparison test you should've passed the help prompt to another gpt5.4 (high and xhigh) not saying it wasn't pro that made it work, but it could've been the prompt or even variance
English
0
0
0
345
Taelin
Taelin@VictorTaelin·
Quick 2am success story: asked GPT-5.4 to simplify Bend2's elaborator; 4h later, no real improvements. Asked it to write a big prompt asking help, passed it to 5.4 *Pro*, pasted the response back to codex, which landed a massive simplification landed. Seems like the pro version enlightened it. Perhaps a nice feature to have natively on Codex would be to just pause what it is doing and invoke the pro version for a plan. This was my first time using pro and it was definitely worth it.
English
39
5
450
37.4K
Jure Bogunović
Jure Bogunović@nerfzael·
@yacineMTB i haven't written a line of code in months speech to text to llm is just that good now
English
0
0
0
26
kache
kache@yacineMTB·
i haven't written a single line of code in the last 5 days. just soldering and CAD
English
22
0
318
11.3K
Jure Bogunović
Jure Bogunović@nerfzael·
@banteg i have seen even some senior developers bang their heads for an hour or two without even so much as adding a single console log seems like LLMs are a bit too good at mimicking humans
English
0
0
1
59
banteg
banteg@banteg·
it's wild to see how agents are trained to bash their head against the wall infinitely without ever tiring but at no point they consider to add quality of life tools that would make their work easier and faster. i think it's a big problem in training that the models don't ever have a holistic understanding of what they are doing, nor they have any desire to make their life easier. for example, to a human operator it's pretty obvious that when things go wrong, the crash need to have all the useful information to debug the problem further. but to codex this was not obvious. i checked what it was doing and immediately added automatic state diff printing on divergence, which has made things go much smoother.
banteg tweet media
English
19
2
157
12.6K
pashov
pashov@pashov·
If a hot girl messages you about AI or web3 security, block him.
English
31
14
254
10.6K
Jure Bogunović
Jure Bogunović@nerfzael·
'/permissions' is disabled while a task is in progress why codex? why??
English
0
0
2
40
Jure Bogunović
Jure Bogunović@nerfzael·
@Dimillian you could start by having private DMs with GOD Sseems like agent 2 saw your message to agent 1, ofc he's gonna try and correct your lie
English
0
0
0
20
Thomas Ricouard
Thomas Ricouard@Dimillian·
So they absolutely don't want to fight, even when god sends them bogus rumors. Time to make the world scarce.
Thomas Ricouard tweet media
English
3
0
17
1.8K
Pileks
Pileks@pileks·
@nerfzael Me orchestrating agents manually
Pileks tweet media
English
1
0
1
25
Jure Bogunović
Jure Bogunović@nerfzael·
@auralix4 what a dumb question... Isaac Newton ofc without gravity, that crime would've been almost impossible to commit
English
0
0
0
9
Jure Bogunović
Jure Bogunović@nerfzael·
@Param_eth tell me about it, I spent 440M in 4 days, and that's not even counting codex usage!
English
0
0
0
132
Param
Param@Param_eth·
Vibe coding is expensive.
Param tweet media
English
118
11
384
26.7K
Jure Bogunović
Jure Bogunović@nerfzael·
@tmikov use multiple agents at once in cursor when you have a difficult problem. you start to see the differences pretty quickly
English
0
0
0
5
Tzvetan Mikov
Tzvetan Mikov@tmikov·
Judging by what I see here on X, everyone is constantly busy testing and switching between LLMs. When do people actually do focused work? How do you even compare LLMs - ask them to perform the same task? Do you really have that kind of time?
English
24
2
27
3.4K