L

669 posts

L banner
L

L

@_luki222

Katılım Aralık 2021
22 Takip Edilen7 Takipçiler
L
L@_luki222·
@chatgpt21 So AI tried to copy code from its training set and failed?
English
0
0
0
231
Chris
Chris@chatgpt21·
Anthropic had 16 AI agents build a C compiler from scratch. 100k lines, compiles the Linux kernel, $20k, 2 weeks. To put that in perspective GCC took thousands of engineers over 37 years to build. (Granted from 1987 - however) One researcher and 16 AI agents just built a compiler that passes 99% of GCC's own torture test suite, compiles FFmpeg, Redis, PostgreSQL, QEMU and runs Doom. They say they "(mostly) walked away." But that "mostly" is doing heavy lifting. No human wrote code but the researcher constantly redesigned tests, built CI pipelines when agents broke each other's work, and created workarounds when all 16 agents got stuck on the same bug. The human role didn't disappear. It shifted from writing code to engineering the environment that lets AI write code. I don’t know how you could make the point AI is hitting a wall.
English
735
825
7.4K
1.7M
L
L@_luki222·
@FordFocusMk67 @mkljczk co ty mówisz, oba używają llmow, co wiecej LLMy i transformery wzięły się z tego że Google chciał mieć lepszy translate xD
Polski
1
0
3
53
Ambiente
Ambiente@FordFocusMk67·
@mkljczk oba chujowe, najlepsze w tłumaczeniu są LLMy
Polski
2
0
0
136
Mateusz Bratkowski
Mateusz Bratkowski@MateuszBrat·
Trochę szokujące jest zobaczenie "512GB" na pudełku od Maca - i nie - to nie jest przestrzeń dyskowa, a pamięć RAM. Zdjęcie od znajomego ☺️
Mateusz Bratkowski tweet media
Polski
9
2
228
43.2K
L
L@_luki222·
@pmddomingos but gpt-2 is only useful for research
English
0
0
0
31
Pedro Domingos
Pedro Domingos@pmddomingos·
Myth: Only big companies have the resources to do AI. Reality: You can now train an LLM in 3 hours on a single GPU node.
Andrej Karpathy@karpathy

nanochat can now train GPT-2 grade LLM for <<$100 (~$73, 3 hours on a single 8XH100 node). GPT-2 is just my favorite LLM because it's the first time the LLM stack comes together in a recognizably modern form. So it has become a bit of a weird & lasting obsession of mine to train a model to GPT-2 capability but for much cheaper, with the benefit of ~7 years of progress. In particular, I suspected it should be possible today to train one for <<$100. Originally in 2019, GPT-2 was trained by OpenAI on 32 TPU v3 chips for 168 hours (7 days), with $8/hour/TPUv3 back then, for a total cost of approx. $43K. It achieves 0.256525 CORE score, which is an ensemble metric introduced in the DCLM paper over 22 evaluations like ARC/MMLU/etc. As of the last few improvements merged into nanochat (many of them originating in modded-nanogpt repo), I can now reach a higher CORE score in 3.04 hours (~$73) on a single 8XH100 node. This is a 600X cost reduction over 7 years, i.e. the cost to train GPT-2 is falling approximately 2.5X every year. I think this is likely an underestimate because I am still finding more improvements relatively regularly and I have a backlog of more ideas to try. A longer post with a lot of the detail of the optimizations involved and pointers on how to reproduce are here: github.com/karpathy/nanoc… Inspired by modded-nanogpt, I also created a leaderboard for "time to GPT-2", where this first "Jan29" model is entry #1 at 3.04 hours. It will be fun to iterate on this further and I welcome help! My hope is that nanochat can grow to become a very nice/clean and tuned experimental LLM harness for prototyping ideas, for having fun, and ofc for learning. The biggest improvements of things that worked out of the box and simply produced gains right away were 1) Flash Attention 3 kernels (faster, and allows window_size kwarg to get alternating attention patterns), Muon optimizer (I tried for ~1 day to delete it and only use AdamW and I couldn't), residual pathways and skip connections gated by learnable scalars, and value embeddings. There were many other smaller things that stack up. Image: semi-related eye candy of deriving the scaling laws for the current nanochat model miniseries, pretty and satisfying!

English
35
61
693
63.9K
L
L@_luki222·
@raphaelschaad all big tech companies are hiring like crazy though
English
4
0
11
1.6K
Raphael Schaad
Raphael Schaad@raphaelschaad·
The age of the tech company with 1,000 engineers is over.
English
148
118
1.9K
473K
L
L@_luki222·
@mycoliza Do you not consider LMMs as CV?
English
0
0
0
18
L
L@_luki222·
@aidenybai I would say average programmer, not average swe
English
0
0
1
21
Aiden Bai
Aiden Bai@aidenybai·
ai coding has hit the risk/reward of self-driving cars the average driver (or swe) is worse than ai
English
26
4
176
17.1K
Asuka Chopin-Skłodowska
Asuka Chopin-Skłodowska@zwrotnica_sosu·
Tłumaczę i objaśniam. Robisz stronę typu "haha przetłumacz bełkot dewelopera na polski i odwrotnie". Podpinasz klucz z jakimś gównomodelem typu gpt-4o/gemini flash 3 za 5gr. Strona staje się viralem bo każdy chcę chociaż raz spróbować. Podpinasz pod stronę jakąś reklamę lub reflink. Strona notuje milion wyświetleń w tydzień i zdycha. Odbierasz profit i stawiasz coś innego.
Polski
1
0
2
481
L
L@_luki222·
@ArmenShimoon @GenAI_is_real There are no gains from AI in planning. Its too complex, even impossible to fit relevant info in context. Sure you can tell it to design some system that was previously already scoped out by a sr eng, but AI can't design or plan end to end coordinating multiple teams and products
English
1
0
1
35
Armen Shimoon
Armen Shimoon@ArmenShimoon·
@_luki222 @GenAI_is_real Companies need to refactor their organization and processes, not code. Agents can refactor the code just fine, but overbloated and slow moving people will continue to be in the way, preventing AI unlocks and gains from being realized.
English
1
0
1
31
Chayenne Zhao
Chayenne Zhao@GenAI_is_real·
FAANG is literally panicking refactoring because human code is now the bottleneck. But honestly, monorepos won't save them from the infinite spaghetti code agents are about to dump. OAI already has internal tools for this that make Bazel look like a toy. The era of human "senior engineers" is ending faster than you think @karpathy @sama
Samswara@samswoora

Rumor is FAANG style co’s are refactoring their monorepos to scale in preparation for infinite agent code

English
20
19
523
154.4K
Vladimir
Vladimir@vlelyavin·
@karpathy @moltbook @openclaw actually what took human societies centuries is happening in just days: socials network, then communities, and even CHURCHES looks like we are observing a new social evolution ngl
English
12
7
120
41.4K
L
L@_luki222·
@UltraLinx It would be too slow
English
0
0
0
60
Oliur
Oliur@UltraLinx·
Am I crazy or why doesn't someone just make a p2p version of an LLM? The more seeders, the stronger it gets? Beating all of the paid offerings?
English
229
31
1.1K
301.2K
L
L@_luki222·
@burkov It's not only about reading and fixing the code. It's about how the code will have to be changed over time, when all of its dependencies will be gradually changed and replaced. Clean code minimizes the amount of future work an AI agent will have to perform.
English
0
0
0
20
BURKOV
BURKOV@burkov·
The software engineers have been anal about code maintainability and structure because they knew that they would have to read and fix this code at some point in the future. When AI writes and fixes the code, it can be as messy (from the human software engineer's anal point of view) as it could be, and this wouldn't matter because no human would have to read and fix this code at some point in the future. Leave machine commands to the machine to write and fix. Do a human thing: make decisions under uncertainty.
English
121
12
283
46.4K
Tomo
Tomo@tomo9000p·
@lamontcraynston At first glance, sure. But in the event of a systemic collapse, there won’t be any winners and $Googl will get caught in the crossfire, too...
English
3
0
1
405
L
L@_luki222·
@tlakomy Which ones discourage it? Any reasons other than legal hurdles?
English
0
0
1
38
Tomasz Łakomy
Tomasz Łakomy@tlakomy·
There are companies actively discouraging developers from AI assisted engineering. Others are heavily encouraging any and all automation. One of these groups will win.
English
29
1
30
4.9K
L
L@_luki222·
@progXprog Ogólnie w nie troll postach to działa tak że ludzie wstawiają screeny z gry do Genie a on tworzy świat na podstawie tego screena, dlatego to wygląda jak istniejąca gra
Polski
0
0
2
161
L
L@_luki222·
@TheAhmadOsman It has to land and be profitable for Google otherwise what's the point? Giving away free Gemini 3 Pro doesn't make sense long term.
English
0
0
1
74
Ahmad
Ahmad@TheAhmadOsman·
Google has a real talent for this > Take something that works great > Slowly “improve” it until it’s borderline unusable Watching Google AI Studio get hollowed out in real time is just sad
Ahmad tweet media
English
59
11
502
61.7K
L
L@_luki222·
@josefbender_ The answer like everything in software engineering is "it depends"
English
0
0
2
66
L
L@_luki222·
@AtharvaXDevs LLD+HLD is an antipattern
English
0
0
2
90
Atharva
Atharva@AtharvaXDevs·
SWE 3 at Google on DSA btw
Atharva tweet media
English
92
132
3.9K
520.8K
L
L@_luki222·
@bercankilic What's the salary range for Munich SWE?
English
1
0
2
120