Luís Marnoto

528 posts

Luís Marnoto banner
Luís Marnoto

Luís Marnoto

@luispl77

Computer engineer

Lisbon, Portugal Katılım Ağustos 2015
318 Takip Edilen69 Takipçiler
Luís Marnoto
Luís Marnoto@luispl77·
I believe in myself like others believe in their Gods
English
0
0
0
9
Luís Marnoto retweetledi
Tibo
Tibo@thsottiaux·
Do people like this? We don't do this for codex because it exists to help you and it's important that you remain the owner and accountable for your work without AI taking credit. At the same time it does mean that you can't trace how popular codex is among repos.
Yuchen Jin@Yuchenj_UW

I noticed something interesting: Claude Code auto-adds itself as a co-author on every git commit. Codex doesn’t. That’s why you see Claude everywhere on GitHub, but not Codex. I wonder why OpenAI is not doing that. Feels like an obvious branding strategy OpenAI is skipping.

English
593
76
2.6K
360K
Luís Marnoto retweetledi
Fuentes Updates
Fuentes Updates@FuentesUpdates·
Nick Fuentes stresses the importance of THINKING, rather than being concerned with reading a ton of books "It's important to be able to think systematically... ponder. And people are always so caught up on reading, and that's so wrong."
English
97
336
4.5K
177.1K
Luís Marnoto retweetledi
Novak Djokovic
Novak Djokovic@DjokerNole·
Lost for words 🇦🇺
Novak Djokovic tweet media
English
12.1K
24.9K
244.5K
5.6M
Luís Marnoto retweetledi
Alex Hormozi
Alex Hormozi@AlexHormozi·
Don’t stop trying because it didn’t work. It never works the first time.
English
629
752
7.4K
159.6K
Luís Marnoto retweetledi
Jaynit
Jaynit@jaynitx·
This guy literally explained why you are "way smarter" than you realize:
English
99
2.1K
11.7K
193.5K
Luís Marnoto retweetledi
Charlie Kirk
Charlie Kirk@charliekirk11·
Congratulations to Novak Djokovic for winning the Australian Open! This year he wins the trophy and will become the No. 1 men’s tennis player in the world. Last year the Australian government deported him for not being vaccinated. Freedom is always worth fighting for.
English
111
692
6.2K
209.2K
Luís Marnoto
Luís Marnoto@luispl77·
@flowersslop The stagnation is incredible. The "big" inovation this year was a simple MODEL ROUTER. Deep Learning winter begins today
English
0
0
3
438
Flowers ☾
Flowers ☾@flowersslop·
I dont wanna read anything about exponential progress anymore.
English
54
31
878
39.7K
Luís Marnoto
Luís Marnoto@luispl77·
GPT-5 will be a disappointment. It's just a model router, which will help it finally, at most, match Claude 4 Sonnet. Today marks the day everyone realises scaling the vanilla transformer is not going to cut it. Missing Continuous Learning, Long Term Memory, Infinite Context.
English
0
0
1
95
Luís Marnoto
Luís Marnoto@luispl77·
@mark_k The GPT-5 architecture will win, because they have multiple models all built into one. And this includes one or more Specialized Coding Model, INSIDE GPT-5. One for web dev, one for backend. etc. Sonnet isn't a specialized coding model, and that's why it will fall short.
English
0
0
1
99
Mark Kretschmann
Mark Kretschmann@mark_k·
GPT-5 will easily beat Claude 4 with coding, and if the pricing is ok the good times will be over for Anthropic.
English
33
16
299
16.7K
Làzaro
Làzaro@lazaro45ive·
You never really finish learning 3D. At some point, you just get comfortable being lost in the unknown #b3d
English
37
211
3.4K
181.2K
Jesse Mu
Jesse Mu@jayelmnop·
I recently moved to the Code RL team at Anthropic, and it’s been a wild and insanely fun ride. Join us! We are singularly focused on solving SWE. No 3000 elo leetcode, competition math, or smart devices. We want Claude n to build Claude n+1, so we can go home and knit sweaters.
Anthropic@AnthropicAI

Introducing the next generation: Claude Opus 4 and Claude Sonnet 4. Claude Opus 4 is our most powerful model yet, and the world’s best coding model. Claude Sonnet 4 is a significant upgrade from its predecessor, delivering superior coding and reasoning.

English
44
21
714
157K
Luís Marnoto
Luís Marnoto@luispl77·
@karpathy It's called continuous learning and long term memory, and you are correct
English
0
0
0
4
Andrej Karpathy
Andrej Karpathy@karpathy·
We're missing (at least one) major paradigm for LLM learning. Not sure what to call it, possibly it has a name - system prompt learning? Pretraining is for knowledge. Finetuning (SL/RL) is for habitual behavior. Both of these involve a change in parameters but a lot of human learning feels more like a change in system prompt. You encounter a problem, figure something out, then "remember" something in fairly explicit terms for the next time. E.g. "It seems when I encounter this and that kind of a problem, I should try this and that kind of an approach/solution". It feels more like taking notes for yourself, i.e. something like the "Memory" feature but not to store per-user random facts, but general/global problem solving knowledge and strategies. LLMs are quite literally like the guy in Memento, except we haven't given them their scratchpad yet. Note that this paradigm is also significantly more powerful and data efficient because a knowledge-guided "review" stage is a significantly higher dimensional feedback channel than a reward scaler. I was prompted to jot down this shower of thoughts after reading through Claude's system prompt, which currently seems to be around 17,000 words, specifying not just basic behavior style/preferences (e.g. refuse various requests related to song lyrics) but also a large amount of general problem solving strategies, e.g.: "If Claude is asked to count words, letters, and characters, it thinks step by step before answering the person. It explicitly counts the words, letters, or characters by assigning a number to each. It only answers the person once it has performed this explicit counting step." This is to help Claude solve 'r' in strawberry etc. Imo this is not the kind of problem solving knowledge that should be baked into weights via Reinforcement Learning, or least not immediately/exclusively. And it certainly shouldn't come from human engineers writing system prompts by hand. It should come from System Prompt learning, which resembles RL in the setup, with the exception of the learning algorithm (edits vs gradient descent). A large section of the LLM system prompt could be written via system prompt learning, it would look a bit like the LLM writing a book for itself on how to solve problems. If this works it would be a new/powerful learning paradigm. With a lot of details left to figure out (how do the edits work? can/should you learn the edit system? how do you gradually move knowledge from the explicit system text to habitual weights, as humans seem to do? etc.).
English
716
1K
10.4K
1.5M
Luís Marnoto
Luís Marnoto@luispl77·
@Citrini7 This is false. If you look at equivalent open-source models such as DeepSeek R1, the electricity costs are nowhere near the API pricing
English
0
0
0
13
Citrini
Citrini@citrini·
people don’t realize how heavily subsidized AI is right now your o3 prompts probably cost more in electricity alone than your subscription fee remember when uber was awesome and cheaper than a taxi? yeah, that.
English
464
827
16.8K
967K
Luís Marnoto
Luís Marnoto@luispl77·
@ns123abc Show an aerial image of a bunch of houses in a neighborhood. Tell it to count the houses. 100% fail rate
English
0
0
1
70
NIK
NIK@ns123abc·
show me a problem that you think LLMs can’t solve and I'll correctly prompt show you why you're wrong
English
579
59
2.2K
405.2K
Luís Marnoto
Luís Marnoto@luispl77·
@kalomaze Even if the goal is balance, if you don't switch domains its expected that some experts are not used. The model was trained to balance experts on trillions of tokens, yours is only 120k tokens
English
0
0
3
416
kalomaze
kalomaze@kalomaze·
the QwenMoE router distributions are... VERY biased even the 30b MoE seems quite prunable
kalomaze tweet media
English
40
19
365
80.2K
Louis Arge
Louis Arge@louisvarge·
i would probably pay like $500/mo for o3 on @GroqInc
English
5
7
81
89.6K
Luís Marnoto
Luís Marnoto@luispl77·
@shaoruu @cursor_ai A way to never have to start a new conversation. Long-term memory. That's where the money is at, infinite and efficient context. Too bad it might not depend on you, but the models themselves.
English
0
0
0
12
ian
ian@shaoruu·
what should i add to @cursor_ai that would make u 10x more productive? feature requests, tiny nits, anything :)
English
1.4K
65
2.7K
884.8K
Luís Marnoto
Luís Marnoto@luispl77·
@IterIntellectus IKR, what’s with these random 100k likes posts that make no sense? They’re not even funny or interesting
English
0
0
4
373
vittorio
vittorio@IterIntellectus·
what's wrong with people?
vittorio tweet media
English
11
1
70
11.4K