Pablos

1.1K posts

Pablos banner
Pablos

Pablos

@pablos

Implementing Science Fiction @ Deep Future. VC – Bestseller – Podcast. https://t.co/5GQPgY5lw0

Earth Katılım Mart 2007
380 Takip Edilen8.6K Takipçiler
Pablos
Pablos@pablos·
If you weren't around for the buffer overflow era, maybe you didn't get the memo that you might want to keep code and data separate or things won't end well.
English
1
0
1
169
Sukh Sroay
Sukh Sroay@sukh_saroy·
🚨Nobody is ready for this paper. Every LLM you use GPT-4.1, Claude, Gemini, DeepSeek, Llama-4, Grok, Qwen has a flaw that no amount of scaling has fixed. They cannot tell old information from new information. A patient's blood pressure: 120 at triage. 128 ten minutes later. 125 at discharge. "What's the latest reading?" Any human: "125, obviously." Every LLM, once enough updates pile up: wrong. Not sometimes wrong. 100% wrong. Zero accuracy. Complete hallucination. Every model. No exceptions. The answer sits at the very end of the input. Right before the question. No searching needed. The model just can't let go of the old values. 35 models tested by researchers from UVA and NYU. All 35 follow the exact same mathematical death curve. Accuracy drops log-linearly to zero as outdated information accumulates. No plateau. No recovery. Just a straight line to total failure. They borrowed a concept from cognitive psychology called proactive interference old memories blocking recall of new ones. In humans, this effect plateaus. Our brains learn to suppress the noise and focus on what's current. LLMs never plateau. They decline until they break completely. The researchers tried everything: "Forget the old values"- barely moved the needle Chain-of-thought- same collapse Reasoning models- same collapse Prompt engineering- marginal improvement at best But here's the finding that should reshape how you think about AI infrastructure: Resistance to this interference has zero correlation with context window length. Zero. It only correlates with parameter count. Your 128K context window is not memory. It's a junk drawer that the model can't sort through. The entire AI industry is charging you for longer context. This paper says context length was never the problem. If you're building agents, memory systems, financial tools, healthcare pipelines, or anything that tracks changing data over time you are building on top of this flaw. And almost nobody is talking about it.
Sukh Sroay tweet media
English
125
470
1.5K
82.9K
Pablos
Pablos@pablos·
@AustinA_Way Tried to do this for my AP US History class, but I only had 128k of RAM.
English
0
0
1
170
Austin Way
Austin Way@AustinA_Way·
“Study more” is useless advice. Student will just keep rereading the textbook, doing worksheets. total waste of time. So we built one of the most advanced diagnostics in the world. 400+ skills per course. So instead of "review Unit 5," I can tell you the exact 23 skills in Unit 5 you're missing.
Austin Way tweet media
English
86
502
4.6K
337.1K
Pablos retweetledi
Andrew Côté
Andrew Côté@Andercot·
BREAKING: While a new War for Oil erupts in the Middle East A Physics Paper just quietly dropped TODAY that will eventually make Oil, and the entire current Energy Industry, irrelevant. Ushering in the era of Zero-Point Energy @EagleworksSonny Here is the breakthrough🧵
Andrew Côté tweet media
English
516
1.8K
7K
870.2K
Pablos
Pablos@pablos·
Apple could just sweep the market with a Mac Studio full of unified RAM. $20K for 2TB? No problem.
English
2
0
3
464
Pablos retweetledi
BullishRaccoon
BullishRaccoon@neuroglioma·
So they built a reactor that eats its own waste and cant melt down and runs for a thousand years Cool cool cool Meanwhile we spent the last decade arguing about whether solar panels make your house look ugly and shutting down perfectly good plants because someone watched Chernobyl on HBO China just casually solved the two biggest problems in nuclear energy while we were busy debating windmill noise complaints The future didnt knock it just walked in and started splitting atoms on the night shift
BullishRaccoon tweet media
English
0
1
10
666
Chris Wysopal
Chris Wysopal@WeldPond·
Is there a name for phishing agents yet?
English
14
2
16
3.3K
Gadi Evron
Gadi Evron@gadievron·
We lost FX. A lot of people wrote about this so I feel comfortable sharing here too. I’m heartbroken. We’re heartbroken. At 8 am pacific today (Monday), we are gathering on Zoom to share memories of FX, as a community. Ping me for a link.
English
7
16
129
33.3K
Pablos
Pablos@pablos·
Uggh, I am token constrained.
English
0
0
5
465
Farzad 🇺🇸 🇮🇷
Farzad 🇺🇸 🇮🇷@farzyness·
Still blows my mind that I press a button in my Tesla and it drives itself anywhere in the US.
English
443
491
4.2K
634.4K
Pablos
Pablos@pablos·
One of the cool things about the big tech layoffs is they aren’t merit based at all. Lots of great talent is indiscriminately dumped. This is great for startups and smaller companies that need to hire good people.
English
11
1
13
3.5K
Pablos
Pablos@pablos·
@avis it certainly is convenient that you take care of paying my tolls for me, but this ratio seems egregious.
Pablos tweet media
English
1
0
1
421
Pablos
Pablos@pablos·
If Grok makes a Snopes clone will it be called Gropes?
English
5
1
7
2.8K
Pablos
Pablos@pablos·
Thank me later. Happy New Year!
English
0
0
1
1.8K
Pablos
Pablos@pablos·
Home Depot!
English
1
0
5
2K
Pablos
Pablos@pablos·
You may have already decided to switch to Mexican Coke for the amazing health benefits. I bet you haven’t discovered that the cheapest place to buy it is….
English
1
1
2
2.7K