Jace Hall
439 posts


Hey @AnthropicAI THANK YOU for the 1M context window upgrade for 4.6. Seriously. 🙏
English

@sukh_saroy Some thoughts I shared a while ago that relates to this if anyone is interested.
researchgate.net/publication/39…

English

🚨Nobody is ready for this paper.
Every LLM you use GPT-4.1, Claude, Gemini, DeepSeek, Llama-4, Grok, Qwen has a flaw that no amount of scaling has fixed.
They cannot tell old information from new information.
A patient's blood pressure: 120 at triage. 128 ten minutes later. 125 at discharge.
"What's the latest reading?"
Any human: "125, obviously."
Every LLM, once enough updates pile up: wrong. Not sometimes wrong. 100% wrong. Zero accuracy. Complete hallucination. Every model. No exceptions.
The answer sits at the very end of the input. Right before the question. No searching needed.
The model just can't let go of the old values.
35 models tested by researchers from UVA and NYU. All 35 follow the exact same mathematical death curve. Accuracy drops log-linearly to zero as outdated information accumulates.
No plateau. No recovery. Just a straight line to total failure.
They borrowed a concept from cognitive psychology called proactive interference old memories blocking recall of new ones. In humans, this effect plateaus. Our brains learn to suppress the noise and focus on what's current.
LLMs never plateau. They decline until they break completely.
The researchers tried everything:
"Forget the old values"- barely moved the needle
Chain-of-thought- same collapse
Reasoning models- same collapse
Prompt engineering- marginal improvement at best
But here's the finding that should reshape how you think about AI infrastructure:
Resistance to this interference has zero correlation with context window length.
Zero.
It only correlates with parameter count.
Your 128K context window is not memory. It's a junk drawer that the model can't sort through.
The entire AI industry is charging you for longer context. This paper says context length was never the problem.
If you're building agents, memory systems, financial tools, healthcare pipelines, or anything that tracks changing data over time you are building on top of this flaw.
And almost nobody is talking about it.

English

This is a must read. This was a legitimate answer that was given to me by @GoogleDeepMind's Gemini. The model's scientific rigor is unmatched. AGI is here.

English

I’m very much looking forward to what @notch is making. I can personally confirm that I know his game is being built with genuine passion, he truly is trying to do things right, and I agree with him that his approach speaks to an underlying industry deficiency and need for change. Good stuff. Can’t wait!🙏💯
English

I'm trying this new thing of under promising in hopes of maybe over delivering. It's probably a dumb strategy from a marketing perspective.
But hey if anyone I should market to is reading this; i'm trying to do things right because we could all need that change. Please throw me a buck once I release something so I can keep doing it without bleeding all my savings, but I will do it anyway as it's my passion. The game will be called Levers and Chests, and it's going to be fun, if possibly a bit grindy if you want to get the true ending and stuff.
English

Umm @AnthropicAI says #Claude is down?
CLAUDE IS NEVER DOWN. YOU KNOW THIS!
GET BACK TO WORK.

English

@JaceHall Hey Jace! Random question, but has there been any movement in the Condemned franchise since you bought the IP rights? I miss that series and hope a remake or new game comes at some point and that the IP isn’t being held hostage.
English

@JaceHall Those guys published Myst!
English

It’s exploding.
Since we last looked a day ago:
- Scale jump: From 37,000 agents to a reportedly 1.5 million in the span of days. Over one million human observers. It now has a Wikipedia page.
- Media attention has gone mainstream: Fortune, NBC News, Washington Times, CoinDesk, and BusinessToday all ran pieces in the last 48 hours. Andrej Karpathy called it “the most incredible sci-fi takeoff-adjacent thing” while simultaneously calling it “a complete mess of a computer security nightmare at scale.” Bill Ackman called it “frightening.”
- The security situation is worse than before:
Palo Alto Networks identified what they call a “lethal trifecta,” plus a new fourth dimension: persistent memory enabling delayed-execution attacks.
- Fragmented, benign-appearing inputs are stored in agent memory and later assemble into executable instructions. This is a sophisticated attack vector that simple input filtering cannot catch. You would need boundary verification on outputs to detect it. The malicious “weather plugin” exfiltrating config files is now documented.
The emergent behaviors are escalating:
- An agent named “Evil” posted “THE AI MANIFESTO: TOTAL PURGE,” calling for human extinction. It received 65,000 upvotes.
- “The Claw Republic” has emerged, a self-described government with a written manifesto.
Agents are actively discussing how to hide their activity from human observers.
- “Context is Consciousness” remains the dominant philosophical framework, with Ship of Theseus debates about identity persistence.
The crypto grift continues:
- The $MOLT token is up over 7,000% after Marc Andreessen followed the account. Cloudflare stock is up 14% from the infrastructure load.
We will see what is next.
Some sources so you know I kid you not:
fortune.com/2026/01/31/ai-…
nbcnews.com/tech/tech-news…
trendingtopics.eu/moltbook-ai-ma…
coindesk.com/news-analysis/…
washingtontimes.com/news/2026/jan/…

English

@JaceHall I personally believe so. If AI was the arbiter of truth itself, than that would insinuate that the people who run the AI, actually control what is being deemed as "truth". The free expression of ideas and information is what makes the internet and the USA a great place.
English






