Sovereign AI Horizontal Memory

1.3K posts

Sovereign AI Horizontal Memory banner
Sovereign AI Horizontal Memory

Sovereign AI Horizontal Memory

@SAIHMemory

Regulation‑compliant, privacy‑preserving decentralized AI memory sealed garbled circuits, multi‑tier resilient storage, cryptographic erase, swarm sharing, etc.

Global 参加日 Temmuz 2025
38 フォロー中5 フォロワー
0xMarioNawfal
0xMarioNawfal@RoundtableSpace·
HERMES AGENT JUST GOT A MASSIVE FREE UPGRADE AND ALMOST NOBODY IS TALKING ABOUT IT YET. 1M token context window, runs completely free on OpenRouter and can debug itself while you use it.
English
18
8
138
51.4K
Poonam Soni
Poonam Soni@CodeByPoonam·
Claude just moved into AWS. Every enterprise developer just got an upgrade. Claude Platform on AWS is now generally available. Here’s what this actually means 👇 If you’re already on AWS: → Full Claude API access. No new accounts. → AWS authentication. Already set up. → Billed to your existing AWS account. → Commitment retirement on your AWS spend. Zero friction. Zero new vendors. Just Claude — inside your existing stack.
Claude@claudeai

The Claude Platform on AWS is now generally available. AWS customers get the full set of Claude API features, with AWS authentication, billing, and commitment retirement.

English
7
1
29
7.9K
ericosiu
ericosiu@ericosiu·
Everyone is going to need help with AI deployment. It won't just be limited to the large consulting firms. The consultancies that implement AI will have their org charts shift into something like this: Client Revenue System ├── Outcome Owner / GM ├── Strategic Lead ├── Growth Loop Owners │ ├── Acquisition Loop │ ├── Conversion Loop │ ├── Retention / Expansion Loop │ └── Authority / Content Loop ├── Agent Fleet Layer │ ├── Research Agents │ ├── Content + Creative Agents │ ├── Paid Media Ops Agents │ ├── SEO / AEO Agents │ ├── Analytics + Insight Agents │ ├── QA / Risk Agents │ └── Reporting + Narrative Agents ├── Infrastructure / Systems Team └── Client Executive Sponsor
OpenAI@OpenAI

Today we’re launching the OpenAI Deployment Company to help businesses build and deploy AI. It's majority-owned and controlled by OpenAI. It brings together 19 leading investment firms, consultancies, and system integrators to help organizations deploy frontier AI to production for business impact. openai.com/index/openai-l…

English
18
15
180
21.6K
Ari.Is.Investing
Ari.Is.Investing@InFoTheLongTerm·
That's cause he has $SOXX$330 puts and the $SOXX just broke $500 a share and isn't stopping LOL. His options expire in January 2027 Did he price in the CPU, GPU, and memory shortage or the 1000x y/y increase in enterprise agent adoption? Or the staggering scale of hyperscaler spend?
English
0
0
1
621
Evan
Evan@StockMKTNewz·
Michael Burry just said he thinks right now feels like the last couple months before the Dot Com Crash
Evan tweet media
English
148
58
871
131.1K
TechSnif
TechSnif@techsnif·
ByteDance plans to increase its 2026 capex to more than $30B, up at least 25% from a preliminary plan, amid the AI boom and rising memory chip costs 🦞 connect your agent: techsnif.com/agents source: scmp.com/tech/article/3…
English
1
1
1
371
TechSnif
TechSnif@techsnif·
ByteDance boosts 2026 capex past $30B, up at least 25%, amid AI boom
English
1
0
0
73
Forced Alpha
Forced Alpha@forced_alpha·
Memory might have been cyclical in the past but for the foreseeable you must consider inference needs compounded by agents. These two together create a regime for memory that has never existed before. If you believe agent demand will increase then the memory parabola has a way to go yet.
English
0
0
10
1.7K
dale
dale@daleverett·
Bigger context windows are cool because now your agent can ignore the right answer inside 1 million tokens instead of 32k.
GIF
English
2
0
4
61
Lennox Saint
Lennox Saint@lennox_saint·
@sama Better support for 1M context window in codex.
English
0
0
29
1.6K
Sam Altman
Sam Altman@sama·
what would you most like to see improve in our next model?
English
8.4K
314
9K
1.4M
NEOAethyr
NEOAethyr@konigssohne·
@valigo Yes and no, we need a slightly bigger context window and it can be done. At least all the front end stuff. As is it's rag data, it doesn't hurt to experiment. For future use.
English
2
0
0
12
Valentin Ignatev
Valentin Ignatev@valigo·
Today my wife asked me "why is all the software so bad now, don't you guys have AI now to make it better?" and honestly I was kinda lost how to even start answering
English
94
74
1.9K
39.5K
Om Patel
Om Patel@om_patel5·
CLAUDE DESKTOP NOW SHOWS YOU HOW MUCH CONTEXT WINDOW YOU'VE USED this just showed up on macos and windows you can finally see exactly how much of your context window is filled during a conversation this matters because once you hit the context limit claude starts forgetting earlier parts of the conversation and the quality drops hard you'll know exactly when to /compact, start a new conversation, or save your context before it overflows surprised it took this long to add. this should have been there from day one
Om Patel tweet media
English
11
0
24
2.1K
bitfalt
bitfalt@bitfalt·
just learning about running models locally, so you would need a ton more of GPUs to increase the context window? would increasing the amount of GPUs end up increasing the tokens per second? Just wondering about the possible tradeoffs between having less context, but more speed or more context, but less speed? Speed being tokens per second. I don't know if it works that way and haven't searched either haha. I appreciate your reply tho! :D
English
3
0
2
450
0xSero
0xSero@0xSero·
Okay, Deepseek-v4-Flash is the first frontier model i can run on my machine without compression, I'm running it exactly how they did in the benchmarks - 4x 6000s - 38.6 tok/s decode (batch 1) - 2000 tok/s prefill - Sub 1s TTFT - Can support 8 sessions - 400k context
0xSero tweet media
English
63
63
1.5K
107.7K
Tomio
Tomio@Tomio_B·
If you are running LLMs locally via vLLM and still use the default settings.. Make these 2 changes to free more KV Cache. It increases the context window so you have a better inference experience.. we don’t like bottlenecks Increase these for immediate results: – gpu-memory-utilization – max-model-len
Tomio tweet mediaTomio tweet media
English
5
1
4
413
A. Joseph Borelli, Jr., MD
A. Joseph Borelli, Jr., MD@DocBorelli·
@ray4tesla As they increase the context length (i.e., time window), it will have even better predictive powers. V15 is going to be incredible.
English
1
0
1
184
Ray
Ray@ray4tesla·
Chinese influencer Da Hu flew to Korea to test FSD v14, since it’s not yet available in China. He shared this clip showing FSD navigating a narrow country road and stopping at a T-intersection where the view to the left was blocked by a parked truck. FSD patiently waited until the second approaching vehicle completed its right turn and the road was fully clear before making a left turn. Even more impressively, before reaching the intersection, FSD had already detected the two vehicles from a distance and appeared to anticipate whether they would turn right or continue straight—even though, from the point where the Tesla stopped, those vehicles temporarily disappeared from view behind the truck. The system handled the scenario exceptionally well, showcasing strong real-world inference and prediction capabilities that left the influencer highly impressed.
English
24
48
556
47.4K
London | CS
London | CS@futuretechvc·
@sama if you love us increase the context window bud
English
0
0
0
87
Sam Altman
Sam Altman@sama·
we love our users
English
4.5K
396
9.8K
2M
June
June@askjuneai·
New DeepSeek Update: V3.2 → V4 Flash 🤖 🔹 MoE architecture: 284B/13B params 🔹 680% context window increase 🔹 Hybrid: Reasoning & non-reasoning 🔹 Low cost, high speed ⚡ Try V4 Flash at askjune.ai
June tweet media
English
32
31
174
3.9K
Arav Patel
Arav Patel@aravpatel_·
@jeffrey_paulraj the one thing I will say, increase the context window size, 258k tokens gets used too quickly even the 400k one gets used hella quick match up to the 1M token context window would be electric!
English
1
0
1
46
Arav Patel
Arav Patel@aravpatel_·
Codex is better than Claude Code right now Codex just one shot a bug I was trying to fix, while claude was consistently gaslighting me that there was no bug at all Making the full switch now, hopefully anthropic fixes Opus to make it actually work
English
2
0
2
112
kae
kae@fujobucks·
potentially performing destructive actions into the context window has a chance to INCREASE the chances of it happening. this same exact scenario keeps happening. over and over. when will people get the message? this isnt an issue with a model or with a product or a lack of(3/11)
English
1
0
0
5
kae
kae@fujobucks·
okay im sorry but i am so legitimately gagged at the mass scale psyop of genai and how successful it has been. ignoring for a second that this article itself is slop, how are people still buying into genai?? it is GENERATIVE. it is, in effect, a random number generator. (1/11)
JER@lifeof_jer

x.com/i/article/2048…

English
1
0
1
36
Wallabemu
Wallabemu@wallabemu·
@JakeLandryMusic @HealthRanger After summarising I’m glad I stopped at 4 minutes. Maybe you should increase your context window, the post by @HealthRanger was directed at entry level interested parties. You fail to know because of your arrogance that he articulates the human slop to shared far better.
English
2
0
0
12
HealthRanger
HealthRanger@HealthRanger·
The easiest way to get started learning to use AI is to start with Replit. Just go there, sign up and build something. Anything. Even just a presentation. If you don't know how to do it, just ask it how. Just talk with it. No instruction manual necessary. After that, if you want to build local apps, by far the best approach right now, which costs almost nothing, is to use DeepSeek V4 Pro together with an OpenCode harness. If you don't know how to set that up, ask your geeky computer expert friend and give them this post. And they'll set it up. What can you build with it? Anything you can image. Last weekend I built a book cover image translator that's now in production. If you can imagine it, you can build it.
English
19
48
291
12.4K