Pascal2_22./

1.2K posts

Pascal2_22./ banner
Pascal2_22./

Pascal2_22./

@Pascal2_22

Accelerating @Gradient_HQ Advancing OIS stack: ./ Open AI for Sovereign Future. Make it you against the matrix

Accra, Ghana Joined Ağustos 2023
266 Following259 Followers
Goku ./
Goku ./@gokuoldskool·
DeepSeek V4 flash is now available on @commonstack_ai DeepSeek V4 Flash matches GPT-5.4 (High) performance but at a fraction of the cost: > Input: $0.14/M vs $30/M (214x cheaper) > Output: $0.28/M vs $180/M (643x cheaper) Both support 1M context. A real bargain for high-intelligence workloads.
Goku ./ tweet media
English
1
0
5
157
Pascal2_22./
Pascal2_22./@Pascal2_22·
The right charts show exactly how constrained Labs redesign attention to need less HBM. DeepSeek didn't solve long context by throwing more memory at it. They redesigned how attention accumulates memory so the KV cache stays flat instead of growing linearly. That's architectural innovation under resource constraint not hardware brute force as Frontier Labs approach it. The left Chart shows: Performance of DeepSeek V4 Pro Max, beating or matching Claude Opus 4.6, GPT-5.4 and Gemini 3.1 Pro across nearly every benchmark. Knowledge, reasoning, agentic tasks. The performance gap between V4 and frontier closed source models is either marginal or nonexistent on most tasks. On the Right chart, the Efficiency of Deepseek V4 Pro runs at 3.7x lower FLOPs than V3.2 at long context. V4 Flash runs at 9.8x lower FLOPs. KV cache — the memory that explodes as context grows — is 9.5x to 13.7x smaller. Same benchmark performance. Fraction of the compute and memory cost. Frontier labs scale infrastructure to match model demands. DeepSeek scales architecture to outrun the hardware bill.
Pascal2_22./ tweet media
English
1
0
14
312
Yuan ./
Yuan ./@yuangao·
Ok, which one should I try first?
English
19
0
22
1.5K
Hexx ./
Hexx ./@HexxRL·
Deepseek V4 Pro Max gives you GPT 5.4 max & Opus 4.6 max at a fraction of the price. Whale dropped after everyone else dropped this month. Open source will win. 🐋
Hexx ./ tweet media
DeepSeek@deepseek_ai

🚀 DeepSeek-V4 Preview is officially live & open-sourced! Welcome to the era of cost-effective 1M context length. 🔹 DeepSeek-V4-Pro: 1.6T total / 49B active params. Performance rivaling the world's top closed-source models. 🔹 DeepSeek-V4-Flash: 284B total / 13B active params. Your fast, efficient, and economical choice. Try it now at chat.deepseek.com via Expert Mode / Instant Mode. API is updated & available today! 📄 Tech Report: huggingface.co/deepseek-ai/De… 🤗 Open Weights: huggingface.co/collections/de… 1/n

English
3
3
35
1.3K
𝓁ℯ𝒾𝓁𝒶 ./
𝓁ℯ𝒾𝓁𝒶 ./@Leila1113514·
DeepSeek V4 is actually insane. Just saw the benchmarks and it's outperforming frontier models in coding and reasoning while being a fraction of the cost. @deepseek_ai seeking deep!
𝓁ℯ𝒾𝓁𝒶 ./ tweet media
English
1
0
6
275
Pascal2_22./
Pascal2_22./@Pascal2_22·
@tryParallax is the ultimate operating system that turns mismatched hardware into one unified service. Model too big for one machine? Parallax shards it across your laptop, a lab GPU, a teammate's workstation, orchestrated seamlessly opening up a wide range of ways to host and run AI apps and agents on your own infrastructure. Sovereign. Yours. @Gradient_HQ
rw ./@gradientintern

Awesome to see @tryParallax’s distributed framework for heterogeneous machines being implemented and serving up inferences! Build and customize your own clusters for AI like never before 🤖 ./ LFG @Gradient_HQ

English
2
0
7
98
Pascal2_22./
Pascal2_22./@Pascal2_22·
Kimi K2.6 just dropped and it's sitting at #4 on the Artificial Analysis Intelligence Index—trailing only the big three (Anthropic, Google, OpenAI at 57). Here's what actually matters: The agentic performance is wild. Elo jumped from 1309 (K2.5) to 1520 on GDPval-AA. This thing handles real knowledge work like presentations, analysis, code execution with tools. 96% on τ²-Bench Telecom puts it in the same tier as frontier models for tool use. Hallucination rate dropped hard. 39% (down from 65% in K2.5). It's learned to shut up when it doesn't know something instead of making stuff up. That's the gap between useful and dangerous in production. Token usage is high (~160M for the full benchmark) but so is everyone else at this level. Claude Sonnet 4.6 uses ~190M. GPT 5.4 uses ~110M. You're paying for reasoning quality, not efficiency.
Pascal2_22./ tweet media
English
2
0
9
115
Pascal2_22./
Pascal2_22./@Pascal2_22·
Win rate is a terrible proxy for long-horizon AI performance. 6-player No-Limit Hold'em results -- PPO: 25% win rate, +5.6 BB/100 (wins small pots, loses value) -- ISO-LLM: 22% win rate, +15.8 BB/100 (strategic value optimization) -- GPT-4o: 14% win rate, -8.2 BB/100 (zero-shot fails) ISO by @Gradient_HQ achieves 3x better long-term returns while being harder to exploit. Strategic foresight > greedy optimization
Pascal2_22./ tweet mediaPascal2_22./ tweet media
English
2
0
12
161
Pascal2_22./
Pascal2_22./@Pascal2_22·
Most AI agents optimize for win rate and fail in long-horizon games because they treat non-stationarity as pure noise instead of predicting strategic regimes. ISO by @Gradient_HQ introduces prediction-aware learning: agents forecast strategic context and adapt policies in real-time. Here's how it works: Pipeline: Expert demos → SFT → Strategic Reward Model → ISO-GRPO Runtime: Predict context → Route policy → Update learner Result: 10.6x cheaper regret, equilibrium convergence scaling with prediction accuracy not horizon length. Win rate ≠ long-term value. Strategic foresight wins.
Pascal2_22./ tweet media
English
0
1
18
178
itsrisss_
itsrisss_@itsrisss_plg·
It’s been a year since I started diving into AI, and honestly, it’s been an eye-opening journey. I’m also grateful to have been part of the @Gradient_HQ community for the past year being surrounded by people who share the same curiosity and passion makes the whole experience even more meaningful. What began as simple curiosity turned into a deep interest in understanding how AI works and how fast it’s evolving. Every day feels like there’s something new to learn from tools, to concepts, to real-world applications. It hasn’t always been easy, but the process of figuring things out step by step is what makes it rewarding. I’m still learning, still adapting, and still exploring the endless possibilities in AI. Big thanks to @HexxRL and @VinoisAsian for the support and for being part of this amazing experience 🤝
itsrisss_ tweet media
English
11
0
18
342