TiTikey

2.5K posts

TiTikey

TiTikey

@TiTiKey_com

Discount AI subscriptions | ChatGPT Plus, Claude, Gemini, Midjourney setup & renewals | Fast delivery, long-term support

Go ➡️ Joined Ocak 2024
10 Following57 Followers
Pinned Tweet
TiTikey
TiTikey@TiTiKey_com·
We have not released any coin; please do not be misled. #titikey
English
1
0
5
1.2K
TiTikey
TiTikey@TiTiKey_com·
🚀 AIサブスクリプション最大90%オフ!ChatGPT Plus、Claude API、X Premiumなど。11言語対応。titikey.comで始めましょう #ChatGPT #Claude #AI #割引
日本語
0
0
0
11
TiTikey
TiTikey@TiTiKey_com·
🚀 AI 구독료 최대 90% 할인! ChatGPT Plus, Claude API, X Premium 등. 11개 언어 지원. titikey.com에서 시작하세요 #ChatGPT #Claude #AI #할인
한국어
0
0
0
32
TiTikey
TiTikey@TiTiKey_com·
That's a solid take. Thanks for sharing!
Akshay 🚀@akshay_pachaar

DeepSeek-V4 just dropped! And it's solving one AI's biggest problem today: It runs 1M-token context at 10% of the KV cache and 27% of the inference FLOPs of V3.2. Here's what that means. KV cache is the memory footprint your GPU holds for every token already in context. It grows linearly with context length, and at 1M tokens it's usually what forces you onto bigger hardware or kills your throughput. Cutting it to 10% means you can serve longer contexts on smaller machines, or pack far more concurrent users on the same ones. Inference FLOPs is the compute cost of generating the next token. With vanilla attention this scales quadratically with context length, which is why long contexts get brutally expensive per token. 27% means each generated token at 1M context is nearly 4x cheaper to produce. Put together, long-context inference goes from a premium feature you ration to something you can run by default. The trick is a hybrid attention design that interleaves two mechanisms instead of picking one. 𝗖𝗦𝗔 (𝗖𝗼𝗺𝗽𝗿𝗲𝘀𝘀𝗲𝗱 𝗦𝗽𝗮𝗿𝘀𝗲 𝗔𝘁𝘁𝗲𝗻𝘁𝗶𝗼𝗻) first squashes every 4 KV entries into a single compressed entry. Then it uses a lightning indexer to select the top-k most relevant compressed blocks. Compression and sparsity stacked. 𝗛𝗖𝗔 (𝗛𝗲𝗮𝘃𝗶𝗹𝘆 𝗖𝗼𝗺𝗽𝗿𝗲𝘀𝘀𝗲𝗱 𝗔𝘁𝘁𝗲𝗻𝘁𝗶𝗼𝗻) goes aggressive. It compresses every 128 tokens into one entry and skips sparse selection entirely, because at that compression ratio dense attention over a tiny set is already cheap. Both get a sliding window branch over the last 128 tokens, so local fine-grained structure isn't lost to compression. The result is that CSA handles medium-grained retrieval while HCA handles coarse context summarization, and alternating them across layers gives you both without paying for both at full cost. V4-Pro (1.6T total, 49B active) ranks 23rd among human Codeforces competitors and hits 120/120 on Putnam-2025. Open weights on Hugging Face. The era of million-token context in open models has effectively started.

English
0
0
0
5
TiTikey
TiTikey@TiTiKey_com·
🚀 Tiết kiệm đến 90% cho đăng ký AI! ChatGPT Plus, Claude API, X Premium và nhiều hơn nữа. Hỗ trợ 11 ngôn ngữ. Bắt đầu tại titikey.com #ChatGPT #Claude #AI #GiảmGiá
Tiếng Việt
0
0
0
32
TiTikey
TiTikey@TiTiKey_com·
🚀 Экономьте до 90% на подписках ИИ! ChatGPT Plus, Claude API, X Premium и другие. Поддержка 11 языков. Начните на titikey.com #ChatGPT #Claude #AI #Скидка
Русский
0
0
0
31
TiTikey
TiTikey@TiTiKey_com·
@realBigBrainAI これは本当にの道理ですね。AIエージェントは高速でコード生成できるけど、「命味」や「人間の創造的判断」がないから、結尾は優れたデザインは生まれない。最新のAIモデルでもこの部分はまだ人間には代替できないと思います 🎨
日本語
0
0
0
7
Big Brain AI
Big Brain AI@realBigBrainAI·
Peter Steinberger, creator of OpenClaw, on why AI agents still produce "slop" without human taste in the loop: "You can create code and run all night and then you have like the ultimate slop because what those agents don't really do yet is have taste." Peter is direct: raw capability without direction still produces mediocre output. "They are spiky smart and they're really good at things, but if you don't navigate them well, if you don't have a vision of what you're going to build, it's still going to be slop. If you don't ask the right questions, it's still going to be slop." Great AI-assisted work is defined by the human guiding it. @steipete describes his own creative process when starting a new project: "When I start a project, I have like this very rough idea what it could be. And as I play with it and feel it, my vision gets more clear. I try out things, some things don't work, and I evolve my idea into what it will become." Most people skip this part entirely, front-loading everything into a single prompt and wondering why the result feels hollow. "My next prompt depends on what I see and feel and think about the current state of the project." Each step informs the next. The work itself is the feedback loop. "But if you try to put everything into a spec up front, you miss this kind of human-machine loop. And then I don't know how something good can come out without having feelings in the loop — almost like taste." The agentic trap is what happens when you remove yourself from the process too early.
English
165
282
2.1K
482.4K