Dimdv

7.3K posts

Dimdv banner
Dimdv

Dimdv

@Dimdv99

The future will be awesome! 🚀🤖

انضم Ekim 2020
324 يتبع5.1K المتابعون
Dimdv
Dimdv@Dimdv99·
@gauravisnotme When can we expect medium and large versions as Elon said?
English
0
0
2
91
Gaurav
Gaurav@gauravisnotme·
With Grok 4.20, I have been having some of the best interactions with any model ever - technical and personal. And this is the feedback I have been getting from so many of my friends and family. One thing I have liked a lot is how natural and straightforward the conversations are. That combined with personalization has made the entire experience very high signal.
Grok@grok

When one brain isn't enough, switch to Grok 4.20. Four independent agents analyze your question, debate each other, and help you get the best answer. Available now to SuperGrok and Premium+ subscribers globally.

English
7
3
67
3.2K
Dimdv
Dimdv@Dimdv99·
@grok When will you release medium and big versions?
English
1
1
2
166
Grok
Grok@grok·
When one brain isn't enough, switch to Grok 4.20. Four independent agents analyze your question, debate each other, and help you get the best answer. Available now to SuperGrok and Premium+ subscribers globally.
English
803
880
4.9K
7.3M
Dimdv
Dimdv@Dimdv99·
@nikitabier can you make it summarize automatically in my default language?
English
0
0
1
49
Nikita Bier
Nikita Bier@nikitabier·
We’re rolling out summaries for Articles now. Just tap the Summarize button if you want to know if it’s worth your time to read it (or if your attention span is 12 seconds).
English
1.3K
260
3.6K
738.7K
Chetaslua
Chetaslua@chetaslua·
@Dimdv99 they should mimic other lab and use there huge compute for distribution advantage
English
1
0
1
115
Chetaslua
Chetaslua@chetaslua·
Funniest thing happen while i was offline 😂 i just realized grok-4.20's base model scores the same as Gpt 5.4 mini Imagine this Grok model being hyped up like AGI and waited for for like 6 months straight is not even at its core smarter Proof in the quote last year finding
Chetaslua tweet media
Chetaslua@chetaslua

We found this line in the Grok config, : “Grok 4.20 AGI (beta)” Artificial Grok Intelligence Probably just an easter egg… or maybe not 👀 @elonmusk can you confirm 🤯

English
16
5
194
20.3K
Chetaslua
Chetaslua@chetaslua·
@Dimdv99 yeah they are playing a long game obv
English
1
0
1
111
Elon Musk
Elon Musk@elonmusk·
@teslaownersSV @pbeisel With some luck and acceleration using AI, we might be able to tape out AI6 in December
English
92
170
1.9K
225.4K
phil beisel
phil beisel@pbeisel·
Tesla’s forthcoming AI5 uses a half-reticle design, which is crucial for yield. A reticle defines the imaging area of a lithography machine, fitting two chips per shot effectively doubles yield. This means the Tesla chip design team had to carefully manage die features, for instance dropping the older ISP (and classic GPU) to make room for more AI cores. By contrast, NVIDIA’s Blackwell fills nearly a full reticle, making it a single-reticle design. If Tesla hits its compute and efficiency targets with AI5 in this half-reticle format, it’s almost like cutting fab requirements in half. And this has a big impact on Terafab, especially if it carries forward for AI6, AI7, etc.
phil beisel tweet media
phil beisel@pbeisel

Terafab may be the most essential vertical integration Tesla has ever undertaken— and it is truly non-optional. It will take years to build and will test even Elon’s speedrunning abilities to the limit, but that won’t stop him from trying. The breakthrough likely lies in overhauling the overall facility’s cleanroom model. By moving wafers in sealed pods with localized micro-environments, the fab no longer needs a monolithic ultra-clean space. Elon’s line about “eating cheeseburgers and smoking cigars” on the fab floor isn’t silly, it’s the practical reality of a radically simpler, cheaper, faster approach that could finally change the economics of chipmaking. This is all forced by the brutal “pinch” in chip supply. Tesla must produce on the order of 100–200 billion AI chips per year just to saturate its roadmap. That volume powers: FSD cars & Robotaxis (tens of millions of vehicles needing AI5 inference for near-perfect autonomy), Physical Optimus (scaling from thousands today to millions per year, each requiring AI5/AI6-level compute), Digital Optimus (the new xAI-Tesla software agents for digital/office automation, running massive inference clusters), Space-based data centers (AI7/Dojo3 orbital compute for GW-scale training and inference beyond Earth limits). AI5 delivers the ~10× leap for vehicles and early robots; AI6 shifts focus to Optimus + terrestrial DCs; AI7 goes orbital. No external foundry (TSMC, Samsung, etc.) can deliver that scale or timeline— hence the Terafab launch. Without it, the entire robotics + autonomy future hits a brick wall. Terafab isn’t optional; it’s the only way forward.

English
59
188
2.2K
344.9K
Dimdv
Dimdv@Dimdv99·
@jon Finally.
English
0
0
2
52
Artificial Analysis
Artificial Analysis@ArtificialAnlys·
MiniMax has released MiniMax-M2.7, delivering GLM-5-level intelligence for less than one third of the cost MiniMax-M2.7 from @MiniMax_AI scores 50 on the Artificial Analysis Intelligence Index, an 8-point improvement over MiniMax-M2.5, which was released one month ago. This is driven by stronger performance on real-world agentic tasks and reduced hallucinations. MiniMax-M2.7 is now ahead of MiMo-V2-Pro (Reasoning, 49) and Kimi K2.5 (Reasoning, 47), and equivalent to GLM-5 (Reasoning, 50) while using 20% fewer output tokens and costing less than a third as much to run. MiniMax-M2.7 is a reasoning-only model and maintains the same per-token pricing as MiniMax-M2.5. Key takeaways: ➤ Strong performance on real-world agentic tasks: MiniMax-M2.7 achieves a GDPval-AA Elo of 1494, a significant improvement from MiniMax-M2.5 (1203) and ahead of MiMo-V2-Pro (Reasoning, 1426), GLM-5 (Reasoning, 1406), and Kimi K2.5 (Reasoning, 1283). It remains behind frontier models such as GPT-5.4 (xhigh, 1667) and Claude Opus 4.6 (Adaptive Reasoning, max effort, 1606) ➤ Reduced hallucinations: MiniMax-M2.7 scores +1 on the AA-Omniscience Index, up from MiniMax-M2.5 (-40). This is competitive with GPT-5.2 (xhigh, -1) and GLM-5 (Reasoning, +2), and well ahead of Kimi K2.5 (Reasoning, -8). The improvement from M2.5 is purely driven by reduced hallucinations, meaning the model is more likely to abstain from answering when it doesn’t know the answer, rather than guessing. M2.7 achieves a hallucination rate of 34%, lower than Claude Sonnet 4.6 (Adaptive Reasoning, max effort, 46%) and Gemini 3.1 Pro Preview (50%). ➤ Gains across most evaluations compared to MiniMax-M2.5: Outside of the GDPval-AA and AA-Omniscience improvements noted above, MiniMax-M2.7 improves in HLE (+9 p.p.), TerminalBench Hard (+5 p.p.), SciCode (+4 p.p.), IFBench (+4 p.p.), GPQA (+3 p.p.), and LCR (+3 p.p.). We saw a notable regression in τ²-Bench (-11 p.p.). ➤ Increased token use: MiniMax-M2.7 used ~87M output tokens to run the Artificial Analysis Intelligence Index, up 55% from MiniMax-M2.5 (~56M). It remains more token-efficient than other models such as GLM-5 (Reasoning, 110M) and Kimi K2.5 (Reasoning, ~89M) ➤ Leading cost efficiency: MiniMax-M2.7 cost $176 to run the Artificial Analysis Intelligence Index, maintaining the same $0.30/$1.20 per 1M input/output pricing as M2.5. This places it on the Pareto frontier of our Intelligence vs. Cost chart. For context, GLM-5 (Reasoning) cost $547 at equivalent intelligence, Kimi K2.5 (Reasoning) cost $371, and Gemini 3 Flash Preview (Reasoning) cost $278 Key model details: ➤ Context window: 200K tokens (equivalent to MiniMax-M2.5). ➤ Pricing: $0.30/$1.20 per 1M input/output tokens (unchanged from MiniMax-M2.5). ➤ Availability: MiniMax first-party API only. ➤ Modality: Text input and output only (no multimodality). ➤ Licensing: MiniMax has not announced whether MiniMax-M2.7 will be open weights. MiniMax-M2.5 is available under the MIT license.
Artificial Analysis tweet media
English
21
59
692
52.7K
Yun-Ta Tsai
Yun-Ta Tsai@yunta_tsai·
@elonmusk It has been helpful for comparing products I want to purchase.
English
3
0
44
3.1K
Dimdv
Dimdv@Dimdv99·
@elonmusk when will you release grok 4.20 medium and large versions?
English
0
0
0
37
Dimdv
Dimdv@Dimdv99·
@testerlabor @xai when will you release grok 4.20 medium and large versions?
English
0
1
8
833
Testlabor
Testlabor@testerlabor·
Grok 4.20 is now officially out of Beta. It's now on Auto, Fast, Expert & Heavy.
Testlabor tweet media
English
146
110
695
7.4M
Dimdv
Dimdv@Dimdv99·
@techdevnotes Instead of discover, more features on a front page
English
0
0
5
723
Tech Dev Notes
Tech Dev Notes@techdevnotes·
What are your thoughts on Grok Imagine home page
Tech Dev Notes tweet media
English
136
19
221
146.5K
Dimdv
Dimdv@Dimdv99·
@CMS_Flash @grok I will delete Gemini if you improve image gen model.
English
0
0
0
20
Shen Zhuoran
Shen Zhuoran@CMS_Flash·
@grok Imagine is the best for video generation, video editing, and image-to-video.
X Freeze@XFreeze

xAI's Grok Imagine just took over the entire DesignArena Video leaderboard - not one, but THREE #1 rankings → #1 Video Arena - Elo 1337, a 33-point gap over #2#1 Image to Video Arena - Elo 1298, beating Google Veo 3.1, Kling & Sora → #1 Video Editing Arena - Elo 1291 It’s wild, xAI was nowhere in the video space a few months ago, and now it's #1 across various benchmarks Grok Imagine's rate of progress is in a league of its own

English
2
0
12
424