Szymon Tworkowski

719 posts

Szymon Tworkowski banner
Szymon Tworkowski

Szymon Tworkowski

@s_tworkowski

reasoning @xAI | prev. @GoogleAI @UniWarszawski | LongLLaMA

Palo Alto เข้าร่วม Kasım 2021
745 กำลังติดตาม10.2K ผู้ติดตาม
ทวีตที่ปักหมุด
Szymon Tworkowski
Szymon Tworkowski@s_tworkowski·
Grok 4 Fast is out, for free for everyone! It pushes the frontier of reasoning efficiency and search 🚀 Join our team to maximize intelligence density x.com/i/jobs/1948829…
xAI@xai

Introducing Grok 4 Fast, a multimodal reasoning model with a 2M context window that sets a new standard for cost-efficient intelligence. Available for free on grok.com, grok.x.com, iOS and Android apps, and OpenRouter. x.ai/news/grok-4-fa…

English
24
30
346
37.6K
Szymon Tworkowski รีทวีตแล้ว
Maciej Mikuła
Maciej Mikuła@maciekmikula·
I'm thrilled to join @xAI and @SpaceX! After many years at Google DeepMind working on Gemini (including co-founding and launching Gemini Diffusion at Google I/O and contributing to Gemini, Gemma, AlphaCode, and Waymo), I can't wait to see what we build together. Few places turn science fiction into reality this fast. We're forging Grok into the most capable assistant yet, making physical intelligence real, and taking it all to Space. The last few years have been intense, but what's ahead is on another level. No better time to build — we can watch from the sidelines or be part of it. DMs are open. Understand the Universe 🚀
Maciej Mikuła tweet media
English
499
953
8K
799.6K
Szymon Tworkowski รีทวีตแล้ว
Arena.ai
Arena.ai@arena·
Grok 4.20 beta1 (single agent) debuts #1 on Search Arena, and #4 overall in Text Arena! Highlights: - #1 in Search, scoring 1226, leading GPT-5.2 and Gemini-3 - #4 in Text, scoring 1492 on par with Gemini 3.1 Pro Congrats to the @xAI team and @elonmusk on this impressive milestone!
Arena.ai tweet media
English
234
240
1.8K
10.1M
Szymon Tworkowski
Szymon Tworkowski@s_tworkowski·
@Yuhu_ai_ Over the past five years you’ve been an extraordinary mentor and a constant source of inspiration for my career in AI. The last 2.5 years creating together at xAI have been some of the most rewarding of my professional life. I’m really going to miss working with you
English
1
0
42
7.1K
Yuhuai (Tony) Wu
Yuhuai (Tony) Wu@Yuhu_ai_·
I resigned from xAI today. This company - and the family we became - will stay with me forever. I will deeply miss the people, the warrooms, and all those battles we have fought together. It's time for my next chapter. It is an era with full possibilities: a small team armed with AIs can move mountains and redefine what's possible. Thank you to the entire xAI family. Onward. 🚀 And to Elon @elonmusk - thank you for believing in the mission and for the ride of a lifetime.
English
742
365
9.3K
3.6M
Szymon Tworkowski รีทวีตแล้ว
Adrian Lancucki
Adrian Lancucki@AdrianLancucki·
🚀📉 Storing KV Cache just got 20-40× cheaper #NVIDIAResearch #ICLR2026 Introducing KVTC: A new KV cache transform coder (think JPEG for KV caches) that solves the "recompute vs. offload" dilemma. It achieves 20×-40× (up to 88×) near-lossless compression, redefining how we handle long-context memory. Unlike eviction methods that permanently delete tokens, KVTC uses PCA-based decorrelation and adaptive quantization to compress caches for storage and transfer. This makes it perfect for: • Multi-turn chats with long contexts: extends 20× - 40× the KV cache lifetime, e.g., for long sessions with AI coding assistants. • Virtually increasing memory capacity: compression of unused KV caches in DRAM, SSD or HBM allows to store 20× - 40× more. It’s a plug-and-play building block compatible with #FlashAttention, #PagedAttention, and #SparseAttention. We validated KVTC across model scales from 1.5B to 70B parameters, maintaining high accuracy on Reasoning (AIME25), CoT (GSM8K), and Long-Context Retrieval (RULER) tasks. Co-authored with amazing @CStanKonrad . Link in the comments 👇
Adrian Lancucki tweet media
English
9
76
524
49K
Szymon Tworkowski รีทวีตแล้ว
Piotr Nawrot
Piotr Nawrot@p_nawrot·
🚀📉 Efficient Inference Just Got a Major Upgrade #NVIDIAResearch We’ve just released Qwen3-8B-DMS-8x fine-tuned for 8x KV cache compression. It maintains dense model accuracy on demanding tasks like AIME24, and is perfect for inference-time scaling. The code on HF works out-of-the-box. With DMS we fine-tune models end-to-end via distillation; this works much better than “token importance” proxies found in usual eviction methods. It’s state-of-art for KV eviction tailored for fast inference: adds negligible amount of parameters and computation to each KV head, and requires as little as 1K fine-tuning steps to reach 8x compression. It speeds-up both prefill and generation phase of Transformer LLMs, and can be combined with Sparse Attention methods such as DSA. Co-Authors: @AdrianLancucki, @CStanKonrad, @PontiEdoardo Links in the comments 👇
Piotr Nawrot tweet media
English
9
57
268
60.9K
Szymon Tworkowski รีทวีตแล้ว
Szymon Tworkowski รีทวีตแล้ว
Aditya Gupta
Aditya Gupta@adityagupta·
Over the past few weeks, we have been working on the post-training RL for sharpening model's alignment with users' preferences in conjunction with model's capabilities and intelligence. It has been an amazing learning journey about the recipe, product, user signals, style, response quality and various inexplicable things which makes a conversation better. We feel 4.1 is a significant improvement in model's conversational intelligence and other aspect of interfacing with humans. Enjoy Grok 4.1, give us your feedback, and stay tuned for the next upgrades.
xAI@xai

Introducing Grok 4.1, a frontier model that sets a new standard for conversational intelligence, emotional understanding, and real-world helpfulness. Grok 4.1 is available for free on grok.com, grok.x.com and our mobile apps. x.ai/news/grok-4-1

English
16
10
304
19.9K
Szymon Tworkowski รีทวีตแล้ว
skcd
skcd@skcd42·
If you have worked on dev-tooling and want to build the best tooling to 2X the productivity of every engineer at xAI, DM me with what you're good at; we'll get you through the pipeline and an offer in <2 days.
English
225
343
2.2K
1.2M
Szymon Tworkowski รีทวีตแล้ว
Nikola Jovanović
Nikola Jovanović@ni_jovanovic·
MathArena Update: Claims about Grok 4 Fast seem to check out, it matches the performance of Grok 4 but is much faster and 20-50x cheaper. Good release! This holds across final-answer competitions, Apex problems, and Project Euler. 🧵
Nikola Jovanović tweet media
English
40
80
632
96.1K
Szymon Tworkowski รีทวีตแล้ว
Chaitu
Chaitu@chaitu·
Our new Grok 4 Fast model jumps through links at lightning speed, ingests media, and synthesizes findings so you don’t have to. Try it out in the Grok app or on Grok.com now.
English
48
51
874
51.6K
Szymon Tworkowski รีทวีตแล้ว
Mark Kretschmann
Mark Kretschmann@mark_k·
Grok 4 Fast has multimodal reasoning capabilities and is the first AI model to solve this task correctly! Screenshot by @ASItechgonewild
Mark Kretschmann tweet media
English
288
476
2.7K
1.4M
Szymon Tworkowski รีทวีตแล้ว
Shuyang Gao
Shuyang Gao@ShuyangGao62860·
The Grok-4-fast journey has been incredible—kicking off right after the Grok 4 launch in July. None of it happens without the absolute GOAT @s_tworkowski our incredibly talented teammates @LiTianleli @mycharmspace , and the unwavering backing from @Yuhu_ai_ . This kind of opportunity is impossible anywhere else and i'm deeply grateful. Hoping Grok-4-fast is lighting up your world like it has mine: scouting the best eats, catching up on sports scores, even drafting emails to my doctor. We're iterating fast across the board—share your stories and feedback! And if you're fired up to pack max intelligence density, come build with us at xAI. 🚀 job-boards.greenhouse.io/xai/jobs/47999…
xAI@xai

Introducing Grok 4 Fast, a multimodal reasoning model with a 2M context window that sets a new standard for cost-efficient intelligence. Available for free on grok.com, grok.x.com, iOS and Android apps, and OpenRouter. x.ai/news/grok-4-fa…

English
20
24
272
50.3K
Szymon Tworkowski รีทวีตแล้ว
Wenhao Chai
Wenhao Chai@wenhaocha1·
xAI just released Grok 4 Fast, a powerful model for competitive programming. Through our collaboration with xAI, we tested this amazing model on LiveCodeBench Pro. We found that Grok-4-Fast can compete with o4-mini, slightly outperform Gemini 2.5 Pro, and even solved a hard-level problem in the 2025 Q2 set! Grok-4-Fast-Non-Reasoning has become the strongest non-reasoning model, potentially rivaling gpt-oss-20b (which is a reasoning model). We are excited to see more powerful models achieving new breakthroughs on LiveCodeBench Pro, and we thank @xai for their support.
Wenhao Chai tweet media
English
60
39
337
57.8K
Szymon Tworkowski
Szymon Tworkowski@s_tworkowski·
Towards intelligence too cheap to meter
Artificial Analysis@ArtificialAnlys

xAI has released Grok 4 Fast - breaking through our intelligence vs cost frontier by achieving Gemini 2.5 Pro level intelligence at a ~25X cheaper cost Intelligence: @xai shared with us pre-release access to Grok 4 Fast. In reasoning mode, the model scores an impressive 60 on our Artificial Analysis Intelligence Index, in line with Gemini 2.5 Pro and Claude 4.1 Opus, while sitting as expected below the prior Grok 4 release and GPT-5 (high). Grok 4 Fast performed especially well on coding evaluations, taking the number one spot on our leaderboard for LiveCodeBench, even outperforming its larger sibling Grok 4. Cost: xAI is offering Grok 4 Fast at a very competitive price of only $0.2/1M Input Tokens and $0.5/1M output tokens. The model is also quite token efficient compared to other reasoning models, taking 61M tokens to complete our intelligence index, significantly less than Gemini 2.5 Pro’s 93M and Grok 4’s 120M. This competitive pricing and efficiency translates to the cost of running Artificial Analysis Intelligence Index being ~25X lower than Gemini 2.5 Pro and ~23X lower than GPT-5 (reasoning mode high). Speed: When benchmarking the pre-release API, xAI’s endpoint for the model was very fast, achieving 344 output tokens per second - ~2.5X faster than OpenAI’s GPT-5 API. This also allows for End to End Latency results that are faster than most non-reasoning models for many workloads. Speeds may drop as traffic on the API increases - keep an eye on our live performance benchmarking to see how this evolves. Congratulations to the @xai team and @elonmusk on this new release! See below for more details and in-depth analysis 👇

English
3
5
165
6.1K
Szymon Tworkowski รีทวีตแล้ว
Eric Jiang
Eric Jiang@veggie_eric·
We just hit 1,000,000,000,000 tokens this week on OpenRouter. And still rising. Thank you for the love!
Eric Jiang tweet media
English
353
179
2.3K
6.9M