๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ

1.6K posts

๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ banner
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ

๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ

@ehsanmok

Mojo๐Ÿ”ฅ maximalist @Modular. Teacher at heart. https://t.co/WvqWAgfsLP, https://t.co/yx2pGCIcYw. Into powerlifting. Used to know some Math. Opinions are mine.

Vancouver, British Columbia Bergabung Temmuz 2014
204 Mengikuti571 Pengikut
Tweet Disematkan
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Hadi Partovi
Hadi Partovi@hadipยท
@PeteHegseth โ€œBack to the Stone Ageโ€ was also the US slogan during the Vietnam War, which lasted 20 years and ended in defeat. Your original goal was liberating a civilization, not ending it.
English
9
60
592
59.7K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Nazanin Boniadi
Nazanin Boniadi@NazaninBoniadiยท
The brave Iranian people have risked everything to rid themselvesโ€”and the worldโ€”of the blight of the Islamic Republic. Threats to send Iran โ€œback to the Stone Ageโ€ are inhumane. This regime shows no regard for the destruction of Iranโ€”in fact, it has driven it.
English
1
533
2.3K
315.8K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
Gemma 4 is live on Modular Cloud, day zero, with the fastest performance on both NVIDIA and AMD. Our MAX inference framework delivers 15% higher throughput vs. vLLM on B200, and weโ€™re the only inference provider to ship @googlegemma 4 on a framework we built ourselves. Two multimodal models live now: Gemma 4 31B (dense, 256K context) and 26B A4B (MoE, only 4B params active per pass). Both SOTA on Modular Cloud: modular.com/blog/day-zero-โ€ฆ Modular Cloud runs on MAX, our inference framework that unifies GPU kernels, graph compilation, and high-performance serving in a single hardware-agnostic stack. New weights to SOTA deployment in days, on two hardware platforms: modular.com/request-demo?uโ€ฆ
GIF
English
0
7
50
4K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Chris Lattner
Chris Lattner@clattner_llvmยท
Google Deep Mind's impressive fully-open Gemma 4 is live day-zero on Modular Cloud. Modular provides the fastest performance on NVIDIA Blackwell and AMD MI355X, thanks to MAX and Mojo๐Ÿ”ฅ. The team took this impressive new model to production inference in days.๐Ÿš€
English
9
33
364
23.9K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
FA4 on Blackwell: 14 ops, 5 hardware units, 28 dependency edges. One wrong sync = a race condition sanitizers won't catch. We built a constraint solver that derives the pipeline schedule automatically, in Mojo ๐Ÿ”ฅ Part 1 of our series is out โ†’ modular.com/blog/software-โ€ฆ
English
4
11
100
30.5K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
130 lines instead of 870. That's the difference between our conv2d implementation on Blackwell and CUTLASS's. We broke kernels into three swappable pieces: one for moving data, one for coordinating the pipeline, one for compute. When you need a new kernel, you only change the piece that actually needs to change. Part 3 of our Structured Mojo Kernels series walks through the details: modular.com/blog/structureโ€ฆ
English
4
18
121
14.5K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
2 days ago we shipped image generation in <1s ๐Ÿ”ฅ Today, we make that <300ms ๐Ÿคฏ NVIDIA + AMDโšก๏ธ Full demo below โฌ‡๏ธ
English
10
12
162
11.7K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Masih Alinejad ๐Ÿณ๏ธ
Masih Alinejad ๐Ÿณ๏ธ@AlinejadMasihยท
Bobby Green, just watched your this, the way your voice broke when you spoke about the execution of Iranian wrestler Navid Afkari, it broke me. You were right. Itโ€™s heartbreaking. Today Saleh Mohamadi, 19-year old wrestler executed for protesting. We need your voice again. ๐Ÿ’”
Masih Alinejad ๐Ÿณ๏ธ@AlinejadMasih

Today, in Iran, in the middle of a war, the regime executed a 19-year-old national wrestling champion for the crime of joining January protests. ๐Ÿ’” After signaling to the world, including President @realDonaldTrump, that they would halt executions of protesters, the regime has done the exact opposite. Three young protesters, Saleh Mohammadi, Mehdi Ghasemi, and Saeed Davoudi, were hanged in Qom after a sham trial. Reports indicate torture. Forced confessions. No access to chosen lawyers. Closed-door proceedings. No right to appeal. I call on @GlobalAthleteHQ to stand with Iranian athletes who are being silenced, imprisoned, and executed simply for raising their voices. This is not just about sports. This is about human dignity.

English
122
2.4K
11.5K
730.9K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
The top three learning resources we're sharing with attendees at our @NVIDIAGTC booth this year: 1. Structured Mojo kernels blog series, 2. Mojo GPU puzzles, and 3. MAX LLM book. Links in thread ๐Ÿงต ๐Ÿ‘‡
English
1
6
28
1.8K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Morad Vaisi
Morad Vaisi@RezaVaisiยท
@elonmusk Dear Elon, In these critical days, the people of Iran need internet access. Please help them stay connected as they struggle to reclaim their country from the rule of the mullahs. Stand with the Iranian people. Iran will not forget its friends.
McNair, VA ๐Ÿ‡บ๐Ÿ‡ธ English
20K
69.5K
80.7K
2.2M
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
vixhaโ„“
vixhaโ„“@TheVixhalยท
Computer science is gradually returning to the domain of physicists, mathematicians, and electrical engineers as large language models automate much of what we currently call software engineering. The fieldโ€™s center of gravity is shifting away from manual code writing and toward deeper theoretical thinking, mathematical insight, and systems-level reasoning.
English
328
1.7K
15.4K
955.5K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
GPU kernel development doesn't have to mean thousands of lines of interleaved pipeline, memory, and compute logic. Structured Mojo Kernels Part 2 walks through how separating those concerns into three components with hard boundaries simplifies the codebase, makes kernels easier to maintain, and keeps the same structure working across NVIDIA and AMD hardware generations. modular.com/blog/structureโ€ฆ
Modular tweet media
English
0
9
75
3.7K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
You shouldn't have to choose between peak GPU performance and code you can actually maintain. We built Structured Mojo ๐Ÿ”ฅ Kernels to fix that. Performance, usability, and portability without the tradeoff. 14k to 7k lines. ~1.8k TFLOPS held. We wrote a 4-part series on how. Part 1 is up modular.com/blog/structureโ€ฆ
Modular tweet media
English
3
8
94
35.3K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Modular
Modular@Modularยท
Our fave slide: 2026 is the year of Mojo! 1.0 and compiler open sourcing are on the horizon. ๐Ÿฅณ
Modular tweet media
English
1
4
35
2.2K
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Andrej Karpathy
Andrej Karpathy@karpathyยท
It is hard to communicate how much programming has changed due to AI in the last 2 months: not gradually and over time in the "progress as usual" way, but specifically this last December. There are a number of asterisks but imo coding agents basically didnโ€™t work before December and basically work since - the models have significantly higher quality, long-term coherence and tenacity and they can power through large and long tasks, well past enough that it is extremely disruptive to the default programming workflow. Just to give an example, over the weekend I was building a local video analysis dashboard for the cameras of my home so I wrote: โ€œHere is the local IP and username/password of my DGX Spark. Log in, set up ssh keys, set up vLLM, download and bench Qwen3-VL, set up a server endpoint to inference videos, a basic web ui dashboard, test everything, set it up with systemd, record memory notes for yourself and write up a markdown report for meโ€. The agent went off for ~30 minutes, ran into multiple issues, researched solutions online, resolved them one by one, wrote the code, tested it, debugged it, set up the services, and came back with the report and it was just done. I didnโ€™t touch anything. All of this could easily have been a weekend project just 3 months ago but today itโ€™s something you kick off and forget about for 30 minutes. As a result, programming is becoming unrecognizable. Youโ€™re not typing computer code into an editor like the way things were since computers were invented, that era is over. You're spinning up AI agents, giving them tasks *in English* and managing and reviewing their work in parallel. The biggest prize is in figuring out how you can keep ascending the layers of abstraction to set up long-running orchestrator Claws with all of the right tools, memory and instructions that productively manage multiple parallel Code instances for you. The leverage achievable via top tier "agentic engineering" feels very high right now. Itโ€™s not perfect, it needs high-level direction, judgement, taste, oversight, iteration and hints and ideas. It works a lot better in some scenarios than others (e.g. especially for tasks that are well-specified and where you can verify/test functionality). The key is to build intuition to decompose the task just right to hand off the parts that work and help out around the edges. But imo, this is nowhere near "business as usual" time in software.
English
1.6K
4.8K
37.2K
5.1M
๐”ˆ๐”ฅ๐”ฐ๐”ž๐”ซ me-retweet
Chris Lattner
Chris Lattner@clattner_llvmยท
The Claude C Compiler is the first AI-generated compiler that builds complex C code, built by @AnthropicAI. Reactions ranged from dismissal as "AI nonsense" to "SW is over": both takes miss the point. As a compiler๐Ÿ‰ expert and experienced SW leader, I see a lot to learn: ๐Ÿ‘‡
Chris Lattner tweet media
English
81
345
2.1K
409.2K