Emilz

749 posts

Emilz banner
Emilz

Emilz

@emilheap

All I want is 100 bytes of heap in my brain · idea guy · https://t.co/1QTNwcnHWL

Brain เข้าร่วม Aralık 2025
249 กำลังติดตาม50 ผู้ติดตาม
ทวีตที่ปักหมุด
Emilz
Emilz@emilheap·
Blink and you’re behind. Here's your February AI Recap: • DeepMind Genie 3 • Voxtral Transcribe 2 • Kling 3.0 • Opus 4.6 • Codex 5.3 • Recursive Language Models (RLM) • Prompt to binary • Context rot • Seedance 2.0 • ElevenLabs v3 Expressive Mode • GLM-5 • MiniMax 2.5 • Gemini 3 Deep Think • GPT-5.3-Codex-Spark • Claude PowerPoint • Qwen 3.5 • Grok 4.20 • Sonnet 4.6 • Double prompt (Left to Right) • Cursor Plugins • Phoenix-4 • Gemini 3.1 Pro • Taalas HC1 (17k tokens/s) • Claude Code Security • US-China Distillation Wars • FDM-1 • Mercury 2 • Cursor Demos • Claude Code Remote • MatX One Chip • Perplexity Computer • QuiverAI Arrow 1.0 • Moonlake World Model • Claude Cowork Scheduled Tasks • Nano Banana 2 • Claude Agent Teams • WarClaude • France Gov Spending Data MCP • ClaudeCode Auto-Memory • Tzafon RL (4B) • Apple Xcode 26.3 with Claude/Codex • OpenAI 110B Funding round • while True: Cursor = "dead" Cursor = "back" • FactoryAI Droid Missions • Imbue Evolver • GPT 5.4 Leak • Anthropic - Military Disagreement • OpenAI - Military Agreement • Claude Import memory • US Claude assisted Iran strikes What did I miss? See you in next week's recap.
Emilz tweet media
English
1
0
3
727
Emilz
Emilz@emilheap·
@0xSero What about the planet man
English
0
0
0
43
0xSero
0xSero@0xSero·
This is how I end my nights. Today: 1.5 Billion tokens in Codex 22M tokens GLM 51M tokens Kimi 41M tokens Claude 14M tokens MiniMax
0xSero tweet media
English
20
1
128
5.9K
Emilz
Emilz@emilheap·
@bruvimtired Cmd + shift + P Switch to light mode
English
0
0
1
87
Tibo
Tibo@thsottiaux·
Codex will take us places
English
49
9
278
6.9K
Emilz
Emilz@emilheap·
When you've used up all your Cursor Cloud Agents free credits, but you're addicted.. bugbot /fast when?
Emilz tweet media
English
0
0
0
9
Emilz
Emilz@emilheap·
@gabriel1 /prompts: so good, why deprecated
English
0
0
0
22
gabriel
gabriel@gabriel1·
im just copy pasting now with codex 5.4 high (NOT xhigh, NOT medium) before every pr. it's beautiful
English
3
0
44
4.6K
Emilz
Emilz@emilheap·
@fchollet English business virtuoso fails at easy Full chinese test
English
0
0
0
116
François Chollet
François Chollet@fchollet·
This is more evidence that current frontier models remain completely reliant on content-level memorization, as opposed to higher-level generalizable knowledge (such as metalearning knowledge, problem-solving strategies...)
Lossfunk@lossfunk

🚨 Shocking: Frontier LLMs score 85-95% on standard coding benchmarks. We gave them equivalent problems in languages they couldn't have memorized. They collapsed to 0-11%. Presenting EsoLang-Bench. Accepted to the Logical Reasoning and ICBINB workshops at ICLR 2026 🧵

English
129
252
2.4K
191.9K
Emilz
Emilz@emilheap·
Composer 2 only needs to be almost as good as the previous opus 4.5 model. If the benchmark is real and felt in the code, it's GG. The only thing that made people bearish of Cursor, was the cost of their plans vs the mega subsidized ones from OpenAI and Anthropic. But if they OWN a good enough model, they can start to truly compete with the BIG 2/3 I feel like Auto mode is about to have crazy limits
Cursor@cursor_ai

Composer 2 is now available in Cursor.

English
0
0
1
85
Cursor
Cursor@cursor_ai·
Composer 2 is now available in Cursor.
Cursor tweet media
English
442
695
7.9K
2.6M
Nikita Bier
Nikita Bier@nikitabier·
No, there’s nothing over there. Come back to the screen.
English
1.6K
203
3.5K
163.2K
Emilz
Emilz@emilheap·
When will we have /fast bugbot man
Emilz tweet media
English
0
0
0
11
Emilz
Emilz@emilheap·
Lesson learned. Don't give Opus a second chance at backend.. I need cursor bugbot /fast version ASAP. Waiting 10 min for it's review is way too long
Emilz tweet media
English
0
0
0
46
Emilz
Emilz@emilheap·
@Tom_Grab @OpenAI They probably are working on a 5/6-o model to satisfy the 4o crowd
English
0
0
0
28
Tom Grabowski
Tom Grabowski@Tom_Grab·
@emilheap @OpenAI The best price-to-intelligence ratio is still 4o-mini. Not only price per token is lower, but it’s a very capable non-thinking model so the output tokens are way lower than the new models. I think OpenAI hit its peak in the 4o and o3 era.
English
1
0
0
42
OpenAI
OpenAI@OpenAI·
GPT-5.4 mini is available today in ChatGPT, Codex, and the API. Optimized for coding, computer use, multimodal understanding, and subagents. And it’s 2x faster than GPT-5 mini. openai.com/index/introduc…
OpenAI tweet media
English
533
679
6.2K
1.5M
Brian Sky
Brian Sky@BrianSOTELI·
@emilheap @JamesTakesOnAI @OpenAI It was previously priced basically at cost, or even a loss. Now it’s being priced to at least break even, or to a profit. It’s preparing to go public. Similar to the 3 year VC funded run of Uber being the best thing on the planet. Cheaper than owning your own car. 🚗
English
2
0
1
40
Emilz
Emilz@emilheap·
@thsottiaux Tomorrow you’ll release quantum agents. The day after, GPT-AGI-1 will be released which is SoTA in everything. And the day after that we’ll all be sub agents
English
2
0
12
3.3K
Tibo
Tibo@thsottiaux·
Yesterday we launched subagents in Codex. Today we released GPT-5.4-Mini, which is SoTA in its category. Coincidence or genius move?
English
152
36
1.6K
78.1K
Emilz
Emilz@emilheap·
@brianakaka @OpenAI Yeah insane value for mass data processing with good enough reasoning
English
0
0
1
93
Brian Akaka
Brian Akaka@brianakaka·
@emilheap @OpenAI This was my first question. Thanks. imo GPT5mini is best of cheap/fast/smart LLMs, and still has value.
English
1
0
0
127
Emilz
Emilz@emilheap·
@JamesTakesOnAI @OpenAI Yeah true, though they change purpose the purpose of the mini models as time goes on. The -mini are amazing for large data processing at low cost and good intelligence. But the output price more than doubling kinda destroys that use case
English
2
0
0
285
James' AI Takes
James' AI Takes@JamesTakesOnAI·
thats the real moat strategy tho. make the best model expensive, then release mini/nano versions that are "good enough" at lower cost. developers get locked into the API ecosystem chasing the flagship benchmarks, then settle for the cheaper tier in production. classic price discrimination playbook
English
3
0
2
376
Emilz
Emilz@emilheap·
@X_Arcadian @OpenAI Already optimized to the max.. so probably would need to check out specialized open source models that don’t suck at reasoning
English
1
0
0
63
Christopher Hartmann
Christopher Hartmann@X_Arcadian·
@emilheap @OpenAI Yeah I guess if you had a use cases for the old nano for 1 off llm calls that's not going to fly anymore. Wonder if you could male up for it a bit with a smaller prompt amd less Guardrails.
English
1
0
0
96
shb
shb@himbodhisattva·
@emilheap @OpenAI probably better to compare 5.4-nano to 5-mini, assuming that comes out
English
1
0
2
495