-‘

657 posts

-‘

-‘

@blank_usrname

Bosnia and Herzegovina Katılım Aralık 2016
137 Takip Edilen12 Takipçiler
Linus ✦ Ekenstam
Linus ✦ Ekenstam@LinusEkenstam·
You spent months engineering around context collapse. Better prompts. Smarter chunking. Careful retrieval. I've seen this before. Teams doing a lot of work, just so that one update later, all of that work gets obsolete. This is one such update. Sub-quadratic sparse-attention architecture, thats a mouthful, but 12 million token context window!?
Alexander Whedon@alex_whedon

Introducing SubQ - a major breakthrough in LLM intelligence. It is the first model built on a fully sub-quadratic sparse-attention architecture (SSA), And the first frontier model with a 12 million token context window which is: - 52x faster than FlashAttention at 1MM tokens - Less than 5% the cost of Opus Transformer-based LLMs waste compute by processing every possible relationship between words (standard attention). Only a small fraction actually matter. @subquadratic finds and focuses only on the ones that do. That's nearly 1,000x less compute and a new way for LLMs to scale.

English
5
6
39
8.3K
-‘
-‘@blank_usrname·
@hasantoxr Love paid partnership…
English
0
0
0
272
Hasan Toor
Hasan Toor@hasantoxr·
This is the biggest LLM breakthrough in years. SubQ IS the first frontier model with a fully sub-quadratic sparse-attention architecture (SSA) + a 12 MILLION token context window. The numbers are insane: → 52x faster than FlashAttention at 1M tokens → Less than 5% the cost of Opus → Nearly 1,000x less compute by ignoring the 99% of token relationships that don’t matter No more quadratic waste. No more context hacks. Early access is open right now (plus their new coding agent SubQ Code). This is how LLMs actually scale next. Get it → subq.ai
Alexander Whedon@alex_whedon

Introducing SubQ - a major breakthrough in LLM intelligence. It is the first model built on a fully sub-quadratic sparse-attention architecture (SSA), And the first frontier model with a 12 million token context window which is: - 52x faster than FlashAttention at 1MM tokens - Less than 5% the cost of Opus Transformer-based LLMs waste compute by processing every possible relationship between words (standard attention). Only a small fraction actually matter. @subquadratic finds and focuses only on the ones that do. That's nearly 1,000x less compute and a new way for LLMs to scale.

English
14
28
261
69.6K
Ounka
Ounka@OunkaOnX·
Trump’s ‘gold card’ visas were supposed to solve the $39 trillion national debt. They’ve sold only one
Ounka tweet mediaOunka tweet media
English
413
1.3K
32.8K
1.9M
-‘
-‘@blank_usrname·
@Tazerface16 Most dont, even worse some think they are sentient 🤣
English
0
0
0
4
Christopher David
Christopher David@Tazerface16·
People understand that LLMs aren't actually "thinking," right?
Drexel-Alvernon, AZ 🇺🇸 English
1.7K
705
15.7K
846.5K
-‘
-‘@blank_usrname·
@andrewchen Haha WHAT to build is decided by someone who has money and not by somenone who flips states in jira tickets
English
0
0
1
16
andrew chen
andrew chen@andrewchen·
bullish on the PM role quietly becoming the most important role in tech again when anyone can build, the person who decides WHAT to build becomes the bottleneck
English
284
176
2.2K
225.4K
-‘
-‘@blank_usrname·
@DeepakNesss I would be more afraid of mossad which already has all your data than Chinese
English
1
0
1
216
DeepakNess
DeepakNess@DeepakNesss·
Genuine question: If I use the DeepSeek V4 Pro API via OpenCode, what data could the Chinese mine from me at worst? My entire codebase? Basic info about my computer? Just trying to understand how exposed I am in terms of privacy and security.
English
150
2
558
107.8K
-‘
-‘@blank_usrname·
@RobertJBye Y but even small model can do this! Whats the advantage then of using opus giga max
English
0
0
0
421
-‘
-‘@blank_usrname·
@satyanadella I'm 101% sure that 99.999% licences wont be used xD
English
0
0
0
67
-‘ retweetledi
Wall Street Mav
Wall Street Mav@WallStreetMav·
36,000 engineering students from IT programs were tested in India. 95% were determined to be unfit for actually working in the field. Only 1.4% could write functionally correct and efficient code. India has a massive issue with cheating and scams in education. Fake degrees
Wall Street Mav tweet media
English
455
3.2K
10.2K
300.2K
-‘
-‘@blank_usrname·
@DearS_o_n 2 weeks until Iran has nuclear weapons
English
0
0
0
6
Dear Son.
Dear Son.@DearS_o_n·
Name a huge scam that has been normalised?
English
14.1K
611
8.8K
11.6M
Maine
Maine@TheMaineWonk·
What the fuck is this?
English
598
778
6.6K
1.3M
-‘
-‘@blank_usrname·
@yasinaktimur Nice, opened for 10 seconds and closed.
English
0
0
1
1.9K
-‘
-‘@blank_usrname·
@heart_ Sometimes they ask AI whats their favorite planet 🤣
English
0
0
0
84
Heart
Heart@heart_·
I still have no idea what people do all day with their $200/mo “AIs” except making dashboards and cleaning e-mails
English
204
22
1.3K
80.8K
Claude
Claude@claudeai·
Introducing Claude Design by Anthropic Labs: make prototypes, slides, and one-pagers by talking to Claude. Powered by Claude Opus 4.7, our most capable vision model. Available in research preview on the Pro, Max, Team, and Enterprise plans, rolling out throughout the day.
English
4.1K
15.1K
148.9K
63.3M
-‘
-‘@blank_usrname·
@petergyang We tried locally 35B for coding. To be honest, its unfortunately useless
English
0
0
0
172
Peter Yang
Peter Yang@petergyang·
What is the sweet spot in open source model size? Are 35B models enough for local agentic workflows? Trying to decide how much RAM I need in a new computer.
Qwen@Alibaba_Qwen

⚡ Meet Qwen3.6-35B-A3B:Now Open-Source!🚀🚀 A sparse MoE model, 35B total params, 3B active. Apache 2.0 license. 🔥 Agentic coding on par with models 10x its active size 📷 Strong multimodal perception and reasoning ability 🧠 Multimodal thinking + non-thinking modes Efficient. Powerful. Versatile. Try it now👇 Blog:qwen.ai/blog?id=qwen3.… Qwen Studio:chat.qwen.ai HuggingFace:huggingface.co/Qwen/Qwen3.6-3… ModelScope:modelscope.cn/models/Qwen/Qw… API(‘Qwen3.6-Flash’ on Model Studio):Coming soon~ Stay tuned

English
79
3
86
34.6K
-‘
-‘@blank_usrname·
@ElxMapping He should write a book about it 🤣🤣
English
0
0
0
13
Election Enjoyer 🇺🇸
Election Enjoyer 🇺🇸@ElxMapping·
visits the pope → pope dies leads Iran negotiations → talks collapse flies to Hungary to prop up Orbán → Orbán loses in a landslide Man’s got a streak.
Election Enjoyer 🇺🇸 tweet media
English
5.2K
35.8K
268.7K
5M
-‘
-‘@blank_usrname·
@gfodor @kenwheeler Tell that to the guys who wrote attention is all you need paper…
English
0
0
0
27
gfodor.id
gfodor.id@gfodor·
@kenwheeler Nobody knows how AI works, anyone who says they do is an idiot
English
3
0
20
1.2K
patagucci perf papi
patagucci perf papi@kenwheeler·
i deeply and i mean deeply dislike this man. with every fiber of my heart. and i don’t believe this for a second.
English
32
7
431
32.9K
-‘
-‘@blank_usrname·
@TeksEdge That’s just benchmark tuning. I asked gemma to give me craftjs component, got plain react 🤣
English
0
0
0
538
David Hendrickson
David Hendrickson@TeksEdge·
❓ Interesting Benchmark: While Gemini 3.1 Pro sits at #9 on BridgeBench's UI ranking, its own Free and Open Source Gemma 3 31B MoE ranks #4.
David Hendrickson tweet media
English
18
2
165
22.9K
greg
greg@greg16676935420·
This might be a stupid question but why don’t they just dig a new route?
greg tweet media
English
4.6K
481
12.9K
2.8M
-‘
-‘@blank_usrname·
@amjadt25 hahahhahahha who asked you?!
English
0
0
0
1
Amjad Taha أمجد طه
Amjad Taha أمجد طه@amjadt25·
As long as the UAE is not in Islamabad, JD Vance and the Islamic regime in Iran will fail in Pakistan. The UAE does not, and will not, accept the Islamic regime in Iran controlling access to the Strait of Hormuz, nor possessing nuclear weapons, ballistic missiles, drones, or any capabilities that threaten regional security. This regime must issue a formal apology to the UAE for its aggression, acknowledge its actions, provide full compensation for any harm caused to individuals and institutions, and guarantee non-repetition. Otherwise, what comes next will be devastating for this terrorist regime.
English
6.6K
3.4K
17.5K
2.6M