Max Headroom

20.3K posts

Max Headroom banner
Max Headroom

Max Headroom

@CosmicMonad

Project Mayhem: Operation Unity Russian from УССР. Vincit omnia veritas.

Bergabung Nisan 2024
431 Mengikuti456 Pengikut
Tweet Disematkan
Max Headroom
Max Headroom@CosmicMonad·
Max Headroom tweet media
ZXX
5
2
29
5.8K
Max Headroom
Max Headroom@CosmicMonad·
@loktar00 MoE's are always more hare brained than dense models. This is what 27b looks like when researching with similar prompt:
Max Headroom tweet media
English
0
0
1
189
Loktar 🇺🇸
Loktar 🇺🇸@loktar00·
Hmmm little concerned with 3.6 35b... it calls tools well but sucks at research and relies on it's internal knowledge instead of searching hard. 3 out of 5 were good.. but these 2 were terrible. Asked it to research what happened in the local llm world today. > Here's what happened in the local LLM / open-weight space for April 16, 2026: 2. Meta Llama 4 Scout & Maverick — First native multimodal MoE 3. OpenAI Codex open-weight + Unsloth Studio launch OpenAI released Codex open-weight models (20B and 120B variants) available on Ollama under "tools thinking cloud"
Loktar 🇺🇸 tweet media
English
10
1
26
2.4K
Max Headroom
Max Headroom@CosmicMonad·
@LottoLabs What automation harness/framework are you using here?
English
0
0
0
155
Lotto
Lotto@LottoLabs·
Let’s see if qwen 35b can not fumble this Big ask, but mainly a long task to check tool call reliability than quality
Lotto tweet media
English
11
0
34
3.5K
Max Headroom
Max Headroom@CosmicMonad·
Installed sglang 0.5.10rc0 with the mamba memory leak fix. Testing qwen3.5-27b-fp8 now, seems snappier and so far memory is doing better!
English
0
0
0
12
Nous Research
Nous Research@NousResearch·
Tool Gateway is now live in Nous Portal. No separate accounts, no API key juggling. All you need is one subscription, and everything works. A paid Nous Portal subscription now includes access to 300+ models and a growing set of third-party tools. Launching with: → Web scraping → Browser automation → Image generation → Cloud terminal backend → Text-to-speech
English
180
171
1.7K
397.3K
Max Headroom
Max Headroom@CosmicMonad·
@NousResearch Will there be an easy setup mode that avoids cloud everything, and sets up things locally, in docker containers? If anyone is interested, I have a docker-compose.yml that sets this up locally, so far w/o image gen and STT, that's in the works.
English
0
0
1
127
Len Seaside
Len Seaside@LenSeaside·
@CosmicMonad @LottoLabs I was trying to let you know that there were known tool calling problems with Gemma 4 until the release of a new Jinja a few days ago. You seem to have come here for an argument...
English
2
0
2
32
Max Headroom
Max Headroom@CosmicMonad·
@LenSeaside @LottoLabs Do you have benchmarks after jinga fix to show it can reliably call tools? I haven't seen anything that isn't from google itself.
Max Headroom tweet media
English
0
0
0
11
Max Headroom
Max Headroom@CosmicMonad·
@ioannadenisova4 Таблеточку этодругина вам, тут все говорят на своих языках, а не на одном :ь
Русский
0
0
0
2
Рыжая 3,14сечка
Рыжая 3,14сечка@ioannadenisova4·
Кстати, господа вавилоняне, напоминаю что во всех писаниях было указано, что перед приходом антихриста и концом всего живого, все люди снова начнут говорить на одном языке.
GIF
Русский
597
922
14.9K
351.3K
Max Headroom
Max Headroom@CosmicMonad·
@theisraelguys According to one interpretation of a couple of religions out of tens of thousands. Quit using your faiths to justify violence against one another.
English
0
0
1
6
The Israel Guys
The Israel Guys@theisraelguys·
This might sound controversial.... but God is with Israel
English
207
27
329
7.2K
SIGKITTEN
SIGKITTEN@SIGKITTEN·
omg they nerfed opus 4.7
English
25
7
329
19K
Max Headroom
Max Headroom@CosmicMonad·
@sudoingX 46tk/s, 1700 prefill. Qwen3.5-27b-fp8. 3090x2, 9950x, x870e, 64gb@6000.
Max Headroom tweet media
Indonesia
0
0
0
64
Sudo su
Sudo su@sudoingX·
if you own any gpu and you're running local models drop your tok/s, quant, flags, and gpu below. nvidia, amd, laptop, desktop, doesn't matter. every config you share saves someone else 3 hours of head scratching. i'll amplify the best ones and add them to the community benchmark sheet. this is how we build the local ai knowledge base, together
English
4
0
22
1.6K
Max Headroom
Max Headroom@CosmicMonad·
@lucaa_wav @mull3r_1 It's only incredible because of a massive dehumanization and deception campaign waged by various NGOs against Russians. Literally hundreds of billions were poured into conditioning people to think Russians are not human. Barely makes sense in modern world.
English
1
1
7
102
Luca
Luca@lucaa_wav·
@mull3r_1 Incrível como os comentários mais sensatos sao dos russos
Português
1
0
47
1.5K
Fábio 🇾🇪
Fábio 🇾🇪@mull3r_1·
Com essa treta toda, descobri que os russos são legais
Português
119
113
4K
46K
Gabriel
Gabriel@GabeZZOZZ·
Right after Zelensky announced that all military-age Ukrainian men must return to Ukraine for mobilization 🤣
Gabriel tweet media
English
62
141
1.3K
26.2K
DROID
DROID@droidbuilds·
most people get this wrong what’s the difference between 100 MB/s and 100 Mb/s?
DROID tweet media
English
955
179
10.7K
1.4M
Max Headroom
Max Headroom@CosmicMonad·
@LenSeaside @LottoLabs I am. You don't understand how tool calling works, are you sure you want to keep adding to this discussion?
English
1
0
0
14
Max Headroom
Max Headroom@CosmicMonad·
@x0ptimal @bnjmn_marie I haven't really used the MoE qwen3.5 offerings, because the dense works so well. Glad they are also functional!
English
0
0
3
26
++0ptim4l
++0ptim4l@x0ptimal·
@CosmicMonad @bnjmn_marie Can agree from experience. Its straight dog water when it comes to tool calling. Been running qwen3.5 35b and it just works perfectly 👌
English
1
0
1
33
Benjamin Marie
Benjamin Marie@bnjmn_marie·
I published my full analysis comparing Gemma 4 31B vs Qwen3.5 27B. > Best accuracy: Gemma 4 31B > Best token efficiency: Gemma 4 31B > Best raw inference throughput: Qwen3.5 27B > Best memory footprint: Qwen3.5 27B > Best end-to-end latency: task-dependent, with Gemma 4 slightly ahead on harder tasks and Qwen3.5 ahead on simpler ones > Best “fast” mode: Gemma 4 31B with thinking disabled > Best generalization / least benchmark affinity: Gemma 4 31B All the results and data here: kaitchup.substack.com/p/gemma-4-31b-…
English
20
22
278
18.4K
Max Headroom me-retweet
McNasty
McNasty@McNasty·
its crazy how nobody talks about the epstein files anymore
English
446
11.7K
93.6K
837.1K
Max Headroom
Max Headroom@CosmicMonad·
@Logically_JC Every 4 years literally half the country stands against the current president, and gets its way roughly 50% of the time. Not that it matters to the epstein class, they run both candidates who answer to them.
English
0
0
0
8
John Collins
John Collins@Logically_JC·
Yes, you can.
John Collins tweet media
English
1.4K
1.6K
17.2K
198.5K