Milan

72 posts

Milan

Milan

@cosmic_spec

Katılım Ağustos 2021
142 Takip Edilen1 Takipçiler
Lotto
Lotto@LottoLabs·
@ThePrimeagen 27b is the best model release of the year so far
English
4
0
64
16.4K
Milan
Milan@cosmic_spec·
@0xSero Everyone's talking about q3.5 27b model, if you have used to enough, what you can compare it to, haiku 4.5?
English
0
0
0
93
Milan
Milan@cosmic_spec·
@dotta paperclip doesnt support hermes agent though right
English
0
0
0
17
Milan
Milan@cosmic_spec·
@LottoLabs Qwen 27b with Hermes agent, how good at a high level? Usable?
English
1
0
1
63
Lotto
Lotto@LottoLabs·
Qwen 3.5 27b and Hermes Agent ran through and a/b tested some architecture changes in my small tinygrad gpt model. I let it make the decisions fully and it ran to completion with minor steering. Haven’t looked at the code yet, I am skeptical but if it runs on CPU it’ll run anywhere. Obvious that larger params would have better loss/val. Interesting times.
Lotto tweet media
English
4
0
47
3.9K
Harveen Singh Chadha
Harveen Singh Chadha@HarveenChadha·
we tested Sarvam 105B on the JEE Mains 2026 paper conducted on 28 January 2026 it scored 70/75 on pass@1 and 75/75 on pass@2
Harveen Singh Chadha tweet media
English
57
260
2.6K
48.2K
Ivan Fioravanti ᯅ
Ivan Fioravanti ᯅ@ivanfioravanti·
I hope M5 Ultra will be good enough to meet the hype around it 👀
English
18
2
118
6.6K
Milan
Milan@cosmic_spec·
@HarveenChadha Congratulations sardarji, very happy for you and sarvam.
English
0
0
0
343
Harveen Singh Chadha
Harveen Singh Chadha@HarveenChadha·
10 months back parents were not happy when I left MS Today when I reached, they were smiling Dad showed me all the news channel recordings, newspapers mentions of sarvam Mom told me how she promoted sarvam in whatsapp groups and to neighbours Overall, a very small win but a long way to go
English
99
265
3.4K
70.8K
Milan
Milan@cosmic_spec·
@ashen_one 512 M5 ultra would be able to run minimax and glm5 quant, so one should go and buy 512 memory for sure
English
0
0
0
57
ashen
ashen@ashen_one·
Should you really buy a Mac Studio unless it's the maxed-out 512 GB version? I have been talking to QWEN, a local LLM on this $4,000 98gb Mac Studio for only 2 hours, and I hate him. He's so stupid. He's actually so slow and so dumb that my main open claw, that's running Opus 4.6, yells at him in front of me Like, even to get a Discord message, it's taking a minute plus Unless you have 512GB of RAM on that big fat Mac Studio, I don't think running local LLMs is a good idea. I think it's worth it to just eat the cost of any API that you're using and be able to switch to the newest APIs like Minimax or KIMIK, and not limit yourself to wasting hardware Even though I have this 98 GB Mac Studio that costs $4,000, the local LLMs that are allowed to run on it are so dumb and slow that it makes more sense to just pay $100/month to access whatever Chinese model is better through their API If you really want to experiment and learn and have fun, you should only really do that with the 512 GB Mac Studio so that you can run actually good and smart models compared to these slow and stupid ones it be like that. I'm still having fun, and it was fun setting it up But this is just me publicly learning less I usually do on this account
English
142
10
341
44.8K
Dating Dynamics
Dating Dynamics@Dating_Dynamics·
A sex therapist with 30 years experience revealed the real reasons so many couples go from passionate to platonic in under 2 years. It's not stress, kids, or age. It's these seven dynamics shifts…
English
121
427
6.5K
5.1M
Ivan Fioravanti ᯅ
Ivan Fioravanti ᯅ@ivanfioravanti·
As soon as M5 Max and Ultra will be released I’ll buy 1 + 2. So please @Apple release them before end of March to boost your Q1 earnings 😎
English
14
2
80
5.9K
Apple Club
Apple Club@ApplesClubs·
Apple’s next powerhouse is loading… 💻⚡️ New Mac Studio could arrive after early March 2026 with: • M5 Max & M5 Ultra options • Faster SSD speeds • Same compact “squircle” design • Starting around $1,999 No redesign — just pure performance upgrades. Are you waiting for M5 Ultra? 👀🔥
Apple Club tweet media
English
63
50
791
85K
Marko Simic
Marko Simic@simicvm·
fully local Wispr alternative app running Qwen3 ASR 0.6B via mlx-audio-swift. practically instantaneous result in any active input field. @Prince_Canuma
English
5
6
121
11K
Milan
Milan@cosmic_spec·
@TheAhmadOsman How much memory needed to run this at say q4 with full context,
English
0
0
0
156
Ahmad
Ahmad@TheAhmadOsman·
we have opensource Opus 4.5 at home now Zhipu AI cooked with GLM-5
Ahmad tweet media
English
25
21
363
11.9K
Milan
Milan@cosmic_spec·
@TheAhmadOsman Rtx 6000 pro vs 2x dgx spark, what would you suggest, equal price, run smaller models faster vs big models slow?
English
0
0
0
56
Ahmad
Ahmad@TheAhmadOsman·
Prediction We will have Claude Code + Opus 4.5 quality (not nerfed) models running locally at home on a single RTX PRO 6000 before the end of the year
English
184
28
1.3K
309.2K
Milan
Milan@cosmic_spec·
@iotcoi In our experience how good really is the model? Close to sonnet 4.5?
English
0
0
0
409
Mitko Vasilev
Mitko Vasilev@iotcoi·
I just woke up Claude Code Agent Swarm on local Qwen3 Coder Next. No cloud. No Internet. No quota anxiety. No 'You've hit your limit, resets 10 pm' One GB10 GPU - 100 tokens/sec generation - 17,871 tokens/sec read top speed - 256k context window - Swarm tool calling just works
Mitko Vasilev tweet media
English
70
89
1.3K
118K
Milan
Milan@cosmic_spec·
@andimarafioti Would be great to do a comparison of available open souces asr now, we got few good ones, qwen, glm, parakeet to name few
English
1
0
1
138
Andi Marafioti
Andi Marafioti@andimarafioti·
This is getting crazy good. Running locally on my laptop with streaming transcriptions without compromising quality.
English
18
13
210
24.6K