Rok Kovač

464 posts

Rok Kovač banner
Rok Kovač

Rok Kovač

@rokkovach

I do internet stuff. 👨🏻‍💻 Lead Solutions Engineer @syncari

Austria Katılım Eylül 2011
1.1K Takip Edilen362 Takipçiler
Cluseau Investments
Cluseau Investments@blondesnmoney·
Yeah what I think people don't get is it doesn't compress the weights (most of GPU VRAM usage), it just compresses context window. Imagine if AI could read your entire codebase and fit it into the context window? Would become massively more helpful, and you'd be willing to pay more for GPUs since they would be so much more productive.
English
7
2
63
6.2K
siliconmemes
siliconmemes@realmemes6·
I may be an idiot but I was reading the turboquant thing about AI using 6x less memory with no losses and it didn't even enter my mind that would be negative for memory. Was actually confused why $MU was down so much. So that's what I think about that.
English
10
3
56
11.3K
Rok Kovač
Rok Kovač@rokkovach·
@LeonChaland The only way you can read this post seriously if it's satire. It is satire right?
English
0
0
1
98
Léon Chaland 🇪🇺
Léon Chaland 🇪🇺@LeonChaland·
This guy picked money over the continent that raised him. He could have helped improve the situation here, but he decided to sell out to a company from a place that has turned fascist, and is positioning itself as a direct enemy of Europe and its values. He decided to help make the enemy more powerful, because it made him more money, than earn less personally but help his continent succeed. If you think this is about advancing technology for the sake of humanity, you're fooling yourself. “Science without conscience is but the ruin of the soul.” There is no point in advancing technology if you put that technology in the hand of people that seek to destroy what you love. There is no point in being rich if everything you love has been destroyed. I would rather make 100 times less money but help Europe succeed, than leave and help our enemies. I have made that choice before, and I would make it again any time.
English
93
10
170
21.6K
Rok Kovač retweetledi
petra sovdat
petra sovdat@petrasovdat·
ANTIREKLAMA ZA GOLOBA Joc Pečečnik piše Petri Sovdat: »Hej, mačka ... maš naslovnico zame, da pohvalim Goloba« Malokrat se zgodi, da se razumna javna oseba medijsko sama "zadavi", še manjkrat se zgodi, da zraven "zadavi" še predsednika vlade. finance.si/joc-pececnik-p…
Slovenščina
70
258
764
52.3K
0xSero
0xSero@0xSero·
If you're interested in running LLMs at home. Here's a full breakdown of the economics and trade-offs related with various hardware options you have. YouTube video by Monday: x.com/0xSero/status/…
0xSero tweet media
0xSero@0xSero

Top 5 builds for AI inference in 2025-2026 I have spent around 6-9 months researching, building, experimenting, and bench-marking AI models, tools, hardware and costs. Top 3 picks will be the safest, best cost to performance ratios. The last 2 will be more interesting experiments that I think are worthwhile for people to take on. This list will be capped at an entry price of 10k USD for an end engine, I do this to avoid wild variance in performance. ---------------- #1. Mac M3 Ultra 512GB RAM This device costs around 10k USD new in the US, 12k new in EU, and around 6-9k used around the world (not often is it available used, but a snag if you have the interest and budget) When I wanted to get into this, the Mac was my first option. I have owned a Macbook Pro M1 Max for over 2 years (it's 4 years old now) and I really believe it can be my daily driver for 10 years to come if they support it. MLX has improved drastically in the last few months, performance is reaching ~100% of Nvidia 3090s, same bandwidth for inference. It is the cheapest way to run Kimi-K2, Deepseek, GLM-4.6, and Minimax-M2 at full context, without extreme quantization. Specifically and most importantly, the power usage is incredibly low, the whole thing at full throttle is less than 1x 3090 Evga max. Pros: > 10K USD~ 500~ GB of usable relatively high bandwidth memory > 400W at PEAK power usage, less than 1 3090 Evga > MLX is very fast, and there's tons of quantizations out there > If you network 2 of these you can HIT 1TB, you'd need minimum 50k USD to do that with NVidia > Easily available, tiny, clean, beautiful OS Cons: > No CUDA, this is a huge con tbh > Higher starting cost, you need 10k~ up front, with Nvidia and DDR4 you can get 96 VRAM and 128 DDR4 at 5k > MLX still becomes very slow after 64k tokens, vllm and sglang hold performance through to the last token. > You don't learn as much, since it's all abstracted away from you > Can't train or finetune on this build Overall I would recommend this for 90% of people --------------------- #2 Nvidia 8x 3090s || 4x 4090D Super || 8x 4080D This is what I ended up choosing (3090s), I put the Chinese mods on the list, as if you're brave enough they could be worth the risk. Nvidia has a choke hold on this market, their GPUs were the first to support LLMs at scale, so a lot of software was built on top of their hardware. Here you can get: - 192GB VRAM for 8x 3090s (24GB VRAM each) || 4x 4090D Super mods from China (48GB VRAM each) - 256GB VRAM for 8x 4080 mods from China (32GB VRAM each) With this you can run: - GLM-4.6-Reap at Q4 (near losses performance for coding, and tech work) . coding - Minimax-m2 at Q4 (Incredible model) . digital assistant - GLM-4.5 Air at FP8 . writing and coding - GPT-OSS-120B . Math and medical - Hermes-70B . Drug knowledge, no censorship - GLM-4.5V & Qwen-3-235B-VL Pros: > Fastest inference money can buy > Have access to anything ever built AI related > Holding retail value decently for now > Can train on these > Lots of learning Cons: > Less VRAM higher cost > Messy as hell > Guzzles electricity, 1500W for full system IF YOU CAP it at 50% wattage (20% inference performance loss) > Market is drying up, at least where I live > Upgrading beyond 8 is impractical, you need to bring in an electrician --------------------- #3. Ryzen AI Max+ (Framework Desktop or DIY AM5 Mini-ITX) This option gives you a very respectable amount of inference RAM for relatively cheap maxing out at about 384GB for 10k USD~ I am a big fan of Framework, and what they're doing but you can DIY this yourself if you want something more custom. You get a fast, quiet, lowish power draw rig which is expandable in 128GB increments at around 3k USD (so you can start with 128GB) You can run almost all the same model at relatively decent speeds, with a still maturing ecosystem of software. this gets you past 4x 3090s for nearly half the price and 1/10th the hassle, but you got some flaws. This is the same build Gosucoder runs, I would check this channel out for a demo with GLM and Qwen Pros: > Cheaper than most options out there > Very impressive amount of RAM > Has stable enough support > Quiet, lower power draw Cons: > No CUDA > Stuck with ROCm, which is not so bad, but under developed > Requires a lot more configuration and finicky software > Supply shortages. > Slower than Nvidia and MLX for raw TPS ----------------------- #4. Nvidia RTX 6000 Blackwell 96GB For those with more disposable income, and are looking for maximum longevity, support, dependability, and upgradability this is the best Nvidia option. They cost between 7-10K USD each, are very clean, small, have high vram per card, run blackwell, and are the best mix of VRAM/Cost/Speed for Nvidia With one of these you can GLM-4.5-Air-4 bit, GPT-OSS-120B, multiple small smart models at the same time at blistering speeds, you get vllm, sglang, and practically every tool out there. You can upgrade to 2, for 192gb VRAM it costs 2x the 3090s for same vram, but at less than half the power draw, with hardware that'll be good for at least 5 years to come. Pros: > Huge VRAM per size/card > Very very fast > Nvidia ecosystem > Modern, clean, has longevity > Upgradability has a good outlook, you can get up to 8 of these for 768GB VRAM on a household circuit > Training beast Cons: > Highest price per GB of VRAM > Assuming you can save 6-10k usd a year it would take you 1 year to add more to your cluster ------------------------ #5. Huawei Atlas 300I Duo 96 During my search for more VRAM I came across these chips from Huawei, they have 400gb/s bandwidth, cost 1600$ USD pre tariffs pr card, and have 96GB VRAM With 10k USD you can get 480GB VRAM, at half the speed of 3090s, which would be incredible for running the big boys. This shouldn't be something you consider seriously if you're not interested in debugging for days, translating Chinese forums, and having your github tickets stuck for months. Pros: > So much VRAM per $ > Decent speeds > Easy to start, try 1, and resell if you don't like it (1.5-2k USD for 96gb vram) > Support for Chinese cards is only growing > You can make some BANGER content, given not many people have touched these cards in the west > vLLM SUPPORT! Cons: > Low bandwidth, lowest on this list > Probably not doable for US citizens > Very finnicky software and hardware > Little information online about this > Gotta order from China

English
17
26
318
30.2K
N
N@LatentFreedom·
A lot of people are buying the Apple Mac Mini, but why not just buy the Nvidia DGX Spark?
N tweet mediaN tweet media
English
3
1
19
7.2K
Max Weinbach
Max Weinbach@mweinbach·
@wongmjane Unfortunately with local you’ll either get fast and bad, just ok and sorta fast, or good and slow + limited context window M5 helped the prefill which is the biggest slowdown Your best bet today is a DGX Spark in a corner or cloud models imo
English
5
0
63
5.8K
Jane Manchun Wong
Jane Manchun Wong@wongmjane·
I tried running a local model for coding agent on my M4 Max laptop through MLX but it’s not very fast (compared to the hosted models out there) Is this normal? What local models do y’all use (if at all)? Thanks!
Jane Manchun Wong tweet media
English
47
2
220
46.9K
0xSero
0xSero@0xSero·
If this works we will have GLM-4.7 At - 90gb - 108gb <- this is done already - 125gb If I have credits left after that I will try and quantize Deepseek.
English
14
2
150
10K
Žiga Turk
Žiga Turk@ZigaTurk·
Jutri pride Fujifilm X-T30 III. Idealen aparat za tiste, ki bi se radi začeli ukvarjati s fotografijo. So pa dovolj dobri tudi starejši modeli, ki jim cena pada zelo počasi. Povsem drugače kot pri računalnikih in telefonih je tukaj 10 let stara roba še vedno konkurenčna, živ je pa tudi trg rabljene opreme. Trajnostno!
FUJIFILMXseriesJapan@FujifilmJP_X

間もなく登場するXシリーズ新製品にご期待ください! 新しいカメラとレンズとともに、あなたの物語が今始まる。 2025年10月23日 午後2時(日本時間)にお会いしましょう! #富士フイルム #FUJIFILM #Xシリーズ #fujifilm_xseries

Slovenščina
2
1
9
3K
Jacob Bank
Jacob Bank@jebank·
Our marketing team is just me and ~40 AI agents. I finally got around to putting them into an "org chart", and it's actually really cool to see! Plus, laying them out this way by sub-function (social media, blog, email, community, partners, etc) has given me a bunch of ideas of other agents I want to build. If you're interested in the full version (and the templates/screenshots for each agent), let me know and I'll send it over to you.
English
598
66
1.3K
319.3K
Rok Kovač
Rok Kovač@rokkovach·
@johnrushx What’s your preferred infra to host agent’s code?
English
0
0
1
177
John Rush
John Rush@johnrushx·
Web Agents are automating the entire class of jobs: > fill out forms (file complaints) > signup for things (apply for a visa) > browse web (buy tickets) > collect data These AI agents are actually good for humanity - they remove the most boring jobs. Tools I use to build them:
English
82
70
993
124.2K
eu/acc
eu/acc@euacc·
Germany has begun Gestapo-style raids on citizens posting "offensive" memes
English
37
79
1.1K
101.7K
Rok Kovač
Rok Kovač@rokkovach·
@frayedcollar This slaps hard. I got the black wool one, goes with everything.
English
0
0
1
284
Kev
Kev@frayedcollar·
Barbour / J. Crew / Paraboot
Kev tweet media
English
41
26
2.6K
158.9K
Rok Kovač
Rok Kovač@rokkovach·
@novikoff Damn, thank God it's just 5 times the price. 😂 Really tempted, no video stuff, right?
English
1
0
0
20
Dimitri Novikov 🇺🇦
Dimitri Novikov 🇺🇦@novikoff·
Figured out I can directly connect XY to my iPhone and just record video.
English
5
1
38
4K