друже не дуже

11.9K posts

друже не дуже banner
друже не дуже

друже не дуже

@rofh

той ближче до неба, кому нічого не треба

Bucha, Ukraine Katılım Eylül 2009
214 Takip Edilen219 Takipçiler
Maks Mykytyn
Maks Mykytyn@myrkytyn·
Гарний сьогодні день. Сонячний. Цей субстрат такий класний на дотик і на запах.
Maks Mykytyn tweet mediaMaks Mykytyn tweet mediaMaks Mykytyn tweet media
Українська
4
0
15
651
друже не дуже
@myrkytyn іронічно тільки фото оригінальних, якраз на метр ширини, але після останнього сезона трохи звузив і зараз воно під снігом. тобто оце я робив шоб між перцями поміщались баклажани і воно було ок але не дуже зручно, краще грядки де є тільк ти на два ряда
друже не дуже tweet media
Українська
1
0
2
7
Maks Mykytyn
Maks Mykytyn@myrkytyn·
@rofh А маєш фото як виглядає?)
Українська
1
0
0
16
Maks Mykytyn
Maks Mykytyn@myrkytyn·
Що ви думаєте про таку композицію городу? Три високі грядки, 1м х 2м, висотою 20 см. Орієнтація грядок схід-захід. Попри паркан - кукурудза. з інтервалом 30см. В грядках - цибуля, кабачок, гарбуз, перець, баклажан, салат, базилік, томати. Має сенс?
Maks Mykytyn tweet mediaMaks Mykytyn tweet media
Українська
5
0
2
491
друже не дуже
прікол шо я навіть не чув нічого. ніндзя-зеленбуд, ну або мабуть дтек
Українська
0
0
1
9
друже не дуже
ні воно не настільки заважало проводам і я його кожної весни підрізаю. дякую шо не пеньок залишили
Українська
1
0
1
12
друже не дуже
мені було сьогодні років коли я дізнався що комунальники можуть приїхати і без дозволу обрізати ваше власне дерево
друже не дуже tweet mediaдруже не дуже tweet media
Українська
1
0
2
32
друже не дуже
@myrkytyn о, брендівайн і сержант пеппер теж купив ще минулого року, мають бути цікаві сорти, з цього магазина все інше топ було
Українська
1
0
1
6
Maks Mykytyn
Maks Mykytyn@myrkytyn·
Насіння знайшлося)
Maks Mykytyn tweet media
Українська
1
0
2
53
Maks Mykytyn
Maks Mykytyn@myrkytyn·
І тут я зрозумів, що насіння з пошти не забрав 🙈 Було багато посилок і в одній дві позиції. Порахував кількість коробок і кількість посилок - зійшлося. Ой
Maks Mykytyn@myrkytyn

@Bilka_lisova 😅😅😅

Українська
1
0
4
318
Дивовижний Аромоксамит
Ми живемо в психоісторії
Поліна Вязовська@poly_viazovska

@PinkTetrapode Можливо збіг, але саме в 21 ст корчинський сформулював основний постулат теорії психополітики: шлях до слави лежить через близький схід. І тепер США бомблять там. Змушує задуматись.

Українська
2
0
17
1.6K
Roman
Roman@roman01la·
@rofh Not in browser unfortunately, in traditional sense. Isolated web apps is pretty new and only works in Chrome, and requires enabling multiple experimental flags.
English
1
0
1
19
Roman
Roman@roman01la·
Pushing browser tech further. Now 100ms glass to glass over UDP, skipping WebRTC. This time using Isolated Web Apps in Chrome, which is more like privileged installable web app, with Direct Sockets API available developer.chrome.com/docs/iwa/direc…
Roman tweet media
English
2
0
9
977
друже не дуже
@avsm people use them as personal assistants, like to automatically analyze new twitter bookmarks, fetch articles and analyze those for users personal interests/goals, even though that doesn’t always work very well :)
друже не дуже tweet media
English
0
1
4
775
Anil Madhavapeddy
Anil Madhavapeddy@avsm·
I've got tens of thousands of concurrent connections hitting my oxcaml webserver, so I hacked in some logs and it turns out most of them are from coding user-agents. Are people pointing their claudes directly at my blog or something?
English
4
2
35
3.7K
друже не дуже
@myrkytyn півгодини часу і можна зробити свого Дімона. він мені тепер клонує репозиторії які мені цікаві і робить ревою, менеджить податки і розсаду, все через телегу
друже не дуже@rofh

є у мене персональний штучка інтелект помічник Даймон для складних задач, але для сімʼї я зробив меншого тупішого брата, Дімона і додав його в чат. з усіма відключеннями повний піздос і холодіна, але хоча б Дімон старається бути корисним

Українська
0
0
1
41
Maks Mykytyn
Maks Mykytyn@myrkytyn·
Хочу дійти того рівня автоматизації конфігураційного менеджменту, щоби на сервер заходити взагалі не треба було, а зміни можна було би пушнути з телефону в git)
Українська
2
0
4
372
друже не дуже
@KuittinenPetri @0xSero hey Petri had a quick search on your account so I can see you actually use it. Are you happy with it, is it quiet? Wanted to preorder as soon as it was announced but got blocked by initial backlash on forums and all those faulty units
English
0
0
0
7
Petri Kuittinen
Petri Kuittinen@KuittinenPetri·
@0xSero I don't understand how Apple can be cheaper than Ryzen AI in $/GB. It must the be massive Apple price difference between USA and EU. Apple computers cost nearly twice here vs USA. I am currently using Beelink GTR9 Pro, which is the cheapest AI computer with 128 GB LPDDR5X.
Petri Kuittinen tweet media
English
3
0
1
810
0xSero
0xSero@0xSero·
If you're interested in running LLMs at home. Here's a full breakdown of the economics and trade-offs related with various hardware options you have. YouTube video by Monday: x.com/0xSero/status/…
0xSero tweet media
0xSero@0xSero

Top 5 builds for AI inference in 2025-2026 I have spent around 6-9 months researching, building, experimenting, and bench-marking AI models, tools, hardware and costs. Top 3 picks will be the safest, best cost to performance ratios. The last 2 will be more interesting experiments that I think are worthwhile for people to take on. This list will be capped at an entry price of 10k USD for an end engine, I do this to avoid wild variance in performance. ---------------- #1. Mac M3 Ultra 512GB RAM This device costs around 10k USD new in the US, 12k new in EU, and around 6-9k used around the world (not often is it available used, but a snag if you have the interest and budget) When I wanted to get into this, the Mac was my first option. I have owned a Macbook Pro M1 Max for over 2 years (it's 4 years old now) and I really believe it can be my daily driver for 10 years to come if they support it. MLX has improved drastically in the last few months, performance is reaching ~100% of Nvidia 3090s, same bandwidth for inference. It is the cheapest way to run Kimi-K2, Deepseek, GLM-4.6, and Minimax-M2 at full context, without extreme quantization. Specifically and most importantly, the power usage is incredibly low, the whole thing at full throttle is less than 1x 3090 Evga max. Pros: > 10K USD~ 500~ GB of usable relatively high bandwidth memory > 400W at PEAK power usage, less than 1 3090 Evga > MLX is very fast, and there's tons of quantizations out there > If you network 2 of these you can HIT 1TB, you'd need minimum 50k USD to do that with NVidia > Easily available, tiny, clean, beautiful OS Cons: > No CUDA, this is a huge con tbh > Higher starting cost, you need 10k~ up front, with Nvidia and DDR4 you can get 96 VRAM and 128 DDR4 at 5k > MLX still becomes very slow after 64k tokens, vllm and sglang hold performance through to the last token. > You don't learn as much, since it's all abstracted away from you > Can't train or finetune on this build Overall I would recommend this for 90% of people --------------------- #2 Nvidia 8x 3090s || 4x 4090D Super || 8x 4080D This is what I ended up choosing (3090s), I put the Chinese mods on the list, as if you're brave enough they could be worth the risk. Nvidia has a choke hold on this market, their GPUs were the first to support LLMs at scale, so a lot of software was built on top of their hardware. Here you can get: - 192GB VRAM for 8x 3090s (24GB VRAM each) || 4x 4090D Super mods from China (48GB VRAM each) - 256GB VRAM for 8x 4080 mods from China (32GB VRAM each) With this you can run: - GLM-4.6-Reap at Q4 (near losses performance for coding, and tech work) . coding - Minimax-m2 at Q4 (Incredible model) . digital assistant - GLM-4.5 Air at FP8 . writing and coding - GPT-OSS-120B . Math and medical - Hermes-70B . Drug knowledge, no censorship - GLM-4.5V & Qwen-3-235B-VL Pros: > Fastest inference money can buy > Have access to anything ever built AI related > Holding retail value decently for now > Can train on these > Lots of learning Cons: > Less VRAM higher cost > Messy as hell > Guzzles electricity, 1500W for full system IF YOU CAP it at 50% wattage (20% inference performance loss) > Market is drying up, at least where I live > Upgrading beyond 8 is impractical, you need to bring in an electrician --------------------- #3. Ryzen AI Max+ (Framework Desktop or DIY AM5 Mini-ITX) This option gives you a very respectable amount of inference RAM for relatively cheap maxing out at about 384GB for 10k USD~ I am a big fan of Framework, and what they're doing but you can DIY this yourself if you want something more custom. You get a fast, quiet, lowish power draw rig which is expandable in 128GB increments at around 3k USD (so you can start with 128GB) You can run almost all the same model at relatively decent speeds, with a still maturing ecosystem of software. this gets you past 4x 3090s for nearly half the price and 1/10th the hassle, but you got some flaws. This is the same build Gosucoder runs, I would check this channel out for a demo with GLM and Qwen Pros: > Cheaper than most options out there > Very impressive amount of RAM > Has stable enough support > Quiet, lower power draw Cons: > No CUDA > Stuck with ROCm, which is not so bad, but under developed > Requires a lot more configuration and finicky software > Supply shortages. > Slower than Nvidia and MLX for raw TPS ----------------------- #4. Nvidia RTX 6000 Blackwell 96GB For those with more disposable income, and are looking for maximum longevity, support, dependability, and upgradability this is the best Nvidia option. They cost between 7-10K USD each, are very clean, small, have high vram per card, run blackwell, and are the best mix of VRAM/Cost/Speed for Nvidia With one of these you can GLM-4.5-Air-4 bit, GPT-OSS-120B, multiple small smart models at the same time at blistering speeds, you get vllm, sglang, and practically every tool out there. You can upgrade to 2, for 192gb VRAM it costs 2x the 3090s for same vram, but at less than half the power draw, with hardware that'll be good for at least 5 years to come. Pros: > Huge VRAM per size/card > Very very fast > Nvidia ecosystem > Modern, clean, has longevity > Upgradability has a good outlook, you can get up to 8 of these for 768GB VRAM on a household circuit > Training beast Cons: > Highest price per GB of VRAM > Assuming you can save 6-10k usd a year it would take you 1 year to add more to your cluster ------------------------ #5. Huawei Atlas 300I Duo 96 During my search for more VRAM I came across these chips from Huawei, they have 400gb/s bandwidth, cost 1600$ USD pre tariffs pr card, and have 96GB VRAM With 10k USD you can get 480GB VRAM, at half the speed of 3090s, which would be incredible for running the big boys. This shouldn't be something you consider seriously if you're not interested in debugging for days, translating Chinese forums, and having your github tickets stuck for months. Pros: > So much VRAM per $ > Decent speeds > Easy to start, try 1, and resell if you don't like it (1.5-2k USD for 96gb vram) > Support for Chinese cards is only growing > You can make some BANGER content, given not many people have touched these cards in the west > vLLM SUPPORT! Cons: > Low bandwidth, lowest on this list > Probably not doable for US citizens > Very finnicky software and hardware > Little information online about this > Gotta order from China

English
17
26
320
30.1K