vume

56 posts

vume

vume

@vachasss

infosec

Katılım Şubat 2025
48 Takip Edilen1 Takipçiler
0xSympathy
0xSympathy@SympathyLabs·
@bob_hw_store @sudoingX I have been interested in these too. Where are you sourcing? 32gb versions on AliExpress all seem to be closer to $1,000 USD.
English
1
0
1
35
Sudo su
Sudo su@sudoingX·
many ask me where to find a rtx 3090 for $900 or $1,200 in 2026. the one place i suggest you look is facebook marketplace. nothing else comes close, ebay sits 30% above and retail is fantasy.
English
14
2
68
6.2K
Uncanny Cassandra 🏳️‍⚧️
Apparently Marvel thinks Professor X is smarter than the mutant who's whole power is knowledge, he's also the only 7 intelligence I can find on marvel.com outside of actual gods
Uncanny Cassandra 🏳️‍⚧️ tweet mediaUncanny Cassandra 🏳️‍⚧️ tweet mediaUncanny Cassandra 🏳️‍⚧️ tweet mediaUncanny Cassandra 🏳️‍⚧️ tweet media
English
46
24
946
1.1M
vume
vume@vachasss·
@alxfazio What is the issue with refurb? Lol
English
0
0
1
82
alex fazio
alex fazio@alxfazio·
it’s insane that it’s basically impossible to buy a new windows mini pc, not refurbished and not used, for under $100. the mac mini neo is going to wipe windows away from earth
English
4
2
49
2.7K
vume
vume@vachasss·
@distributedkv gwern-mandated zyns oneshotting a generation of nootropicmaxxers one by one
English
0
0
1
48
tenso
tenso@distributedkv·
get out of permanent underclass starter pack
tenso tweet media
English
41
13
217
98K
Vinicius S N
Vinicius S N@ViniciusRRSouza·
@PrimeiroFront A casa é dos pais, quem manda é eles! Vc só vai decidir alguma coisa quando pagar um boleto ou ter seu próprio canto. Seja os pais bons ou não,isso não tira o poder deles de mandar e a autoridade sobre o lar. Quer direitos, vai ter seu canto. Povo fraco e encostado kkk
Português
9
0
4
3.6K
vume
vume@vachasss·
@pixningopain A gente ta tipo o Goku no comeco do arco boo, estamos mortos mas em 2027 vamos voltar super sayajin 3.
Português
0
0
0
44
paiN Pixnin ◤✠◢
paiN Pixnin ◤✠◢@pixningopain·
A timeline pegando fogo e nem um fiozinho de paiN Gaming envolvida. Resta aceitar nosso papel como meros NPCs no cenário a partir de agora, personagens secundários totais, daqueles que só fazem número. De Luffy a Usopp. De Naruto a Kiba. De Goku a Kuririn. Não sobrou nada, o famoso BRUTAL.
Português
52
27
1.6K
44.8K
vume
vume@vachasss·
@zekramu What made u like it over nvim?
English
1
0
1
444
zek
zek@zekramu·
I am so Zed pilled it’s not even funny anymore
zek tweet media
English
27
20
464
14.1K
vume
vume@vachasss·
@txhno @LottoLabs 32gb vram is plenty for that. Subagents are also super pricey (oauth still always obv win but it's not gonna be that price forever)
English
0
0
1
16
txh
txh@txhno·
@LottoLabs even with Hermes if you dont have a big enough context window and are unable to spin up subagents you're not nearly sota. it also doesn't help that most of my potential automations require it to be able code. i do agree the tool use on the model has been pretty great for its size
English
4
0
0
60
txh
txh@txhno·
dont get me wrong, im a huge open weights advocate. but the way people romanticize open models like theyre plug and play replacements for frontier models just isnt reality. were just not there yet. youre not running opus on a macbook on a flight and its not gonna cost any less
Lotto@LottoLabs

How Apple mfrs think this goes >be me >drop $1600 on two RTX 3090s used off eBay >"48GB VRAM, I'm basically a datacenter now" >they arrive in anti-static bags that look like they've been through a war >plug them into my motherboard and it sounds like a jet engine taking off >neighbors probably think I'm mining crypto again >install llama.cpp, download qwen3.6-27b quantized >"Q4_K_M, only 16GB, totally fits" >start LM Studio on port 1234 >type "hello" into the chat box >GPU fans spin up to 100% instantly >wait 8 seconds for a response >>"Hello! How can I assist you today?" >I've seen faster responses from my grandma reading a text aloud >try Q8_0 quantization because "quality matters" >OOM error, obviously >spend three hours tweaking n_gpu_layers and n_ctx like it's some kind of dark art >finally get it running at 4 tokens per second >ask it to write me a poem about my GPUs >>"Two cards of silicon and light / They hum through the endless night" >"bro this is actually fire" >show it to someone on Discord >”why are you running LLMs locally when you could just use an API for free" >explain that the joy isn't in the output, it's in watching 94% VRAM usage and knowing nobody else has access to my model >they don't understand >close Discord, open LM Studio again >"let's try a longer context window" >crash

English
1
0
1
668
vume
vume@vachasss·
@LottoLabs $800 for a rtx 3090…. If only
English
0
0
1
279
Lotto
Lotto@LottoLabs·
How Apple mfrs think this goes >be me >drop $1600 on two RTX 3090s used off eBay >"48GB VRAM, I'm basically a datacenter now" >they arrive in anti-static bags that look like they've been through a war >plug them into my motherboard and it sounds like a jet engine taking off >neighbors probably think I'm mining crypto again >install llama.cpp, download qwen3.6-27b quantized >"Q4_K_M, only 16GB, totally fits" >start LM Studio on port 1234 >type "hello" into the chat box >GPU fans spin up to 100% instantly >wait 8 seconds for a response >>"Hello! How can I assist you today?" >I've seen faster responses from my grandma reading a text aloud >try Q8_0 quantization because "quality matters" >OOM error, obviously >spend three hours tweaking n_gpu_layers and n_ctx like it's some kind of dark art >finally get it running at 4 tokens per second >ask it to write me a poem about my GPUs >>"Two cards of silicon and light / They hum through the endless night" >"bro this is actually fire" >show it to someone on Discord >”why are you running LLMs locally when you could just use an API for free" >explain that the joy isn't in the output, it's in watching 94% VRAM usage and knowing nobody else has access to my model >they don't understand >close Discord, open LM Studio again >"let's try a longer context window" >crash
English
84
87
2.5K
198.5K
vume
vume@vachasss·
@VictorTaelin lower is better but the wider the bar the better 😵‍💫 (but holy hell i love the colors!!!)
English
0
0
0
20
Taelin
Taelin@VictorTaelin·
Introducing LamBench . . . You asked me to make a benchmark, so I made it. It is a simple, old style Q&A consisting of 120 fresh λ-calculus programming questions. Some are easy, like "implement add for λ-encoded nats". Some are harder, like "derive a generic fold for arbitrary λ-encodings". It measures: - intelligence (% tasks completed) - elegance (BLC-length of solutions) - speed (completion time) Basically what I care about, other than long context. I made it today because I was excited about GPT 5.5. It didn't do too well ): (My first-day impression is that I can't tell the difference between GPT 5.5 and GPT 5.4. I would be lying if I said otherwise. I'd not be able to distinguish in a blind test. I need more time. It is much faster though.) This is a new, simple bench, so expect be bugs. Specially on OpenRouter models. I'll retest soon. Also, it was born saturated. V2 will be harder... ↓ Link and more charts below ↓
Taelin tweet media
English
57
50
899
49.6K
vume
vume@vachasss·
@VictorTaelin Do they want to allocate that much compute to… anything other than improving their models?
English
0
0
0
28
vume
vume@vachasss·
@jufxfgbvcfgv @Adianu4 @BrooklynGuyNBA o de amanha passa no meu tambem, o que nao vai passar eh o de domingo. mas tb acho que no league pass brasileiro tem mais jogo do que o dos EUA
Português
0
0
0
393
Joey
Joey@aijoey·
Qwen 3.6 27B Mac Mini 24g
Joey tweet media
Magyar
11
3
82
8.7K
vume
vume@vachasss·
@yacineMTB it’s the year of no desktop. Phone and voice typing only. Linux homelab tho
English
0
0
0
261
vume
vume@vachasss·
@akhsurgin I got 10% cashback on amazon so it ended up being much cheaper, prime visa + young adult cash back, also no time to tinker rn classes lowk kicking my ass
English
1
0
1
60
Akh Surgin
Akh Surgin@akhsurgin·
@vachasss It was just an example screenshot, you can get it faster with newegg, bestbuy, microcenter, etc.
English
1
1
1
162
Akh Surgin
Akh Surgin@akhsurgin·
You don't need a $5,000 PC to run Qwen3.6-27B. 2x RTX 5060 TI 16 GB GPUs cost <$1,000 and they give you the same 32 GB of VRAM as a 5090. In LMstudio, 27B will run at 20+ tokens/second with 200k of context, 35B MoE will run at 110+ tokens/second.
Akh Surgin tweet media
English
3
0
2
1.2K