XpeGj0

63 posts

XpeGj0 banner
XpeGj0

XpeGj0

@224Shroud

Katılım Eylül 2022
26 Takip Edilen1 Takipçiler
Gustavo
Gustavo@gugadotmd·
Yeah it’s crazy. I got a new macbook to replace my old M1, I got an 24gb M5. Running a 9b q4 locally is slow af, I can’t. Then I just send requests to to my rig remotely, running qwen3.6 27b q8 and there is no comparison in the speed. GPUs are on another level. As usual, the buy a gpu guy was right lol I did not get the mac to do inference or anything but I thought it was faster. A bit disappointing
English
2
0
5
547
Ahmad
Ahmad@TheAhmadOsman·
Please for the love of God don't buy MacBooks to cluster them for LLMs That shit that's trending this website is pure performative slop that should be muted / blocked
English
44
15
275
10.9K
alphaXiv
alphaXiv@askalphaxiv·
"Thinking Without Words: Efficient Latent Reasoning with Abstract Chain-of-Thought" Do reasoning models really need to think in words? This paper replaces long verbal CoT with a short learned sequence of abstract tokens that acts like a latent scratchpad. Warmed up from verbal CoT, then distilled and improved with RL. Resulting up to 11.6x fewer reasoning tokens while staying competitive with standard CoT.
alphaXiv tweet media
English
7
51
289
13.8K
Mustafa
Mustafa@oprydai·
if diffusion models denoise latent space, what is the equivalent process in human intuition?
English
11
1
20
2.7K
XpeGj0
XpeGj0@224Shroud·
@SergioHidalAERO Exijimos otro video hablando de los nuevos videos que publicaron 🙂
Español
0
0
2
100
Sergio Hidalgo
Sergio Hidalgo@SergioHidalAERO·
Veo muchisma gente hablar de OVNIS y creo que muchos aun no conocen la historia en la que varios pilotos de caza se toparon con uno Preparaos para alucinar youtu.be/fF_MvxOxqCE?is…
YouTube video
YouTube
Español
5
5
77
8K
World of Science
World of Science@Science_TechTV·
Isaac Newton's college notebook, handwritten by the genius himself around 1664.
World of Science tweet mediaWorld of Science tweet mediaWorld of Science tweet media
English
7
152
1K
59.4K
XpeGj0
XpeGj0@224Shroud·
@agusbuilds @gabmfrl Es una pena. Aumentan el tamaño del modelo y miran que habilidades emergen despues del entrenamiento. Es la única manera segura y fácil de mejorar el modelo y poder competir en un mercado tan competitivo. Así no podemos adaptar la AI en la sociedad.
Español
0
0
1
41
Gabriel Merlo
Gabriel Merlo@gabmfrl·
Supongo que pueden haber muchas variables que no vemos explicando parte de estos resultados. Pero ojalá sigan sacando evidencia que vaya solidificando el caso de qué tan capaz es realmente Claude Mythos aplicado en el mundo real. Y además no es un gráfico de esos engañosos que se sacan algunos labs a veces (ejem), la diferencia realmente ha sido muy pronunciada.
Alex Albert@alexalbert__

With the help of Claude Mythos Preview, the Firefox team fixed more security bugs in April than in the past 15 months combined.

Español
3
2
15
1.7K
XpeGj0
XpeGj0@224Shroud·
@agusbuilds @gabmfrl No las cumplira, va a estar carisimo y aparte de eso, estara limitado por el computo. Simplemente están inflando el valor del modelo.
Español
1
0
2
29
Agus 🛠️ IA Aplicada
@gabmfrl Menudas expectativas está generando este modelo. Cuando lo lancen para todos, veremos si las cumple
Español
2
0
1
81
XpeGj0
XpeGj0@224Shroud·
@fbjoe @alex_whedon It better be true, I want open source models to adopt this architecture 🫩
English
0
0
0
46
joez.eth
joez.eth@fbjoe·
Look I would love to use it first before I call BS but so far it's empty promise just like you said and I am not disagreeing with you BUT Remember how Deepseek able to train their stuff simply by adjusting some params that are so simple to achieve desired outcome? I am saying what they promised might be possible but remain skeptical until I see it and try it out.
English
1
0
0
77
Alexander Whedon
Alexander Whedon@alex_whedon·
Hey, folks! We have been blown away by the response to SubQ and the SSA breakthrough over the last 48 hours. It is awesome to see how many people are responding to our mission of creating more efficient algorithms to create better models. We are working hard to firm up our release timeline and will share more very soon. We will also share additional data and third-party validation in our model card next week. If you have questions, please post them in the thread, and I'll do my best to respond! Above all, THANK YOU! The support, feedback, and discussion from this community have been inspiring.
English
76
24
451
48.7K
XpeGj0
XpeGj0@224Shroud·
@fbjoe @alex_whedon To overvalue their company. How long will it take the big Ai companies to copy and invest in this architecture? If it’s true, then this startup won’t have money to compete with the other models at the end it’s lose or lose. When they lose, they would lose but with a bag of money
English
1
0
0
65
joez.eth
joez.eth@fbjoe·
@224Shroud @alex_whedon looks like a larp announcement why announce it without giving ppl to try is this like trying to raise money? for VCs doing DD on this please actually try the product and do the benchmark runs.
English
1
0
0
168
90S KID
90S KID@epochster·
@digitalix 144GB HBM on PCIe. Impressive specs until you realize interconnect bandwidth is what actually limits distributed training at scale.
English
4
0
14
3.7K
Alex Ziskind
Alex Ziskind@digitalix·
New "GPU" with 144GB HBM - AMD Instinct MI350P, BUT IT's PCIe!
Alex Ziskind tweet media
English
55
37
682
53.8K
joez.eth
joez.eth@fbjoe·
@alex_whedon Question for anyone on X : has anyone got their hands on this subq coding plan yet?
English
3
1
1
3K
Ahmad
Ahmad@TheAhmadOsman·
Let me make local AI easy for you Give Codex Cli the tweet below & tell it: - Infer the right Inference Engine from your hardware + tweet content below - Use uv+venv - Pick the right kernels - Tune flags, batching, KVCache, etc - Optimize for your hardware & chosen model Enjoy
Ahmad@TheAhmadOsman

You don’t pick an Inference Engine You pick a Hardware Strategy and the Engine follows Inference Engines Breakdown (Cheat Sheet at the bottom) > llama.cpp runs anywhere CPU, GPU, Mac, weird edge boxes best when VRAM is tight and RAM is plenty hybrid offload, GGUF, ultimate portability not built for serious multi-node scale > MLX Apple Silicon weapon unified memory = “fits” bigger models than VRAM would allow but also slower than GPUs clean dev stack (Python/Swift/C++) sits on Metal (and expanding beyond) now supports CUDA + distributed too great for Mac-first workflows, not prod serving > ExLlamaV2 single RTX box go brrr EXL2 quant, fast local inference perfect for 1/2/3/4 GPU(s) setups (4090/3090) not meant for clusters or non-CUDA > ExLlamaV3 same idea, but bigger ambition multi-GPU, MoE, EXL3 quant consumer rigs pretending to be datacenters still CUDA-first, still rough edges depending on model > vLLM default answer for prod serving continuous batching, KV cache magic tensor / pipeline / data parallel runs on CUDA + ROCm (and some CPUs) this is your “serve 100s of users” engine > SGLang vLLM but more systems-brained routing, disaggregation, long-context scaling expert parallel for MoE built for ugly workloads at scale lives on top of CUDA / ROCm clusters this is infra nerd territory > TensorRT-LLM maximum NVIDIA performance FP8/FP4, CUDA graphs, insane throughput multi-node, multi-GPU, fully optimized pure CUDA stack, zero portability (And underneath all of it: Transformers → model architecture layer → CUDA / ROCm / TT-Metal → compute layer) What actually happens under the hood: > Transformers defines the model > CUDA / ROCm executes it > TT-Metal (if you’re insane) lets you write the kernel yourself The Inference Engine is just the orchestrator (simplified) When running LLMs locally, the bottleneck isn’t just “VRAM size” It isn’t even the model It’s: - memory bandwidth (the real limiter) - KV cache (explodes with long context) - interconnect (PCIe vs NVLink vs RDMA) - scheduler quality (batching + engine design) - runtime overhead (activations, graphs, etc) (and your compute stack decides all of this) P.S. Unified Memory is way slower than VRAM Cheat Sheet / Rules of Thumb > laptop / edge / weird hardware → llama.cpp > Mac workflows → MLX > 1–4 RTX GPUs → ExLlamaV2/V3 > general serving → vLLM > complex infra / long context / MoE → SGLang > NVIDIA max performance → TensorRT-LLM

English
9
21
343
29.4K
Aditya Bawankule
Aditya Bawankule@legorobotdude·
@224Shroud @0xSero Yeah the real issue is the thinking loop. Gemma won't do this since they baked in a system prompt but that's a trade off with other effects.
English
1
0
1
312
Hugging Models
Hugging Models@HuggingModels·
Ever wanted to generate high-quality images from text prompts? Meet Dustoevsky/pornmasterZImage_turboV35Fp8, a powerful text-to-image model from the US. It's built for fast, creative image generation, perfect for artists and developers alike.
Hugging Models tweet media
English
14
6
106
11.3K
Lotto
Lotto@LottoLabs·
>be me >have $2000 and no life >buy two RTX 3090s on eBay >"used but good condition" my ass one has thermal paste the color of mayonnaise from 2018 >open a window, close every other door in the house >plug them into my motherboard like I'm wiring a bomb >turn it on >noise level: commercial jet taking off >wifey: "what is that sound?" >"it's just... science" >she doesn't come back for 3 days >now have 48GB VRAM total >can run models that weigh more than my car >lm studio running at 103.69.27.87:1234 >serve cold LLM responses from a room that's basically an oven in July >electric bill arrives >stare at it for 20 minutes >"it'll be fine" >want to remote into my little homelab without exposing ports >install tailscale on everything >now I can SSH into my GPU rig from literally anywhere >my laptop connects like magic, no router config needed >mDNS, exit nodes, funnel — all working out of the box >"I'm basically a hacker now" >sit in airport at 2am running benchmarks from my phone >some random guy watches me typing furiously >he thinks I'm doing illegal stuff >I'm just waiting for qwen3.5-27b to finish inference on a poem about his mom >send him the output anyway >mfw I have more VRAM than most datacenters but less sleep than a college freshman
English
36
25
823
131.7K
XpeGj0
XpeGj0@224Shroud·
@TheAhmadOsman How about we talk about making the best LLM as a community!?!? And get rig of the big techs!?!?
English
0
0
1
216
Ahmad
Ahmad@TheAhmadOsman·
In the Bay Area for the next couple of weeks If you’re around and wanna grab food, coffee, or yap about GPUs / local AI / inference engines / the future of owning the stack Hit me up 🤙
English
24
4
141
16.6K
XpeGj0
XpeGj0@224Shroud·
@sven2401 @TheAhmadOsman Can’t scale it to 1T parameters, you would need it to be quantized at least 2 bits for the transistor to fit on a silicon warfare
English
0
0
0
12
sven2401
sven2401@sven2401·
@TheAhmadOsman It might be good for at home inference where you say okay GLM-5 / kimi k2.5 is good enough for my purposes and you might upgrade after 5-10 years or so like your pc. Assuming it scales to 1T models. But that would be a deliberate decision not wanting the newest best things
English
1
0
0
32
XpeGj0
XpeGj0@224Shroud·
@TheAhmadOsman Lowkey. Open source will only win if the community unites to make a model
English
0
0
0
150
Ahmad
Ahmad@TheAhmadOsman·
GPT-5.5 Pro is very impressive ngl I make this model take a pass at anything critical because it really can come up with useful feedback (Yes, I am still saying Opensource WILL WIN, these 2 things don't contradict each other)
English
23
2
183
8K
Alex Ziskind
Alex Ziskind@digitalix·
tough choice. sell or keep.
Alex Ziskind tweet media
English
51
2
140
13.5K
XpeGj0
XpeGj0@224Shroud·
@TheAhmadOsman Man 😭 I haven’t understand what’s the point of the dgx spak, bandwidth of a macbook pro.
English
0
0
0
65