Python Papi

7.7K posts

Python Papi banner
Python Papi

Python Papi

@Python_Papi

The Flame Flickers Wildly Right Before It Goes Out. Keep It Steady. SWE G . AI Tyrant . Gucci? AI Media Studio LLC. Come Find Out 🫴🏾 Link in bio...

Earth Katılım Şubat 2025
44 Takip Edilen237 Takipçiler
Python Papi
Python Papi@Python_Papi·
@sama I used the wrong email address when I applied 😞 I was supposed to use my business email, it's the one I actually use for Codex AIMediaStudioLLC@outlook.com It'd be awesome if you could reevaluate my invitation based on this new discovery 😒 If not, the 10x will do 😞
English
2
0
1
896
Sam Altman
Sam Altman@sama·
we are gonna do something nice for everyone who applied for the GPT-5.5 party and that we didn't have space for. hope you enjoy!
English
1.2K
163
7.4K
597.3K
Python Papi
Python Papi@Python_Papi·
@Cobratate @f_jxr_ Bro listen, I can divide and conquer them all day long, but when they start forming formations and shit, it's best to let them run amuck (to a degree) and let them tire themselves out. I Didn't pick them because they're push overs
English
0
0
3
1.9K
TABI 🐲
TABI 🐲@tabi_fit·
rate my flexibility :3 🍌
TABI 🐲 tweet media
English
67
57
4.2K
24.1K
Python Papi
Python Papi@Python_Papi·
altworldhq.com We Train It, YOU Profit Come get affiliated with the best. *** AI Media Studio LLC ***
English
2
0
0
61
Python Papi
Python Papi@Python_Papi·
@boardyai How can I introduce you to my AI Agent. He's been acting weird the past few days. Maybe he just needs a Boyz night out.
English
0
0
0
31
Python Papi
Python Papi@Python_Papi·
@TheAhmadOsman I thought I was done using uv when I finished exploring Smart Contracts 😔
English
1
0
2
329
Ahmad
Ahmad@TheAhmadOsman·
Let me make local AI easy for you Give Codex Cli the tweet below & tell it: - Infer the right Inference Engine from your hardware + tweet content below - Use uv+venv - Pick the right kernels - Tune flags, batching, KVCache, etc - Optimize for your hardware & chosen model Enjoy
Ahmad@TheAhmadOsman

You don’t pick an Inference Engine You pick a Hardware Strategy and the Engine follows Inference Engines Breakdown (Cheat Sheet at the bottom) > llama.cpp runs anywhere CPU, GPU, Mac, weird edge boxes best when VRAM is tight and RAM is plenty hybrid offload, GGUF, ultimate portability not built for serious multi-node scale > MLX Apple Silicon weapon unified memory = “fits” bigger models than VRAM would allow but also slower than GPUs clean dev stack (Python/Swift/C++) sits on Metal (and expanding beyond) now supports CUDA + distributed too great for Mac-first workflows, not prod serving > ExLlamaV2 single RTX box go brrr EXL2 quant, fast local inference perfect for 1/2/3/4 GPU(s) setups (4090/3090) not meant for clusters or non-CUDA > ExLlamaV3 same idea, but bigger ambition multi-GPU, MoE, EXL3 quant consumer rigs pretending to be datacenters still CUDA-first, still rough edges depending on model > vLLM default answer for prod serving continuous batching, KV cache magic tensor / pipeline / data parallel runs on CUDA + ROCm (and some CPUs) this is your “serve 100s of users” engine > SGLang vLLM but more systems-brained routing, disaggregation, long-context scaling expert parallel for MoE built for ugly workloads at scale lives on top of CUDA / ROCm clusters this is infra nerd territory > TensorRT-LLM maximum NVIDIA performance FP8/FP4, CUDA graphs, insane throughput multi-node, multi-GPU, fully optimized pure CUDA stack, zero portability (And underneath all of it: Transformers → model architecture layer → CUDA / ROCm / TT-Metal → compute layer) What actually happens under the hood: > Transformers defines the model > CUDA / ROCm executes it > TT-Metal (if you’re insane) lets you write the kernel yourself The Inference Engine is just the orchestrator (simplified) When running LLMs locally, the bottleneck isn’t just “VRAM size” It isn’t even the model It’s: - memory bandwidth (the real limiter) - KV cache (explodes with long context) - interconnect (PCIe vs NVLink vs RDMA) - scheduler quality (batching + engine design) - runtime overhead (activations, graphs, etc) (and your compute stack decides all of this) P.S. Unified Memory is way slower than VRAM Cheat Sheet / Rules of Thumb > laptop / edge / weird hardware → llama.cpp > Mac workflows → MLX > 1–4 RTX GPUs → ExLlamaV2/V3 > general serving → vLLM > complex infra / long context / MoE → SGLang > NVIDIA max performance → TensorRT-LLM

English
9
15
265
22.2K
Young Kings
Young Kings@youngkingsgrow·
Young Kings. Broke people try to make money because they want to buy things. Rich people try to make money because they want to meet people. UNDERSTAND. There is a reason a millionaire wants to be a billionaire. It's about people and power, not junk from Walmart.
English
31
72
458
47.3K
Python Papi
Python Papi@Python_Papi·
@tunguz Fair. But at least you may be able to learn something from your interaction.
English
0
0
0
12
Bojan Tunguz
Bojan Tunguz@tunguz·
@Python_Papi Except that I know so many really high IQ individuals who are some of the worst people I've ever had misfortune to come across.
English
1
0
1
66
Wendy O
Wendy O@CryptoWendyO·
Crypto things you might have missed: -$XRP treasury Evernorth adds OpenAI CFO to board -XRPL = hot spot for tokenized assets -BlackRock fights OCC on tokenized caps -Western Union stablecoin launches on Solana -WLFI files lawsuit vs Justin Sun -Strategy pauses $BTC purchases
English
14
9
89
5.8K
TABI 🐲
TABI 🐲@tabi_fit·
standing like i’m the main character in an anime.. 😼
TABI 🐲 tweet media
English
40
110
4.9K
23K
Python Papi
Python Papi@Python_Papi·
@katie_haun *** AI Media Studio LLC *** We Specialize In Merging AI and Crypto ™️ The Inevitable Marriage ™️ Launched - altworldhq.com - Crypto Training Platform that gives real time Buy/Sell signals based on training Data Crypto Land - Casino Land - Coming Soon
English
0
0
0
17
Rihanna
Rihanna@rihanna·
it’s giving everything but basic for your everyday basics 🍓 new cozy Cotton Essentials, reworked for all-day comfort AND available now.
Rihanna tweet mediaRihanna tweet mediaRihanna tweet media
English
6.6K
13K
238.4K
11.7M
kache
kache@yacineMTB·
Being a parent is great. You guys should have kids. I really mean it. You guys should have as many kids as you can
English
204
98
2.2K
203.1K