Yves crypto

3.5K posts

Yves crypto banner
Yves crypto

Yves crypto

@cryptoyves1

professional MOD/project manager / check #Brothersmarketing for good services

Beigetreten Kasım 2015
1.3K Folgt716 Follower
Angehefteter Tweet
Yves crypto
Yves crypto@cryptoyves1·
My trading bot for Solana @u55nv" target="_blank" rel="nofollow noopener">app.telemetry.io/@u55nv
English
0
0
0
36
Ace
Ace@AceAgain_·
Cheers to more life 🥂
Ace tweet media
English
130
4
134
2.9K
Ace
Ace@AceAgain_·
Gn 🌖
GIF
121
9
114
2.4K
Ace
Ace@AceAgain_·
Because you always active and I see you “Sorry I don’t have my wallets on me”
Ace tweet media
Kelly of web3@kellyofweb3_

@AceAgain_ Ggs Ace Kindly send me a sol so I can go again into the trenches FUuzRJjQRVxnLvW6Dfp9k3MVpZZfx9jnTbwQ4NaRMZDr

English
208
13
247
14.9K
PathtoMonad 🦖
PathtoMonad 🦖@openyourmind569·
@BitcoinChina_ Take a look at @YeeErc20 Haven‘t seen a better narrative since $Pepe Study $Yee The return of the dino is coded for billions. 🦖🫶 STUDY THE TRILOGY 🐶🐸🦖
English
10
9
15
81
Bitcoin China中国🇨🇳
Bitcoin China中国🇨🇳@BitcoinChina_·
Which memecoin has the potential to hit a billion dollar market cap? Shill me
English
82
5
44
5.5K
Ace
Ace@AceAgain_·
GIF
ZXX
50
3
69
1.9K
Ace
Ace@AceAgain_·
Gm ☀️
Ace tweet media
196
8
169
3.3K
zak.eth
zak.eth@0xzak·
Ok, what up y'all. I'm back in action. Took some time off. About a week and a half. First real break in a while. Spent the entire thing building my own AI inference cluster because apparently that's what I do for fun now. Picked up a pair of used NVIDIA DGX Stations. Each one has 4x Tesla V100 GPUs, 64GB VRAM, NVLink, 256GB RAM. Originally $69,000 each. Got both used for a less than $10k total. Connected the first one to my main workstation (Threadripper 3970X, RTX 3090) over a direct 10GbE ethernet link with jumbo frames. 8 V100s, one RTX 3090, 160GB of combined GPU memory across three machines sitting on my desk. Two of them are former enterprise servers that a Fortune 500 company probably depreciated off their books years ago. The model lineup running right now: - Qwen3 30B (MoE, 3B active params): 21 tokens/sec - GPT-OSS 20B (MoE, 3.6B active): 13 tokens/sec - Qwen2.5 Coder 32B: 3 tok/s, best open-source coding model at this size - DeepSeek R1 32B: 3 tok/s, reasoning model that benchmarks above o1-mini - Llama 3.3 70B: 1.5 tok/s, strongest overall open model MoE architectures are the move for this hardware. Only 3B parameters active per forward pass means you get 21 tok/s out of a 30B model. Dense 32B models crawl at 3 tok/s on V100s. The 70B is usable for batch work but you're not having a conversation at 1.5 tok/s. Biggest lesson I learned the hard way is that vLLM (the standard high-performance inference engine) straight up does not work on V100 GPUs. AWQ quantization requires compute capability 7.5+. V100 is 7.0. GPTQ kernel is documented as "buggy" and hung indefinitely. Marlin needs Ampere or newer. Three quantization backends, three failures. Spent days debugging before pivoting entirely to Ollama with GGUF models. llama.cpp just works on everything. NVIDIA also dropped V100 driver support starting at version 550. Locked to the legacy R535 branch forever. Enterprise hardware depreciates like a brick and the software ecosystem moves on without it. For comparison, a Mac Mini M4 Pro with 48GB unified memory ($2,000) runs 30B models at 12-18 tok/s and MoE models at up to 83 tok/s via MLX. Faster than a single DGX for single-model inference and uses 30 watts instead of 1,200. The Mac wins on efficiency, but it can't run 70B models, which is what I need for the research I'm doing. A 70B Q4 needs 42GB, leaving almost nothing for context in 48GB. One DGX handles it with room to spare. Two of them with 128GB of combined VRAM opens up models the Mac can't even attempt. Cost math: - Claude Max: $100-200/mo - ChatGPT Pro: $200/mo - Heavy API usage (Opus): Can be thousands/mo - Two DGX Stations electricity at full load: $180-300/mo for unlimited local inference across both Unlimited requests, complete data privacy, zero dependence on third-party APIs. The hardware pays for itself within a year and then it's just electricity forever. Local models do not replace Claude or GPT-4 (yet) for complex multi-file agentic coding. That gap is still massive and I'll be real about it. I use local for code completion, quick questions, reasoning tasks, RAG pipelines, and anything I don't want touching a third-party server. Cloud for the heavy agentic loops. Optimization stack, for anyone building something similar: - CPU governor locked to performance mode - Transparent hugepages enabled, swap disabled, vm.swappiness at 1 - NCCL for NVLink GPU communication - Models pinned in VRAM for 24 hours (no cold-start penalty) - 10GbE jumbo frames (MTU 9000) between machines - UFW firewall locked to only accept requests from my workstation Every knob I could find, turned. Next step is clustering the two DGX Stations together. 128GB VRAM across 8 GPUs. 10GbE is too slow for tensor parallelism across machines (you need InfiniBand) but pipeline parallelism and model routing work fine over ethernet. Run different models on different boxes, route requests to whichever one has capacity. This is me getting back to the grind. I'm diving back into agentic research, specifically the intersection of decentralized AI and Ethereum. If individuals can assemble competitive inference clusters from depreciated enterprise hardware, the economics of AI access change permanently. You don't need a data center. You need a couple used servers, some ethernet cable, and about a week of fighting NVIDIA drivers. The models are open, the hardware is cheap, and the software exists. Sovereign compute is the play. Decentralized AI starts in your garage. We just have to build it. Will be sharing some cool new stuff here in the coming days, so stay tuned!
zak.eth tweet media
English
19
1
58
4.1K
Dylan Rossi
Dylan Rossi@Dylan_HODL·
gm. what's the loudest community?
Dylan Rossi tweet media
English
72
2
34
26.1K
Ace
Ace@AceAgain_·
Cheers bros 🥂
Ace tweet media
English
80
7
87
2K
Cara X 🦖
Cara X 🦖@tronbullet·
Yeehaww I use a lot of ai agents for work, by grok I was able to generate clear videos in a couple of seconds. @OpenAI should learn a lot from @elonmusk TY For @grok
English
8
8
11
218
TassHub
TassHub@TassHubCrypto·
TASSHUB PITCH PIT HOSTED BY: @Mystik_Crypto🧙‍♂️ Tomorrow, Wednesday Mar. 4th, 2:00PM (EST) / 7:00PM (UTC) x.com/i/spaces/1jGXg… Don't miss our next Pitch Pit episode, featuring: @YeeErc20 @MeanPippin @Whaleminersol Join your favorite hosts and tune in live with TASSHUB for our 3x weekly Spaces, where we showcase promising projects and talk some Tass! Dive deeper into Tasshub: tasshub.com
English
68
75
146
10.4K
B1 🦖
B1 🦖@BrotherMKT·
The one sequence of letters and numbers that should be priority right now. 0x9ac9468e7e3e1d194080827226b45d0b892c77fd
B1 🦖 tweet media
English
28
51
211
23.6K
buytech 🤖
buytech 🤖@BuyBotTech·
@Gemini The one sequence of letters and numbers that should be priority right now. 0x9ac9468e7e3e1d194080827226b45d0b892c77fd
buytech 🤖 tweet media
English
12
11
23
268