Alex

865 posts

Alex banner
Alex

Alex

@wadeAlexC

Oolong maximalist. Bureaucracy minimalist. | Prev: breaking things @eigenlayer, @ConsenSysAudits, @EtherealVC

[REDACTED] 가입일 Temmuz 2018
311 팔로잉1.9K 팔로워
Alex
Alex@wadeAlexC·
@alpeh_v "made it not bad for you" no, i'll find a way. trust.
English
1
0
1
50
ʕ •ᴥ•ʔ
ʕ •ᴥ•ʔ@alpeh_v·
theres something so hopeful about the slot machine rougelike and the poker rougelike, like yes we took this super popular gambling game and made it not bad for you or extractive while also being more fun
English
6
0
19
1.1K
Alex
Alex@wadeAlexC·
ChatGPT's interface is hot garbage. There's very little contrast between different UI elements - it's black, on dark grey, on charcoal. It's peak modern design - everything's round and bubbly in a way that gives icons and elements a samey vibe. I really like Claude's interface. It's warmer overall, but they darken the users' message background so it pops out. They use a serif'ed font for Claude that makes it really easy to read (to me). They also use more straight lines in their icons/elements that create clearer borders to divide up the screen.
Alex tweet mediaAlex tweet media
English
1
1
1
104
StErMi
StErMi@StErMi·
I've reflected about it and the way ChatGTP answer is really overwhelming. There have been times that I closed the chat because I was lost in the answer. I wonder if I made some customizations way back in time and that's the result of it.
English
1
0
1
111
StErMi
StErMi@StErMi·
I have uploaded a blood test result to both @ChatGPTapp and @claudeai. ChatGPT really needs to improve drastically the way it "render" the output. Claude's UI/UX is orders of magnitude better than ChatGPT. The way ChatGPT writes back to you is always so disorganized and visually confusing, it's like it's puking a wall of text back to me with random bullet points and emojis.
English
1
0
5
694
Alex
Alex@wadeAlexC·
@0xClandestine that's wild. i wonder if it'd be faster if it also used RAM
English
1
0
1
20
clandestine.eth 🦇🔊
clandestine.eth 🦇🔊@0xClandestine·
@wadeAlexC Funny enough it doesn’t even use much of my 64gb ram, few gbs. Flash-moe aggressively streams the entire model from ssd.
English
1
0
0
33
Alex
Alex@wadeAlexC·
@Duffaluffaguss Not sure what you're asking. I haven't personally gotten into Openclaw, but you could for sure hook up a local model to it 😊
English
0
0
1
12
Alex
Alex@wadeAlexC·
I ditched cloud LLMs for self-hosted last year. 24/7 availability. My phone. My laptop. My GPU. It's faster than ChatGPT. Almost as smart. It's not being used by the DoD to conduct warrantless surveillance. No data leaves my server. I own everything. When things break, I can fix them myself. I can ask anything I want and get an answer back 10-20x faster than I can read. It costs pennies to run. Why isn't everyone doing this?
English
4
0
8
379
Alex
Alex@wadeAlexC·
A good way to test would be to try out some of the target models you'll want to run via OpenRouter. With 256-512 GB, you'll probably want to look at these: - Qwen3.5-397B-A17B - Minimax M2.5 - GLM 5 (you'll need quantized versions of these, and im not personally familiar with their game; I stick to models < 100B params)
English
0
0
0
28
Alex
Alex@wadeAlexC·
@Duffaluffaguss @LLMJunky I'm not 100% sure what your usecase is. You said "self-maintaining business with ongoing research" -- sounds like you don't care if your llm is slow, it just needs to be smart? Mac Studios are, what, 256-512 GB unified memory? Yeah, that'll let you run some really great models.
English
1
0
1
18
Alex
Alex@wadeAlexC·
@Duffaluffaguss Local models are capable of that, but you need an interface that feels intuitive to use and hides a lot of that complexity from the user. Ex: Qwen3.5 can't output images. If I wanted to add that feature, I would need my interface to do some fancy routing to image-gen models.
English
2
0
0
16
Alex
Alex@wadeAlexC·
@Duffaluffaguss For generic chat, 95% of the time you won't notice a difference between local and frontier. Frontier is only decent because their interface has a lot of features. For example, you go to chatgpt, and it can ingest and output images, do web search, write/execute code.
English
1
0
0
22
Alex
Alex@wadeAlexC·
@TheCesarCross Qwen3.5-35B-A3B is my go to. Absolute beast. I posted more details about my setup in the other replies :)
English
0
0
0
21
cCross
cCross@TheCesarCross·
@wadeAlexC Which models are you running on your setup?
English
1
0
0
8
Alex
Alex@wadeAlexC·
Models: Qwen3.5-35B-A3B and Qwen3.5-27B Unsloth Q4 quants for both. Both easily fit <24 GB VRAM at 100k context. Device: I have an NVIDIA RTX 5090 and keep everything on-GPU. However, with these models/quants, you can accomplish the same even with a 4090. I haven't tried any Mac products. I hear good things, but my understanding is you're getting slower inference (but maybe you can run larger models). Whether you want an NVIDIA setup or a Mac setup probably comes down to your usecase. - If you mainly want to support a "chat app" or "personal assistant", you want the nicest GPU you can get, because you're gonna want the speed. - If you want like... home assistant/automation/coding agents, maybe opt for something with better capacity for larger models? Mac Studio, DGX Spark, things like that. I can't advise as much here because I went the GPU route.
English
1
0
1
46
duffaluffaguss.eth
duffaluffaguss.eth@Duffaluffaguss·
@wadeAlexC What models and on what device? I'm trying to figure out what I need and how well it will perform and I feel like everything I read suggests something different. Mac studio with 256 ram, custom Nvidia setup. Which models. Ty ser
English
1
0
1
51
Alex
Alex@wadeAlexC·
@auryn_macmillan Qwen3.5-35B-A3B and Qwen3.5-27B Unsloth Q4 quants for both. Both easily fit <24 GB VRAM at 100k context.
English
1
0
1
31
Auryn
Auryn@auryn_macmillan·
@wadeAlexC Which models are you running?
English
1
0
0
53
Alex
Alex@wadeAlexC·
@0xClandestine Qwen3.5-35B-A3B is the goat for both speed and intelligence. Handles most of my daily usage.
English
0
0
1
47
Alex 리트윗함
Neeraj K. Agrawal
Neeraj K. Agrawal@NeerajKA·
He was found guilty on the charge that directly contradicts FinCEN's guidance. Unbelievable.
English
29
124
713
92.1K
Alex
Alex@wadeAlexC·
@nicht_tintin God damn that's pretty! Saved as inspiration :D
English
0
0
1
48