Dimasya

3.2K posts

Dimasya banner
Dimasya

Dimasya

@satanworker

Previously @aave and @bgdlabs @safe_space_app - encrypted iOS media folder Currently @ValantisLabs opinions are my own

انضم Ekim 2014
2K يتبع563 المتابعون
تغريدة مثبتة
Dimasya
Dimasya@satanworker·
I really can’t believe, but SafeSpace 1.6.0 is out. Although it’s not an impressive number, it’s a massive milestone for us. Now it’s possible to store videos! Let it sink for a while how far mobile technologies advanced.
English
2
0
23
0
Микрофасолина
Микрофасолина@tireduckquack·
Кто бросал курить? Расскажите как смогли
Русский
430
2
141
43.9K
Silly Swap
Silly Swap@sillyswap_xyz·
Private swaps on Hyperliquid just got an upgrade > Light mode added > Smart Routing for multi-hop swaps > New markets added: $ZEC, $SOL, $ETH, $BTC > $USDH quote route added > New @silhouette_ex fee model added What else is missing?
English
4
7
33
4.8K
Dimasya
Dimasya@satanworker·
I just claimed my .agent domain and joined the .agent community! get yours now and help shape the future of autonomous agents #RWLZHZG2" target="_blank" rel="nofollow noopener">agentcommunity.org/join#RWLZHZG2 @agentcommunity_
English
0
0
0
52
Dimasya
Dimasya@satanworker·
Opened twitter from vacation today and the timeline looks absolutely wicked
English
0
0
0
58
Bobby Bobak ☕️🇵🇱
Bobby Bobak ☕️🇵🇱@bobek_balinek·
I was the same, ditched Realm for CoreData as soon as MongoDB has announced their acquisition in 2019. It was a crippling decision in hindsight. I've had to work around so many weird constraints. What I've realised is that the frustration with CD/SwiftData is common enough that a project like SQLiteData has every right to reach stable, long-term commitment as an OSS.
English
1
0
9
3.7K
Dimasya
Dimasya@satanworker·
@0xSero I find Droid just sometimes hang when starting implementation, and also it spawns headless chromium which prevents my Mac from sleep :D
English
1
0
0
875
0xSero
0xSero@0xSero·
Here's my current stack ---------- Models: 1. GPT-5.4 - Research + Programming + Parchi + OC 2. Kimi-K2.5 - Frontend Dev + Computer Use + Parchi 3. Opus-4.6 - Writing + Review GPT + Fallback 4. Gemini-3.1-Pro - Writing + Research + Reviews 5. Secret Model - Very fast + Devops + Cheap ----------- Harnesses: 1. Codex - It's the highest value play rn 2. Droid - Best multi-model harness & missions 3. Parchi - Browser use (I genuinely think this is the best one on the market) ------------ Apps: 1. Codex / Opencode - Kitty and OC for OSS models 2. KittyLitter - Mobile Codex access ------------- Terminals & IDEs: 1. Cmux - When I need to read while I prooompt 2. Zed - When I want to read the code 3. Ghostty - GOAT simple terminal -------------- Browser: 1. Brave - I am open to Helium if u wanna tell me what makes it better -------------- Infra: 1. Tailscale - Securely linking all your devices 2. VibeProxy - Use your subscriptions anywhere 3. Amphetamine - Keep computer awake 24/7 -------------- Quality of life: 1. Superwhisper - AI voice transcription 2. Raycast - Clipboard / Calculator / shortcuts 3. OBS - I record a lot of stuff and don't pay :p
0xSero tweet media
English
64
68
1K
60.1K
Thomas Paul Mann
Thomas Paul Mann@thomaspaulmann·
Glaze apps 🤝 MCP servers Left: An app to monitor your system resources Right: @claudeai asking "which process is eating memory?" via MCP Simple example but powerful concept.
Thomas Paul Mann tweet media
English
21
7
139
9.6K
Dimasya
Dimasya@satanworker·
The sweet gap between how things suppose to work and how things really work
English
0
0
0
54
Dimasya أُعيد تغريده
bunny
bunny@ConejoCapital·
what another day of prompting feels like
English
36
190
2.1K
83.7K
Dimasya
Dimasya@satanworker·
@0xSero wait did you realy moved to opencode GUI?
English
1
0
0
287
0xSero
0xSero@0xSero·
I think a lot about how I can run Kimi in a recognizable form at home, at over 20 tokens/s I don't think models smaller than 1T will be able to provide high quality consistent responses across a wide range of topics, the math just doesn't work. I would mark MiniMax as the golden standard for how much you can stuff into an LLM, it's very good at it's narrow set of work, it makes mistakes but it works and is good enough. But to get there you have to sacrifice multi-modal, world knowledge, hallucination rate increases, and multi-lingual. There's a lot of world knowledge in there that is important. ----------- I tried a bunch of things, for example extreme REAPs, it would just collapse after 60% reduction. I tried Sherry ternary quantisation, LoRA to recover accuracy, 1 bit quants. The closest we can get is the 1 bit unsloth quants but vllm doesn't run it fast or at all so what's the point. ------------- New idea: What if you can provide a gigantic dataset to the model and watch it's outputs, then cluster all the weight activations for certain topics like: - Language - Coding - Agentic reasoning - Image description - Multi-Lingual - Casual - Assistant - Roleplay There will be a lot of overlap and fuzzy places, but nothing 1 LoRA can't fix. Now you map your topics to expert activations, next you load your model architecture into vllm, without any weights. You create 10 REAPs each one zeros out everything but the topics associated experts, so you can get it down to 1.25-2 bits pre on average pre-8bit quantisation. Your router will the load and unload expert weights, process the context, then get unloaded, you can basically maintain 16 slices of experts of the model, that get loaded and unloaded dynamically with very little latency if stored on CPU, each slice maintains the original 8 bit weights of the specific topic. So basically you introduce latency only between switching of experts, and you can make that more or less of a problem based on how small you want to go. This means you can run the original 8bit weights on 3.75% of the original FP16 without losing any fidelity in model responses. Will this work over a few turns? it surely will over 1 turn, I've demoed this. the latency is negligible as long as I am not switching the experts too often.
0xSero tweet media
English
9
6
76
5.7K
@sitnik_ru@mastodon.social
@[email protected]@andrey_sitnik·
Очень милый пост, напоминающий нам, что несмотря на трудности эмиграции, детская версия вас очень гордилась бы вашим путём instagram.com/p/DVaqHgvDUDA/
@sitnik_ru@mastodon.social tweet media
Русский
5
4
152
7.2K
Dimasya
Dimasya@satanworker·
that's what I pay internet for
English
0
0
1
69
Dimasya أُعيد تغريده
Valantis Labs
Valantis Labs@ValantisLabs·
Introducing Valantis Prime. Unifying the HyperEVM and HyperCore via a singular account layer:
Valantis Labs tweet media
English
8
12
116
7.2K