stableAPY.hl

2.1K posts

stableAPY.hl banner
stableAPY.hl

stableAPY.hl

@stableAPY

Building HyperFolio - HyperEVM Portfolio Tracker

Katılım Şubat 2025
233 Takip Edilen2K Takipçiler
Sabitlenmiş Tweet
stableAPY.hl
stableAPY.hl@stableAPY·
I just released a new feature for Hyperfolio and the Hyperfolio API: Yield. You can now browse thousands of yield opportunities on HyperEVM across nearly 30 protocols
English
10
5
31
1.9K
stableAPY.hl
stableAPY.hl@stableAPY·
Composer 2 Fast is insane for code reviews + fixes need to try this model for normal dev see what it can do
English
0
0
0
39
stableAPY.hl
stableAPY.hl@stableAPY·
I think I'll get some GPUs soon—your posts are giving me so much FOMO. I want to build a 100% local personal assistant for privacy using Hermes: - 1x 3090 → Qwen 3.5 35B A3B or 27B depending on the tasks - 1x 3060 → Qwen 3.5 4B (maybe 9B) for Honcho memory models + embedding model I might have some good results without leaking my personal informations to OpenAI or Anthropic
English
0
0
0
69
stableAPY.hl
stableAPY.hl@stableAPY·
spent 3 days trying to fine-tune a Qwen 3.5 2B into a terminal assistant and here's what happened I wanted some sort of small LLM living in my terminal. The idea was to have a local macOS assistant that takes natural language and outputs shell commands for this qwen 3.5 2B 8bit seemed the best fit on my m1 pro, 32gb ram. low RAM usage while keeping decent intelligence to evaluate the fine tuning I use an automated agentic benchmark this 60 real world queries run 0 raw model scored 202/300 (67%) on 60 real-world queries. not great but worked for basic stuff and this validated that this was possible to push Qwen3.5 2B run 1 first LoRA pass, 613 training examples, 500 iters. score: 197/300, it got worse runs 2-5 added QA examples, deduped, fixed edge cases. score: 240/300, slow gains run 6 last night I left Opus 4.6 with a set of instructions in order to fine tune this model this morning: 783 corrective examples generated from benchmark failures, a new fine tuning with 600 iters, batch_size 1 score: 274/300. 91% success rate ain't bad for such a small model results on my 60-query benchmark: 67% to 91% thanks to fine tuning, sounds great right? then i tested on queries NOT in the training set, to make sure I was not overfitting and result was 100% The fine-tune wins: no more infinite loops, proper macOS commands, clean output, even tho the base model sometimes picks smarter commands, idk why fine-tuning a 2B model doesn't teach it to reason, it teaches it to pattern-match your dataset and in my case pattern-matching is exactly what I needed, the base model outputs garbage loops and linux commands on macOS and fine-tununing mostly fixed 783 examples is enough to fix format, but not enough to teach judgment, so I'm going further what i'm doing next: scaling to 1500+ diverse examples generated by sonnet 4.6 subagents, with dedup and quality filtering it was fun to watch one AI teaching another AI to use a terminal where I'm at: score: from 202 to 274/300 intent matching: 22% to 78% ~20 to 30 min per training run on my M1 pro. finally, I think the real enemy isn't overfit, it's that 2B params is a really tiny brain, this way every pattern you teach it seems to erases another one Opus tells me that 783 examples is enough so I'm scaling to 1500+ diverse examples generated by sonnet 4.6 subagents, with dedup and quality filtering 3 days ago I knew nothing about fine tuning, model training and was not even aware I could do that locally on my macbook Idk if'll ever reach my goal and if I'm not just overfitting the fuck out of the model, but at least it's pretty fun to do and I'm learning tonnes of stuff
stableAPY.hl tweet media
English
1
1
3
266
stableAPY.hl
stableAPY.hl@stableAPY·
@_weiping wait this model is Kimi k2.5 level on multiple benchmarks?
English
0
0
1
308
Wei Ping
Wei Ping@_weiping·
🚀 Introducing Nemotron-Cascade 2 🚀 Just 3 months after Nemotron-Cascade 1, we’re releasing Nemotron-Cascade 2: an open 30B MoE with 3B active parameters, delivering best-in-class reasoning and strong agentic capabilities. 🥇 Gold Medal-level performance on IMO 2025, IOI 2025, and ICPC World Finals 2025: • Capabilities once thought achievable only by frontier proprietary models (e.g. Gemini Deep Think) or frontier-scale open models (i.e. DeepSeek-V3.2-Speciale-671B-A37B). • Remarkably high intelligence density with 20× fewer parameters. 🏆 Best-in-class across math, code reasoning, alignment, and instruction following: • Outperforms the latest Qwen3.5-35B-A3B (2026-02-24) and even larger Qwen3.5-122B-A10B (2026-03-11). 🧠 Powered by Cascade RL + multi-domain on-policy distillation: • Significantly expand Cascade RL across a much broader range of reasoning and agentic domains than Nemotron-Cascade 1, while distilling from the strongest intermediate teacher models throughout training to recover regressions and sustain gains. 🤗 Model + SFT + RL data: 👉 huggingface.co/collections/nv… 📄 Technical report: 👉 research.nvidia.com/labs/nemotron/…
Wei Ping tweet media
English
28
96
571
55.6K
stableAPY.hl
stableAPY.hl@stableAPY·
trying Opus 4.6 generated interactions rather than Sonnet, let's see
stableAPY.hl tweet media
English
0
0
0
33
stableAPY.hl
stableAPY.hl@stableAPY·
last training on 1500 example make the model worst I'll try to make Opus generate the interaction dataset not Sonnet, maybe I'll get better results also I think my benchmark is too naive, maybe running it multiple times and average the results for more accurate measurements
stableAPY.hl tweet media
stableAPY.hl@stableAPY

spent 3 days trying to fine-tune a Qwen 3.5 2B into a terminal assistant and here's what happened I wanted some sort of small LLM living in my terminal. The idea was to have a local macOS assistant that takes natural language and outputs shell commands for this qwen 3.5 2B 8bit seemed the best fit on my m1 pro, 32gb ram. low RAM usage while keeping decent intelligence to evaluate the fine tuning I use an automated agentic benchmark this 60 real world queries run 0 raw model scored 202/300 (67%) on 60 real-world queries. not great but worked for basic stuff and this validated that this was possible to push Qwen3.5 2B run 1 first LoRA pass, 613 training examples, 500 iters. score: 197/300, it got worse runs 2-5 added QA examples, deduped, fixed edge cases. score: 240/300, slow gains run 6 last night I left Opus 4.6 with a set of instructions in order to fine tune this model this morning: 783 corrective examples generated from benchmark failures, a new fine tuning with 600 iters, batch_size 1 score: 274/300. 91% success rate ain't bad for such a small model results on my 60-query benchmark: 67% to 91% thanks to fine tuning, sounds great right? then i tested on queries NOT in the training set, to make sure I was not overfitting and result was 100% The fine-tune wins: no more infinite loops, proper macOS commands, clean output, even tho the base model sometimes picks smarter commands, idk why fine-tuning a 2B model doesn't teach it to reason, it teaches it to pattern-match your dataset and in my case pattern-matching is exactly what I needed, the base model outputs garbage loops and linux commands on macOS and fine-tununing mostly fixed 783 examples is enough to fix format, but not enough to teach judgment, so I'm going further what i'm doing next: scaling to 1500+ diverse examples generated by sonnet 4.6 subagents, with dedup and quality filtering it was fun to watch one AI teaching another AI to use a terminal where I'm at: score: from 202 to 274/300 intent matching: 22% to 78% ~20 to 30 min per training run on my M1 pro. finally, I think the real enemy isn't overfit, it's that 2B params is a really tiny brain, this way every pattern you teach it seems to erases another one Opus tells me that 783 examples is enough so I'm scaling to 1500+ diverse examples generated by sonnet 4.6 subagents, with dedup and quality filtering 3 days ago I knew nothing about fine tuning, model training and was not even aware I could do that locally on my macbook Idk if'll ever reach my goal and if I'm not just overfitting the fuck out of the model, but at least it's pretty fun to do and I'm learning tonnes of stuff

English
1
0
0
114
stableAPY.hl
stableAPY.hl@stableAPY·
@0xSero by end of the year half of those people will be running your quantized models on consumer hardware I believe
English
0
0
0
29
0xSero
0xSero@0xSero·
20 million impressions
0xSero tweet media
English
9
1
101
3.9K
stableAPY.hl
stableAPY.hl@stableAPY·
@AndDegen step 1: ask any intelligent LLM (opus, codex) to do research on best practice, science paper on the best way to do it on your config step 2: vibe with the IA and ask a lot of questions when you don't understand and be critical worked for me
English
0
0
1
24
stableAPY.hl retweetledi
HyperFlow
HyperFlow@HyperFlow_fun·
HyperFlow just got a major upgrade. 1/ HyperFlow Execution Relay → Power-user fee rates for everyone. Lower fees on every trade, automatically. 2/ Split Orders (Core + EVM) → One trade, two engines. We split across Hyperliquid Core and EVM to guarantee the best rate. 3/ Portfolio Dashboard → All positions, All Chains, PnL, History. Everything you need. One screen. 4/ Swap-Bridge UI → Cross-chain swaps and bridging. One clean interface. $760M+ volume. No incentives. Just better trading. → alpha.hyperflow.fun
English
7
2
16
790
stableAPY.hl
stableAPY.hl@stableAPY·
fine-tuning a small model on your own data is the fastest way to learn you don't have enough data three attempts on my m1, three worse outputs, base model won every time at least my 32gb ram got a workout
stableAPY.hl@stableAPY

this second fine tuning was meh generated a dataset with 30/70 real interactions/synthetic data, let's see if I can have better results this way pretty fun to fine tune qwen 3.5 2B locally on my M1 Pro 32gb of ram, the process eats all my ressources tho

English
0
0
2
134
stableAPY.hl retweetledi
Honcho
Honcho@honchodotdev·
The best thing? Honcho gets even better long term.
stableAPY.hl@stableAPY

@3rosika feels like it helps a lot! Hermes has stopped forgetting stuff it used to before, let's see over the weeks

English
1
2
2
523
stableAPY.hl
stableAPY.hl@stableAPY·
Hermes agent with glm-5-turbo is dang good
stableAPY.hl@stableAPY

after seeing @sudoingX heavily advocating for Hermes, I've finally switched from OpenClaw I paired it with the new GLM-5-Turbo from my Max plan for now I'm looking for a cheap 3090 to set up a local personal assistant using Qwen 3.5 35B A3B, and 27B

English
0
0
0
208
stableAPY.hl
stableAPY.hl@stableAPY·
@3rosika feels like it helps a lot! Hermes has stopped forgetting stuff it used to before, let's see over the weeks
English
0
0
2
549
stableAPY.hl retweetledi
0xSero
0xSero@0xSero·
Putting out a wish to the universe. I need more compute, if I can get more I will make sure every machine from a small phone to a bootstrapped RTX 3090 node can run frontier intelligence fast with minimal intelligence loss. I have hit page 2 of huggingface, released 3 model family compressions and got GLM-4.7 on a MacBook huggingface.co/0xsero My beast just isn’t enough and I already spent 2k usd on renting GPUs on top of credits provided by Prime intellect and Hotaisle. ——— If you believe in what I do help me get this to Nvidia, maybe they will bless me with the pewter to keep making local AI more accessible 🙏
0xSero tweet media
Michael Dell 🇺🇸@MichaelDell

Jensen Huang is loving the new Dell Pro Max with GB300 at NVIDIA GTC.💙 They asked me to sign it, but I already did 😉

English
135
327
2.7K
485K
stableAPY.hl
stableAPY.hl@stableAPY·
@AndDegen Idk about the fine tuning, but yes on the 16gb you can run small models
English
0
0
1
38