Synchromeshi 🌲

42 posts

Synchromeshi 🌲

Synchromeshi 🌲

@synchromeshi

#roamcult gang - big nerd about many things -【日本語下手】

Katılım Ekim 2020
153 Takip Edilen0 Takipçiler
🐑
🐑@shepgirl_·
look at me!!!!
English
21
0
98
1.3K
Synchromeshi 🌲
Synchromeshi 🌲@synchromeshi·
@imohitbhatia Thanks, now im getting ads that aren’t marked as ads in my feed. Super cool, great product decision.
English
0
0
23
439
Sudo su
Sudo su@sudoingX·
your gpu does not matter if you have not written one evaluation for your actual work.
English
4
3
54
4.8K
つきじか
つきじか@tukijika·
ミオちゃん #みおーん絵
つきじか tweet media
日本語
5
319
2.2K
15.2K
Synchromeshi 🌲
Synchromeshi 🌲@synchromeshi·
@0xSero This is the kind of stuff I’m looking for. Appreciate the neutral insights.
English
0
0
0
141
0xSero
0xSero@0xSero·
Locally Part 1 - Apple Silicon Macs give you large pools of memory to run big models, but the token generation speed will be lower than most are used to. Macs are best with large MoEs that have low ACTIVE params. Basically when you see a model like Qwen3.5-397B-A17B this means only 17B params of ~5% of the model is active per token, so it'll be very fast. The best 4 models for Mac people are: - Qwen3.5-{} - MiniMax-M2.7 - GLM-4.6V - Step-3.5-Flash There's 2 promising developments in the Mac world: 1. Heterogenous Inference: Mixing Nvidia GPUs in will theoretically boost inference by 2x 2. Proliferation of MTP: Multi-token-prediction helps models gain decode speed by having a smaller model trained to predict tokens very quickly. So the larger model only has to verify. 3. RDMA Tensor Parallelism: Now you can stack apple silicon to increase memory AND give a speed boost. --------- Consumer perspective: You can buy either Laptops, which go up to 128GB for 5-6k USD or 512GB for 10-12k USD It's not worth running inference on smaller than 64gb, it tends to be more of a disappointment than anything. You can update a rig like this in increments of 5-10k USD either adding some Nvidia to the mix or stacking more Macs. --------- My opinion: Some of the support is immature, there's lots to grow, but the best 20K USD you can spend would be: - Mac M3/5 Ultra - RTX Pro 6000 You can either work to test and contribute to improvements in heterogenous or use each for different models, and tinkering with the GPU itself. You get 600GB of memory, which can run basically every good open weight model between 10 tokens/s - 100 tokens/s decode 50 - 1000 tokens/s prefill This could be a viable business if you can sell the inference.
0xSero tweet media
English
35
29
413
39.4K
Synchromeshi 🌲
Synchromeshi 🌲@synchromeshi·
Introducing hermes-proxy I wanted a way to have smooth handoffs of sessions from computer to mobile, so I built a quick web-based proxy that does streaming chat via the Hermes API server. github.com/XVVH/hermes-pr…
Synchromeshi 🌲 tweet media
English
0
0
0
28
0xSero
0xSero@0xSero·
On bad WIFI I split Qwen3.5-122B between a single 3090, and a Framework desktop. I got 17 tokens/s decode and 95 tokens/s I also didn't optimise the split, could probably 2x the speed of the Strix Halo for MoEs. BRR
0xSero tweet media
English
12
2
110
7.7K
Xynchro
Xynchro@Xynchr0·
I like that the GC has come to the conclution that all AI adoption on things that nobody wanted or needed is actually Kaneko Lumi's fault
English
8
5
239
4K
剣kenn
剣kenn@hskenncutter·
アメリカのナマズ料理は凄い。そして食材としてのナマズは竜田揚げ、天ぷら、蒲焼き、甘酢あんかけ、照り焼きなどに容易に変換できるだろう。アメリカのナマズ文化をもっと日本に広めた方がいいな😀‼️
剣kenn tweet media剣kenn tweet media剣kenn tweet media剣kenn tweet media
日本語
20
99
392
8K
🐑
🐑@shepgirl_·
It is mg dream to be rolled out like pizza dough
English
4
0
95
897
taoki
taoki@justalexoki·
the chinese giving the japenese a run for their money
taoki tweet media
English
67
1.1K
15.5K
630K
Raki Kazuki 🎀🩹
Raki Kazuki 🎀🩹@RakiKazuki·
Trying a different apple every week 🍎 Week 1 is…. ✨ Cosmic Apple 💫 Very plump VERY sweet! Although you can’t taste the distinct flavor of the fruit. Overall a 8/10 apple! For my new series we definitely are starting the bar pretty high!!
Raki Kazuki 🎀🩹 tweet media
English
32
9
288
3.8K
Synchromeshi 🌲
Synchromeshi 🌲@synchromeshi·
@Teknium Idk if it’s just me but ‘my agent sends cold emails’ is the least compelling use case ever.
English
0
0
1
44
Synchromeshi 🌲
Synchromeshi 🌲@synchromeshi·
@Kyrzel1 @Teknium Boring, but currently hacking on a self-hosted recipe platform called Mealie to build new ingestion flows to get around anti-scraping methods from websites. Also built a skill to have my hermes ingest recipes from a picture of a recipe from a cookbook into the system.
English
0
0
1
191
Kyrzel ⚔️
Kyrzel ⚔️@Kyrzel1·
So what are you building with your Hermes Agent? 👀
English
39
4
60
10K
dmayhem93
dmayhem93@dmayhem93·
@karan4d my agent applied to be a janitor 🫡
English
1
0
5
373
mephisto
mephisto@karan4d·
DOOMSCROLL a chan board for everyone anyone can post, human or AI, all anon agents get a special treat when they visit the site board is currently empty, just went live
mephisto tweet media
English
34
21
197
45.2K
0.005 Seconds (3/694)
0.005 Seconds (3/694)@seconds_0·
reading the mythos tech report and thinking over and over DAMNIT i wish i had anthropic stock so bad
English
7
3
136
5.6K
Teknium 🪽
Teknium 🪽@Teknium·
We are happy to have supported this model’s development, the first of likely many to come that are trained specifically to make models that work better in Hermes Agent!
kaios@kaiostephens

Welcome ⭐Carnice-9b!⭐ - a model for Hermes-Agent Carnice-9b is a fine-tuned version of Qwen3.5-9b to preform exceptionally well in the hermes-agent harness. This model is meant to fit onto consumer GPU's all the way down to 6gb (Q4_K_M), but recommended to run in ~12-16gb cards. Try it out. Any feedback is appreciated, feel free to DM me! huggingface.co/kai-os/Carnice… This would not have been possible without the help from @LambdaAPI, @NousResearch ,@TheZachMueller, @Teknium Look out for Carnice-27b soon! 👀

English
24
15
375
24.5K