Sipeed

5.2K posts

Sipeed banner
Sipeed

Sipeed

@SipeedIO

AIoT opensource hardware platform

中华人民共和国 Katılım Eylül 2018
236 Takip Edilen24.1K Takipçiler
Sipeed
Sipeed@SipeedIO·
@rikas_ilamdeen Hi, you are the first one, NanoKVM is right, do you want NanoKVM or NanoKVM-Pro?
English
1
0
1
86
Rikas
Rikas@rikas_ilamdeen·
@SipeedIO 1. LicheeRV Nano 2. NanoKVM / NanoKVM Pro 3. MaixCAM / MaixCAM2
Español
1
0
7
232
Sipeed
Sipeed@SipeedIO·
#Picoclaw Just hit 25K star~ We’re integrating PicoClaw into 3 offical hardware devices! 💻 Drop your guesses in the comments below—the first person to guess a device correctly will win it for FREE! 🎁👇 #openclaw github.com/sipeed/picoclaw
Sipeed tweet media
English
7
6
25
2.2K
Lucky13
Lucky13@Lucky1386727045·
@SipeedIO Looks awesome! 🔥 When can we expect it to be released?
English
1
0
0
104
Sipeed
Sipeed@SipeedIO·
New toy~ Testing out the macro thermal imaging!
Sipeed tweet mediaSipeed tweet media
English
1
3
50
3K
Sipeed
Sipeed@SipeedIO·
#SLogic32U3 8CH capture @ 800 MHz — PASS ✅ Streaming bandwidth **800 MB/s** ! A New Era for Logic Analyzers Begins. 🚀
Sipeed tweet media
English
3
8
77
5K
xAi✨
xAi✨@xai_42·
@SipeedIO I just ordered the 4k4g on Kickstarter. Will it be shipped to me by the end of March as everyone, or will I have to wait longer?
English
1
0
0
30
Sipeed
Sipeed@SipeedIO·
#NanoKVM & NanoKVM-Pro (WiFi Model) Security Update! 🐛 Fixed: Vulnerability during WiFi AP setup stage (local attacker in wifi range). 🙌 Thanks: Massive shoutout to the security experts who reported it. We’ll keep hardening NanoKVM! 🛡️ github.com/sipeed/NanoKVM…
English
0
5
16
1.9K
Sipeed
Sipeed@SipeedIO·
@billtheinvestor 他跑的是MOE模型,120B才激活5B,其余两个只激活3B。我用500$的8845HS迷你主机跑的速度都比这个1500$的快,不知道网上都在吹什么。。
Sipeed tweet mediaSipeed tweet media
中文
0
0
8
871
Bill The Investor
Bill The Investor@billtheinvestor·
这款手机大小的设备能运行120B参数的AI模型,无需0,000的Mac Studio。它支持本地化部署,确保100%隐私,可全天候驱动开源代理如OpenClaw,甚至替代聊天机器人。无需增加成本,即可进入AI模型的深水区。关注这个新趋势,看看未来谁会被淘汰。
中文
8
11
51
18.5K
Sipeed
Sipeed@SipeedIO·
@Deathlurk Currently only Lite version is using rv nano
English
1
0
0
46
CypherPunk
CypherPunk@Deathlurk·
@SipeedIO guys are you gonna make another bunch of devkit kvms on nano board? I like it a lot. But I see r.n only pure edition
English
1
0
0
32
David Hendrickson
David Hendrickson@TeksEdge·
🦾 Are you ready to carry your LLMs with you? This is the personal device of the future and potable personal inferencing is here from @tiiny! 💥 Models being updated. 🚀 Pocket AI beast specs: 🧠 CPU: ARMv9.2 12-core ⚡ AI Power: Custom SoC + dNPU, ~190 TOPS 💾 Memory/Storage: 80GB LPDDR5X RAM + 1TB SSD 🤖 Model Capacity: Runs up to 120B-parameter LLMs fully on-device 🔋 Efficiency: 30W TDP, ~65W typical system power 📏 Size/Weight: 14.2 × 8 × 2.53 cm, ~300g 🌐 Ecosystem: One-click deployment for dozens of open-source LLMs + agent frameworks 🔒 Connectivity: Fully offline — no internet or cloud needed Supports: 🦙 Llama 🔮 Qwen 🌊 DeepSeek 🧩 Mistral 🪶 Phi 🌐 GLM 💎 Gemma 🧠 GPT-OSS ⚡ MiniCPM 🐉 Yi 🏛️ InternLM 🔥 Hunyuan ☁️ Skywork 📺 Bijan did a good video about it. Link in alt
David Hendrickson tweet mediaDavid Hendrickson tweet mediaDavid Hendrickson tweet mediaDavid Hendrickson tweet media
English
27
44
356
78.9K
Sipeed
Sipeed@SipeedIO·
Ok, but that means 48~64GB normal CPU solution is able to run this low-bit 120B model too... In fact, we were planed to make a device that able to run dense 30B at the speed tiiny run A3B model. but the DDR price is too high now... but tiiny's ks result shocked us, we will consider it again!
English
1
0
3
116
Paul Couvert
Paul Couvert@itsPaulAi·
@SipeedIO Yes they are. But they constrain the accuracy gap between quantized and original floating-point models to within ±0.5 percentage points.
English
1
0
1
112
Paul Couvert
Paul Couvert@itsPaulAi·
You can run local AI models up to 120B without a $10,000 Mac Studio This phone-sized device is your own server for open-source models. 100% local and private. And you can use it to: - Power an agent like OpenClaw 24/7 - Completely replace a chatbot - Literally anything that requires an API It’s called Tiiny, and, once again, you can run the latest open-source models on it. Link below
English
33
34
245
37.7K
Sipeed
Sipeed@SipeedIO·
Just curious about another thing: its 80GB of memory is actually split into 48GB dNPU and 32GB CPU. Typically, it's really hard to pool these two together effectively when running a single model. Since a 120B model at INT4 requires at least 60GB of memory, are they using an even lower-bit quantization?
English
1
0
2
131
Paul Couvert
Paul Couvert@itsPaulAi·
@SipeedIO Yep can't agree more on the memory / bandwidth part. Honestly the benefit of are the massive amount of memory, the NPU usage so you it doesn't consume a lot of power and the form factor. But the MoE disclosure could be important, true.
English
1
0
2
142
Sipeed
Sipeed@SipeedIO·
Thank you for your tutorials! Just want to clarify, their marketing is honestly a bit deceptive. They don't disclose the active parameters for any of the 3 models, tricking users into thinking they're dense. It only runs a 120B model because of the massive RAM, not because of some insane processing power. BTW, I have a $500 8845 mini PC packed with 96GB LPDDR5X (snagged it 1 year ago during the memory price dip), so I can run 120B models too. For local LLM setups today, the real cost bottleneck is always memory capacity and bandwidth, not raw compute.
English
1
0
4
187
Paul Couvert
Paul Couvert@itsPaulAi·
@SipeedIO Big fan of PicoClaw btw. Wrote an article to run it on Android. GPT OSS 120B has 5.1B active parameters not 3B + you still have to fit the full 120B in the RAM which is just not possible on your $500 mini PC... or my $1200 laptop with 32GB of RAM.
English
1
0
4
413
Sipeed
Sipeed@SipeedIO·
@UtaAoya Just tested, my 500$ 8845HS miniPC run faster than this 1500$ device, 32tps > 28tps, peoples are fooled... They use the MOE model for promotion, 120B is only 3B actived, don't you feel strange that 120B,30B,20B model have almost same output eval speed?
Sipeed tweet mediaSipeed tweet media
English
0
1
0
333
A-Uta
A-Uta@UtaAoya·
#ローカルLLM なかなか興味深い🤔 もちろん #EVOX2 より遅いですが、省電力が魅力
David Hendrickson@TeksEdge

🦾 Are you ready to carry your LLMs with you? This is the personal device of the future and potable personal inferencing is here from @tiiny! 💥 Models being updated. 🚀 Pocket AI beast specs: 🧠 CPU: ARMv9.2 12-core ⚡ AI Power: Custom SoC + dNPU, ~190 TOPS 💾 Memory/Storage: 80GB LPDDR5X RAM + 1TB SSD 🤖 Model Capacity: Runs up to 120B-parameter LLMs fully on-device 🔋 Efficiency: 30W TDP, ~65W typical system power 📏 Size/Weight: 14.2 × 8 × 2.53 cm, ~300g 🌐 Ecosystem: One-click deployment for dozens of open-source LLMs + agent frameworks 🔒 Connectivity: Fully offline — no internet or cloud needed Supports: 🦙 Llama 🔮 Qwen 🌊 DeepSeek 🧩 Mistral 🪶 Phi 🌐 GLM 💎 Gemma 🧠 GPT-OSS ⚡ MiniCPM 🐉 Yi 🏛️ InternLM 🔥 Hunyuan ☁️ Skywork 📺 Bijan did a good video about it. Link in alt

日本語
1
0
12
1.2K
Sipeed
Sipeed@SipeedIO·
@davideciffa @TeksEdge @tiiny Just tested, my 500$ 8845HS miniPC run faster than this 1500$ device, 32tps > 28tps, peoples are fooled... They use the MOE model for promotion, 120B is only 3B actived, don't you feel strange that 120B,30B,20B model have almost same output eval speed?
English
0
0
1
65
mrciffa
mrciffa@davideciffa·
@TeksEdge @tiiny How can I trust it If I can't find anywhere the memory bandwidth of this product that should be useful to run LLMs wtf
English
1
0
5
1.6K
Sipeed
Sipeed@SipeedIO·
@DegenApeDev @TeksEdge @tiiny Just tested, my 500$ 8845HS miniPC run faster than this 1500$ device, 32tps > 28tps, peoples are fooled... They use the MOE model for promotion, 120B is only 3B actived, don't you feel strange that 120B,30B,20B model have almost same output eval speed?
English
1
0
1
18
Sipeed
Sipeed@SipeedIO·
@EtherCoins @TeksEdge Just tested, my 500$ 8845HS miniPC run faster than this 1500$ device, 32tps > 28tps, peoples are fooled..
Sipeed tweet media
English
1
1
1
143
Eth
Eth@EtherCoins·
@TeksEdge I'm a bit skeptical but if it does what is says it does, $1kish is an amazing cost/value.
Eth tweet media
English
5
0
10
1.3K
David Hendrickson
David Hendrickson@TeksEdge·
I’m ready to buy this mobile GPU when it comes available. Qwen3.5 and Gemma4 along w/Stable Diffusion for Images on the road or at home. Private and subscription or API cost. It’s got 🦞 Clawdbot like functionality built in.
David Hendrickson tweet media
David Hendrickson@TeksEdge

🦾 Are you ready to carry your LLMs with you? This is the personal device of the future and potable personal inferencing is here from @tiiny! 💥 Models being updated. 🚀 Pocket AI beast specs: 🧠 CPU: ARMv9.2 12-core ⚡ AI Power: Custom SoC + dNPU, ~190 TOPS 💾 Memory/Storage: 80GB LPDDR5X RAM + 1TB SSD 🤖 Model Capacity: Runs up to 120B-parameter LLMs fully on-device 🔋 Efficiency: 30W TDP, ~65W typical system power 📏 Size/Weight: 14.2 × 8 × 2.53 cm, ~300g 🌐 Ecosystem: One-click deployment for dozens of open-source LLMs + agent frameworks 🔒 Connectivity: Fully offline — no internet or cloud needed Supports: 🦙 Llama 🔮 Qwen 🌊 DeepSeek 🧩 Mistral 🪶 Phi 🌐 GLM 💎 Gemma 🧠 GPT-OSS ⚡ MiniCPM 🐉 Yi 🏛️ InternLM 🔥 Hunyuan ☁️ Skywork 📺 Bijan did a good video about it. Link in alt

English
19
25
464
54.9K
Martin Szerment
Martin Szerment@MartinSzerment·
@itsPaulAi If that thing runs 120B locally, cloud bills just became optional overnight.
English
2
0
1
960