GMI Cloud

1.7K posts

GMI Cloud banner
GMI Cloud

GMI Cloud

@gmi_cloud

AI-native Inference Cloud. Questions: https://t.co/0lGtMMaeY6

Mountain View, CA Katılım Aralık 2023
42 Takip Edilen2.6K Takipçiler
Sabitlenmiş Tweet
GMI Cloud
GMI Cloud@gmi_cloud·
updates for GMI users, GLM-5 → $0.60 in / $1.92 out (40% off) GLM-5.1 → $0.98 in / $3.08 out (30% off) per M tokens. unlimited.
GMI Cloud tweet mediaGMI Cloud tweet media
English
2
0
4
227
GMI Cloud
GMI Cloud@gmi_cloud·
updates for GMI users, GLM-5 → $0.60 in / $1.92 out (40% off) GLM-5.1 → $0.98 in / $3.08 out (30% off) per M tokens. unlimited.
GMI Cloud tweet mediaGMI Cloud tweet media
English
2
0
4
227
GMI Cloud
GMI Cloud@gmi_cloud·
Massive congrats to @lmsysorg and @radixark! always in awe of your contributions to the inference community grateful to have you as a partner 🤝
RadixArk@radixark

Today, we are thrilled to officially launch RadixArk with $100M in Seed funding at a $400M valuation. The round was led by @Accel and co-led by @sparkcapital. RadixArk exists to make frontier AI infrastructure open and accessible to everyone. Today, the systems behind the most capable AI models are concentrated in a small number of companies. As a result, most AI teams are forced to rebuild training and inference stacks from scratch, duplicating the same infrastructure work instead of focusing on new models, products, and ideas. RadixArk was founded to change that. We are building an AI platform that makes it easier for teams to train and serve the best models at scale. RadixArk comes from the open-source community. We started with SGLang, where many of us are core developers and maintainers, and expanded our work to Miles for large-scale RL and post-training. We will continue contributing to both projects and working with the community to make them the strongest open-source infrastructure foundations for frontier AI. We would like to thank our long-term partners, contributors, and the broader SGLang community for believing in this mission. We're also grateful to @Accel and @sparkcapital, NVentures (Venture capital arm of @nvidia), Salience Capital, A&E Investment, @HOFCapital, @walden_catalyst, @AMD, LDVP, WTT Fubon Family, @MediaTek, Vocal Ventures, @Sky9Capital and our angel investors @ibab, @LipBuTan1, Hock Tan, @johnschulman2, @soumithchintala, @lilianweng, @oliveur, @Thom_Wolf, @LiamFedus, @robertnishihara, @ericzelikman, @OfficialLoganK, and @multiply_matrix among others. Thanks for the exclusive interview with @MeghanBobrowsky at @WSJ about our vision.

English
0
2
6
534
GMI Cloud
GMI Cloud@gmi_cloud·
we compared Eleven Labs and Inworld's newest TTS models (Realtime TTS 2) in 7 examples across English, Japanese, Chinese, French, and Spanish Inworld focuses on pronunciation and punctuation, while eleven labs is more fluent
English
0
0
1
157
RadixArk
RadixArk@radixark·
Today, we are thrilled to officially launch RadixArk with $100M in Seed funding at a $400M valuation. The round was led by @Accel and co-led by @sparkcapital. RadixArk exists to make frontier AI infrastructure open and accessible to everyone. Today, the systems behind the most capable AI models are concentrated in a small number of companies. As a result, most AI teams are forced to rebuild training and inference stacks from scratch, duplicating the same infrastructure work instead of focusing on new models, products, and ideas. RadixArk was founded to change that. We are building an AI platform that makes it easier for teams to train and serve the best models at scale. RadixArk comes from the open-source community. We started with SGLang, where many of us are core developers and maintainers, and expanded our work to Miles for large-scale RL and post-training. We will continue contributing to both projects and working with the community to make them the strongest open-source infrastructure foundations for frontier AI. We would like to thank our long-term partners, contributors, and the broader SGLang community for believing in this mission. We're also grateful to @Accel and @sparkcapital, NVentures (Venture capital arm of @nvidia), Salience Capital, A&E Investment, @HOFCapital, @walden_catalyst, @AMD, LDVP, WTT Fubon Family, @MediaTek, Vocal Ventures, @Sky9Capital and our angel investors @ibab, @LipBuTan1, Hock Tan, @johnschulman2, @soumithchintala, @lilianweng, @oliveur, @Thom_Wolf, @LiamFedus, @robertnishihara, @ericzelikman, @OfficialLoganK, and @multiply_matrix among others. Thanks for the exclusive interview with @MeghanBobrowsky at @WSJ about our vision.
RadixArk tweet media
English
75
83
515
235.1K
GMI Cloud
GMI Cloud@gmi_cloud·
tested four newest open source Kimi K2.6 is the fastest, GLM 5.1 the fanciest, DeepSeek V4 is the most comprehensive, and Xiaomi MiMo is the slowest
English
27
26
496
41K
AskClaw 🦀
AskClaw 🦀@GetAskClaw·
@YuLin807 @gmi_cloud 车枪球,比小镇有意思 让agent们比赛,赢了多投钱(怎么有菠菜的苗头)
中文
1
0
1
22
GMI Cloud
GMI Cloud@gmi_cloud·
@bnafOg thank you for the explanation! very helpful! we were just testing them on game design. prompt included in the video
English
0
0
0
96
Bnaf.OG | 🟧
Bnaf.OG | 🟧@bnafOg·
@gmi_cloud Architecture explains the gap: MiMo's MoE runs more active params per token than Kimi K2.6's optimized routing — hence slowest. DeepSeek V4's 'comprehensive' edge is partly MLA: ~75% KV-cache compression makes it far better for long agentic loops. What task were you testing?
English
2
0
3
775
GMI Cloud
GMI Cloud@gmi_cloud·
@bettercallsalva we originally wanna include token burns as well. this is more like speed in completing the task and the visual output for a small game. will be more specific next time
English
0
0
0
65
Thiago Salvador
Thiago Salvador@bettercallsalva·
@gmi_cloud Those rankings imply a single eval but the order changes if you measure throughput, params, or task accuracy. Which one was this?
English
1
0
2
164
Avais Aziz
Avais Aziz@avaisaziz·
@gmi_cloud Impressive benchmark. Kimi K2.6 at 29s for a full playable racer stands out, while GLM 5.1 edges visuals and DeepSeek V4 the feature depth. All four delivering functional games shows solid progress.
English
1
0
1
168
GMI Cloud
GMI Cloud@gmi_cloud·
@ofabdalaX we tested them for a very simple game, the prompt is actually in the video. this is very helpful to know!
English
0
0
1
105
Abdala
Abdala@ofabdalaX·
@gmi_cloud rodei o mesmo set semana passada e o ranking inverte por task: Kimi K2.6 voa em chat curto mas trava em agentic loop com 3+ tools. DeepSeek V4 mais lento mas hallucina menos em decisao. GLM 5.1 unico que fica consistente em pt-BR. qual era a task?
Português
1
0
2
705