Lillian Ma

45 posts

Lillian Ma banner
Lillian Ma

Lillian Ma

@lillian_ma_

head of global partnerships @gmi_cloud

Santa Clara Katılım Mayıs 2018
335 Takip Edilen328 Takipçiler
Lillian Ma
Lillian Ma@lillian_ma_·
@yuqih Snow boss is checking on my token burn 🔥
English
0
0
1
21
Yuqi Hou
Yuqi Hou@yuqih·
@lillian_ma_ I hope Snow was entertained while Claude Code was discombobulating 🐩
English
1
0
0
22
Lillian Ma
Lillian Ma@lillian_ma_·
pov: me & office dog waiting for my Claude Code to finish solving a million $ question
Lillian Ma tweet media
English
1
0
3
76
Yuqi Hou
Yuqi Hou@yuqih·
Hi, I'm Yuqi! I just launched the ambassador program for @gmi_cloud with @venjia_z! Bit about it: - $200–500/mo in credits to build - funding to host your own meetups + events - your demos featured across GMI - swag Let's connect if you're into building with AI, making content/demos, and running discord or irl events
Yuqi Hou tweet media
English
7
0
25
1K
Lillian Ma
Lillian Ma@lillian_ma_·
great benefits when you work at Castro St in MTV so many dessert options during afternoon’s break @yuqih @nicoleegong
Lillian Ma tweet media
English
2
0
5
114
Lillian Ma
Lillian Ma@lillian_ma_·
@yuqih when will you start posting stories around sf hacker community haha
English
1
0
1
98
Yuqi Hou
Yuqi Hou@yuqih·
Hi I’m Yuqi! I recently moved to SF. Bit about me: - live in a growth house - 10k on youtube - used to live in nyc, paris and london - writing a novel on the caltrain - working at an ai startup Let’s connect about agents, ai models, making content and writing fiction
Yuqi Hou tweet media
English
125
3
452
44.6K
Brandon Johnson
Brandon Johnson@bj0hn5on·
Inference Co-design Day was a hit! 💚 S/o to all the @nvidia Cloud Partners and Inference Service Providers that came out for a deep dive on optimizing inferencing. @gmi_cloud @CrusoeAI @friendliai @baseten @togethercompute @nebiusai @DeepInfra @nscale @LightningAI @Vultr @SimplismartHQ @digitalocean @Eigen_AI_Labs @inference_net
Lillian Ma@lillian_ma_

Big day 🚀 As early adopters of TRT-LLM, Dynamo, and NIM, we’re at @nvidia’s Inference Codesign Day meeting the team IRL. As an inference infra provider, partnering with NVIDIA to bring world-class inference to our customers is exactly where we want to be. Heard TRT-LLM is expanding coverage for visual generative models 👀 the road ahead is going to be wild. @gmi_cloud @bj0hn5on @ReneeYao1 @NVIDIAAI @NVIDIAAIDev

English
1
0
5
178
Lillian Ma
Lillian Ma@lillian_ma_·
Come say hi to me if you’re at our @gmi_cloud Scale Program Demo Day
Lillian Ma tweet mediaLillian Ma tweet mediaLillian Ma tweet media
English
1
0
5
126
Lillian Ma
Lillian Ma@lillian_ma_·
What's the market availability for Blackwell? Effectively 0% for new buyers. No kidding. I've personally fielded 10x inquiries in the past month around "do you have X GPU," and I feel a lot of founders and builders still haven't touched the truth: there's just no liquid compute capacity on the market right now. 6 months ago if someone told me H100 1-yr contract pricing would surge 40% in half a year (from $1.70 to $2.35/hr) with 36–52 week lead times and capacity sold out through Aug–Sept 2026, I would've thought they were crazy. And it's not even a money game anymore. The liquidity has been locked down by hyperscalers (~$700B capex in 2026), neoclouds and frontier AI labs already. Anthropic is renting from a direct rival's data center because that's where the electrons are. That's the market we're in. So here comes the trend: -For small OSS providers: fewer cloud providers are going to host you. Attracting customized deals from public endpoints will soon become a top-player's game. -TCO matters more. People who actually handle token economics optimization will be valued more on the market. -Industry players are desperately pushing forward an inference-efficiency-driven playbook. -Hard times ahead for any ecosystem built on top of spot instances. -Founders dreaming of Claude lowering their token price should drop the idea immediately. Yes, Opus 4.5/4.6 came in cheaper than 4/4.1 — but Opus 4.7 kept the headline price flat, the new tokenizer can raise effective costs up to 35%, and now they're paying premium rates to a rival for emergency capacity. If you can't justify your unit economics right now, you'll only burn more money per user in the near future. -TTS models will have a huge advantage in this era to justify their margins. Opinions here are on my own.
English
0
0
2
97
Lillian Ma
Lillian Ma@lillian_ma_·
Big day 🚀 As early adopters of TRT-LLM, Dynamo, and NIM, we’re at @nvidia’s Inference Codesign Day meeting the team IRL. As an inference infra provider, partnering with NVIDIA to bring world-class inference to our customers is exactly where we want to be. Heard TRT-LLM is expanding coverage for visual generative models 👀 the road ahead is going to be wild. @gmi_cloud @bj0hn5on @ReneeYao1 @NVIDIAAI @NVIDIAAIDev
Lillian Ma tweet mediaLillian Ma tweet mediaLillian Ma tweet mediaLillian Ma tweet media
English
1
0
6
357
RadixArk
RadixArk@radixark·
Today, we are thrilled to officially launch RadixArk with $100M in Seed funding at a $400M valuation. The round was led by @Accel and co-led by @sparkcapital. RadixArk exists to make frontier AI infrastructure open and accessible to everyone. Today, the systems behind the most capable AI models are concentrated in a small number of companies. As a result, most AI teams are forced to rebuild training and inference stacks from scratch, duplicating the same infrastructure work instead of focusing on new models, products, and ideas. RadixArk was founded to change that. We are building an AI platform that makes it easier for teams to train and serve the best models at scale. RadixArk comes from the open-source community. We started with SGLang, where many of us are core developers and maintainers, and expanded our work to Miles for large-scale RL and post-training. We will continue contributing to both projects and working with the community to make them the strongest open-source infrastructure foundations for frontier AI. We would like to thank our long-term partners, contributors, and the broader SGLang community for believing in this mission. We're also grateful to @Accel and @sparkcapital, NVentures (Venture capital arm of @nvidia), Salience Capital, A&E Investment, @HOFCapital, @walden_catalyst, @AMD, LDVP, WTT Fubon Family, @MediaTek, Vocal Ventures, @Sky9Capital and our angel investors @ibab, @LipBuTan1, Hock Tan, @johnschulman2, @soumithchintala, @lilianweng, @oliveur, @Thom_Wolf, @LiamFedus, @robertnishihara, @ericzelikman, @OfficialLoganK, and @multiply_matrix among others. Thanks for the exclusive interview with @MeghanBobrowsky at @WSJ about our vision.
RadixArk tweet media
English
83
100
627
346.3K