
We learn to speak before we learn to read. Voice is the most natural interface we have. We just raised a $100M to make building voice AI as easy as a web app.
dsa
8.7K posts

@dsa
early yc, early twitter, early 23andme, late bloomer @livekit

We learn to speak before we learn to read. Voice is the most natural interface we have. We just raised a $100M to make building voice AI as easy as a web app.


"Any feature we release, a competitor could release within two weeks." @MatanSF (@FactoryAI) on why the moat isn't software anymore. @dsa (@livekit) on building the framework for voice, video, and physical AI. @gsivulka (@HebbiaAI) on what it takes to win in vertical AI. They join @jason on This Week in AI, Episode 11: 00:00 Intro & AGI debate 03:30 Factory: autonomy for software engineering 04:29 LiveKit: open source to ChatGPT voice 10:31 Hebbia: AI for capital markets 13:21 SpaceX-Cursor $60 billion deal breakdown 26:28 Moats in the age of vibe coding 38:10 Deterministic agents vs. open chaos 45:56 DeepSeek V4 01:05:23 OpenAI's spend problem 01:12:08 P-doom scores

@agermanidis @runwayml @btaylor @SierraPlatform @AssemblyAI @graceisford @Lux_Capital @jakesaper @emergencecap @_jeff_liu @assort_health @juberti @OpenAI @krandiash @cartesia @omooretweets @a16z .@dsa, founder and CEO, @livekit. All talk. All action. May 6 in San Francisco. Apply now to join us: cerebralvalleyvoice.com

10k stars on livekit/agents We released version 1.0 a year ago. Today, our customers are building agents for healthcare, finance, insurance, education, robotics, and more. It’s been amazing to see our community grow over the past year. Thank you to everyone building with us.








day 2 of building a self-driving power wheel today i officially trained a self driving model from scratch and deploy it on the car by just simply brute forcing everything, I: > made a remote tele-op and remote data collection app built on @livekit infra > feat: 60ms e2e latency between the car and inference compute (car and compute in vietnam with singapore sfu) > feat: data is collected on operator side, baking latency into observation space itself (I expect this made the model more robust against latency) > recorded 30 min of data at 30fps and converted the dataset to lerobot (you can check a sample here) > trained a simple ACT model (3 epoch, batch size 8) to drive the car around my house > deployed the model on the car with remote inference the video explains everything shortly reflection: > the model is ofc bad, idt behavior cloning would work at all for such complex task on such small sample size > it did work in some cases where the observation is well within distribution, even generalizes to back the car when it gets stuck up next: > will hack alpamayo (@nvidia) or @comma_ai ’s e2e to somehow fit this > or train with a llm backbone or a locomotion prior to see if it generalizes