Richard Li

1.7K posts

Richard Li banner
Richard Li

Richard Li

@rdli

Entrepreneur. Advisor. AI and inference. Prev: ceo/founder @ambassadorlabs; product @duosec; corp dev/strategy @rapid7.

Boston, MA Katılım Ocak 2010
196 Takip Edilen658 Takipçiler
Sabitlenmiş Tweet
Richard Li
Richard Li@rdli·
I've been building an #ai application for a little while now, and wrote up my 7 macro takeaways about building an AI app that I didn't know when I started. thelis.org/blog/lessons-f…
English
0
0
0
175
Richard Li
Richard Li@rdli·
#Agents aren’t the future—they’re already here. But building them? That takes a whole new stack. 🧠 AI ⚙️ Durable execution 🧱 Frameworks 🗂 Context 🛠 Actuators Check out the breakdown (with @jflomenb and @Wing_VC): 🔗 wing.vc/content/the-ag…
English
1
1
4
1.4K
Richard Li
Richard Li@rdli·
Context Is King 👑. Smarter agents need richer context—not just prompts. They aggregate context from multiple sources: 🧠 Knowledge (vector databases) 💾 Memory (short/long-term) 🌐 Actuators (APIs, sensors) Check out the full Agentic Runtime Stack: wing.vc/content/the-ag…
English
0
0
0
45
Richard Li retweetledi
Akka
Akka@akka_io_·
Final chance to register for tomorrow's webinar with @InfoQ Learn how to design and implement the next generation of AI-powered services with Tyler and @rdli. See you there! bit.ly/42vzyFa #InfoQ #Java #AI
Akka tweet media
English
0
2
2
508
Richard Li
Richard Li@rdli·
Most apps act fast. Agents don’t. They pause, retry, wait hours or days. That’s why they need durable execution—resilient workflows that persist across failures. Check out our post on the agentic runtime stack: wing.vc/content/the-ag…
English
0
1
3
180
Richard Li retweetledi
Casper Hansen
Casper Hansen@casper_hansen_·
2.1k stars, 2+ million downloads, and 7000+ models on Huggingface later, and I am officially ready to retire my long-time project AutoAWQ ⚡️ Proud to say that AutoAWQ has been adopted by the @vllm_project and will now be maintained by 55+ contributors 🥳
Casper Hansen tweet media
English
14
13
180
44.4K
Richard Li
Richard Li@rdli·
4. There can (and are) viable software inference businesses, but they need to be decoupled from GPU supply.
English
0
0
0
36
Richard Li
Richard Li@rdli·
3. GPU businesses such as CoreWeave should be viewed through more of a finance lens than a tech lens.
English
1
0
0
106
Richard Li
Richard Li@rdli·
Fascinating! I have not thought deeply about the GPU market but clearly @evanjconrad has. GPU market is radically different from CPU market in a bunch of ways.
Latent.Space@latentspacepod

🆕 SF Compute: Commoditizing Compute latent.space/p/sfcompute We're excited for our latest deep dive into the compute market with @evanjconrad of @sfcompute! It should not be normal for the prices of one of the world’s most important resources right now to swing from $8 to $1 per hour (as @picocreator observed) based on drastically inelastic demand AND supply curves - from 3 year lock-in contracts to stupendously competitive over-ordering dynamics for NVIDIA allocations — especially with increasing baseline compute needed for even the simplest academic ML research and for new AI startups getting off the ground. The entire point of SFC is creating liquidity between GPU owners and consumers and making it broadly tradable, even programmable. As we explore, these are the primitives that you can then use to create your own, high quality, custom GPU availability for your time and money budget, similar to how Amazon Spot Instances automated the selective buying of unused compute. The ultimate end state of where all this is going is GPU that trade like other perishable, staple commodities of the world - oil, soybeans, milk. Because the contracts and markets are so well established, the price swings also are not nearly as drastic, and people can also start hedging and managing the risk of one of the biggest costs of their business, just like we have risk-managed commodities risks of all other sorts for centuries. As a former derivatives trader, you can bet that swyx doubleclicked on that… Also to end off, we of course had to ask about how on earth SFCompute manages to have such immaculate vibes.... Timestamps [00:00:05] Introductions [00:00:12] Introduction of guest Evan Conrad from SF Compute [00:00:12] CoreWeave Business Model Discussion [00:05:37] CoreWeave as a Real Estate Business [00:08:59] Interest Rate Risk and GPU Market Strategy Framework [00:16:33] Why Together and DigitalOcean will lose money on their clusters [00:20:37] SF Compute's AI Lab Origins [00:25:49] Utilization Rates and Benefits of SF Compute Market Model [00:30:00] H100 GPU Glut, Supply Chain Issues, and Future Demand Forecast [00:34:00] P2P GPU networks [00:36:50] Customer stories [00:38:23] VC-Provided GPU Clusters and Credit Risk Arbitrage [00:41:58] Market Pricing Dynamics and Preemptible GPU Pricing Model [00:48:00] Future Plans for Financialization? [00:52:59] Cluster auditing and quality control [00:58:00] Futures Contracts for GPUs [01:01:20] Branding and Aesthetic Choices Behind SF Compute [01:06:30] Lessons from Previous Startups [01:09:07] Hiring at SF Compute

English
1
0
0
50
Richard Li retweetledi
Akka
Akka@akka_io_·
The era of #agenticAI is here. #AIagents are replacing manual workflows—making decisions, taking action, scaling fast. Learn how to design and implement the next generation of AI-powered services. 🔗 bit.ly/4hJea3v #Java @InfoQ @rdli
Akka tweet media
English
0
2
3
634
Richard Li retweetledi
Akka
Akka@akka_io_·
🚨 We're 1 day away! 🚨 Agentic AI is reshaping software—but scaling it isn’t easy. → TPS skyrockets → LLMs struggle with latency → Costs add up fast How do you build services that can handle agentic scale without breaking the bank? Join Tyler Jewell & @rdli tomorrow at 10 AM ET as they break it all down. Live Q&A included! 📅 Register here→ bit.ly/43GeLjd #Akka #AppDevelopment #CloudComputing #CloudNative #DevOps #DistributedSystems #SoftwareDevelopment #AgenticAI #LLMs #LLM #AIAgent #AI
English
0
3
6
550
Richard Li
Richard Li@rdli·
My friend @bnfb introduced me to a critical component of product requirements: setting a price you're willing to pay. This seemingly simple change creates focus, reduces risk, and improves communication. More: thelis.org/blog/set-a-pri…
English
0
0
0
26
Richard Li
Richard Li@rdli·
@rekurencja @adpirz @rakyll I also think that strategies like model distillation are already reducing the amount of data plumbing work. Depending on what you’re trying to build, data may very well be less than 80%.
English
0
0
0
103
Piotr Turek
Piotr Turek@rekurencja·
Some of the key pitfalls of ML/AI vs traditional software engineering: 1. A lot of ML/AI is inherently crappy software (by hard engineering standards), starting from the dominant language (Python), all the build & dependency management... the list goes on 2. It does magic... until it doesn't. Binary reality becomes fuzzy (yes/no/maybe). Here the closest to the ML reality is if somebody had prior experience with streaming algorithms, processing data at scale 3. Data is a mess. Working with any sort of ML is 80%+ being a data plumber and at most 20% what people expect they would be doing in ML/AI 4. Super annoyingly slow iteration speed. Forget about rapidly developing software with TDD and trunk based development. Go make a coffee and come back to see that your training failed Expectation vs reality is what often burns engineers out when they switch to AI/ML
English
2
2
15
1.3K
Jaana Dogan ヤナ ドガン
ML / AI is a very hard pivot if you are coming from traditional software or systems engineering. Over the years, I’ve seen extremely brilliant engineers burned out by this pivot. You need to relearn how to do engineering with its dynamics and priorities to be successful.
English
27
31
461
29.1K