Sam Crowder

511 posts

Sam Crowder banner
Sam Crowder

Sam Crowder

@samecrowder

Head of Product, LangSmith at @LangChain 🚀 | prev: @Harvard MS/MBA, @RocksetCloud (acq. OpenAI), @BainCapVC, @ContraryCapital

San Francisco, CA Joined Eylül 2017
583 Following1.1K Followers
Sam Crowder retweeted
LangChain
LangChain@LangChain·
💫 New LangChain Academy Course: Building Reliable Agents 💫 Shipping agents to production is hard. Traditional software is deterministic – when something breaks, you check the logs and fix the code. But agents rely on non-deterministic models. Add multi-step reasoning, tool use, and real user traffic, and building reliable agents becomes far more complex than traditional system design. The goal of this course is to teach you how to take an agent from first run to production-ready system through iterative cycles of improvement. You’ll learn how to do this with LangSmith, our agent engineering platform for observing, evaluating, and deploying agents. Enroll for free ➡️ academy.langchain.com/courses/buildi…
English
3
13
76
5.5K
Sam Crowder retweeted
LangChain
LangChain@LangChain·
Open Models Panel at GTC with Harrison & Jensen: Join us tomorrow, Wednesday March 18th at 12:30pm at GTC for “Open Models: Where We Are and Where We’re Headed”, a panel featuring Harrison, Jensen, and the CEOs of Cursor, Thinking Machines Lab, Perplexity, and more. Add it to your schedule ➡️ nvidia.com/gtc/session-ca…
LangChain tweet media
English
3
9
42
4.6K
Ralph👨🏻‍💻
Ralph👨🏻‍💻@ralphesber·
Just published my first OpenClaw skill 🎉 langsmith-cli — query your LangSmith traces with natural language, right from your AI assistant. ask "what do failing runs have in common?" and get an answer back. Also: cost breakdowns, latency percentiles, before/after diffs. No extra API key. No data leaving your machine. 👉 clawhub.com/skills/langsmi…
English
2
5
20
7.8K
Sam Crowder
Sam Crowder@samecrowder·
we love working with @cogent_security, and their team is hiring. go work with one of the best teams building frontier agents!
LangChain@LangChain

🚀 LangSmith for Startups Spotlight: @cogent_security Cogent is building AI agents that protect the world's largest organizations from cyberattacks. One of the hardest problems in cybersecurity is going from finding a vulnerability to actually fixing it. Cogent is automating that entire process from end-to-end. Cogent is already working with dozens of Fortune 1000 and Global 2000 enterprise customers such as major universities, hospitality brands, and consumer retailers. Cogent uses LangSmith for production tracing and monitoring of our agents. Their team leverages execution traces for usage insight and use-case categorization, self-refinement loops to diagnose eval failures, and online evaluators to flag undesired behavior. Join their team if you want to build frontier AI for mission critical problems 🤝cogent.com/careers

English
0
1
4
2K
Sam Crowder retweeted
LangChain
LangChain@LangChain·
New Conceptual Guide: You don’t know what your agent will do until it’s in production 👀 With traditional software, you ship with reasonable confidence. Test coverage handles most paths. Monitoring catches errors, latency, and query issues. When something breaks, you read the stack trace. Agents are different. Natural language input is unbounded. LLMs are sensitive to subtle prompt variations. Multi-step reasoning chains are hard to anticipate in dev. Production monitoring for agents needs a different playbook. In our latest conceptual guide, we cover why agent observability is a different problem, what to actually monitor, and what we've learned from teams deploying agents at scale. Read the guide ➡️ blog.langchain.com/you-dont-know-…
LangChain tweet media
English
4
7
47
6.2K
Sam Crowder retweeted
LangChain
LangChain@LangChain·
📊 How to evaluate skills❓️ Lots of companies are building skills for coding agents. But how do you know if your skill is actually working? It's tempting to go by vibes, but performance varies a lot across tasks — and coding agents have a huge action space, which makes that variance even harder to predict. We built an evaluation benchmark for our newly released LangSmith and LangChain skills. ➡️ Learn about our findings here: blog.langchain.com/evaluating-ski… ➡️ Check out the benchmark for yourself: github.com/langchain-ai/s…
LangChain tweet media
English
15
35
244
33.4K
Sam Crowder
Sam Crowder@samecrowder·
come join us at Interrupt at learn how the best teams are building reliable agents!
LangChain@LangChain

@coinbase and @Rippling are at Interrupt. Evan Kormos on how Coinbase built a multi-agent system to scale AI-handled support from 20% to 80%. Ankur Bhatt on how Rippling built deep agents to diagnose payroll tax notices across 50 states. May 13-14 · San Francisco ➡️ interrupt.langchain.com

English
0
2
6
2.1K
Sam Crowder
Sam Crowder@samecrowder·
langsmith is the flywheel for automating agentic improvements
Harrison Chase@hwchase17

we're building ai into langsmith not just to be a generic assistant, but to actually help debug agents 🧵here's a real example where it helped me over the weekend: context: I'm building an agent on deepagents (github.com/langchain-ai/d…). It has a bunch of tools for interacting with files issue: I noticed thanks to langsmith monitoring (docs.langchain.com/langsmith/dash…) that ~1% of calls to `ls` were failing. sidenote - this is value of ai native monitoring, we automatically tracking failing tool calls. I clicked into an example run and saw that the model was generating the wrong parameter to `ls` - it was passing `file_path` not `path` at this point, i knew what the issue was, but had no idea WHY it was occurring. the trace here was very long and the prompt was long as well. i suspected that there was something wrong in the prompt - maybe a bad example? i asked polly (docs.langchain.com/langsmith/polly) our in app assistant to help me debug. she investigated, and found that other file tools in deepagents use `file_path`, and `ls` is the only one that uses `path`. see screenshot below I don't know how long it would have taken me to figure this out otherwise everyone is adding assistants into app for basic question/answering. imo really valuable assistants go beyond that - they are purposefully placed in situations where they can augment human intelligence nicely. in this case - reading long traces and prompts is something llms are great at!

English
0
0
1
134
Sam Crowder retweeted
LangChain
LangChain@LangChain·
Tools in the LangSmith Playground 🛝 Define tools once, use them everywhere. - Save & reuse tool definitions alongside your prompts, across playground sessions - Share a tool library across your workspace - Manage tool definitions programmatically with the SDK Learn more: buff.ly/PHFMHCY
English
4
5
48
8.3K
Sam Crowder retweeted
Harrison Chase
Harrison Chase@hwchase17·
Many teams treat evals as a last-mile check. monday.com Service made them a Day 0 requirement for their AI service agents. Using LangSmith, the monday service team has been able to: 🔷Achieve 8.7x faster evaluation feedback loops (from 162 seconds to 18 seconds). 🔷Get comprehensive testing across hundreds of examples in minutes instead of hours 🔷Gain agent observability with real-time, end-to-end quality monitoring on production traces Read more on their eval-driven development here: blog.langchain.com/customers-mond…
Harrison Chase tweet media
English
15
4
40
4.1K
Sam Crowder
Sam Crowder@samecrowder·
Google ADK tracing for LangSmith! Folks wonder all the time if LangSmith only works for our frameworks like LangChain and LangGraph. Given our naming conventions, I can't really blame them. but it couldn't be further from the truth!
LangChain@LangChain

🔎 We shipped native tracing for Google ADK! See how easy it is to get started observing your ADK agents in LangSmith with just a few clicks. LangSmith works natively with over 25 frameworks and providers, and not to mention OpenTelemetry! 🔥 Docs 👉 docs.langchain.com/langsmith/trac…

English
1
1
2
2.6K