Lance Martin

1.2K posts

Lance Martin banner
Lance Martin

Lance Martin

@RLanceMartin

@AnthropicAI. Previously langchain, vision for self-driving trucks / cars 🚘 🤖, phd @stanford 🧪

San Francisco, CA Katılım Mayıs 2009
407 Takip Edilen25.4K Takipçiler
Lance Martin retweetledi
Lydia Hallie ✨
Lydia Hallie ✨@lydiahallie·
if your skill depends on dynamic content, you can embed !`command` in your SKILL.md to inject shell output directly into the prompt Claude Code runs it when the skill is invoked and swaps the placeholder inline, the model only sees the result!
Lydia Hallie ✨ tweet media
English
124
228
2.8K
756.4K
Lance Martin
Lance Martin@RLanceMartin·
more details here: #march-18-2026" target="_blank" rel="nofollow noopener">platform.claude.com/docs/en/releas…
English
0
0
4
1.1K
Lance Martin
Lance Martin@RLanceMartin·
a useful trick: Claude API now programmatically lists capabilities for every model (context window, thinking mode, context management support, etc). just ask Claude Code or use the API directly. platform.claude.com/docs/en/api/py…
English
13
13
122
15.4K
Lance Martin retweetledi
Thariq
Thariq@trq212·
i think we might have undersold 1M context tbh, the performance is so so good, I really just don't clear the context window much these days
Thariq tweet media
English
262
68
2.1K
91K
An Van
An Van@anvanvan·
@RLanceMartin Any data on a context rot curve for the Opus 1M?
English
1
0
2
778
Lance Martin
Lance Martin@RLanceMartin·
a project i've been excited about: make Claude Code better at using the Claude API. just ask Claude Code about Claude API features (e.g., prompt caching, adaptive thinking + effort, tools, etc).
English
19
21
308
31.5K
Lance Martin
Lance Martin@RLanceMartin·
@ajay_sreeram not quite, but this question is helpful b/c we should share some example traces when we talk about PTC (e.g., examples w web search).
English
0
0
0
104
Ajay Sreeram
Ajay Sreeram@ajay_sreeram·
@RLanceMartin But doesn’t Claude need to spend even more tokens to write code in same context? (writing code without knowing tool response is even worse) (Or is writing post processing code is handled as separate agent, even here Claude need to prepare enough context to pass to the agent)
English
1
0
1
154
Lance Martin
Lance Martin@RLanceMartin·
@dav_taylo not quite, Claude will write code that calls the search tool and performs any post processing that it sees fit for the task. so it’s more general than any specific hard coded step (re-rank) etc. Claude decides what to do based on the query.
English
0
0
0
59
David Taylor
David Taylor@dav_taylo·
@RLanceMartin Is the web search dynamic filtering effectively just an addl reranking step, though?
English
1
0
0
75
Lance Martin
Lance Martin@RLanceMartin·
check out the updated skill-creator. i esp like built-in support for test generation (e.g., to measure + optimize tricky things like skill trigger rate). available in Claude Code as plugin, Claude.ai, + Cowork.
English
67
137
1.7K
347.6K
Lance Martin
Lance Martin@RLanceMartin·
claude opus 4.6 + its web search tool is now #1 on @arena's search arena. the web search tool is available in the API, and uses code execution to process intermediate search results (saves tokens / improves performance). docs: platform.claude.com/docs/en/agents… x.com/arena/status/2…
Lance Martin tweet media
Arena.ai@arena

Exciting update: Claude Opus 4.6 is now #1 in Search Arena! 🌐 Highlights: -#1 wide lead by Opus 4.6 scoring 1255, +30pt over Grok-4.20-beta1, GPT-5.2 and Gemini-3 - Sonnet 4.6 ranks #7 on par with GPT-5.1 With this result, @anthropicAI's Claude Opus 4.6 is now #1 across Text, Code, and Search Arena. Congrats to the @AnthropicAI team on this incredible milestone!

English
7
14
157
30.8K
Lance Martin
Lance Martin@RLanceMartin·
ya, this is an interesting and subtle point: sub-agents and PTC both isolate context from an orchestrator. but sub-agents allocate tasks to LLMs. PTC chains actions in code. sub-agents can use PTC. PTC is useful anytime you have actions that you can easily compose in code.
English
0
0
1
76
PrimeLine
PrimeLine@PrimeLineAI·
interesting distinction. haven't tried PTC yet - I've been solving the composition tax with sub-agent delegation instead. built scoring that decides complexity per task, spawns isolated agents in their own context window, only the final result comes back. avoids the round-trip problem differently but I'm curious how PTC handles cases where intermediate results need judgment before the next step. that's where my sub-agents earn their keep.
English
1
0
1
62
VuongNg
VuongNg@Agimon_AI·
@ashvanth_s1 @Ashwinizer @RLanceMartin Built in caching keeps prompts simpler and more stable. The consistent prefix thing is the sneaky hard part. Most "manual optimization" falls apart right there.
English
1
0
1
23
Lance Martin
Lance Martin@RLanceMartin·
@kchdmn @lateinteraction @a1zhang ya, this part is the same as the PTC findings: RLMs can write code in a REPL to call tools as functions. RLMs have some additional properties -- prompt itself lives in the REPL external to LLM and LLM can call itself that are not part of PTC specifically, but are interesting!
English
0
0
2
1.2K