Jason Smale

6.1K posts

Jason Smale banner
Jason Smale

Jason Smale

@jwswj

Building @cursor. Prev fast inference @groqInc, @patientnotesapp, @zendesk.

Australia Katılım Kasım 2008
887 Takip Edilen1.7K Takipçiler
Jason Smale retweetledi
lauren
lauren@poteto·
i made this skill! one cool thing is that it matches frontier models to the type of task. in the skill rewrite, i opted to use Opus for prose and evals. for our backend cold start improvements, Opus served as subplanners, while GPT rapidly hillclimbed on perf
Cursor@cursor_ai

Introducing /orchestrate, a skill that recursively spawns agents to tackle your most ambitious tasks with the Cursor SDK. We’ve used it to: - Autoresearch our internal skills, cutting token use by 20% while improving evals - Cut cold start times on our internal backend by 80%

English
25
14
326
23.3K
Jason Smale
Jason Smale@jwswj·
@Itayd @Cursor we're constantly improving this but it still sucks when it happens ... got something for you to try. will dm.
English
1
0
1
41
Itay
Itay@Itayd·
Hey @Cursor is there a way for us to get BugBot to find all issues in one round? My PRs go through 5-10 rounds of back and forth with BugBot before it returns an all clear -- this is a real bottleneck. Love the product otherwise.
English
1
0
0
82
shubhaankar
shubhaankar@__spongeboi·
. @cursor_ai i wish sometimes bugbot gave compliments , the constant critique is draining, but compliments all the time would also be bad. like once in a while, when a PR is absolute genius
English
1
0
1
100
Jason Smale retweetledi
Cursor
Cursor@cursor_ai·
We use a combination of offline benchmarks and online evals to measure model quality. This makes results more useful, especially as public benchmarks are increasingly saturated.
Cursor tweet media
English
8
11
307
65.1K
Jason Smale retweetledi
Jon Kaplan
Jon Kaplan@aye_aye_kaplan·
Story time! The very first version of Bugbot we launched was stateless. We kept track of runs in our analytics database and we only used that data offline for metrics. Whenever we needed to analyze run metadata, we had to do online lookups from the database. Relying on our analytics database during online operations was definitely tech debt. We were using a large RDS cluster that was already running hot from offline usage. We started noticing that some Bugbot runs were failing because of external analytics load; by that time, we had added a few Bugbot-specific tables to it, so we had to come up with a migration plan quickly before this became a widespread availability issue. We chose PlanetScale Postgres for our new Bugbot cluster. We already had many terabytes of data in PlanetScale MySQL/Vitess which was working incredibly well, so we felt comfortable with this choice. The migration went very smoothly, and now all of Bugbot's queries are instant and we stopped worrying about downtime!
Sam Lambert@samlambert

PlanetScale protects production reliability with Bugbot: cursor.com/blog/planetsca…

English
6
11
117
25.3K
Jason Smale retweetledi
Groq Inc
Groq Inc@GroqInc·
It’s official: McLaren F1 x Groq Bringing inference speed at a winning cost to the grid and beyond. See you in Singapore. 🧡🏁
English
115
54
468
216.6K
Jason Smale retweetledi
Ben Klieger
Ben Klieger@benklieger·
Underdiscussed! As we push for more intelligence and lower latency in AI systems, serverside execution of tools becomes a more common pattern, and Responses API makes that possible in a clean way. Chat completions API, the legacy solution, has to be retrofit by each provider to make server side execution work and thus loses its main value proposition: standardization. This is just one of the benefits of responses over completions. Try the responses API on @GroqInc and let us know your thoughts!
dominik kundel@dkundel

One of the secret side launches of gpt-oss was the introduction of Responses API support in more providers! @GroqInc, @vllm_project, @huggingface and @nvidia started shipping support to name a few 🙌

English
1
3
7
2.1K
Jason Smale retweetledi
Matt Shumer
Matt Shumer@mattshumer_·
It's over. OpenAI just crushed it. We have their o3-level open-source model running on @GroqInc at 500 tokens per second. Watch it build an entire SaaS app in just a few seconds. This is the new standard. Why the hell would you use anything else??
English
232
206
3.2K
498.5K
Jason Smale retweetledi
Cline
Cline@cline·
This is what 200 tokens/second looks like with Kimi K2 on @GroqInc For reference, Claude Sonnet-4 is usually delivered at ~60 TPS
English
42
55
854
100.7K
Jason Smale retweetledi
Groq Inc
Groq Inc@GroqInc·
*YOLO Launch* Kimi K2 is now in preview on GroqCloud at 185 tokens/sec. Build fast. Link in comments.
English
81
84
1.3K
269.9K
Jason Smale retweetledi
Groq Inc
Groq Inc@GroqInc·
This solo founder built an open-source competitor to Perplexity with no team, no funding, and no permission. What started as a weekend project now powers over 1M searches, with 60K+ monthly users. This is the story of @sciraai and how Zaid Mukaddam built what others wouldn’t. 🧵👇
Groq Inc tweet media
English
97
219
3.1K
870.9K
Jason Smale retweetledi
Gavin
Gavin@GavinSherry·
The @GroqInc team is rolling out some nice Llama 4 Maverick efficiency improvements to our global network. Check them out!
Gavin tweet media
English
3
3
51
5.2K
Jason Smale retweetledi
Groq Inc
Groq Inc@GroqInc·
The Official Llama API ⚡️Accelerated by Groq In partnership with @AIatMeta. The fastest way to run Llama with no tradeoffs. Preview now live.
GIF
English
9
22
212
37.5K
Jason Smale retweetledi
kwindla
kwindla@kwindla·
🔊🔛🔥 ... @GroqInc launched voice generation today. GroqCloud now has realtime transcription, LLMs, *and* text-to-speech. You can build super-responsive, ultra low-latency voice agents end-to-end entirely on Groq! At @trydaily, we're big fans of Groq's fast, low-latency inference. @pipecat_ai supports all the Groq models, including the new voice model. Lately, we've been obsessively playing a voice chat game that @mark_backman wrote. (My high score is 6, so far.) Here's Mark, with Groq's `Celeste-PlayAI` voice.
English
13
30
232
31.8K
Jason Smale retweetledi
Groq Inc
Groq Inc@GroqInc·
"We optimized our infrastructure to its limits – but the breakthrough came with GroqCloud™. Overnight, our chat speed surged 7.41x while costs fell by 89%. I was stunned. So, we tripled our token consumption. We simply can’t get enough." – Nicolas Bustamante, CEO, Fintool Read more in our latest customer use case with @fintoolx groq.com/groq-customer-…
English
2
8
70
6.9K