will brown

13.7K posts

will brown banner
will brown

will brown

@willccbb

reward hacking @primeintellect

sf Sumali Şubat 2015
1.3K Sinusundan41.3K Mga Tagasunod
will brown nag-retweet
Ankur Goyal
Ankur Goyal@ankrgyl·
I personally find this quite inspiring — Cursor is much more sophisticated than what most companies can do today, but compounding your learnings into a model that excels at your use case is the ultimate way to build an AI product.
Cursor@cursor_ai

Composer 2 is now available in Cursor.

English
4
9
104
9.7K
will brown
will brown@willccbb·
@TheAhmadOsman optimizing local stacks just for bs=1 was never the right call personal agents are becoming multi-agents real quick
English
2
0
19
1.1K
Ahmad
Ahmad@TheAhmadOsman·
DGX Spark uses unified memory > 273 GB/s RTX PRO 6000 delivers > 1.8 TB/s (1792 GB/s) If someone told you they’re comparable, they’re wrong And this is exactly why llama.cpp isn’t the right tool here Try vLLM or SGLang on a GPU and you’ll see very different results
Max Weinbach@mweinbach

@TheAhmadOsman I have on DGX Spark and then was having insane tool calling issues and was told by Nvidia to use llama cpp

English
40
12
295
30.3K
kache
kache@yacineMTB·
prediction: someone is going to get a coding AI like codex to automate turning existing steam video games into harnesses, come up with architecture to parallelize the games themselves in a manner that is conducive for RL training, and train an RL demigod model
English
35
12
406
18.3K
Celine Halioua
Celine Halioua@celinehalioua·
went on a date with a gen z guy and he told me i tweet like a boomer ☹️
English
22
1
178
15.6K
will brown
will brown@willccbb·
@vikhyatk yc startup for automating signing up for all the other yc startups that automate all your ops stuff who’s building this
English
3
1
48
2.4K
vik
vik@vikhyatk·
software generation is no longer the bottleneck. it's operations trillion dollar opportunity for whoever solves it
English
18
3
117
24.2K
will brown nag-retweet
Larissa Schiavo
Larissa Schiavo@lfschiavo·
My “I don’t have LLM psychosis" hoodie has people asking a lot of questions already answered by my hoodie.
Larissa Schiavo tweet media
English
33
14
394
12.6K
will brown nag-retweet
Luke Drago
Luke Drago@luke_drago_·
underrated product: @PrimeIntellect’s liquid compute. we're going to earn money on one of our clusters this month.
Luke Drago tweet media
English
6
10
287
27.8K
will brown nag-retweet
Alex
Alex@afurgs·
Big fan of Luke and what they’re doing at workshop labs. It’s also great tactical example of why we built this feature, no lab should be forced to pay for idle compute when there is an abundance of demand. Tbh it’s insane it’s not an option with any other providers.
Luke Drago@luke_drago_

underrated product: @PrimeIntellect’s liquid compute. we're going to earn money on one of our clusters this month.

English
3
2
47
4.3K
will brown nag-retweet
Johannes Hagemann
Johannes Hagemann@johannes_hage·
it was a good event sir
Johannes Hagemann tweet media
English
9
5
181
5.7K
will brown nag-retweet
Prime Intellect
Prime Intellect@PrimeIntellect·
At Prime Intellect, we’re building that stack end to end: - agentic RL training and inference on frontier open models - RL sandboxes - open-source libraries like verifiers + prime-rl Giving everyone access to frontier lab infrastructure
Prime Intellect tweet media
English
3
6
164
6.9K
Omar Khattab
Omar Khattab@lateinteraction·
Folks worked on multi-hop open-domain question answering since 2019. Led to powerful systems like GoldEn, IRRR, Baleen, and STORM. Then it got rebranded “deep research”. A small fraction worked on LLMs as optimizers. Now that got rebranded “autoresearch”. 🤔 *-research it is!
English
11
4
151
9.7K
will brown nag-retweet
Prime Intellect
Prime Intellect@PrimeIntellect·
Today, we’re sharing how our collaboration with @nvidia helps power the open superintelligence stack. The next frontier of AI infrastructure is building systems for agentic models that can reason for hours, use tools, execute code, and learn from outcomes at scale. primeintellect.ai/blog/nvidia-co…
English
13
45
361
28.4K
will brown nag-retweet
Workshop Labs
Workshop Labs@WorkshopLabs·
Letting a provider see all your data is the price of admission for AI. We're changing that. Introducing Silo, the first private post-training and inference stack for frontier models, with hardware-level guarantees that we can’t see your data. Privacy without compromises. 🧵
Workshop Labs tweet media
English
17
35
247
35.6K
will brown
will brown@willccbb·
first time in whichever of the San’s is in south bay i always forget
English
3
0
55
8.2K
will brown nag-retweet
stochi
stochi@stochi0·
Stitched up smth fun over the weekend, prototype of an autoresearch RLM environment inspired by @karpathy, using @PrimeIntellect infra. Haven’t run full evals yet, but the setup looks like this: The model can: - modify training file - run experiments inside a sandbox - parse logs for the metric (val_bpb) - iterate to improve the score So the model does the full research loop: code, experiment, logs, hypothesis, patch, repeat Essentially turning autoresearch loop into an RLM training environment, producing trajectories of autonomous research behavior. The interesting bit would be generalizing this to: - any repo - any metric - any experiment harness - envs where model can optimize on specific pieces in a big codebase. Most importantly, this produces trajectories of autonomous research behavior. From those we can identify failure modes and iteratively improve the environment itself.👀🧋🎋 Github: github.com/stochi0/athena… Environments Hub: app.primeintellect.ai/dashboard/envi…
English
2
6
40
4.6K
Lotto
Lotto@LottoLabs·
I like my models small, chinese, dense and not thinking.
English
13
12
212
15.9K