ricky

21.9K posts

ricky banner
ricky

ricky

@rickyflows

ride the tiger mom

Seattle, WA Katılım Mart 2014
619 Takip Edilen2.4K Takipçiler
Sabitlenmiş Tweet
ricky
ricky@rickyflows·
Who the hell is this imposter?
English
6
2
45
4.7K
Tenobrus
Tenobrus@tenobrus·
agreed on chalamet. beyond that, the dune movies are absolutely fucking crippled by the fact that they're *movies* instead of an HBO show. they're forced into this shitty liminal space where they're simultaneously much too long and much too short to effectively tell the story.
roon@tszzl

the dune movies were doomed from the start to be good and not great due to the casting of chalamet as paul. he does not have the gravitas for a child-god and is much better suited for kind of silly coming of age movies

English
19
2
145
12.8K
Mark
Mark@marksnidal·
@jmelahman It’s called we do a little forking
English
2
0
2
45
Charlie Marsh
Charlie Marsh@charliermarsh·
We've entered into an agreement to join OpenAI as part of the Codex team. I'm incredibly proud of the work we've done so far, incredibly grateful to everyone that's supported us, and incredibly excited to keep building tools that make programming feel different.
English
285
146
3.1K
438.9K
ricky
ricky@rickyflows·
@samswoora “did claude write that for you?”
English
1
0
3
509
Samswara
Samswara@samswoora·
I can’t handle another person trying to show me their claude skill file. Its all slop
English
21
5
325
12.6K
ricky
ricky@rickyflows·
@tautologer i wonder if it would still be human readable, probably not right?
English
0
0
1
82
tautologer
tautologer@tautologer·
you can trivially (but expensively) fine-tune a compaction model by using "similarity of weight activation pattern on the uncompacted context vs compaction summary" as a reward signal, right?
English
5
0
36
1.6K
tautologer
tautologer@tautologer·
I feel like LLMs are a local optimum we're stuck at because they're easy to train from scratch because training data already abounds
English
12
1
67
2.5K
Markov
Markov@MarkovMagnifico·
@rickyflows @tautologer by why is the attention transformer the ideal vessel for scaling up compute?
English
1
0
2
45