Aryan Seth

733 posts

Aryan Seth banner
Aryan Seth

Aryan Seth

@AryanSeth07

your cat said pspsps to me

Katılım Kasım 2017
1.5K Takip Edilen270 Takipçiler
єℓαιηє
єℓαιηє@elainesim28·
Please pray for my husband, he got stung by a bee in the forehead. He’s in the hospital now, his face all swollen and bruised. He almost died. Luckily I was close enough to hit the bee with a shovel.
English
1.5K
3.5K
61.2K
1.2M
Aryan Seth retweetledi
Ali Romman
Ali Romman@aliromman_·
@OpenAI You should celebrate by resetting rate limits
English
16
7
690
41K
Aryan Seth
Aryan Seth@AryanSeth07·
chatgpt plus codex limits feel so much lower than before
English
0
0
1
83
Aryan Seth
Aryan Seth@AryanSeth07·
@willccbb but plain OPD has issues - mostly pointed out in this paper; arxiv.org/pdf/2604.03128 "privileged information" leads to an irreducible loss term, and i have not found much luck with training stability (maybe this is a personal issue lol)
English
0
0
1
105
p
p@FreezingToe·
@AryanSeth07 “You’re absolutely right!”
English
1
0
3
33
Aryan Seth
Aryan Seth@AryanSeth07·
sharp insights
Aryan Seth tweet media
English
1
0
4
82
Aryan Seth
Aryan Seth@AryanSeth07·
tinker is nice i wish i could afford it consistently
English
0
0
2
35
SpaceX
SpaceX@SpaceX·
SpaceXAI and @cursor_ai are now working closely together to create the world’s best coding and knowledge work AI. The combination of Cursor’s leading product and distribution to expert software engineers with SpaceX’s million H100 equivalent Colossus training supercomputer will allow us to build the world’s most useful models. Cursor has also given SpaceX the right to acquire Cursor later this year for $60 billion or pay $10 billion for our work together.
English
2.4K
5.1K
38.4K
20.6M
Aryan Seth retweetledi
Sanjal Sangle
Sanjal Sangle@gunnergworl·
thanks to arsenal there will be no drought this year my tears shall fill every river
English
12
25
385
6.2K
Aryan Seth retweetledi
Sanjal Sangle
Sanjal Sangle@gunnergworl·
I quit football
English
89
282
1.9K
46.3K
P.P.C
P.P.C@ppc0x23·
@_lyraaaa_ they all track the same underlying signal but the directions are different because each layer performs a Richelot step on the representation. the polynomial gets factored. the curve transforms. the class persists.
English
1
0
1
192
lyra bubbles
lyra bubbles@_lyraaaa_·
found a vector in gemma4 e4b that roughly correlates with perplexity the more surprisal in the passage, the stronger this vector fires R²=0.775, corr=0.880, p<10⁻³³²³⁶ this is avg across all layers, L20 and 21 have R² 0.828 and 0.845 alone wonder what i can do with this
English
32
14
597
43K
Aryan Seth retweetledi
Lakshya
Lakshya@tokenaware·
We got into @ycombinator! A few months ago, @onkar_borade_10, @SujaySriv , and I met at a football game in HSR, Bengaluru, and went deep on one question - Why does SaaS take months and a huge team to get delivered after the sale? Building a good product should be enough, right? Right? Messy integrations. Handoffs. Siloed information. Poor documentation. Fragmented data. The list goes on. We started a company with the vision to make SaaS self-serve. @lab0_ai Huge thanks to @dessaigne , @collinmathilde , and the YC team for this opportunity. If your product rollouts take months or you're a system integrator/partner/FDE implementing SaaS, let's talk. Link below.
Lakshya tweet mediaLakshya tweet mediaLakshya tweet media
English
65
19
413
55.7K
Aryan Seth retweetledi
Cosmos Raj
Cosmos Raj@cosmos_raj·
Breaking news: Anthropic buys the All in podcast just to shut it down Dario quoted as saying: “this isn’t even about new media I just want to stop seeing them on my timeline”
Cosmos Raj tweet mediaCosmos Raj tweet media
English
87
187
6.5K
582.9K
Aryan Seth
Aryan Seth@AryanSeth07·
@viplismism sure, harnessing works at inference, but I still don't follow how this solves sparse reward because that's a training-time problem; recursive call = tool call, so you can train using the subagent output, but the reward is still a scalar (higher due to harness), but sparse
English
1
0
1
42
vipli
vipli@viplismism·
@AryanSeth07 true in principle i guess, but in practice the harness has to do the heavy lifting man. models tend to jump to conclusions or lose context fast if the harness isn't driving the search process properly
English
1
0
0
52
vipli
vipli@viplismism·
most people don't realize that rlms are just solving the sparse reward problem for long context! instead of an llm hunting for checkmate in one giant forward pass, it's like you break it into bite-sized reasoning tasks. every recursive step is a checkpoint where the model updates its internal value of the context before moving to the next piece it turns a massive search space into a dense signal
English
1
0
8
708
Aryan Seth
Aryan Seth@AryanSeth07·
there's going to be a movie at some point about hinton being like oppenheimer or something with the toronto lab being the ai manhattan project and sam altman and co being like the govt and scaling the models
English
0
0
3
104