Pove (tokyo)

5.3K posts

Pove (tokyo) banner
Pove (tokyo)

Pove (tokyo)

@Pove_iOS

頑張ってね✨

Tokyo-to, Japan Inscrit le Nisan 2023
659 Abonnements226 Abonnés
.
.@frxzenwave·
i wish i could restart my whole life
English
1
0
9
99
Pove (tokyo)
Pove (tokyo)@Pove_iOS·
@craigzLiszt as a developer i don't think benchmarks are valid (especially if you know what they are)
English
0
0
0
10
Craig Weiss
Craig Weiss@craigzLiszt·
ai labs are being very sketchy with how they’re advertising and benchmarking their models
English
23
4
73
2.7K
Pove (tokyo)
Pove (tokyo)@Pove_iOS·
@craigzLiszt i like when it speaks well and knows what i'm saying and doesnt make bad assumptions. accuracy and humanity.
English
0
0
0
37
Craig Weiss
Craig Weiss@craigzLiszt·
what do you find most attractive about your coding agent?
English
66
0
52
3.5K
Pove (tokyo)
Pove (tokyo)@Pove_iOS·
you met me in a very chinese time in my life
Pove (tokyo) tweet media
English
0
0
0
13
★
@midgirlfriend·
can you guys tell im not allowed outside
English
1
2
21
251
Pove (tokyo)
Pove (tokyo)@Pove_iOS·
no, it's just curiosity, the same as academia. curiosity + rigor = explanation. propagandists call independent explorations conspiracy theory, because they deem the state and commercial institutions as the *only necessary and qualified authors of truth*. i'd say your claim is backwards. -those who accept institutional dogma without question do so to make the world to make sense -those who investigate the world in its ambiguity trade certainty for clarity
English
0
0
0
7
Sabine Hossenfelder
i think people are drawn to conspiracy theories because that way they get to make sense of a world that just doesn't make sense
English
871
101
1.4K
85.5K
Luke
Luke@Luke_The_Tako·
🌂
QME
21
345
6.2K
261.4K
Pove (tokyo) retweeté
Notsu
Notsu@Notsu311406·
ZXX
6
152
1.2K
14.5K
Pove (tokyo) retweeté
リポ・トウ
リポ・トウ@riposhima·
寝てね
リポ・トウ tweet media
日本語
2
64
2.1K
21.3K
白まもの
白まもの@shiromamono·
古明地こいしちゃん(1/2) Koishi Komeiji (Touhou Project)
白まもの tweet media
日本語
193
7.7K
92.8K
1.6M
Pove (tokyo)
Pove (tokyo)@Pove_iOS·
@uki_hajime あなたの絵のスタイルがすごく好きです。🌝
日本語
1
0
0
11
☆
@heyReze·
☆ tweet media
ZXX
37
272
1.5K
14.8K
沫柠
沫柠@Moningmeng·
GIF
ZXX
96
5.3K
69.1K
1.2M
💖
💖@twaniimals·
💖 tweet media
ZXX
38
3.7K
24.4K
550.6K
Pove (tokyo)
Pove (tokyo)@Pove_iOS·
@k1rallik i noticed this and it really crushes down development. it feels like it did 6+ months ago, when claude was buggy all the time. it's kinda sad, how we had a good thing, and now it's gone. is there any way to have a good chatbot again?
English
0
0
1
304
BuBBliK
BuBBliK@k1rallik·
🚨do you understand what just happened with Claude.. Anthropic quietly cut Claude's thinking depth by 67%.. didn't announce it.. didn't explain it.. an AMD AI Director had to dig through session logs just to prove it happened. median reasoning dropped from ~2,200 to ~600 characters.. API calls went up 80x.. meaning Claude thinks less, fails more, retries more. and YOU burn more tokens paying for those retries. they added a header that hides Claude's thinking from your logs. so when the model analyzed itself it found blank pages. and concluded it had stopped thinking.. you're paying $200/month for a model that can't read its own diary.. the thinking didn't disappear.. it just became invisible.. and Anthropic said nothing until the numbers went public
ℏεsam@Hesamation

AMD Senior AI Director confirms Claude has been nerfed. She analyzed Claude's session logs from Janurary to March: > median thinking dropped from ~2,200 to ~600 chars > API requests went up 80x from Feb to Mar. less thinking and failed attempts meaning more retries, burning more tokens, and spending more on tokens > reads-per-edit dropped from 6.6x → 2.0x. model stops researching code before touching it. > model tried to bail out or ask "should i continue" 173 times in 17 days (0 times before March 8). > self-contradiction in reasoning ("oh wait, actually...") tripled. > conventions like CLAUDE.md get ignored because there's less thinking budget to cross-check edits > 5pm and 7pm PST are the worst hours, late night is significantly better. this means the thinking allocation is most likely GPU-load-sensitive.

English
176
301
3K
562K