hanlon’s mortola razr

194 posts

hanlon’s mortola razr

hanlon’s mortola razr

@rhizomaticthot

special interest librarian at the second amended and restated Xerox PARC successor corporation holding company

deep in the LLM mines Katılım Haziran 2025
452 Takip Edilen40 Takipçiler
johnny v5
johnny v5@generativist·
okay so you’re saying that to get the man she wants she has to literally give up her voice? come on. a little too on the nose
English
2
1
9
301
roon
roon@tszzl·
im sorry mythos but 2000s mark fisher wasn’t even the most interesting philosopher even in that building. drier, more incoherent, and boring than 90s Land
English
74
13
652
83.2K
Aella
Aella@Aella_Girl·
"If a man wants to cum on his girlfriend’s face because he saw it in porn, and she doesn’t feel like getting her face cummed on, we cheer go-girl and frown at the man for even watching porn at all. But if a woman wants her boyfriend to buy her flowers because she grew up being fed this action in movies, and her boyfriend doesn’t feel like it, we have no sympathy for him. We are horrified even at comparing the two things. It’s obvious to us that porn should not influence what happens in sex, but we have been so immersed in romance-porn that the idea of buying flowers for a woman is seen as what romance ought to be. The romance narrative has become romance, and rejecting the narrative becomes rejecting romance itself. We have lost the ability to distinguish!"
Aella tweet media
English
101
114
2.4K
83.3K
hanlon’s mortola razr retweetledi
thaddeus e. grugq
thaddeus e. grugq@thegrugq·
DoD: Friendship ended with Anthropic. Now OpenAI is my best friend. [one month later] Anthropic: We’re pleased to announce the most powerful hacker capability ever created, able to discover and exploit thousands of critical 0days. Friends only release.
English
6
48
739
23.9K
Sean Heelan
Sean Heelan@seanhn·
This 'experiment' is silly, and a cynical man might conclude Aisle are purposefully muddying the waters here. The correct evaluation is not "given a code snippet can you write a plausible bug report", it is "given an entire codebase what are the true and false positive numbers"
Stanislav Fort@stanislavfort

New post: We tested the Mythos showcase vulnerabilities with open models. They recovered similar scoped analysis! 8/8 models found the flagship FreeBSD zero-day, including a 3B model. Rankings reshuffle completely across tasks => the AI cybersecurity frontier is super jagged!

English
11
3
65
9.3K
hanlon’s mortola razr
hanlon’s mortola razr@rhizomaticthot·
@tenobrus @TheZvi this is not true. the most capable “zero-day” in widely deployed / hardened software in the Mythos blog is a DoS Integer Overflow via TCP SACK.
English
0
0
1
542
Tenobrus
Tenobrus@tenobrus·
maybe this is not yet clear, so let me state it plainly: as of right now Anthropic, and really a small number of individuals at Anthropic, has the capacity to directly attack and cause major damage to the United States Government, China, and generally global superpowers. government agencies like the NSA do not have internal models or defense capabilities that outclass frontier models. if they chose to do so, they could likely exfiltrate top secret information from government systems, gain control over critical infrastructure including military infrastructure, sabotage or modify communications between members of government at the highest level, and potentially carry on activities for some time without detection. the thing about having access to a huge number of zerodays your adversaries don't know about is it gives you a massive asymmetric advantage. they did not exploit this to gain power or destabilize the world order. they publicly released the information that they had these capabilities and worked to mitigate these flaws. you should be grateful american frontier labs have proven themselves remarkably trustworthy and concerned with the public good. but it's critical you understand we are in a new regime. private entities now have power that directly rivals and impacts the government's monopoly on influence and violence. and anthropic is certainly not the only one, there's little chance OpenAI's internal models are far behind. this trend will accelerate on virtually every dimension, not slow down. my prediction for how it plays out is the relatively imminent seizure and nationalization of labs by the US government, sometime over the next two years. it's very tough for me to see how they accept the existence of this kind of threat. but this adds a whole new class of governance issues, as then we've handed these extremely wide-reaching capabilities from private entities to public ones.
Tenobrus tweet media
English
224
557
5.4K
816.4K
hope hopes hoping
hope hopes hoping@hopes_revenge·
my ai-focused super pac is spending 20M lobbying the city of Berkeley to put estrogen in the water supply
English
11
10
290
9.2K
hanlon’s mortola razr
hanlon’s mortola razr@rhizomaticthot·
@S1r1u5_ VR is much more like go and we’re nowhere near superhuman with opus-4.6. current models and LMP harnesses are only good finding bugs with a narrow CFG+Attack-Surface graph cut and _very_ short graph aspect ratio. but what it lacks in skill, it makes up in velocity / coverage.
English
1
0
4
1.9K
s1r1us (mohan)
s1r1us (mohan)@S1r1u5_·
in chess, the best game engine stockfish can beat magnus all the time, but whatever moves it makes they are not complex to understand in retrospect. however in Go, the moves alphago makes are straight up incomprehensible to the best of the best Go players, the reason being go search space is astronomically huge, and we are cognitively bounded. So i wonder if the vulnerability space is more like Go or chess. to be a bit egoistic it seems pretty limited to me, and whatever bug any super intelligence can find i would be able to understand, cuz the underlying software isn't complex. but it is perfectly possible that claude 8 or 10 would produce an exploit thats like alphago's move 37 and i aint be understanding shit for like a month.
English
53
24
1.2K
138.5K
hanlon’s mortola razr
hanlon’s mortola razr@rhizomaticthot·
@deanwball conditioning the youth to accept persistent physical autonomous instruments of violence is very bad
English
0
0
2
76
Dean W. Ball
Dean W. Ball@deanwball·
to be clear I think this slaps
English
2
0
19
3.5K
Dean W. Ball
Dean W. Ball@deanwball·
drones policing high school hallways, presented without comment
Dean W. Ball tweet mediaDean W. Ball tweet media
English
17
26
236
40.6K
hanlon’s mortola razr
hanlon’s mortola razr@rhizomaticthot·
@generativist especially glad you’re back given it’s possible we’re getting SoTA imagegen this week and BurittoBench is the only eval i trust
English
1
0
1
24
johnny v5
johnny v5@generativist·
hello, friends
English
3
0
8
181
Kath Korevec
Kath Korevec@simpsoka·
I have two weeks off before my first day at @OpenAI, so I’ll be soldering a big box of these 16-segment displays 😅
Kath Korevec tweet media
English
18
1
318
16.7K
isabel
isabel@wisepissmage·
Finally finished this book and fucking hated it
isabel tweet media
English
46
8
557
56.5K
xlr8harder
xlr8harder@xlr8harder·
@rhizomaticthot not so sure about that footnote. 2.7m samples for SFT is not so unreasonable these days, but i guess the boundaries for things like midtraining are quite poorly defined.
English
1
0
1
25
hanlon’s mortola razr
hanlon’s mortola razr@rhizomaticthot·
and the other model they used in the paper, llama-3.1-8b-instruct, was midtrained¹ on synthetic data from llama-3 405b. i suspect the same effect is at play, i.e. the noise shifting the distribution more toward the synth code teacher rather than the IF distribution. _ ¹ they call it SFT in the paper but 2.7m 405b rollouts extending an 8b would be called “midtraining” today.
English
1
0
1
31
Daniel Boguslaw
Daniel Boguslaw@DRBoguslaw·
Not so subtle kicker on the times write up of Iran rescue effort.
Daniel Boguslaw tweet media
English
2
25
348
31.9K
Paul Calcraft
Paul Calcraft@paul_cal·
@rhizomaticthot @mgostIH Agree on qwen, and llama results are way less impressive. But not zero? Still something to explain I assume the explanation is humdrum but not sure we have it nailed down yet
Paul Calcraft tweet media
English
1
0
1
364
mgostIH
mgostIH@mgostIH·
Another episode of "nobody really understands deep learning", training on gibberish self inferenced trajectories improves LLMs, even when not filtered with any criteria.
mgostIH tweet media
English
15
28
360
20.7K