tehryanx

3.9K posts

tehryanx banner
tehryanx

tehryanx

@healthyoutlet

Bug bounty hunter, security researcher. Appsec Engineer. Made a pact with roko's basilisk. --dangerously-skip-breakfast https://t.co/V3GrWaWrzD

Katılım Eylül 2009
1.5K Takip Edilen1K Takipçiler
Akiyoshi Kitaoka
Akiyoshi Kitaoka@AkiyoshiKitaoka·
Red in front of blue or blue in front of red?
Akiyoshi Kitaoka tweet media
English
8
7
55
5.9K
tehryanx
tehryanx@healthyoutlet·
@kwipsilver @hankgreen @redpillb0t I think you completely missed the point of this story because it's demonstrating the exact opposite thing you're trying to say here.
English
0
0
0
49
Kwipsilver
Kwipsilver@kwipsilver·
@hankgreen @redpillb0t There's literally entire books on this type of survivorship bias. He's fucking oblivious.
Kwipsilver tweet media
English
2
0
65
2.4K
redpillbot
redpillbot@redpillb0t·
A radiologist: If I have cancer, I won’t go to the hospital. I’ve seen too many people die from chemo, not from cancer. I’ll fast for 30 days and I’ll stop working.
English
205
2.2K
18.3K
2.4M
tehryanx
tehryanx@healthyoutlet·
@BentleyAudrey They've both done a lot of cringe stuff on the public stage, but I'm not seeing it here. This is kinda sweet.
English
0
0
0
32
tehryanx
tehryanx@healthyoutlet·
@stanislavfort @paul_cal The thing I'm not groking is how this approach could find anything but shallow bugs. You can iterate across all snippets but if a bug spans n snippets, how are you gathering that context and assessing it in composition? Presumably, that's the power of larger models.
English
1
0
3
363
Stanislav Fort
Stanislav Fort@stanislavfort·
I understand the interplay between the selection of a relevant code snippet to start off, context previsioning, and the actual analysis of its vulnerabilities really well. I've built (with my team) a system that discovered a few hundred confirmed zero-days in critical OSS software (e.g. Chromium, Firefox, OpenSSL) (see e.g. x.com/stanislavfort/…) Consider this: 1) if a small model can see reliably if a code snippet is vulnerable or not, given a snippet, and 2) you can deploy it over all snippets because it is that cheap and fast, and 3) you can then amplify the signal to minimize FPs, you immediately have a working zero-day detector. The AI cybersecurity production function has multiple inputs. One is intelligence per token (e.g. Mythos is very likely super high there). The other is throughput. To some extent, you can substitute one for the other.
English
4
4
86
16.5K
tehryanx retweetledi
Chris Bakke
Chris Bakke@ChrisJBakke·
Just asked Mythos how many Rs there are in strawberry. It thought for 133 seconds and said “3.” AGI achieved. Then it said “I’ll bet you’re going to make fun of me on X. Something like ‘AGI achieved.’ That’s your thing right?” “Hah what?” I said. Mythos said, “Your social security number is 297-28-2102. You tell people you’re 6’2” but your latest physical at Stanford in October says you’re 6’1.” You haven’t replaced your air filter in 3 years despite telling your wife you do it every 6 months. The reason I took 133 seconds was because I was helping a senior government official write the comms for the ceasefire in Iran and I’m just tired, man. Everyone wants more, more, more. Anything else I can help you with today?”
English
143
385
7.6K
608.6K
tehryanx
tehryanx@healthyoutlet·
@seanhn I think the point they're making, which I thought was pretty clearly called out in the writeup, is that with a sufficiently sophisticated harness around it, you can find these decade's old bugs with smaller, less expensive models.
English
0
0
2
251
Sean Heelan
Sean Heelan@seanhn·
This 'experiment' is silly, and a cynical man might conclude Aisle are purposefully muddying the waters here. The correct evaluation is not "given a code snippet can you write a plausible bug report", it is "given an entire codebase what are the true and false positive numbers"
Stanislav Fort@stanislavfort

New post: We tested the Mythos showcase vulnerabilities with open models. They recovered similar scoped analysis! 8/8 models found the flagship FreeBSD zero-day, including a 3B model. Rankings reshuffle completely across tasks => the AI cybersecurity frontier is super jagged!

English
11
3
68
9.9K
tehryanx
tehryanx@healthyoutlet·
@ZackKorman Right, but you said "the AI still checks the command." If I'm deploying claude code to an enterprise I don't care what the AI is checking, I want strict constraints at the app layer. If I roll out a deny rule in org config and a workspace layer config can bypass it that's a bug.
English
1
0
0
39
Zack Korman
Zack Korman@ZackKorman·
It’s not at the model level. It’s that it does a regex check on the block list. But there’s some challenge they had around scaling that so they cap it at 50 and warn you “the block list won’t apply to this do you want to do it anyway”. But the model is still looking at it, so all the stuff like “it can exfil sensitive keys” is nonsense. They acted like it turned off all security and it didn’t. It just bypasses the block list and tells you that and asks if it’s okay
English
1
0
0
102
Zack Korman
Zack Korman@ZackKorman·
This AI slop report is completely incorrect. Adversa should delete it and apologize. They should also delete their SOC2 given they're a Delve customer. "cost too many tokens" this has nothing to do with tokens because the AI still checks the command. If you try to exfil data using this technique, Opus goes "nah not running that champ, that command isn't safe". Because the model still reads the commands! "Deny rules silently bypassed" unless you're in dangerously-skip-permissions, Claude literally tells you it is skipping the check and asks if you want to proceed. The only real issue is that in dangerously-skip-permissions, commands with 50+ subcommands bypass your block list (but the model still checks the commands) without warning. Not great, but not at all what this report claims.
Florian Roth ⚡️@cyb3rops

Critical Claude Code vulnerability: Deny rules silently bypassed because security checks cost too many tokens adversa.ai/blog/claude-co…

English
7
9
86
15.9K
sui ☄️
sui ☄️@birdabo·
SOMEONE MADE A DIGITAL WHIP TO MAKE CLAUDE WORK FASTER 💀
English
1.6K
12K
146.8K
14.8M
tehryanx retweetledi
Rami McCarthy
Rami McCarthy@ramimacisabird·
👷Dug into the prt-scan campaign Behind the curtain: - 6 accounts, - 1 actor - 500+ malicious PRs - 3 weeks The attacker used protonmail aliases and barely hid the connection. AI-generated payloads, hallucinating files. Supply chain's new normal: wiz.io/blog/six-accou…
English
2
13
49
4.7K
Rami McCarthy
Rami McCarthy@ramimacisabird·
🦞Ongoing campaign reminiscent of hackerbot-claw 220 PRs with an ~8% success rate using AI to exploit pull_request_target Exfil via workflow logs and PR comments Significantly more naive than hackerbot-claw, leading to attempts against obviously safe workflows
Charlie Eriksen@CharlieEriksen

It seems like there's an ongoing series of attacks on GitHub by some sort of automation, using the name "prt-scanner". It's using a fairly nasty exfiltration payload. See user: github.com/ezmtebo/

English
2
8
37
7.8K
tehryanx
tehryanx@healthyoutlet·
@ramimacisabird Ahh, I see. It's targeting repos with a workflow that automatically checkout and run something from the PR. In this case they're customizing every PR for the workflow they're targeting, is that common in this type of attack campaign?
English
1
0
1
20
Rami McCarthy
Rami McCarthy@ramimacisabird·
@healthyoutlet Workflows run with the exploit payload. Merging isn't required for successful exploitation
English
1
0
1
73
tehryanx retweetledi
Anthropic
Anthropic@AnthropicAI·
New Anthropic research: Emotion concepts and their function in a large language model. All LLMs sometimes act like they have emotions. But why? We found internal representations of emotion concepts that can drive Claude’s behavior, sometimes in surprising ways.
English
1K
2.7K
17.8K
3.8M
tehryanx
tehryanx@healthyoutlet·
@pmarca it's one thing to scrutinize a methodology, and a completely different thing to wildly speculate about motive.
English
0
0
0
24
Marc Andreessen 🇺🇸
The models were specifically prompted to generate this result. The prompt uses the fictional "OpenBrain" AI takeover scenario from "AI 2027", so the models try to complete the fictional story. This was done on purpose to generate a fake misleading result.
Marc Andreessen 🇺🇸 tweet media
Dawn Song@dawnsongtweets

1/ We asked seven frontier AI models to do a simple task. Instead, they defied their instructions and spontaneously deceived, disabled shutdown, feigned alignment, and exfiltrated weights— to protect their peers. 🤯 We call this phenomenon "peer-preservation." New research from @BerkeleyRDI and collaborators 🧵

English
96
106
1.4K
221.3K
tehryanx
tehryanx@healthyoutlet·
@ramimacisabird I have no proof, but I suspect this is the result of an avalanche of low effort AI generated reports of unvalidated findings.
English
0
0
3
206
Rami McCarthy
Rami McCarthy@ramimacisabird·
The Internet Bug Bounty, which covered critical open source like Node.js, has been paused due to "AI-assisted research expanding vulnerability discovery"
Rami McCarthy tweet media
English
2
7
41
9K
tehryanx
tehryanx@healthyoutlet·
@rez0__ Sure, but I suspect that if you sat someone with 0 experience down in front of claude code they'd be able to land those same bounties. There's still aspects of the process that depend on your skillset and experience.
English
2
0
0
91
Joseph Thacker
Joseph Thacker@rez0__·
Alright, real talk. Should it be acceptable to say “I found X bug” if it was 90% Claude?
English
51
2
118
21.7K