Michael Allen

7.6K posts

Michael Allen

Michael Allen

@_Dark_Knight_

Building at the intersection of AI + offensive security::https://t.co/5DiZGjk6VJ

Seattle, WA Katılım Mart 2009
277 Takip Edilen1.5K Takipçiler
Val Smith
Val Smith@mvalsmith·
@Dave_Maynor @zeroxjf @trq212 I've suggested having a vetted cyber-researcher option or something but have been ignored. I canceled most of the LLM subscriptions, bought some rtx 6000s, and am using local LLMs with cyber friendly system prompts now. Done with nanny no privacy LLMs.
English
1
0
0
47
johnny
johnny@zeroxjf·
The new cyber-abuse guardrails in Opus 4.6 are likely to drive a mass exodus of researchers from the platform. They give option to submit a form to prove legitimate research, but for me got no confirmation of its submission last week and no way of knowing its status 🤷‍♂️ @trq212
English
11
5
59
22.6K
johnny
johnny@zeroxjf·
@EvanKlein338226 @trq212 2 weeks?! 🤦🏻‍♂️ it’s like they want people to flock to codex even more rapidly
English
1
0
0
448
ᴅᴀɴɪᴇʟ ᴍɪᴇssʟᴇʀ 🛡️
MCP is confusing. Some think it died when CLIs took over, but you can also view them as complimentary. Like MCP is how you stay fresh on what's available in the tool/service, and how to use it, and CLI is how you actually execute.
English
11
4
38
5K
Michael Allen
Michael Allen@_Dark_Knight_·
@mattpocockuk ok kewl...was running into some weirdness with processing each issue -- think I it was due to the updates I made to the script -- eventually worked but was curious if I was missing something -- thanks
English
0
0
1
3
Michael Allen
Michael Allen@_Dark_Knight_·
@mattpocockuk I was curious on what your ralph-loop looks like to work through the issues created from the /prd-to-issues skill. Right now I do something like 5. Push and create a PR, then merge it. 6. Close the issue. But what is your approach?
English
1
0
0
37
Michael Allen
Michael Allen@_Dark_Knight_·
so I thought I had a bug and man was claude confident! Gave it to codex to review with all the reports and context and it said..do not report this..went back to claude who then said...yeah I overstated somethings
English
0
0
2
119
Bo Bleyl
Bo Bleyl@BleylDev·
Claude auth having issues this morning. Claude Code triggers login screen. Login screen hangs for ~20 seconds. Problem is, Claude Code has a 15 second time-out for waiting on auth. So by the time the success message comes through, claude code ignores it due to timeout. Lovely.
English
1
0
6
934
Tib3rius
Tib3rius@0xTib3rius·
You may not like it, but this is what peak hacking looks like.
Tib3rius tweet media
English
39
36
721
36.3K
Stefan Boesen
Stefan Boesen@stefanboesen·
Love that codex knows Ruby is a superior language to Python
Stefan Boesen tweet media
English
1
0
1
57
Michael Allen retweetledi
Stefan Boesen
Stefan Boesen@stefanboesen·
OpenAI released Symphony this week. I tried implementing the same pattern to see how it behaves. Most of the work ended up in review loops, artifacts, visibility, and PRs. Notes: blog.boesen.me/posts/lessons-…
Stefan Boesen tweet media
English
0
2
5
194
Michael Allen
Michael Allen@_Dark_Knight_·
[BLOG] mykalseceng.github.io/posts/agentic-… | ODYSSEUS: Building an Agentic Pentest Platform -- How I built a multi-stage agentic pentest pipeline, what it found and missed, and how to use the approach in your workflows
English
0
0
0
99
Michael Allen
Michael Allen@_Dark_Knight_·
"I'm not giving you the answer — I'm grilling you. Let me reframe:" -- wait what?
English
0
0
0
42
STÖK ✌️
STÖK ✌️@stokfredrik·
@_Dark_Knight_ @HackingLZ At the same time they have a junior dev, the marketing team, the cfo all smashing metrics, building stuff and being productive using personal accounts. The disconnect is real and compliance and legal will have a hard time catching up.
English
1
0
2
877
STÖK ✌️
STÖK ✌️@stokfredrik·
I once said: AI is not going to take your job as a pentester or bugbounty hunter. I was wrong.
English
95
50
766
119.9K
Michael Allen
Michael Allen@_Dark_Knight_·
@HackingLZ @stokfredrik "For example, many wouldn't allow external or even internal people to run their data through frontier models during offensive testing." -- 100% this
English
1
0
1
752
Justin Elze
Justin Elze@HackingLZ·
I have no doubt AI tooling will augment testing in lots of ways, so if that means fewer OffSec jobs, I get it. We were in a period where OffSec was "easy" and people forgot the job was supposed to get harder over time instead, boot camps told people they would make X after 12 weeks. The nature of most organizations security programs is a little more complex than their public facing bug bounty programs. The leap most people are making is that AI will close the gap to 80% and someone with no domain knowledge will drive that 80%. There is also a whole "replace a job vs. tasks" argument that all of AI land is currently having.| Another somewhat useful point bug bounties largely avoid the data protection requirements companies have. For example, many wouldn't allow external or even internal people to run their data through frontier models during offensive testing. The greater tipping point in the replacement discussion will come when local models reach a certain capability threshold, because it will allow companies to maintain safeguards while still meeting compliance and regulatory requirements. In that same space, there's also a lack of training data for internal pentesting and other areas compared to much of the bug bounty landscape.
English
5
8
76
31K
Matt Pocock
Matt Pocock@mattpocockuk·
I have an AI that writes stuff for me. Here are all the phrases I've banned it from using: real power wake-up call fundamentally changes key benefit cut through the noise key insight the irony the good news the reality it's kind of like here's the thing: hard truth uncomfortable truth
English
58
7
329
28.3K