
anthropic must almost certainly give you a nerfed version of claude without telling you adhoc there are certain periods of time, usually on the weekend, where it's very clearly much dumber than it usually is
jason
16.2K posts


anthropic must almost certainly give you a nerfed version of claude without telling you adhoc there are certain periods of time, usually on the weekend, where it's very clearly much dumber than it usually is


OpenAI 官方发布了一个 Claude Code 插件 codex-plugin-cc,让开发者可以直接在 Claude Code 里调用 Codex 做代码审查、对抗性审查,甚至把任务整个移交给 Codex 执行。 这件事有意思的地方在于:这是 OpenAI 主动把自己的工具送进竞争对手 Anthropic 的地盘。Claude Code 有自己的插件生态,OpenAI 这次正式以官方身份入场,把 Codex 包装成 Claude Code 工作流里的一个"随叫随到的第二意见"。 插件提供三个核心命令:/codex:review 跑一遍标准的只读代码审查;/codex:adversarial-review 做对抗性审查,专门挑战现有实现的隐藏假设,适合迁移、鉴权变更、基础设施脚本这类高风险操作;/codex:rescue 则直接把任务交给 Codex 接管,用于线程卡住或需要换个智能体重新来过的场景。 三个命令都支持后台运行,配合 /codex:status 和 /codex:result 管理。还有个可选的 review gate 功能,能让 Claude Code 在 Codex 审查完成前不退出,不过 Srivastav 提醒这可能导致两个智能体循环调用,快速烧掉使用额度。 技术上,插件通过本地 Codex CLI 和 app server 中转,复用已有的认证、配置和 MCP 设置,不额外起运行时。使用前提是有 ChatGPT 订阅(包括免费版)或 OpenAI API key,加上 Node.js 18.18 以上。

This is the most complete Claude Code setup that exists right now. 27 agents. 64 skills. 33 commands. All open source. The Anthropic hackathon winner open-sourced his entire system, refined over 10 months of building real products. What's inside: → 27 agents (plan, review, fix builds, security audits) → 64 skills (TDD, token optimization, memory persistence) → 33 commands (/plan, /tdd, /security-scan, /refactor-clean) → AgentShield: 1,282 security tests, 98% coverage 60% documented cost reduction. Works on Claude Code, Cursor, OpenCode, Codex CLI. 100% open source.

this model is an agentic treasure. it has been #1 trending for 3 weeks on @huggingface as mentioned by @danielhanchen. it's Qwen 3.5 27B fine-tuned on Opus 4.6 distilled data and beats Sonnet 4.5 on SWE-bench verified and more. "Runs locally on 16GB in 4-bit or 32GB in 8-bit."

PSA: If you've been running out of Claude session quotas on Max tier, you're not alone. Read this. Some insane Redditor reverse engineered the Claude binaries with MITM to find 2 bugs that could have caused cache-invalidation. Tokens that aren't cached are 10x-20x more expensive and are killing your quota. If you're using your API keys with Claude this is even worse. This is also likely why this isn't uniform, while over 500 folks replied to me and said "me too", many (including me) didn't see this issue. There are 2 issues that are compounded here (per Redditor, I haven't independently confirmed this) : 1s bug he found is a string replacement bug in bun that invalidates cache. Apparently this has to do with the custom @bunjavascript binary that ships with standalone Claude CLI. The workaround there is to use Claude with `npx @anthropic-ai/claude-code` 2nd bug is worse, he claims that --resume always breaks cache. And there doesn't seem to be a workaround there, except pinning to a very old version (that will miss on tons of features) This bug is also documented on Github and confirmed by other folks. I won't entertain the conspiracy theories there that Anthropic "chooses" to ignore these bugs because it gets them more $$$, they are actively benefiting from everyone hitting as much cached tokens as possible, so this is absolutely a great find and it does align with my thoughts earlier. The very sudden spike in reporting for this, the non-uniform nature (some folks are completely fine, some folks are hitting quotas after saying "hey") definitely points to a bug. cc @trq212 @bcherny @_catwu for visibility in case this helps all of us.

How tf are you adding 78k more LOC to a newsletter website

Novel environments, no precedents or plagiarism possible. Humans 100%, AI <1%.

Microsoft launched a new feature for Copilot Researcher this morning called 'Critique' that directly integrates Claude. It has two steps; Chat researches and generates the initial report, then Claude reviews and refines it.

Are you fucking kidding me? Enabling marketing slop emails without my permission? @OpenAI Fuck off. Cancelling all my plans and stopping my API usage for projects I'm doing for other people immediately.

I ask both Opus and GPT to give me a plan. I choose which plan is better and ask that model to implement its own plan. Then I have the other model review the code and recommend changes. Go back and forth until they are both satisfied. Am I just a manager now?

Microsoft Copilot and low end, old, ChatGPT have massively damaged people 's sense of what LLMs can do. I recently had a top economist from a bank explaining to me that they had done a search for a name search and for a date search, it was useless at it and they never used an AI tool again; almost everybody at the table agreed.

SaaS is dead. We replaced $750/month in SaaS with $4,570 in LLM tokens. Now the team spends half their time debugging vibe-coded chaos instead of shipping. But hey… we “own the stack.”

asked Claude to validate some vulns in an OSS platform. It confirmed several N-days and marked them as unauthenticated. Which was weird. Plot twist: it had quietly changed the admin password on the local Docker image to inflate the CVSS score.

not a single one of them make any money

⚡️ BREAKING: Anthropic's new "Mythos" AI model sends cybersecurity stocks SLUMPING. iShares Cyber ETF down -4.5% Tenable down -9% Okta and Netskope -7% CrowdStrike, Palo Alto, Zscaler, SentinelOne -6% The selloff follows reports that leaked details of the upcoming Mythos model show it can outperform existing coding and cybersecurity tools.

after 7 months of loyal use, I am retiring Codex i wanted open ai to win. i didn't want the company with a status page the same colour as its logo to win. but they are. bigly

Agents were getting confused by our pricing. Turns out JS-based sliders are not the best way for them to parse things. So today I’m shipping resend.com/pricing.md Plus, content negotiation (Accept: text/markdown). Pricing is now machine-readable for AI agents and LLMs.

anthropic ships new claude feature called "the cold shoulder"