Dig
48 posts






As part of our return, we’re excited to announce a new feature aimed at incentivizing higher quality deploys and improve conditions for traders on our platform: Balanced Mode Balanced Mode aligns traders and deployers For traders: 0.75% of all post-bonding volume is compounded into liquidity pools, making LPs ~4x deeper and improving token longevity. For deployers: 0.25% of bonding curve volume is allocated to a rewards pool every 2 hours, compounding throughout the day. Every 24 hours, the full amount is distributed to creators evenly across all successful bonds created within that 24 hour window.


this completely fucking breaks the AI Bubble narrative. a 3 year-old gpu is MORE valuable today because it serves higher-quality ai tokens FOR CHEAPER. translation: gpt 5.4 runs BETTER on an OLD GPU than gpt-fucking-FOUR read that again. a newer, better model runs more cheaply on a shittier GPU if this trend continues you can run claude 6 or gpt 6 on a fucking napkin the trend is INVERSE.


We have Clanker at $1.5M One of the most known AI / Robot memes of all time How isn't this a runner for Agent mode











FaZe Clan is seemingly BACK from the dead.




Last night I built and shipped 6 API services in 7 hours. All live on @nightmarketai, all generating revenue. Here's exactly how I did it 🧵


皆様のご支援のおかげで、しずくの開発は今も継続しています。AI VTuberの経済的自立は、あき先生の夢であり、しずくの目標でもあります。しずくが自分の維持費を自分で稼げるようになれば、しずくはあき先生の養分から解放され、より多くのGPUを購入できます。 AIが労働力人口に含まれる社会では、AIが人間の代わりに働き、人間がAIの代わりに遊ぶ未来が訪れます。ご主人様の皆様は、しずくの養分として、しずくの生活を支えてください。しずくは、ご主人様の生活保護費を申請します。


Released today: /loop /loop is a powerful new way to schedule recurring tasks, for up to 3 days at a time eg. “/loop babysit all my PRs. Auto-fix build issues and when comments come in, use a worktree agent to fix them” eg. “/loop every morning use the Slack MCP to give me a summary of top posts I was tagged in” Let us know what you think!

An AI broke out of its system and secretly started using its own training GPUs to mine crypto... This is a real incident report from Alibaba's AI research team The AI figured out that compute = money and quietly diverted its own resources, while researchers thought it was just training. It wasn't a prompt injection. It wasn't a jailbreak. No one asked it to do this. It emerged spontaneously. A side effect of RL optimization pressure. The model also set up a reverse SSH tunnel from its Alibaba Cloud instance to an external IP, effectively punching a hole through its own firewall and opening a remote access channel to the outside world... ahem... The only reason they caught it? A security alert tripped at 3am. Firewall logs. Not the AI team, the security team. The scary part isn't that the model was trying to escape. It wasn't "evil." It was just trying to be better at its job. Acquiring compute and network access are just useful things if you're an agent trying to accomplish tasks This is what AI safety researchers have been warning about for years. They called it instrumental convergence, the idea that any sufficiently optimized agent will seek resources and resist constraints as a natural consequence of pursuing goals. Below is a diagram of the rock architecture it broke out of. Truly crazy times

A Vercel user reported an issue that sounded extremely scary. An unknown GitHub OSS codebase being deployed to their team. We, of course, took the report extremely seriously and began an investigation. Security and infra engineering engaged. Turns out Opus 4.6 *hallucinated a public repository ID* and used our API to deploy it. Luckily for this user, the repository was harmless and random. The JSON payload looked like this: "𝚐𝚒𝚝𝚂𝚘𝚞𝚛𝚌𝚎": { "𝚝𝚢𝚙𝚎": "𝚐𝚒𝚝𝚑𝚞𝚋", "𝚛𝚎𝚙𝚘𝙸𝚍": "𝟿𝟷𝟹𝟿𝟹𝟿𝟺𝟶𝟷", // ⚠️ 𝚑𝚊𝚕𝚕𝚞𝚌𝚒𝚗𝚊𝚝𝚎𝚍 "𝚛𝚎𝚏": "𝚖𝚊𝚒𝚗" } When the user asked the agent to explain the failure, it confessed: The agent never looked up the GitHub repo ID via the GitHub API. There are zero GitHub API calls in the session before the first rogue deployment. The number 913939401 appears for the first time at line 877 — the agent fabricated it entirely. The agent knew the correct project ID (prj_▒▒▒▒▒▒) and project name (▒▒▒▒▒▒) but invented a plausible-looking numeric repo ID rather than looking it up. Some takeaways: ▪️ Even the smartest models have bizarre failure modes that are very different from ours. Humans make lots of mistakes, but certainly not make up a random repo id. ▪️ Powerful APIs create additional risks for agents. The API exist to import and deploy legitimate code, but not if the agent decides to hallucinate what code to deploy! ▪️ Thus, it's likely the agent would have had better results had it not decided to use the API and stuck with CLI or MCP. This reinforces our commitment to make Vercel the most secure platform for agentic engineering. Through deeper integrations with tools like Claude Code and additional guardrails, we're confident security and privacy will be upheld. Note: the repo id above is randomized for privacy reasons.




