Andrés
1.3K posts

Andrés
@eigenoid
AI | Systems | https://t.co/CjvBlk0PEn | opinions are my own


🚨 AWS DevOps Agent is finally here! On March 31, 2026, AWS DevOps Agent became generally available. This is actually a big deal. It can: → Generate CI/CD pipelines → Debug failed deployments → Suggest infrastructure changes → Analyze logs and incidents → Help with Terraform & CloudFormation → Recommend cost optimizations → Explain AWS architecture issues Basically, it’s like having a junior DevOps engineer inside AWS. DevOps is slowly becoming AI-assisted operations. Source: aws.amazon.com/devops-agent/

🚨 CRITICAL: Active supply chain attack on axios -- one of npm's most depended-on packages. The latest axios@1.14.1 now pulls in plain-crypto-js@4.2.1, a package that did not exist before today. This is a live compromise. This is textbook supply chain installer malware. axios has 100M+ weekly downloads. Every npm install pulling the latest version is potentially compromised right now. Socket AI analysis confirms this is malware. plain-crypto-js is an obfuscated dropper/loader that: • Deobfuscates embedded payloads and operational strings at runtime • Dynamically loads fs, os, and execSync to evade static analysis • Executes decoded shell commands • Stages and copies payload files into OS temp and Windows ProgramData directories • Deletes and renames artifacts post-execution to destroy forensic evidence If you use axios, pin your version immediately and audit your lockfiles. Do not upgrade.

Ollama is now updated to run the fastest on Apple silicon, powered by MLX, Apple's machine learning framework. This change unlocks much faster performance to accelerate demanding work on macOS: - Personal assistants like OpenClaw - Coding agents like Claude Code, OpenCode, or Codex

@SocketSecurity UPDATE in case you missed it earlier: This is bigger than initially reported. Both axios@1.14.1 AND axios@0.30.4 were compromised – the attacker poisoned the 1.x and 0.x branches within 39 minutes of each other, maximizing blast radius across projects using caret ranges.



what the actual fuck is he talking about

Introducing TurboQuant: Our new compression algorithm that reduces LLM key-value cache memory by at least 6x and delivers up to 8x speedup, all with zero accuracy loss, redefining AI efficiency. Read the blog to learn how it achieves these results: goo.gle/4bsq2qI








