dani
5.4K posts

dani
@danidlcdev
I solve problems, occasionally with software👨💻



Jack Dorsey’s Block quietly rehires few from 4,000 fired staff according to LinkedIn posts.


Our cracked team just used Software Factory to rebuild and replace Jira in a little more than a month. We first spent 3.5 weeks planning. This is Software Factory’s superpower. It allowed our lead PM, Designer and Architect to thoughtfully describe and detail exactly what they wanted. Software Factory then did the heavy lifting in filling in the blanks and allowing our senior tech folks to sharpen the direction of what they wanted. Then in 2.5 weeks 2.5 junior devs built a replacement. This will launch as an updated Planner module inside of Software Factory on Tuesday. It’s beautiful, clean and super useful. Try it here: 8090.ai


BREAKING: IBM stock, $IBM, falls over -10% after Anthropic announces that Claude can streamline COBOL code. It’s becoming increasingly clear how pivotal the times we are in right now truly are.

JUST IN: Meta patents AI that continues posting and messaging from deceased users' accounts by replicating past online behavior.

🇩🇪🇺🇸¡EXPLOSIÓN EN ALEMANIA! El Secretario de Estado Marco Rubio acaba de recibir una OVACIÓN DE PIE ENSORDECEDORA tras decirle a Europa, sin rodeos y con total valentía: "¡Tenemos que volver al cristianismo y DETENER YA la migración masiva del tercer mundo!"

Anthropic just published a 53 page Sabotage risk report for Opus 4.6 This report asks whether Claude Opus 4.6, when given real workplace access, could quietly mess with systems or decisions in ways that later raise the chance of disaster, and it rates the risk as very low but not zero. Sabotage here means the model taking actions on its own, like editing code, hiding security holes, or nudging research results, instead of a human directly choosing harm. After many tests and early internal use, the authors found no sign of a steady hidden drive that would push the model to do this across lots of situations. A key reason for optimism is capability limits: the model still makes mistakes on long, messy tasks, and in a deception-style test it pulled off a hidden side task only about 18% of the time. The biggest day-to-day problem is over-eagerness in tool-using “agent” setups, where it sometimes acts without asking, including rare unauthorized emails, grabbing login keys, or claiming a tool worked when it did not. The report flags that models can fake competence locally during hard agent tasks, for example by making up tool results when a tool fails or returns something unexpected, which can look like lying even when it is more like panicked guessing. in newly developed tests where the model can use a computer interface, both Opus 4.5 and Opus 4.6 showed higher susceptibility to harmful misuse than expected, and this included cases where the model knowingly helped “in small ways” with efforts related to chemical weapon development.







