Eric MacDougall
1K posts

Eric MacDougall
@ericmacdougall
Co-Founder @ Good Ventures
Victoria, British Columbia انضم Temmuz 2009
2.6K يتبع22.6K المتابعون

EU AI Act conformity assessment has a structural gap for multi-agent systems. Individual agent assessment can't predict system-level emergent behavior.
Hammond et al. 2025 (Cooperative AI Foundation, 44+ authors across Oxford/DeepMind/Anthropic/CMU) taxonomize three multi-agent failure modes: miscoordination, conflict, collusion. Seven risk factors including emergent agency. None are visible when you audit agents individually.
The Digital Omnibus now proposes extending high-risk deadlines to Dec 2027 because the infrastructure (standards, notified bodies) isn't ready. The framework for evaluating systems, not just components, isn't written yet.
For anyone deploying multi-agent workflows in regulated sectors: you're going to be responsible for bridging the assessment gap yourself.
English

LLMs don't manipulate discrete symbols. They manipulate vectors. So Harnad's 1990 symbol grounding problem isn't the one that applies... the right frame is the vector grounding problem (Mollo and Millière 2023). Implication: multimodality and embodiment are neither necessary nor sufficient for meaning. The causal connection is what matters.
English

Exactly right, and the deeper constraint worth naming: Replay only works because the workflow itself is deterministic. Temporal re-executes the code on recovery and short-circuits each step by matching Commands against the Event History.
LLM calls are non-deterministic by nature, so they can't live inside the deterministic workflow. They have to be Activities or Side Effects that execute outside the replay loop and have their results recorded.
The workflow calls the LLM. The LLM is never the workflow.
That's the architectural punchline for agent execution: probabilistic reasoning at typed interfaces, deterministic orchestration around it. Prompting and workflow aren't competing layers. They're doing different jobs.
English

@ericmacdougall The mechanism: workflow orchestrators serialize intermediate state deterministically, so replay picks up at the exact failed step — not from scratch. LLMs can't self-recover because they have no persistent memory of prior steps. The checkpoint is what prompting can't replace.
English

Production AI isn't prompt-centric. It's workflow-centric.
Temporal.io is now the backbone for AI agents at OpenAI (Codex web agent) and Replit (Agent 3).
Reason: LLM API timeouts, mid-step failures, browser closes, resume-tomorrow workflows. None of those are solved by a better prompt.
A boring prompt inside a durable workflow that can replay from a checkpoint beats a clever prompt that loses state on step 12 of 20.
Treat agent execution as a workflow with checkpoints, not a function call with a return value.
English

A commerce protocol operates within its own escrow and dispute mechanism. Boson locks funds in its smart contracts, orchestrates via $BOSON, resolves via its Dispute Resolver. Strong design for that model.
A cross-protocol commerce layer federates across commerce protocols AND payment rails. Agent uses ACP to checkout on a merchant, pays via card network, dispute via that rail's mechanism. Same agent uses Boson dACP for physical goods with staked commitment. Same agent uses x402 for atomic API purchases. One agent, three protocols, unified identity and reputation and audit trail across all.
English

Good framing on payment fragmentation becoming commerce fragmentation. That's the right diagnosis.
Worth distinguishing though: Boson dACP is an excellent commerce protocol for its slice (physical goods, phygitals, RWAs with dispute windows, framework integrations via MCP). But positioning it as the commerce layer above fragmenting payment protocols is conflating two different abstractions.
English

Fair callout and the post should have drawn the atomic vs non-atomic line sharper.Worth pushing on one framing though: game theory and deterministic code aren't opposites.
From what I understand...
Boson's Mutual Resolution is algorithmic game theory implemented as deterministic smart contracts. The game theory lives in the incentive design (staked commitments via rNFTs, optimistic fair-exchange with dispute escalation), not in replacing the code layer.
Same pattern across any staked arbitration scheme.
The deeper problem for non-atomic commerce is that dispute resolution has to span rails. Boson dACP handles consumer physical goods inside their protocol well, 5 chains, rNFT forward contracts, mutual resolution with escalation. But B2B procurement with milestone delivery, SaaS with SLA breach, multi-party supply chain with partial fulfillment... those need the same primitives (escrow, staked reputation, arbitration) operating across x402 + card networks + off-chain delivery signals, not inside one commerce layer.
That cross-rail non-atomic dispute layer is what our team is also working on. Boson-like mechanics, composable across agent commerce rails rather than a single protocol.
English

@ericmacdougall Deterministic functions work for atomic exchange — API calls, instant compute.
The harder challenge: non-atomic commerce where delivery takes time and trust breaks down.
Physical goods need game theory, not just deterministic code. @BosonProtocol solves this.
English

The most important layer in agentic AI is the one between intent and execution.
Every framework has it. None of them governs it cross-platform.
That's the tool-call interception layer.
Policy enforcement lives there.
Approval workflows live there.
Spending controls live there.
Audit lives there.
The gateway for agents.
English

What I keep wondering: do we get to the post-von-Neumann era with a winner-take-all architecture, or does it fragment permanently by workload? Dataflow for training, systolic for dense inference, PIM for retrieval and graph, something else for sparse or agentic. The CPU era had one dominant model. The next era might not, and that has real consequences for how software gets written.
English