
Boaz Avital ⚓
2K posts

Boaz Avital ⚓
@bx
Head of Product and Founding Engineer at @Anchorage Digital. Formerly @Twitter. Tweets are my own opinion, and that opinion is that infrastructure is a product.


Honest question: Why are token launches still being managed on spreadsheets?

Digital assets are entering a new phase. What once ran in parallel to existing financial systems is increasingly being applied to solve practical, real-world needs — often behind the scenes – from cross-border remittances to B2B money transfers. This creates new opportunities to add value in how money moves globally. Today, we introduced the Mastercard Crypto Partner Program — a global initiative that brings together more than 85 crypto-native companies, payments providers, and financial institutions. Together, we're creating a forum for meaningful dialogue and collaboration as this space continues to mature.


How I think about "security": The goal is to minimize the divergence between the user's intent, and the actual behavior of the system. "User experience" can also be defined in this way. Thus, "user experience" and "security" are thus not separate fields. However, "security" focuses on tail risk situations (where downside of divergence is large), and specifically tail risk situations that come about as a result of adversarial behavior. One thing that becomes immediately obvious from the above definition, is that "perfect security" is impossible. Not because machines are "flawed", or even because humans designing the machines are "flawed", but because "the user's intent" is fundamentally an extremely complex object that the user themselves does not have easy access to. Suppose the user's intent is "I want to send 1 ETH to Bob". But "Bob" is itself a complicated meatspace entity that cannot be easily mathematically defined. You could "represent" Bob with some public key or hash, but then the possibility that the public key or hash is not actually Bob becomes part of the threat model. The possibility that there is a contentious hard fork, and so the question of which chain represents "ETH" is subjective. In reality, the user has a well-formed picture about these topics, which gets summarized by the umbrella term "common sense", but these things are not easily mathematically defined. Once you get into more complicated user goals - take, for example, the goal of "preserving the user's privacy" - it becomes even more complicated. Many people intuitively think that encrypting messages is enough, but the reality is that the metadata pattern of who talks to whom, and the timing pattern between messages, etc, can leak a huge amount of information. What is a "trivial" privacy loss, versus a "catastrophic" loss? If you're familiar with early Yudkowskian thinking about AI safety, and how simply specifying goals robustly is one of the hardest parts of the problem, you will recognize that this is the same problem. Now, what do "good security solutions" look like? This applies for: * Ethereum wallets * Operating systems * Formal verification of smart contracts or clients or any computer programs * Hardware * ... The fundamental constraint is: anything that the user can input into the system is fundamentally far too low-complexity to fully encode their intent. I would argue that the common trait of a good solution is: the user is specifying their intention in multiple, overlapping ways, and the system only acts when these specifications are aligned with each other. Examples: * Type systems in programming: the programmer first specifies *what the program does* (the code itself), but then also specifies *what "shape" each data structure has at every step of the computation*. If the two diverge, the program fails to compile. * Formal verification: the programmer specifies what the program does (the code itself), and then also specifies mathematical properties that the program satisfies * Transaction simulations: the user specifies first what action they want to take, and then clicks "OK" or "Cancel" after seeing a simulation of the onchain consequences of that action * Post-assertions in transactions: the transaction specifies both the action and its expected effects, and both have to match for the transaction to take effect * Multisig / social recovery: the user specifies multiple keys that represent their authority * Spending limits, new-address confirmations, etc: the user specifies first what action they want to take, and then, if that action is "unusual" or "high-risk" in some sense, the user has to re-specify "yes, I know I am doing something unusual / high-risk" In all cases, the pattern is the same: there is no perfection, there is only risk reduction through redundancy. And you want the different redundant specifications to "approach the user's intent" from different "angles": eg. action, and expected consequences, expected level of significance, economic bound on downside, etc This way of thinking also hints at the right way to use LLMs. LLMs done right are themselves a simulation of intent. A generic LLM is (among other things) like a "shadow" of the concept of human common sense. A user-fine-tuned LLM is like a "shadow" of that user themselves, and can identify in a more fine-grained way what is normal vs unusual. LLMs should under no circumstances be relied on as a sole determiner of intent. But they are one "angle" from which a user's intent can be approximated. It's an angle very different from traditional, explicit, ways of encoding intent, and that difference itself maximizes the likelihood that the redundancy will prove useful. One other corollary is that "security" does NOT mean "make the user do more clicks for everything". Rather, security should mean: it should be easy (if not automated) to do low-risk things, and hard to do dangerous things. Getting this balance right is the challenge.


NEW! Browser extension for Porto users, now live. Execute DeFi transactions from self-custody in seconds and switch between wallets and networks directly from the browser, without sacrificing institutional-grade security or control.

Morpho is now available on @Anchorage! Anchorage’s institutional clients can access Morpho Vaults directly through a $50B+ AUM, federally regulated crypto bank, inside a familiar environment built for institutional custody, risk, and compliance. Morpho’s noncustodial, open infrastructure is clearing the bar for the most regulated institutions.

Tether Announces the Launch of USA₮, the Federally Regulated, Dollar-Backed Stablecoin, Made in America 🇺🇲🚀 Read more: tether.io/news/tether-an…




New: When it comes to token management, we’re making spreadsheets, manual follow-ups, and second-guessing security a thing of the past.

Welcome @HedgeyFinance to the @Anchorage team, our second acquisition this week. @goforlindsey and the whole Hedgey team have already hit the ground running and I’m very excited to work with them. Protocols can now automate token management and cap table workflows, giving both their teams and investors a seamless experience they love. Over the years, @Anchorage Digital has helped dozens of protocols sprint towards TGE. Navigating token management is one of the areas where teams spend the most time on tasks that could be automated.