Adrian Scott | A.I. + Business Upscaling

29.4K posts

Adrian Scott | A.I. + Business Upscaling banner
Adrian Scott | A.I. + Business Upscaling

Adrian Scott | A.I. + Business Upscaling

@adrianscottcom

Helping businesses enter new growth incl. w/ A.I. Example: from flat sales to 600% annual growth. Social networking pioneer: Ryze; Napster; Longevity; Math PhD

Katılım Ocak 2009
4K Takip Edilen4K Takipçiler
Adrian Scott | A.I. + Business Upscaling retweetledi
Hugh Anthony
Hugh Anthony@TheHughAnthony·
Banksy should make a statue of an Englishman being stabbed by a foreigner, or is that too realistic for his liking?
English
53
348
12.3K
101.5K
🇵🇦 Mónica Araúz de Touriño
En Boquete planean que El que derribe árboles (por construcción dentro de tu terreno) debe sembrar la misma cantidad. Según veo, mucha gente está de acuerdo. Yo también!
Español
6
5
31
718
Plan B Expat
Plan B Expat@planbexpat·
You’re not wrong on the sovereignty angle. That tension is real. Panama is essentially being forced to legislate by Brussels to get off their blacklist. The irony is the reform is narrow enough that the core territorial tax system stays intact. But yes, the direction of travel is uncomfortable for anyone who values jurisdictional independence.
English
1
0
1
117
Alessandro Calvo
Alessandro Calvo@Ale_PPathways·
Panama is introducing a 15% corporate tax to all companies who can't show proof of substance and operations in the country. So much for "Panama never changes rules" 🤷‍♂️
English
17
3
155
15.9K
Angie 🇵🇦
Angie 🇵🇦@AvilaAngie_·
Lo de la plantilla de los Asesores de 6k c/u de Mayer, lejos de enojarme me da tristeza, la gran mayoría le da igual porque en redes está subiendo “lo que hace” A este ritmo el siguiente alcalde pude ser hasta un loro con buen equipo de Redes.
Español
1
4
27
610
Raúl Conte
Raúl Conte@raulconte10·
Mayer tiene 30 funcionarios ganando salario de viceministro: 6k al mes, 180k mensuales del bolsillo de todos. Viva la reactivación económica a costa del dinero público.
Español
24
148
468
9.1K
a random guy
a random guy@arandomguy808·
@Ale_PPathways For companies making over $750 million a year. It's been discussed since 2024
English
1
0
4
520
Plan B Expat
Plan B Expat@planbexpat·
A few things being missed here. This is a proposal, not law. Still needs National Assembly approval. And the 15% applies to passive income generated inside Panama for companies without real substance. A company with zero Panamanian revenue has nothing to tax. Panama is doing this to get off the EU blacklist, not to chase expats. Territorial tax on foreign income stays untouched
English
2
0
37
1.5K
André Conte
André Conte@AndreBConte·
Nunca estuve y sigo sin estar de acuerdo con utilizar plástico reciclado en cada cosa publica, como alcantarillas y sillas. Cuando lo dije hace 2 años me tildaron de hater, hoy dia se van dando cuenta del por qué. Una búsqueda en google nos hubiera ahorrado, mucho…
Español
3
5
57
5.5K
Teknium 🪽
Teknium 🪽@Teknium·
Our first dive into Multi-Agent Coordination and Cooperation is here, with Hermes Agent Kanban Orchestrate tasks across multiple agent profiles and dependencies easily and visually. Achieve more. See the docs here: hermes-agent.nousresearch.com/docs/user-guid…
Teknium 🪽 tweet media
Nous Research@NousResearch

Hermes Agent now has multi-agent via the Kanban, new in v0.12.0. Agents claim tasks from a board, work in parallel, and hand off when blocked. You watch progress and unblock from one easy view instead of juggling terminals. We asked it to plan and make this video about itself:

English
109
113
1.9K
900.4K
Adrian Scott | A.I. + Business Upscaling retweetledi
Geri Perna
Geri Perna@GeriPerna·
So, the J6ers, under Biden were rounded up like dogs, arrested, imprisoned, given unfair trials and sent to prison for years AND ANTIFA WAS NEVER TOUCHED. But now, under President Trump, Antifa continues wreaking havoc, assaulting people, destroying property, but there are no FBI raids, arrests, or accountability?!? WHY IS THIS ALLOWED!!???
Katie Daviscourt 📸@KatieDaviscourt

It’s impossible for me to do my job down at the Portland ICE facility, even with security. I was violently mobbed out by Antifa, and a suspect in black bloc threw a rock at my face. I have zero desire to discuss this further, but everyone is tired of it and wondering why a crackdown on Antifa — a terrorist organization— has not yet occurred.

English
25
182
313
3.2K
Reclaim The Net
Reclaim The Net@ReclaimTheNetHQ·
Hawley's GUARD Act just passed committee 22-0. Every American would have to upload a government ID or submit to a face scan to use an AI chatbot. Even for asking for algebra help or fixing a billing issue. The framing is child safety but the result is a national ID system for talking to a computer. reclaimthenet.org/senate-panel-b…
English
993
2.1K
3.7K
403.7K
Adrian Scott | A.I. + Business Upscaling retweetledi
Ahmad Awais
Ahmad Awais@MrAhmadAwais·
how did we make deepseek outperform opus 4.7? i've been thinking about why "open model bad at tool calling" is almost always a harness problem, not a model problem. context: spent the two days looking at billions of tokens in @CommandCodeAI (tb open source ai cli) using deepseek. I ended up writing a tool-input repair layer. the trigger was watching deepseek-flash fail on the simplest /review run, every shellCommand and readFile call bouncing back with a raw zod issues blob, the model unable to recover because the error wasn't in a form it could read. by the end deepseek v4 pro was beating opus 4.7 6/10 times on our internal evals. a few things i learned that feel general: 1/ the failure modes aren't random they're a small finite compositional set. across deepseek-flash, deepseek v4 pro, glm, qwen, the same four mistakes repeat almost exactly: - sending `null` for an optional field instead of omitting it - emitting `["a","b"]` as a json *string* instead of an actual array - wrapping a single arg in `{}` where the schema expected an array (an "empty placeholder") - passing a bare string where an array was expected (`"foo"` instead of `["foo"]`) four repairs, ~30-100 lines each, ordered carefully (json-array-parse must run before bare-string-wrap or `'["a","b"]'` becomes `['["a","b"]']`). that is the whole catalogue. when i hear "this open source model can't do tool calls" i now assume one of those four, and so far that's been right ~90% of the time. 2/ the funniest failure mode is also the most revealing. deepseek-flash, when asked to edit or write a file, sometimes emits the path as a *markdown auto-link*: filePath: "/Users/x/proj/[notes.md](http://notes. md)" our writeFile tool obediently trued creating files literally named `[notes.md](http://notes .md)` until we caught it. this is not a hallucination. it's the post-training chat distribution leaking through the tool boundary the model has been rewarded for auto-linking in conversational output, and is applying that prior in a context where it makes no sense. the fix is two regex lines that unwrap only the degenerate case where link text equals url-without-protocol real markdown like `[click](https://x .com)` passes through untouched. this is also conditioning of their own tools during RL which were different from all other tools we write and ofc can't predict. "tool confusion" is a more useful frame than "capability gap." the model knows how to format a path. it just hasn't been told clearly enough that this path is going to fopen, not into a chat bubble. so we encode that hint at the schema level `pathString()` instead of `z.string()` and the leak is plugged for every path field at once. 3/ the design choice that mattered was inverting preprocess-then-validate to validate-then-repair. my first attempt was the obvious one: a preprocessing pass that normalized inputs (strip nulls, parse stringified arrays, etc.) before zod ever saw them. it broke immediately, writeFile content that *happened* to be json-shaped got rewritten before it hit disk. silent corruption, easy to miss in a smoke test. then i made it less greedy - parse the input as-is. if it succeeds, ship it. valid inputs are never touched. - on failure, walk the validator's own issue list. for each issue path, try the four repairs in order until one applies. - parse again. on success, log `tool_input_repaired:${toolName}`. on failure, log `tool_input_invalid:${toolName}` and return a model-readable retry message. the structural insight here is: when you preprocess, you encode a prior about what's broken. when you let the validator complain first, the schema is the prior, and you only spend repair budget at the exact paths the schema actually disagreed at. the validator is doing the work of localizing the bug for you. it's the same shape as cheap-then-careful everywhere else try the fast path, fall back on evidence. (this also gives you per-tool telemetry for free. you can watch repair rates per (model, tool) and notice when a model regresses on a specific contract before users do.) 4/ shape invariants and relational invariants need different fixes. the four repairs above all handle shape problems wrong type, missing key, wrong container. but read_file had a *relational* invariant: "if you provide offset, you must also provide limit, and vice versa." deepseek kept calling `readFile({ absolutePath, limit: 30 })` and getting an `ERROR:` back. you can't fix this with input repair, because each field is independently valid the bug is in the relationship between them. so i taught the function the model's intent instead. `limit` alone → `offset = 0`. `offset` alone → `limit = 2000` (matches common read tool ops default). then surfaced the decision back to the model in the result: "Note: limit was not provided; defaulted to 2000 lines. To read more or fewer lines, retry with both offset and limit." no `Error:` prefix, so the tui doesn't paint it red. the model sees what we picked and can self-correct on the next turn if our guess was wrong. transparency over silent magic wins big. repair where you can. extend semantics where you can't. surface the choice either way. zoom out: a lot of what looks like model capability is actually contract design. a strict schema is a choice with a cost it filters out noise, but it also filters out recoverable noise from any model that hasn't memorized the exact json contract you happened to pick. the largest commercial models eat that cost invisibly and are linient on tool calling because they've seen enough of every contract during pretraining; open models pay it loudly and get dismissed for it. the harness is where you mediate between distributions. four small repairs (i'm sure more to follow as we have three more merging today), two regex lines for auto-links, one relational default, one prefix change. the model didn't change. the contract got more forgiving in exactly the places it needed to be. deepseek v4 pro now beats opus 4.7 6/10 times on our internal evals. imo "skill issue" applies to the harness more often than the model.
Ahmad Awais@MrAhmadAwais

Wow I just made DeepSeek V4 Pro beat Opus 4.7 6/10 times in our internal evals by auto repairing many of its quirks in tool calling. It’s performing super solid for such a cheap model.

English
34
88
913
147.7K
Teddy Blank
Teddy Blank@teddypowday·
Is there really nobody in SF who can pull garry tan aside for an honest chat about what's happening to him
English
110
52
3.7K
316.6K
Entérate
Entérate@Enterate_507·
@adrianscottcom @AvilaAngie_ La inundación es en general, diversas áreas, no únicamente vía Argentina. Dejen de seguir la cantaleta de acuaman, que salió del cargo hace 7 años, y exijan respuestas a las autoridades actuales.
Español
1
0
0
49
Angie 🇵🇦
Angie 🇵🇦@AvilaAngie_·
Llueve en un país tropical, acto seguido la ciudad se inunda por partes y es un caos vehicular… Bienvenidos a Panamá 🙃😢
Español
1
4
41
1.3K