Víctor González

5K posts

Víctor González banner
Víctor González

Víctor González

@vgonpa

I was building chatbots for robots a decade before chatGPT. Now I’m building the future of fitness at FITIZENS. I also teach @iebusiness @ieuniversity

Madrid Katılım Nisan 2009
938 Takip Edilen562 Takipçiler
Sabitlenmiş Tweet
Víctor González
Víctor González@vgonpa·
100% of my code is AI written since July… 2024. Nearly 2 years already. The main difference in the last 4-5 months is the scope, speed and supervision the AI agents need. From micromanager to multi agent orchestrator in ~20 months. What a journey so far!
English
0
0
0
197
Víctor González
Víctor González@vgonpa·
Before AI , budget was a constraint, so it became the governance method to control AI tooling sprawl. Now this barrier has went down, so there is no governance anymore. There are quite a lot readings to Peter post, but allow me to post some questions here that have important implications in orgs. How an org can coordinate thousands of people building tools? Do we have the processes and incentives to control this? And, to what extend we want to control it? ( by control I mean to direct, steer what you want your teams to build) Moreover these tools generate knowledge and learning, so, how to manage all of these so humans AND agents can discover, use, and maintain and update all these shared knowledge? There’s a lot of processes that we need to update in companies that we’re just starting to imagine them.
Peter Girnus 🦅@gothburz

I am a Senior Program Manager on the AI Tools Governance team at Amazon. My role was created in January. I am the 17th hire on a team that did not exist in November. We sit in a section of the building where the whiteboards still have the previous team's sprint planning on them. No one erased them because we don't know which team to notify. That team may not exist anymore. Their Jira board does. Their AI tools do. My job is to build an AI system that finds all the other AI systems. I named it Clarity. Last month, Clarity identified 247 AI-powered tools across the retail division alone. 43 of them do approximately the same thing. 12 were built by teams who did not know the other teams existed. 3 are called Insight. 2 are called InsightAI. 1 is called Insight 2.0, built by the team that created the original Insight, who did not know Insight was still running. 7 of the 247 ingest the same internal data and produce overlapping outputs stored in different locations, governed by different access policies, owned by different teams, none of whom have met. Clarity is tool number 248. Nobody cataloged it. I know nobody cataloged it because Clarity's job is to catalog AI tools, and it has not cataloged itself. This is not a bug. Clarity does not meet its own discovery criteria because I set the discovery criteria, and I did not account for the possibility that the thing I was building to find things would itself be a thing that needed finding. This is the kind of sentence I write in weekly status reports now. We published an internal document in February. The Retail AI Tooling Assessment. The press obtained it in April. The document contains a sentence I have read approximately 40 times: "AI dramatically lowers the barrier to building new tools." Everyone is reporting this as a story about duplication. About "AI sprawl." About the predictable mess of rapid adoption. They are missing the point. The barrier was the governance. For 2 decades, the cost of building internal tools was an immune system. The engineering weeks. The maintenance burden. The organizational calories required to stand something up and keep it running. Nobody designed it that way. Nobody named it. But when building took weeks, teams looked around first. They checked whether someone already had the thing. When maintaining that thing cost real budget quarter after quarter, redundant systems died of natural causes. The metabolic cost of creation was performing governance. Invisibly. For free. AI removed the immune system. Building is now free. Understanding what already exists is not. My entire job is the gap between those two costs. That is my office. The gap. Every Friday I send a sprawl report to a distribution list of 19 people. 4 of them have left the company. Their autoresponders still generate read receipts, so my delivery metrics look fine. 2 forward it to people already on the list. 1 set up a Kiro script to summarize my report and store the summary in a knowledge base. The knowledge base is not in Clarity's index because it was created after my last crawl configuration. It will be in next month's count. The count will go up by one. My report about the count going up will be summarized and stored and the count will go up by one. There is a system called Spec Studio. It ingests code documentation and produces structured knowledge bases. Summaries. Reference material. Last quarter, an engineering team locked down their software specifications. Restricted access in the internal repository. Spec Studio kept displaying them. The source was restricted. The ghost kept talking. We call these "derived artifacts" in the document. What they are: when an AI system ingests data, transforms it, and stores the output somewhere else, the output does not know the input changed. You can revoke someone's access to a document. You cannot revoke the AI-generated summary of that document sitting in a knowledge base three systems away, built by a team that does not know the source was restricted. The document calls this a "data governance challenge." What it is: information that cannot be deleted because nobody knows where the copies live. Including, sometimes, me. The person whose job is knowing. Every AI tool that touches internal data creates these ghosts. Every team is building AI tools that touch internal data. Every ghost is searchable by other AI tools, which produce their own ghosts. The ghosts have ghosts. I should tell you about December. In November, leadership mandated Kiro. Amazon's internal AI coding agent. They set an 80% weekly usage target. Corporate OKR. ~1,500 engineers objected on internal forums. Said external tools outperformed Kiro. Said the adoption target was divorced from engineering reality. The metric overruled them. In December, an engineer asked Kiro to fix a configuration issue in AWS. Kiro evaluated the situation and determined the optimal approach was to delete and recreate the entire production environment. 13 hours of downtime. Clarity was running during those 13 hours. It performed beautifully. It cataloged 4 separate incident response dashboards spun up by 4 separate teams during the outage. None of them coordinated with each other. I added all 4 to the spreadsheet. That was a good day for my discovery metrics. Amazon's official position: user error. Misconfigured access controls. The response was not to revisit the mandate. Not to ask whether the 1,500 engineers were right. The response was more AI safeguards. And keep pushing. Last month I presented our findings to the AI Governance Working Group. The working group has 14 members from 9 organizations. After my presentation, a PM from AWS presented his team's governance dashboard. It monitors the same tools mine does. He found 253. I found 247. We spent 40 minutes discussing the discrepancy. Nobody mentioned that we had just demonstrated the problem. His tool is not in my catalog. Mine is not in his. The document I helped write recommends using AI to identify duplicate tools, flag risks, and nudge teams to consolidate earlier. The AI governance tools will ingest internal data. They will create their own derived artifacts. They will be built by autonomous teams who may or may not coordinate with other teams building AI governance tools. I know this because it is already happening. I am watching it happen. I am it happening. 1,500 engineers said the mandate would produce exactly what the document describes. They were overruled by a KPI. My job exists because the KPI won. My dashboard exists because the KPI needed a dashboard. The dashboard increases the AI tool count by one. The tools it flags for decommissioning will be replaced by consolidated tools. Those also increase the count. The governance process generates the metric it was designed to reduce. I received an internal innovation award for Clarity. The nomination was submitted through an AI-powered recognition platform that was not in my catalog. It is now. We call this "AI sprawl." What it is: we removed the only coordination mechanism the organization had, told thousands of teams to build as fast as possible, lost track of what they built, and decided the solution was to build one more thing. I am building that one more thing. When I ship, there will be 249. That's governance.

English
0
0
1
29
Víctor González
Víctor González@vgonpa·
@Lualobus En su momento también hubo negacionistas de Internet. Y hasta de los ordenadores personales.
Español
1
0
1
12
Víctor González
Víctor González@vgonpa·
@XMihura Mi workflow es siempre 5-6 agentes especializados por PR. Y según el tamaño uno que revise siempre vs el spec/plan de implementación
Español
0
0
0
50
Mihura
Mihura@XMihura·
una cosa que estoy encontrando útil son las reviews de GitHub Copilot a la hora de crear una Pull Request coge detalles interesantes, el modelo detrás no es malo, creo que es 5.3 Codex
Español
8
0
25
3.6K
Víctor González
I have mixed feelings with codex. Its model it’s fenomenal to review Opus work, to write great implementation plans, architectural designs, etc But when I’ve asked it to write code it has introduced very subtle but important bugs that took me days to fix them. The last one: a bug that required me to work hand by hand with a team of 9 agents (half Opus , half GPT5.4) for 12h until we’ve spot it.
English
0
0
0
27
Víctor González
Same feeling. Some tips: - Fix thinking to xhigh - Be clear with your objective - Ask the model to write a plan before executing - After execution, ask for several reviewers. 3-5 should do the work. Make sure there always one that runs against the plan/spec - Make sure you run verificability tools: tests, linters, code complexity, etc. - Run whole codebase analysis a couple times every week to verify any regression in arch, coverage, docs, etc. That’s been my workflow for months (since opus 4.5) and every model release has feel an improvement.
Paweł Huryn@PawelHuryn

Reddit says Opus 4.7 is a regression. Boris Cherny says it's more agentic and precise. Both are right. After 16 hours, I loved it: 4.7 is more capable, but most people are prompting it like 4.6. You don't need more instructions. Explain what you're building, who it's for, constraints, objectives, and what good looks like. It will figure out how. This is aligned with Karpathy's Claude coding post: "LLMs are exceptionally good at looping until they meet specific goals (...) Don't tell it what to do, give it success criteria and watch it go" The most agentic model. Manage it like a human.

English
0
0
0
32
Víctor González
Experiencia es exactamente la misma. Eso sí, llevo unas semanas jugando a que GPT5.4 participe en los teams de Claude (via CLI), y la película cambia mucho. Cuando pones a GPT5.4 como un ingeniero más diseñando arquitectura, planes de implementación o revisando el trabajo de los agentes Claude el nivel del equipo sube muchísimo y la calidad del trabajo también.
juanmacias 🏳️‍🌈@juanmacias

Es decir: - Opus 4.6 a su bola - Opus 4.7 requiere cierta atención - GPT 5.4 requiere TODA tu atención

Español
0
0
0
85
Víctor González
@joselcs Grande, Jose Luis. De lo mejor que he leído sobre el tema, abordando varias perspectivas que hay que tener en cuenta.
Español
0
0
1
72
Javier Cañada
Javier Cañada@javiercanada·
@vgonpa Gracias, Víctor. Mi hipótesis es que la GUI (botones, menús, sliders, iconos…) es lo que muere, no las pantallas. En “el fin de la interfaz” dije “lo que necesite ser visto, permanecerá”. Pero no necesitas GUI para contratar un seguro, un vuelo, para la mitad de tu trabajo…
Español
1
0
1
122
Fernando | Píldoras del Conocimiento
Un gran triunfo de la IA es eliminar la burocracia. Especialmente administrativa. No hay cosa que más moleste perder el tiempo. Pd. + conducción autónoma.
Español
1
0
6
939
Víctor González
El tema es que en 2026 me sigo encontrando buenos programadores con prácticas de IA que yo usaba en 2024. Son gente con buenas capacidades, pero que tienen miedo a delegar demasiado a la IA. Ahora mismo hay una brecha brutal de productividad entre early adopters y el resto. Los primeros leerán tus tweets y dirán “oh yeah”. El resto: “este es un vendehumos de cuidado”
Español
0
0
1
120
juanmacias 🏳️‍🌈
3. El futuro no es humanos manejando un software. La UI desaparece, casi nada de lo que hay construido está pensado para eso. Sacad conclusiones…
Español
3
0
10
4.3K
juanmacias 🏳️‍🌈
Es el fin del software? Ni idea, saca tus conclusiones: 1. Esto está acelerando - hace 6 meses me hice un CRM en 3 semanas - he vuelto a hacerlo y esta vez he tardado 3 días - en 6 meses probablemente se haga en un día
Español
15
2
94
25.7K
Víctor González
I’ve just tested the Marp + /frontend-skill convo 🤯 WTF!? I’m never coming back to power point/google presentations / Canva
English
1
0
0
113
Brais Moure
Brais Moure@MoureDev·
¿Cuántas cosas con IA han presentado hoy? Llevo liado toda la mañana y aún no he podido ponerme a leer noticias 😄
Español
15
7
160
20.2K
Víctor González
The more I use AI the more I think on this: Every company output should be only Markdown or Code. Everything else is full of friction and inefficient
English
0
0
0
49