Sam Elrad

291 posts

Sam Elrad

Sam Elrad

@SamElrad

Leading Enterprise AI Strategy | AI is changing everything and most of the coverage is noise. I break down what actually matters and why.

Katılım Temmuz 2025
200 Takip Edilen79 Takipçiler
Sabitlenmiş Tweet
Sam Elrad
Sam Elrad@SamElrad·
AI is the most important technology shift of our lifetime and it’s almost impossible to keep up with. I’m going to try to make that easier. Follow along if you want sharp takes on what’s happening in AI and what it means for business. No hype. No fear. Just signal.
English
1
0
3
470
Sam Elrad
Sam Elrad@SamElrad·
Really interesting article here. It is so much easier for an individual to understand a visual vs a markdown file, even if it is in plain text. I predict that the future of what LLMs consume best is also what humans consume best, and HTML files are clearly helping pave the way.
Thariq@trq212

x.com/i/article/2052…

English
0
0
0
20
Sam Elrad
Sam Elrad@SamElrad·
When these packages are build, I am curious what kind of impact this will have on the emerging vertical AI companies in the space. Will be interesting to see how a company like Rogo competes or evolves with these new offerings in their market.
Josh Kale@JoshKale

Anthropic just automated the first-year analyst job at every bank on Wall Street. They released these 10 AI agents for finance: → Pitch builder → Meeting preparer → Earnings reviewer → Model builder → Market researcher → Valuation reviewer → GL reconciler → Month-end closer → Statement auditor → KYC screener The analyst pyramid just got a lot flatter.

English
0
0
0
45
Sam Elrad
Sam Elrad@SamElrad·
This is what will unlock exponential opportunity for Enterprises. If you can solve the context and memory problem, you will become a truly AI native company.
Y Combinator@ycombinator

Company Brain @t_blom Every company has critical know-how scattered across people's heads, old Slack threads, support tickets, and databases, and AI agents can't operate like that. We think every company in the world is going to need a new primitive: a living map of how the company works that turns its own artifacts into an executable skills file for AI.

English
0
0
1
32
Sam Elrad
Sam Elrad@SamElrad·
@claudeai Is this supposed to be a more flexible version of Databricks dashboards?
English
0
0
1
3K
Claude
Claude@claudeai·
In Cowork, Claude can now build live artifacts: dashboards and trackers connected to your apps and files. Open one any time and it refreshes with current data.
Claude tweet media
English
666
1.6K
19.4K
6.5M
Sam Elrad
Sam Elrad@SamElrad·
@claudeai How many minutes until OpenAI releases their new model? 😂
English
0
0
0
571
Claude
Claude@claudeai·
Introducing Claude Opus 4.7, our most capable Opus model yet. It handles long-running tasks with more rigor, follows instructions more precisely, and verifies its own outputs before reporting back. You can hand off your hardest work with less supervision.
Claude tweet media
English
4.8K
10.2K
81.1K
13.8M
Sam Elrad retweetledi
Aaron Levie
Aaron Levie@levie·
Another week on the road meeting with a couple dozen IT and AI leaders from large enterprises across banking, media, retail, healthcare, consulting, tech, and sports, to discuss agents in the enterprise. Some quick takeaways: * Clear that we’re moving from chat era of AI to agents that use tools, process data, and start to execute real work in the enterprise. Complementing this, enterprises are often evolving from “let a thousand flowers bloom” approach to adoption to targeted automation efforts applied to specific areas of work and workflow. * Change management still will remain one of the biggest topics for enterprises. Most workflows aren’t setup to just drop agents directly in, and enterprises will need a ton of help to drive these efforts (both internally and from partners). One company has a head of AI in every business unit that roles up to a central team, just to keep all the functions coordinated. * Tokenmaxxing! Most companies operate with very strict OpEx budgets get locked in for the year ahead, so they’re going through very real trade-off discussions right now on how to budget for tokens. One company recently had an idea for a “shark tank” style way of pitching for compute budget. Others are trying to figure out how to ration compute to the best use-cases internally through some hierarchy of needs (my words not theirs). * Fixing fragmented and legacy systems remain a huge priority right now. Most enterprises are dealing with decades of either on-prem systems or systems they moved to the cloud but that still haven’t been modernized in any meaningful way. This means agents can’t easily tap into these data sources in a unified way yet, so companies are focused on how they modernize these. * Most companies are *not* talking about replacing jobs due to agents. The major use-cases for agents are things that the company wasn’t able to do before or couldn’t prioritize. Software upgrades, automating back office processes that were constraining other workflows, processing large amounts of documents to get new business or client insights, and so on. More emphasis on ways to make money vs. cut costs. * Headless software dominated my conversations. Enterprises need to be able to ensure all of their software works across any set of agents they choose. They will kick out vendors that don’t make this technically or economically easy. * Clear sense that it can be hard to standardize on anything right now given how fast things are moving. Blessing and a curse of the innovation curve right now - no one wants to get stuck in a paradigm that locks them into the wrong architecture. One other result of this is that companies realize they’re in a multi-agent world, which means that interoperability becomes paramount across systems. * Unanimous sense that everyone is working more than ever before. AI is not causing anyone to do less work right now, and similar to Silicon Valley people feel their teams are the busiest they’ve ever been. One final meta observation not called out explicitly. It seems that despite Silicon Valley’s sense that AI has made hard things easy, the most powerful ways to use agents is more “technical” than prior eras of software. Skills, MCP, CLIs, etc. may be simple concepts for tech, but in the real world these are all esoteric concepts that will require technical people to help bring to life in the enterprise. This both means diffusion will take real work and time, but also everyone’s estimation of engineering jobs is totally off. Engineers may not be “writing” software, but they will certainly be the ones to setup and operate the systems that actually automate most work in the enterprise.
English
255
646
5.3K
1.8M
Sam Elrad
Sam Elrad@SamElrad·
Very interesting and insightful approach to what we might see in the future with LLMs
Andrej Karpathy@karpathy

LLM Knowledge Bases Something I'm finding very useful recently: using LLMs to build personal knowledge bases for various topics of research interest. In this way, a large fraction of my recent token throughput is going less into manipulating code, and more into manipulating knowledge (stored as markdown and images). The latest LLMs are quite good at it. So: Data ingest: I index source documents (articles, papers, repos, datasets, images, etc.) into a raw/ directory, then I use an LLM to incrementally "compile" a wiki, which is just a collection of .md files in a directory structure. The wiki includes summaries of all the data in raw/, backlinks, and then it categorizes data into concepts, writes articles for them, and links them all. To convert web articles into .md files I like to use the Obsidian Web Clipper extension, and then I also use a hotkey to download all the related images to local so that my LLM can easily reference them. IDE: I use Obsidian as the IDE "frontend" where I can view the raw data, the the compiled wiki, and the derived visualizations. Important to note that the LLM writes and maintains all of the data of the wiki, I rarely touch it directly. I've played with a few Obsidian plugins to render and view data in other ways (e.g. Marp for slides). Q&A: Where things get interesting is that once your wiki is big enough (e.g. mine on some recent research is ~100 articles and ~400K words), you can ask your LLM agent all kinds of complex questions against the wiki, and it will go off, research the answers, etc. I thought I had to reach for fancy RAG, but the LLM has been pretty good about auto-maintaining index files and brief summaries of all the documents and it reads all the important related data fairly easily at this ~small scale. Output: Instead of getting answers in text/terminal, I like to have it render markdown files for me, or slide shows (Marp format), or matplotlib images, all of which I then view again in Obsidian. You can imagine many other visual output formats depending on the query. Often, I end up "filing" the outputs back into the wiki to enhance it for further queries. So my own explorations and queries always "add up" in the knowledge base. Linting: I've run some LLM "health checks" over the wiki to e.g. find inconsistent data, impute missing data (with web searchers), find interesting connections for new article candidates, etc., to incrementally clean up the wiki and enhance its overall data integrity. The LLMs are quite good at suggesting further questions to ask and look into. Extra tools: I find myself developing additional tools to process the data, e.g. I vibe coded a small and naive search engine over the wiki, which I both use directly (in a web ui), but more often I want to hand it off to an LLM via CLI as a tool for larger queries. Further explorations: As the repo grows, the natural desire is to also think about synthetic data generation + finetuning to have your LLM "know" the data in its weights instead of just context windows. TLDR: raw data from a given number of sources is collected, then compiled by an LLM into a .md wiki, then operated on by various CLIs by the LLM to do Q&A and to incrementally enhance the wiki, and all of it viewable in Obsidian. You rarely ever write or edit the wiki manually, it's the domain of the LLM. I think there is room here for an incredible new product instead of a hacky collection of scripts.

English
0
0
0
25
Sam Elrad
Sam Elrad@SamElrad·
@claudeai At what point are Claude Chat and Claude Cowork going to merge together? I get the UX differences now, but every time I see a feature update… it makes me feel like we are just building toward one unified app.
English
1
0
2
1.6K