ibby

2.5K posts

ibby

ibby

@StatueofIBBertY

CEO @coterahq - AI agents for the enterprise.

nyc Katılım Ocak 2012
673 Takip Edilen899 Takipçiler
ibby
ibby@StatueofIBBertY·
@ilyasu Create a sign up flow and use the info to DDoS them back
English
0
0
4
179
Ilya Sukhar
Ilya Sukhar@ilyasu·
My personal website on Vercel is getting DDoS'd for some reason. Besides turning on "bot protection" and "attack mode", what else should I do?
English
4
0
6
3.3K
ibby
ibby@StatueofIBBertY·
We took a customer who was scraping reddit and LinkedIn for ~$0.90 per page, and got the cost down to half a cent per page. I routinely jump on calls with people who complain "I'm using Claude to do web scraping for 'X', but it's really expensive. Can your product help?" I then ask what they're doing, and they're always really proud to show me (which I appreciate!): 1. They give Claude access to their laptop 2. It then uses their browser to scrape sites that are blocked (e.g. reddit) 3. Each page is like a million tokens of raw html 4. It does each page 1 by 1 I then show them what happens when you use a dedicated tool to scrape a webpage - the cost is IMMEDIATELY down 98%. Then you add in the fact that an open source model is just as good at scraping - that brings the cost down another 90%. Lastly... it's much faster if you can parallelize it. And no, not doing batches of 5 at a time. Doing 1000 at a time. I don't think software engineering jobs are going anywhere - I've seen proof multiple times in the last few weeks.
English
2
0
9
636
ibby
ibby@StatueofIBBertY·
The think that confuses me about web scraping also - is I *assume* it's not random. So why not an intermediary layer that *reduces* the size of the corpus you're trying to scrape? E.g. reddit. A post with a web scraper attached to my laptop's chrome is 1.8M tokens. With the reddit api it's 35k tokens. If using Opus, the raw scraping is almost $10 of tokens PER PAGE. The API brings that down to 17.5 cents. On Deepseek that's half a penny.
ibby tweet media
English
0
0
0
40
Doug Colkitt
Doug Colkitt@0xdoug·
@zephyr_z9 Why would you run web scraping through at Opus at $5/mtok instead of DeepSeek Flash at $0.15/mtok. Any time I’ve done web scraping, Opus is way *way* oversized
English
3
0
19
2.6K
ibby
ibby@StatueofIBBertY·
@matt_slotnick Agree with this, I don't think people are actually using AI to anywhere near how they could. though not sure oai/ant are the ones that win that market
English
1
0
2
160
Matt Slotnick
Matt Slotnick@matt_slotnick·
i think many are underestimating the size of the AI market by orders of magnitude because they still see AI as "person asks AI for help" rather than "AI is always working in the background and asks people when it needs help”
English
15
12
119
6.1K
ibby
ibby@StatueofIBBertY·
The responses to this tweet are hilarious. Everyone is like "THE FRAMEWORK IS WRONG" while not offering any retort outside of "people that aren't devs are claude coding!!!" No they aren't. Not in any sort of significant way, revenue wise.
Doug Colkitt@0xdoug

I’m really struggling to see how the back of the envelope math on this works out… There are generously 4 million characterized “software workers” in America. That’s pretty broad and includes a lot of people who aren’t really classical engineers don’t produce that much code. That comes out to nearly $1k per month of average Claude spend across every dev in America. Yes, there’s some international usage, but it can’t be that much. Yes there is some non software Cowork usage, but that doesn’t use that many tokens. Yes, some non engineers are using Claude to vibe code, but I really doubt many are spending hundreds per month on. Even if we assume 50% of all software workers are using Claude, that comes out to $2k spend per month per Claude user. Thats 10X more than the highest tier Max subscription. So almost all of Anthropics revenue has to be API billing So the only explanation is that something like 20%+ of software engineers are not only Claude users but on API billing and regularly spending thousands per month. At $5/m Opus tokens that means the average API user has to be going through something like 25 million tokens per day. *OR* the other possibility is API revenue is heavily power law dominated. Maybe there’s just something like 100k super users who are making up the majority of the revenue. For that to work the typical super user would have to be spending on the order of $50k/month and guzzling nearly 1 billion tokens per day.

English
0
0
2
201
Doug Colkitt
Doug Colkitt@0xdoug·
I’m really struggling to see how the back of the envelope math on this works out… There are generously 4 million characterized “software workers” in America. That’s pretty broad and includes a lot of people who aren’t really classical engineers don’t produce that much code. That comes out to nearly $1k per month of average Claude spend across every dev in America. Yes, there’s some international usage, but it can’t be that much. Yes there is some non software Cowork usage, but that doesn’t use that many tokens. Yes, some non engineers are using Claude to vibe code, but I really doubt many are spending hundreds per month on. Even if we assume 50% of all software workers are using Claude, that comes out to $2k spend per month per Claude user. Thats 10X more than the highest tier Max subscription. So almost all of Anthropics revenue has to be API billing So the only explanation is that something like 20%+ of software engineers are not only Claude users but on API billing and regularly spending thousands per month. At $5/m Opus tokens that means the average API user has to be going through something like 25 million tokens per day. *OR* the other possibility is API revenue is heavily power law dominated. Maybe there’s just something like 100k super users who are making up the majority of the revenue. For that to work the typical super user would have to be spending on the order of $50k/month and guzzling nearly 1 billion tokens per day.
Tannor Manson@Futurenvesting

Anthropic is now showing off $44 BILLION in annual recurring revenue. This is up $14 billion (+46.6%) since last month! BULLISH for AI Infrastructure $NVDA $AMD

English
291
20
492
490.9K
ibby
ibby@StatueofIBBertY·
@sporadica I'm a damn moron, I could have clicked on the link. "According to GWI’s tracking of 250,000 adults across 50+ countries, daily social media use across the developed world fell by about 10% from a peak of ~2.5 hours in 2022 to ~2.3 hours." I guess that was a bit past covid
English
0
0
0
17
ibby
ibby@StatueofIBBertY·
@sporadica The data shows the drop off from 2022/2023? I wish the x axis was wider because I want to see the variability, but my first guess is covid?
English
1
0
0
55
ibby
ibby@StatueofIBBertY·
@0xdoug There's no possible way this lasts I think If it's dominated by CC -- there's not a chance that a majority renew. We've replaced CC w/ Cursor on OSS models. If it's API usage for workflows - open models are good enough This is a race to the bottom. Anthropic dies in 5 years.
English
0
0
1
146
ibby
ibby@StatueofIBBertY·
@0xdoug @Big_Fin_Art I'm also 99% sure these specialized workflows use open source in the long run...so for ant's sake I hope it isn't this
English
0
0
1
33
Doug Colkitt
Doug Colkitt@0xdoug·
@Big_Fin_Art How many of those agents are using Opus though? My experience is most of these “specialized agent workflows” like summarize a meeting definitely do not need Opus scale models
English
5
0
1
2.7K
ibby
ibby@StatueofIBBertY·
@andrewchen Fully agree on this. The best pms I work with now start both the design process and engineering mockups themselves. I think there's probably a market for a product that takes a PM's vibecoded mockup and creates a test suite for engineers to take a real product across the line
English
0
0
1
229
andrew chen
andrew chen@andrewchen·
bullish on the PM role quietly becoming the most important role in tech again when anyone can build, the person who decides WHAT to build becomes the bottleneck
English
279
175
2.2K
224.4K
ibby
ibby@StatueofIBBertY·
@RegularAvgGuy Yeah it's surprising how often it uses bad tools
English
0
0
0
5
Just a Regular Average Guy
Just a Regular Average Guy@RegularAvgGuy·
@StatueofIBBertY I agree. It's why I watch the terminal while Claude Code is running. I can see it going in the wrong direction, and I have to redirect it. I even watch the thinking output in the ChatGPT window for simple cognitive tasks because I see it taking shortcuts that I think are wrong
English
1
0
1
16
ibby
ibby@StatueofIBBertY·
I think this tweet is true, but unsure if I *want* to believe it. On the one hand, one of my customers said this to me this week: "AI is making my smart people smarter and my dumb people dumber" The thing that's missing from AI is systems thinking - Cotera has spent the last year helping businesses automate everything from basic support ticket QA to helping companies develop digital twins of their customers. The thing that I routinely see AI fail at is "creating the system". Take a basic use case where want Claude to tag 5000 support tickets with whether the customer was "happy" or "sad" or "neutral" at the end. The data is too big for context, so chat GPT's answer to this is to write a python script that looks for keywords. It outputs a tagged csv, but when you look closely, 90% of the data is neutral... but it's not actually correct. The real answer is to use a small LLM in parallel to tag each one - running that minimizes the "neutral" tag to around 30% of the data. The thing that worries me about people managing agents "smarter than themselves" is that AI is a shit systems thinker, and it just creates an output that looks right from a 30k foot level, but not when you dive in. At least when you had to do it manually you had to think through the process.
Haseeb >|<@hosseeb

Most people have never managed anyone. Those who have, most have never managed anyone smarter than themselves. This is about to change for everyone. You will be surrounded by agents that are smarter than you, working for you 24/7. In a world of cognitive abundance, your understanding becomes the bottleneck. You can genuinely delegate a lot of cognitive work now. This is not sci-fi. It's already permanently changed the texture of knowledge work. But the less you understand what your agents are doing and how they're doing it, the less you will be able to get out of them. This is why it's still important to understand things like software, coding, economics, math, statistics, game theory. Not because you need to DO them (you don't), but you need to understand what's easy and what's impossible. Try to be as smart as your agents. You will inevitably fail, but you don't need to get all the way there. You just need to become smart enough to manage things smarter than you.

English
1
0
4
124
ibby
ibby@StatueofIBBertY·
Lmfao there's an ex Latham and Watkins employee that realized that Harvey was a chatgpt wrapper on top of some legal apis so he created an open source version and called it MikeOSS
English
1
0
0
134
ibby
ibby@StatueofIBBertY·
I ran the "should I walk or drive to the car wash 100 feet away" question to 14 of the latest AI models. Look, shit on Gemini all you want, but it's the only frontier model that got it right. Yes, Claude 4.7 still got it wrong. All through Cotera, btw
ibby tweet media
English
1
0
3
142
ibby
ibby@StatueofIBBertY·
@bindureddy We also did this - ended up showing customers a collective million dollars a year in cost savings
English
0
0
2
287
Bindu Reddy
Bindu Reddy@bindureddy·
Kimi 2.6 and GLM 5.1 are insanely close to closed AI in performance The only issue is speed. I am hopeful the open source models will soon solve them We are moving to open source on all batch jobs. Closed source APIs are extremely expensive with AI labs maximizing API margins 🤷‍♀️
English
50
20
346
18.8K