Rob Corbidge

428 posts

Rob Corbidge banner
Rob Corbidge

Rob Corbidge

@CorbidgeRob

Content Intelligence at https://t.co/qMq0CmQ9ZR.

Beigetreten Şubat 2022
791 Folgt47 Follower
Rob Corbidge retweetet
Ed Newton-Rex
Ed Newton-Rex@ednewtonrex·
If you stacked up the 7 million books Anthropic pirated, the pile would be 165x taller than the Burj Khalifa. Scale diagram:
Ed Newton-Rex tweet media
English
15
68
292
12.2K
Rob Corbidge retweetet
Nathan Gotch
Nathan Gotch@nathangotch·
I'm still waiting for the day that someone can show me a brand that's doing well in AI search that is NOT doing well in traditional search engines. There is no successful GEO or AEO without effective SEO.
English
36
14
150
8.7K
Rob Corbidge retweetet
jerb
jerb@searobbing·
Worst case scenario we can just make a new strait. Shouldn't take long once we unleash the productivity boosting powers of AI. Maybe a day or two at most.
English
80
1.2K
17.3K
281.6K
Rob Corbidge retweetet
summer 🖕
summer 🖕@salesforcechild·
i am so stressed out from running my pet lizards instagram account i’m about to have a nervous breakdown
English
489
11.1K
69.8K
0
Rob Corbidge retweetet
Caitlin Cook
Caitlin Cook@DeadCaitBounce·
Believing Claude finds your questions insightful is like believing the stripper actually likes you
English
452
881
11.4K
510.8K
Rob Corbidge retweetet
Ed Newton-Rex
Ed Newton-Rex@ednewtonrex·
The UK government is reportedly considering introducing a 'commercial research exception' (CRE) to copyright law that would allow AI training. Here's why it would be totally unworkable. First, what is it? Basically, a CRE would let AI developers train models on copyrighted work without permission or payment. Then, before they brought the model to market, they would have to license the training data they had used. But it doesn't work, because of the 'single-dissenter problem'. Imagine you spend millions of pounds training a model on millions of scraped articles. You like the model, and want to release it. You go and try to license the training data you used. If even *one rights holder* says no, you can't release your model. A single dissenter means you have wasted millions training a model you can't release. It is obvious, in fact, that the only way to make a CRE work is to pair it with *compulsory licensing*: that is, with the government *forcing* rights holders to license their work for AI training. But that will never fly. It is totally unfair - it means creatives being forced to license their work to AI companies that are trying to replace them! Not to mention that it would likely contravene international law. If the government so much as mentioned compulsory licensing for AI training in passing, there would be uproar. So CREs don't work. They are a non-solution. In fact, the actual solution is simple: it is for AI developers to get a licence *before* they train their model. Everyone wins: the developer doesn't waste money training unreleasable models, and the rights holder gets paid their due. (Luckily, this is already the law in the UK.) Don't let big tech tell you a commercial research exception is fair. It is either unworkable or a Trojan Horse for compulsory licensing. Full paper, including other issues with CREs, here: ed.newtonrex.com/commercial-res…
Ed Newton-Rex tweet media
English
8
55
189
9.6K
Rob Corbidge retweetet
Pedro Dias
Pedro Dias@pedrodias·
Because we didn’t have enough confirmation bias: “Unlike hallucinations, which introduce false-hoods, sycophancy is a bias in the selection of the data people see. When AI systems are trained to be helpful, they may inadvertently prioritize data that validates the user’s narrative over data that gets them closer to the truth.” arxiv.org/abs/2602.14270
Pedro Dias tweet media
English
2
1
3
175
Rob Corbidge retweetet
Ed Newton-Rex
Ed Newton-Rex@ednewtonrex·
… then blocks me. I wonder what that model could be trained on.
Ed Newton-Rex tweet media
English
6
11
173
1.8K
Rob Corbidge retweetet
Ed Newton-Rex
Ed Newton-Rex@ednewtonrex·
Google just added Lyria, its music generator, to Gemini. What was it trained on? They haven’t said. They say they have been ‘mindful of copyright’. What does this mean?! It sounds worryingly like intentional misdirection. You should assume it’s trained on copyrighted work without a licence unless they come out and say it’s not.
Ed Newton-Rex tweet mediaEd Newton-Rex tweet media
English
21
109
465
19.2K
Rob Corbidge retweetet
courtsdesk
courtsdesk@courtsdesk·
BREAKING: UK Gov Legal Dept has written to us on behalf of HMCTS/MOJ asking us to pause deletion of our archive and engage in dialogue about a new licence. We are grateful for the opportunity to resolve it and have replied to confirm retention. More soon #OpenJustice #Courtsdesk
English
283
1.4K
5.8K
450.9K
Rob Corbidge retweetet
Jason Kint
Jason Kint@jason_kint·
that moment when you realize the public really doesn't like you. this is the state AG and school districts (MDL) case going to trial in a new months not to be confused with the trials ongoing in LA and New Mexico.
Jason Kint tweet media
English
1
7
23
2K
Rob Corbidge retweetet
Ed Newton-Rex
Ed Newton-Rex@ednewtonrex·
@ajs @FortuneMagazine And all it took was training on millions of people’s copyrighted images without permission
English
1
4
119
1.1K
Rob Corbidge retweetet
Peter Girnus 🦅
Peter Girnus 🦅@gothburz·
I am the Chief Revenue Officer of an AI company that will not exist in 18 months. On Sunday, I bought a Super Bowl ad. $8 million for 30 seconds. We lost $11.5 billion last quarter. The ad cost rounding error. The board approved it in nine minutes. No one asked about profitability because no one has asked about profitability since 2023. The ad shows a child asking our AI a question about the stars. The AI answers beautifully. What the ad does not show is that the same AI told a different child that the sun orbits the earth. We generated 40,000 test renders to find one where it didn't hallucinate. The approved cut was take 39,847. We are not the only ones. Super Bowl LX has more AI ads than beer ads. More AI ads than car ads. Anthropic, OpenAI, Google, Svedka's dancing robot -- all of us, clustered together in the commercial breaks like startups at a WeWork in 2019. This has happened before. Super Bowl XXXIV. January 30, 2000. Fourteen dot-com companies bought ads. Pets.com. OurBeginning.com. Epidemic.com. Companies whose names are now punchlines and whose domains are now parked pages. The NASDAQ peaked 44 days later. That Super Bowl featured an E-Trade ad with a dancing monkey and the tagline: "We just wasted $2 million. What are you doing with your money?" It was the funniest ad of the night. It was also the most honest. Here is what I am doing with your money. $650 billion. That is what Amazon, Google, Meta, and Microsoft will spend on AI infrastructure in 2026. Combined. It is more than the GDP of Sweden. It is 75% more than last year. Amazon alone is spending $200 billion. Alphabet: $185 billion. On what? Data centers. Chips. Cooling systems for the chips. More chips. Custom chips for the chips. Land to put the chips on. Concrete to hold the chips. Generators to power the chips when the grid fails from powering the other chips. I was in the earnings call. An analyst asked when this investment returns a profit. Our CEO said the word "enormous" three times in one sentence and then said "next question." The market dropped the stocks 7.5% in a week. The Nasdaq fell 1.9%. Bitcoin dropped below $70,000. Software companies -- the ones who were supposed to build on our platform -- lost more in five days than some countries produce in a year. The Dow hit 50,000 the same week. Everything is fine. Nothing is fine. Both of these are true at the same time. Time magazine named us "Architects of AI" -- Person of the Year 2025. Here is a statistic I would like you to hold: Companies and business trends named Time Person of the Year see their stocks decline a year later seven out of eight times. That is an 87.5% hit rate. The Magazine Cover Curse has a better track record than our AI model. We know this. We bought the ad anyway. Because here is what no one in my industry will say out loud: OpenAI lost $11.5 billion in a single quarter. Not a year. A quarter. Ninety days. That is $127 million per day. $5.3 million per hour. $88,000 per minute. In the time it takes to watch our Super Bowl ad, we lose $44,000. The projected cumulative losses by 2029: $115 billion. For comparison, Amazon's cumulative losses before it became profitable in its tenth year were $3 billion. We will lose 38 times that and still have no path to profitability. The more customers we acquire, the more money we lose. Every user who loves our product brings us closer to insolvency. Growth is the disease. Revenue is the symptom. "But Amazon survived," the pitch decks say. Yes. Amazon's early losses are a rounding error now. Ours would be a sovereign debt crisis. The hopeful investors are gullible investors. That is not my observation. That is from the MarketWatch analysis published the day before our ad aired. They compared us to the dot-com companies that Super Bowl ads carried to their graves. I read the piece. I forwarded it to my team. Subject line: "Good press." Because in this industry, a piece about your imminent collapse is still a piece about you. Blackstone's president said last week that if AI hype persists unchecked, it risks assuming endless growth. UBS warned that aggressive AI disruption could push private credit defaults to 13%. We are not building the future. We are building the most expensive proof of concept in human history and charging $20 per month for access while burning $88,000 per minute to keep the lights on. The Super Bowl is on. We are all watching. The dancing monkey has been upgraded to a neural network. The tagline hasn't changed.
English
25
41
254
37.4K
Rob Corbidge retweetet
Declaration of Memes
Declaration of Memes@LibertyCappy·
What the internet was really intended for... 😏
English
1
13
47
2.7K
Rob Corbidge retweetet
Ed Newton-Rex
Ed Newton-Rex@ednewtonrex·
Quite a few of my followers are AI skeptics, believing it’s over-hyped, or a scam, or some combination of the two. To those people - I implore you to reconsider. The models aren’t perfect. They make mistakes. But they are already insanely capable, and they are improving. There is a crowd who will cheer you on when you point out a hallucination, or laugh at a failed task, or lament yet another inexplicable use of the word ‘quietly’. But you are avoiding the reality of the situation. Capabilities have increased exponentially over the last few years, and there is every sign this trend will continue. I’m not saying you should get on board with AI. Far from it. I think AI poses huge risks, and is pretty clearly currently on a path to being net negative for society. But denying the obvious reality achieves nothing - and, worse, will make anyone who listens to you less prepared for the huge changes that are coming. I get why it is tempting to hope it is all hype. I genuinely wish it were. And I get why people dislike it so much. I think that dislike is good, is important - it is the thing that will drive people to fight the harms AI causes. But disliking it doesn’t require you to think it is make-believe. You can fight AI’s harms without trying to convince people it is useless, when all the evidence points to the contrary. Believe CEOs when they say they are cutting their workforce because AI can do the work of workers. It is true. This is why they are cutting their workforce. They are not lying. This doesn’t mean you have to accept it is the right path. It isn’t. It is very bad news that so many in power are realising they need fewer workers, instead paying huge companies for subscriptions to AI models. But pick the right fight. The right fight is the urgent fight of addressing AI’s harms - not pretending nothing is happening. AI is very capable, and becoming more so. That should be the starting point of any discussion about this. If you think an invader isn’t real, you won’t protect yourselves until it’s too late.
English
33
12
102
7.5K