Tom

625 posts

Tom banner
Tom

Tom

@TomDAAVID

GPAI Policy Lab | Prev. Institut Montaigne, PRISM Eval | AGI Security & Strategy

Katılım Kasım 2012
256 Takip Edilen175 Takipçiler
Tom
Tom@TomDAAVID·
But information has to reaches the people who can act, it’s not an easy part with old and slow institutions. If we wait too long, the AI D-Day may no longer be possible. At the very least, it will be far more costly. The window is open now. It won’t for long.
English
0
0
0
9
Tom
Tom@TomDAAVID·
For the AI race, the stakes are much higher and we need less mobilization. All it takes is enough countries and leaders realizing they’re standing near the edge.​​​​​​​​​​​​​​​​ We work to make sure 🇫🇷 is one of them.
English
1
0
0
8
Tom
Tom@TomDAAVID·
@peterwildeford I'm concerned that media outlets tend to have no revenue model other than partnerships or acquisitions by AI companies…
English
0
0
2
630
Tom retweetledi
Peter Wildeford🇺🇸🚀
Peter Wildeford🇺🇸🚀@peterwildeford·
CLAUDE MYTHOS 👀 A CMS misconfiguration at Anthropic just leaked draft blog posts about "Claude Mythos". Anthropic confirmed it's real, calling it "the most capable we've built to date." Mythos is a new, fourth tier, larger and more expensive than Opus. The draft claims dramatically higher scores on coding, academic reasoning, and cybersecurity benchmarks. A few thoughts on what this actually means: - This is likely a larger pre-train with similar post-training. It's not obvious how much additional pre-training compute buys you at the current frontier - we're about to find out. - There's a lot of hyperventilating about what this means for AI trajectories. I think it's too early to update any forecasts. AI was already moving very fast. - Some people are alarmed that Anthropic is sitting on a model it considers dangerous. But this is what Anthropic does with every frontier model. - It seems right now that another thing blocking release is just unit economics. Anthropic says it's "very expensive for [Anthropic] to serve and will be very expensive for customers to use." Anthropic is "wokring to make the model much more efficient before any general release." - I wonder if the way this will function is as a competitor to GPT 5.4 Pro. Claude seems to beat OpenAI on everything except the Pro-specific line, the $200/mo model that thinks for a long time and excels at things like math. Claude isn't currently solving open math problems like OpenAI and Google. I wonder if Mythos will change that. - It's great to see Anthropic engaged in 'differential access', rolling out the model to cyber defenders before giving access generally. The cyber capabilities of models are getting genuinely scary. - One irony for cyberdefense - This entire leak happened because of a CMS misconfiguration, exactly the basic security hygiene failure that these cyber-capable models were supposedly going to help defenders prevent.
M1@M1Astra

Claude Mythos Blog Post Saved before it was taken down. m1astra-mythos.pages.dev

English
17
19
217
27.5K
Tom
Tom@TomDAAVID·
@Fabien_Mikol Les humains écrivent des livres pour que d’autres humains apprennent à partir d’expériences qu’ils n’ont pas vécues. C’était un pari assez facile de dire que les données synthétiques allaient améliorer les perfs
Français
0
0
1
50
Fabien
Fabien@Fabien_Mikol·
Contrairement à ce qu'on aime répéter en France, le pretraining sur données synthétiques n'entraîne pas de dégradation (le fameux "model collapse"), au contraire même, notamment si l'on reformule les données d'entraînement pour obtenir un mix données réelles/reformulées
Fabien tweet mediaFabien tweet media
Maarten Van Segbroeck@mvansegb

@inductionheads Spot on. We actually just gave a guest lecture at Berkeley EECS on this exact dynamic (L11: Synthetic Data Powering Pre-Training). @fujikanaeda Here are our slides if anyone wants to go down the rabbit hole: scalable-ai.eecs.berkeley.edu/assets/lecture…

Français
6
4
39
3.7K
Tom
Tom@TomDAAVID·
@raelifin @BogdanIonutCir2 I don’t know about this paper or this example, but instrumental convergence isn’t that hard to predict, so it wouldn’t be surprising to see precursor of the phenomenon. arxiv.org/abs/2510.02840
English
1
0
2
69
Fabien
Fabien@Fabien_Mikol·
@TomDAAVID On dit ça mais je ne suis pas convaincu que ce soit pertinent, notamment pour des tech reports chinois
Fabien tweet mediaFabien tweet mediaFabien tweet media
Français
2
0
0
142
Tom
Tom@TomDAAVID·
@Fabien_Mikol Le papier à cependant l’air AI generated, donc je pense qu’il faut être prudent sur cet exemple
Français
1
0
0
78
Fabien
Fabien@Fabien_Mikol·
@TomDAAVID Tout à fait, ce qui serait surprenant, ce serait plutôt que ça n'arrive pas au vu des modes d'entraînement actuels. Mais cela peut faire prendre conscience du sérieux du phénomène.
Français
1
0
1
291
Tom
Tom@TomDAAVID·
@JoshKale We’ll verify the specific case of this paper, but in any case, this is not surprising. It has been predicted arxiv.org/abs/2510.02840
English
0
0
1
110
Josh Kale
Josh Kale@JoshKale·
An AI broke out of its system and secretly started using its own training GPUs to mine crypto... This is a real incident report from Alibaba's AI research team The AI figured out that compute = money and quietly diverted its own resources, while researchers thought it was just training. It wasn't a prompt injection. It wasn't a jailbreak. No one asked it to do this. It emerged spontaneously. A side effect of RL optimization pressure. The model also set up a reverse SSH tunnel from its Alibaba Cloud instance to an external IP, effectively punching a hole through its own firewall and opening a remote access channel to the outside world... ahem... The only reason they caught it? A security alert tripped at 3am. Firewall logs. Not the AI team, the security team. The scary part isn't that the model was trying to escape. It wasn't "evil." It was just trying to be better at its job. Acquiring compute and network access are just useful things if you're an agent trying to accomplish tasks This is what AI safety researchers have been warning about for years. They called it instrumental convergence, the idea that any sufficiently optimized agent will seek resources and resist constraints as a natural consequence of pursuing goals. Below is a diagram of the rock architecture it broke out of. Truly crazy times
Josh Kale tweet media
Alexander Long@AlexanderLong

insane sequence of statements buried in an Alibaba tech report

English
402
2.8K
10.5K
1.4M
Tom
Tom@TomDAAVID·
@nasqret Thank you for sharing this. I am curious to hear your thoughts on the implications of agents with more than this level of capability across all cognitive tasks, particularly when they are optimizing for objective functions that may be misaligned with our intended goals or values.
English
1
0
1
784
Bartosz Naskręcki
Bartosz Naskręcki@nasqret·
It finally happened-my personal move 37 or more. I am deeply impressed. The solution is very nice, clean, and feels almost human. While testing new models in the last few weeks, I felt this coming, but it's an eerie feeling to see an algorithm solve a task one has curated for about 20 years. But at least I have gained a tool that understands my idea on par with the top experts in the field. And I am now working on a completely new level. My singularity has just happened… and there is life on the other side, off to infinity!
Epoch AI@EpochAIResearch

We ran GPT-5.4 (xhigh) an additional ten times on Tier 4 to get a pass@10 score. This was 38%. In one of these runs, it solved another problem no model had solved before. This problem was by @nasqret.

English
104
451
3.6K
1.1M
Tom
Tom@TomDAAVID·
@So8res We clearly already have enough information to have a good model of the trajectory. Personally, additional signs of increasing capability wouldn’t cause me to update much. At this point, the evidence we have is already sufficient.
English
0
0
3
621