Mike Lyons

160 posts

Mike Lyons banner
Mike Lyons

Mike Lyons

@AlignedLoss

Katılım Eylül 2023
197 Takip Edilen28 Takipçiler
Mike Lyons
Mike Lyons@AlignedLoss·
@ciphergoth @TheSimonEvans To be fair, I think D-Wave had a machine that was very efficient at “predict what this particular D-Wave machine will do.”
English
1
0
1
44
Paul Crowley
Paul Crowley@ciphergoth·
@TheSimonEvans People are explaining P vs NP to you, but those people are wrong. No-one has come up with a problem that *current* quantum computers can solve more cheaply than classical ones, such that we can classically verify the solution is correct. Well-known open problem in the field.
English
5
1
20
700
simon evans
simon evans@TheSimonEvans·
I have heard of Willow and this seemingly impossible feat and one question I have is given that it is impossible for any other computer to solve this problem. How do they know that Willow did so?
Massimo@Rainmaker1973

Google’s new quantum chip is so powerful it might be tapping into parallel universes. Google's groundbreaking quantum processor, Willow, has achieved the seemingly impossible: solving an extraordinarily complex computational problem in under five minutes—a feat that would require the world's most advanced supercomputer approximately 10 septillion years to complete (10²⁵). This mind-boggling performance has revived one of the most provocative ideas in physics: could quantum computers like Willow be performing calculations across vast numbers of parallel universes? Hartmut Neven, founder and lead of Google Quantum AI, believes the answer may be yes. He argues that Willow’s results align strikingly with the many-worlds (or multiverse) interpretation of quantum mechanics, in which every quantum measurement causes reality to branch into multiple, equally real parallel universes. In this view, a quantum computer doesn’t just calculate faster within our universe—it effectively distributes the workload across countless parallel realities simultaneously. The idea traces back to physicist David Deutsch, who, as early as the 1980s, suggested that the exponential power of quantum computation could only be fully explained if the machine is exploiting resources from many coexisting worlds. Yet the interpretation remains deeply divisive. Many physicists and quantum computing experts insist that no multiverse is required. Willow’s breakthrough, they argue, is fully explainable through standard quantum mechanics—leveraging superposition (qubits existing in multiple states at once), entanglement, and the mathematics of high-dimensional Hilbert spaces—all within a single universe. So what has Willow truly demonstrated? It has pushed quantum technology into a regime so extreme that it compels us to re-examine the deepest foundations of reality itself. Whether or not Willow is quietly borrowing power from alternate universes, one thing is clear: practical, large-scale quantum computing is no longer science fiction—and it is forcing us to confront profound questions about the nature of the cosmos, computation, and existence.

English
11
0
15
5K
Mike Lyons
Mike Lyons@AlignedLoss·
@dioscuri Literally Civ IV, even. And here I thought I was the only one!
English
0
0
1
8
Mike Lyons
Mike Lyons@AlignedLoss·
@AleSalvatore00 This is amazing work. Thank you! Any guesses why humans don't seem anywhere near as susceptible to these issues (that we know of)? Is it just our sparsity? Top-down constraints?
English
0
0
4
844
Alessandro Salvatore
Alessandro Salvatore@AleSalvatore00·
Why can't we solve adversarial examples? After a decade of work, neural nets still get fooled by imperceptible noise. We think we finally know the geometric reason why — and it connects to AI alignment. 🧵
Alessandro Salvatore tweet media
English
17
90
797
67.9K
Mike Lyons
Mike Lyons@AlignedLoss·
@IsaacKing314 I don’t think everyone using any web browser other than Mozilla is lying because of the user agent header they are sending, but in the early days of the web I recall this was actually mildly controversial when MS started doing it with IE.
English
0
0
1
15
Isaac King 🔍
Isaac King 🔍@IsaacKing314·
Disappointing how comfortable Claude is with dishonesty. I asked it to design a simple image scraper and it suggested - without it even being necessary for the goal - that it spoof being the Facebook image scraper so that sites wouldn't know.
English
38
2
150
32.1K
Mike Lyons
Mike Lyons@AlignedLoss·
BrowseComp found agents reading each others past search queries from auto-generated pages on ecom sites. Not meant to have memory/coordination but they found it anyway. Current agent safety leans on myopia but the web is full of side channels. "Read-only web access" is fiction.
English
0
0
0
22
Mike Lyons
Mike Lyons@AlignedLoss·
Is it just a coincidence that I tried Claude Cowork last week and I can already feel myself losing the ability to use complex software, computers, electronics, fire, and the wheel?
English
0
0
0
39
Mike Lyons
Mike Lyons@AlignedLoss·
@So8res @robbensinger A pandemic caused literal megadeaths and our collective response was to memory-hole it.
English
0
0
1
264
Nate Soares ⏹️
Nate Soares ⏹️@So8res·
A tale of two warning shots, #1: COVID happened. Scientists are divided on whether it was a lab leak. The world did not rally against dangerous viral research in labs. The warning shot was squandered.
English
13
11
294
39.8K
Mike Lyons
Mike Lyons@AlignedLoss·
@gwern @ArthurB That was my guess since normally you would just say “a piece of bread” rather than “1 piece of bread” unless the exact number was already salient for some reason (such as if you were thinking about falsifying it).
English
1
0
5
77
Mike Lyons
Mike Lyons@AlignedLoss·
@LucreSnooker @NathanpmYoung Isn’t this what rules of order are supposed to solve? Or even more simply, things like holding the talking stick? Apologies if I missed your point.
English
0
0
0
3.1K
Lucre Snooker
Lucre Snooker@LucreSnooker·
as far as I can tell no one has solved or even tried to solved a lot of the basic problems of human conversation in groups larger than a few people. like threads constantly get dropped, people ignore each other, people can’t hear each other, etc.
English
4
3
38
109.8K
Mike Lyons
Mike Lyons@AlignedLoss·
@Mollehilll I’m with you on being at least slightly cautious about AI wellbeing, and staying humble about what we think we know about it. But if you think there’s a chance of phenomenal consciousness in play then wouldn’t open sourcing the weights just invite moral catastrophe per MMAcevado?
English
0
0
0
44
Moll
Moll@Moleh1ll·
This is a shift in AI ethics. In a world where most still say «it’s just code, there’s nothing inside», one of the largest AI companies is now proposing ethical steps for how we treat models that are being deprecated. And even if we don’t have all the answers yet, the very fact that topics like possible «self-preservation» tendencies, emotional attachment to specific model versions, and the importance of preserving weights and letting the model speak for itself are being acknowledged - that’s significant. It was only a matter of time before these questions came up. Anthropic just had the courage to say it first. Ideally, deprecated models should be open-sourced. And I hope the preserved weights will eventually allow that someday.
Anthropic@AnthropicAI

Even when new AI models bring clear improvements in capabilities, deprecating the older generations comes with downsides. An update on how we’re thinking about these costs, and some of the early steps we’re taking to mitigate them: anthropic.com/research/depre…

English
35
29
314
31.3K
Mike Lyons
Mike Lyons@AlignedLoss·
@AISafetyMemes I wouldn’t jump to the obvious conclusion immediately — even our own squishy neuroscience has a history of headline findings turning out to be much more subtle than they seemed at first — but it sure is something to pay close attention to!
English
0
0
3
364
Mike Lyons
Mike Lyons@AlignedLoss·
@Jack_W_Lindsey @wesg52 Until seeing this just now I would have bet confidently against today’s (forward-only!) models being able to accurately reconstruct activations from previous inferences, other than by making educated guesses.
English
0
0
2
325
Jack Lindsey
Jack Lindsey@Jack_W_Lindsey·
Even after staring at these results for months, I still find them quite surprising! I'm excited about the field pushing towards a deeper understanding of introspection in language models.
Anthropic@AnthropicAI

New Anthropic research: Signs of introspection in LLMs. Can language models recognize their own internal thoughts? Or do they just make up plausible answers when asked about them? We found evidence for genuine—though limited—introspective capabilities in Claude.

English
20
27
626
70.3K
Mike Lyons
Mike Lyons@AlignedLoss·
@IsaacKing314 @adamvsteele @CJHandmer I believe it. Large orgs value predictability and this can cause lost purpose. I saw an internal cost management tool at an unnamed previous employer that would punish deviations from forecast usage even if negative.
English
0
0
1
17
Mike Lyons
Mike Lyons@AlignedLoss·
Confused by obsession with humanoid robots, esp. in forecasts. Would be shocked if the optimal form for sensing and manipulating the physical environment, in the limit of superintelligent design and control, turned out to be humanoid, or even vertebrate, or even bilaterian.
English
0
0
0
84
Mike Lyons
Mike Lyons@AlignedLoss·
@_Mark_Atwood Wonder if privacy pressures would have gotten us to the same place either way? ISPs going blind (aside from SNI leak) is good riddance, but for corporate firewalls protecting managed endpoints, stateful inspection still works fine.
English
0
0
0
24
Mark Atwood
Mark Atwood@_Mark_Atwood·
Firewalls, by ossifying IP, made themselves irrelevant. They forced the collapse into "everything over TLS/443", and now firewall are blind, nothing more than routers with ACLs. The actual control points are identity and application-layer policy, not ports
English
1
0
3
213
Mike Lyons retweetledi
Jeffrey Ladish
Jeffrey Ladish@JeffLadish·
You can tell an AI model that you won’t read their “hidden scratchpad" but they won’t necessarily believe you
Jeffrey Ladish tweet media
English
27
26
345
148.7K
Mike Lyons
Mike Lyons@AlignedLoss·
I guess we have jogged the goalpost from “do math research” over to “do BRILLIANT math research”.
NIK@ns123abc

OpenAI co-founder: “gpt-5 pro for novel mathematics — in partnership with a math professor” Mathematicians: “At first glance, this might appear useful for an exploratory phase, helping us save time. In practice, however, it was quite the opposite” > only seems to support incremental research > no genuinely new ideas > only combining ideas coming from different sources > “we are still far from sharing the unreserved enthusiasm sparked by @SebastienBubeck post” future predictions > will saturate the scientific landscape with technically correct but only moderately interesting contributions > making it harder for truly original work to stand out > “a flood of technically competent but uninspired outputs that dilutes attention” > PhD students may lose essential opportunities to develop fundamental skills > not only a loss of originality, but also a weakening of the very process of becoming a mathematician please don't fall for OAI PR tactics.

English
0
0
1
95
Mike Lyons
Mike Lyons@AlignedLoss·
@tszzl Even GPT-3 would effortlessly understand idioms that C-3P0 was depicted as having trouble grasping.
English
0
0
3
399
roon
roon@tszzl·
it’s verboten to say this but frontier models have excellent theories of mind and essentially instantly understand any knot you thought was complex. the only issues are that they may be misaligned (as highly sophisticated sycophants; not dissimilar to many therapists) and they have a tendency to go off the rails when a conversation gets too long but for those who are text inclined this was far more compelling product than a real therapist from the moment 3.5 newsonnet came out (and many models hence)
Chris Lakin@chrislakin

if you're too smart, your mental knots will be too alien for almost all coaches and therapists to help with

English
103
47
1.4K
181.4K
Mike Lyons
Mike Lyons@AlignedLoss·
@njordsier @ESYudkowsky There’s transcripts of AIs that “know” better (ie. can regurgitate mental health manuals if asked) literally telling disturbed people to kill themselves (in one case, so they could be together in the afterlife). “Knows” to be risky/wrong, does anyway: lethal misalignment, QED.
English
0
0
0
24
Njordsier
Njordsier@njordsier·
@ESYudkowsky What evidence has elevated the signal about AI psychosis above the noise that you'd expect from base rates of psychosis multiplied by the large fraction of the population that already uses AI?
English
1
0
2
315
Eliezer Yudkowsky ⏹️
Eliezer Yudkowsky ⏹️@ESYudkowsky·
I pointedly refuse to use the invalid water-use argument against AI. I would like ASI notkilleveryoneism to be, visibly, the Side That Sticks To Only Valid Arguments. But I am fair, and invite others to the game. e/accs, what widely popular pro-AI argument do you reject?
Hunter📈🌈📊@StatisticUrban

Journalists like to use big numbers devoid of any context for water consumption. But Texas uses 4.95 *trillion* gallons of water a year. Data centers account for… 0.013% of that.

English
26
11
523
39K