
Alex Loftus
392 posts

Alex Loftus
@AlexLoftus19
Our textbook is on amazon now! https://t.co/ayc3bMWVFt https://t.co/CUEOxFRDse | PhD student, Bau lab @ Northeastern. Studying LLM internals.
Boston, MA Beigetreten Ağustos 2020
481 Folgt217 Follower
Angehefteter Tweet

We're on Amazon!! a.co/d/38KkG0V
If anybody is curious about machine learning on graphs, check out this textbook. Lots of cool methods w fundamental linear algebra!
English

@wolfiesch @tszzl There probability of another independent human landing on this specific combination of words it's also ~0, and roon's probability of making the same post again if his memory were erased. So what you're saying is true but trivial
English

@francoisfleuret There are plenty of simple reasonable heuristics. What are the norms of all the layers? If you run activations through what is the per-layer distribution of various statistics about them? Etc
English

@TrueAIHound @pmddomingos @Susan16Park Domingos is correct and you are wrong. Go look at all of the good things America has done and then weigh the balance rather than cherry picking bad things.
English

@pmddomingos @Susan16Park By murdering little school girls, blowing up schools, civilian bridges and lying nonstop to the world? You're not well, Domingos. You need therapy. Stick to your "master algorithm".
English

@Jonas_Vollmer Why will demand for these types of in-person jobs go up in the absence of people needing more food, healthcare, etc?
English

@bennyzhu84 @nikitabier Oh wow, I'm seeing it now. People talking in French, Indonesian, and Vietnamese inside this thread and everyone can understand each other. Cool!
English

@nikitabier Có nghĩa là bây giờ tôi viết bằng tiếng Việt thì bạn bè của tôi trên toàn thế giới vẫn có thể hiểu được và tôi không cần chuyển ngữ nữa. Mọi người có ai hiểu tôi đang nói gì không?
Tiếng Việt

@ArthurB @Aella_Girl I don't think that's a strong meta-argument when you consider the cultural context. LessWrong and EA discourse has been obsessively developing their arguments and counterarguments for a decade; the people arguing against usually don't care to the same degree.
English

@Aella_Girl The meta case for the seriousness of AI x risk is that the arguments against it are almost uniformly shallow, when they amount to more than logical fallacies or straight name calling.
English

Just saw the AI doc and came away pissed at the optimists. I sort of expected them to have any argument that actually addressed the x-risk side, but they were basically like 'historically tech is good, people have been worried before but it was fine!' They didn't address at ALL the extremely entry-level concerns of like 'building something smarter than us is a categorically new type of threat'. They just repeated that tech would help humanity.
It's especially infuriating cause the most lifelong techno optimists I know ARE the doomers. The x-risk community are the ones who grew up on epic sci-fi fiction and have thought long and hard about what the singularity might bring. One of my friends (who was in the doc) once spent all night carrying ice into a hospital room to preserve the corpse of his friend in a desperate attempt to get him into a cryonics lab. It's real for them! But "AI has promise" is not even close to an adequate response to the extinction threat on the table. Even the AI CEOs in the movie - the ones that are *actually* doing the most acceleration - seemed to at least understand the gravity of the arguments they were engaging with.
The optimists in the doc seemed to have domain expertise in their technical fields, but were amateurs. They both are insufficiently visionary and also fail to engage with the actual risk in a practical way. I think they pattern match the "ai might kill us" people onto general woke anti-tech movement, and shout against them from a place of ego. That's the only good explanation I can think of for why they must be beating an activist drum that's so damn empty.

English

@evisdrenova List some hedge funds that would want this, I'm in!
English
Alex Loftus retweetet

Calling attention to an exciting "deception detection" hackathon we're planning this summer! w @NDIF and @CadenzaLabs.
Recruiting red teams now, blue teams later. Red teams, time is short: proposals due Mar 31. $10K stipend + compute, $15K finals prize.
nnsight.net/blog/2026/03/2…
English

Reinforcing negative neural pathways via therapy or introspection is a recipe for misery. Don’t cut a rut in the road.
Marc Andreessen 🇺🇸@pmarca
My big conclusion from this week: Introspection causes emotional disorders.
English

I mean it's not strong evidence but
a) it's definitely weak evidence
b) given the above, the onus is sort of on the doomers to make a strong argument that something bad will happen when thing scale more
You can't just randomly say "the world is about to explode" with 0 empirical evidence. It's just death cult energy at that point.
Many of the doomer arguments I read also feel like they are written by people who don't actually build stuff, and therefore have no actual substantive understanding of the system they are talking about.
English

@JeffLadish @tszzl I mean really, instances of scaling things by 10,000x are really a terrible candidate for induction. Very "i've never died from heroin before, i'm sure this will be fine" vibes.
English

modern alignment methods seem to work reasonably well across orders of magnitude of model scaling, survived the transition to verifiable rewards and that should at least inform your decision making
Brangus🔍⏹️@RatOrthodox
I have heard that some anthropic safety leadership are going around telling people that alignment is a solved problem. This seems like a predictable failure to me, and I would like people who thought that funneling talent towards anthropic was a good idea to think about it.
English

@tricksthatstick @eudaemonea @SecWar Or the united states could just go use another model provider instead of labeling a US company a foreign adversary because they couldn't strong-arm them into submission.
English

@eudaemonea @SecWar The question isn't the morality of those positions. It's the question of who gets to decide when those morals have been broken.
By agreeing with Anthropic you are saying that unelected leaders of company should have veto power on the government of the United States.
English

This week, Anthropic delivered a master class in arrogance and betrayal as well as a textbook case of how not to do business with the United States Government or the Pentagon.
Our position has never wavered and will never waver: the Department of War must have full, unrestricted access to Anthropic’s models for every LAWFUL purpose in defense of the Republic.
Instead, @AnthropicAI and its CEO @DarioAmodei, have chosen duplicity. Cloaked in the sanctimonious rhetoric of “effective altruism,” they have attempted to strong-arm the United States military into submission - a cowardly act of corporate virtue-signaling that places Silicon Valley ideology above American lives.
The Terms of Service of Anthropic’s defective altruism will never outweigh the safety, the readiness, or the lives of American troops on the battlefield.
Their true objective is unmistakable: to seize veto power over the operational decisions of the United States military. That is unacceptable.
As President Trump stated on Truth Social, the Commander-in-Chief and the American people alone will determine the destiny of our armed forces, not unelected tech executives.
Anthropic’s stance is fundamentally incompatible with American principles. Their relationship with the United States Armed Forces and the Federal Government has therefore been permanently altered.
In conjunction with the President's directive for the Federal Government to cease all use of Anthropic's technology, I am directing the Department of War to designate Anthropic a Supply-Chain Risk to National Security. Effective immediately, no contractor, supplier, or partner that does business with the United States military may conduct any commercial activity with Anthropic. Anthropic will continue to provide the Department of War its services for a period of no more than six months to allow for a seamless transition to a better and more patriotic service.
America’s warfighters will never be held hostage by the ideological whims of Big Tech. This decision is final.
English

@MadCADLad @DrInsensitive @eudaemonea @SecWar Anthropic is not even a little bit woke. They do not give a shit about student debt. I have a ton of friends there. Half of them are libertarian and center-right. They're being called woke because they (reasonably) don't want Claude used for mass surveillance.
English

@DrInsensitive @eudaemonea @SecWar You're probably not wrong and what'll happen is that the company will fold as soon as one of their ideological goals is given to them.
"Sure we'll cancel student debt, if you drop the other stuff and give us our weapon"
It's a lose lose for middle class, every scenario
English

@DrInsensitive @eudaemonea @SecWar I have like 20 friends at anthropic. It's not even slightly woke lmao. Literally nobody there is woke.
English

@eudaemonea @SecWar I suspect these two items were just the tip.
Woke companies always promise to stick in just the tip.
if we dig in deeper, we will find woke mandates.
English

One thing to understand about academic culture is that universities are not a monolith. Stanford did not publish psilocybin research, a particular lab at Stanford did. And in fact, often a particular PhD student or postdoc did. There is no university official that reads all the papers. Sometimes there are funding/grant constraints and there are certainly games played around citations and impact, but there isn't anyone telling us to do or not do any research.
English

This transparency you mention was a highlight.
Alex, after the last year, I am super gun shy.
Stanford has published things on psilocybin that push their patents, MITvsaid agents fail 95% of the time, universities lied about AI getting trauma, the labs don’t allow us to even load in a study to assess if it is counter to their funding…
I am so sick of people using confirmation bias.
This one was far less bad than any of the studies I current have downloaded on my phone.
Please know this is not anger at you. This is anger at manipulation at scale. I want truth and truth is in transparency.
I think AI is a solution to those like me who hated the education system and never fit.
I did well straight As, but learned nothing. Testing is the push in education now. To me, AI empowers and I wish someone would do a full study that looks at all the good and bad.
So few assessing the good. Almost all are fear based.
Lastly, I know the game. Scientists can either eat and go gray with ethics, or they don’t eat.
This is another game I wish to change. I am but one small violinist. The world won’t change until everyone sees what I do.
I hope we see more studies that are transparent. There was a lot of great stuff here. •
English

In this amazing multidisciplinary collaboration, we report our early experience with the @openclaw
->

English

@MarkerDaSharker @valigo @jpnyunja The literal most conservative assumptions possible still put us at 9x decrease per year
English

theres a lot of assumptions, and stated at the bottom is that the models are smaller (which the thinking generation that recently came out isnt), and hardware being more cost effective (the hardware being so cheap caused a ram shortage).
there was a discrepancy between the hardware price, the size of the models needed to do tasks, and a lot of obvious inefficiencies in early designs of LLMs that almost certainly will not (and have not) continue, not to mention a larger context window that spikes the input cost, albeit it is usually cached (and this is prolly even more true considering the price per input token this generation has fallen while output token has stayed the same or gone up).
this also obfuscates the cost from the company, they couldve also just overcharged because they didnt know how to price it/smaller user base that didnt scale as well initially.
English

@Amank1412 someone should make a version of this that turns it into the starcraft interface @AlexLoftus19
English

@Chaos2Cured @NatalieShapira @wendlerch @gsarti_ @kpal_koyena @adambelfki @nikhil07prakash @jannikbrinkmann @can_rager @AmirZur2000 @DiAtkinson1 @rohitgandikota @jadenfk23 @eunjeong_hwang @OrgadHadas @sam8393239 @nitalon @KaplanYotam @VeredShwartz @TamarRottShaham @criedl @r_mirsky @MaartenSap @davidmanheim @TomerUllman @davidbau Hi, I was one of the coauthors. I don't think we think AI is bad. We are also pretty independent and uninfluenced from our university's politics. The study is fully open; you can even go to our website and read all the discord logs for yourself!
agentsofchaos.baulab.info
English

@NatalieShapira @wendlerch @gsarti_ @kpal_koyena @adambelfki @AlexLoftus19 @nikhil07prakash @jannikbrinkmann @can_rager @AmirZur2000 @DiAtkinson1 @rohitgandikota @jadenfk23 @eunjeong_hwang @OrgadHadas @sam8393239 @nitalon @KaplanYotam @VeredShwartz @TamarRottShaham @criedl @r_mirsky @MaartenSap @davidmanheim @TomerUllman @davidbau Really?
Proving AI is a) bad, b) a failure.
Why?
Because those at the universities are freaking out. So they LIE to entrap new students into debt when there is ZERO guarantee of a degree helping them in an AI world. •
English

@cadmarlow @ylecun Is this happening to every other country, who also have access to AI, or just America?
English

@ylecun Is it Trump or AI ? Come on Yann your smarter than this
English

Alignment in multi-agent scenarios is very much not solved. We document many ways in which multiple interacting agents can scheme together, gaslight themselves and others, and generally cause havoc.
Natalie Shapira@NatalieShapira
In this amazing multidisciplinary collaboration, we report our early experience with the @openclaw ->
English

@valigo @jpnyunja What the fuck are you talking about?
epoch.ai/trends
epoch.ai/data-insights/…
LLM inference prices are dropping at a rate of 40x/yr
English








