Aryeh L. Englander

1.3K posts

Aryeh L. Englander

Aryeh L. Englander

@AryehEnglander

Entrou em Mart 2016
430 Seguindo101 Seguidores
Tyler John
Tyler John@tyler_m_john·
OK since p(doom) is discourse here is my view on communicating risk with probabilities. We should do it because it makes it much clearer to people what you think and is empirically demonstrated good epistemic practice. Also we should hedge to show high-order uncertainty. Some theses: 1. Probabilities give people more insight into what you think. If you use vague, qualitative language instead of numbers, people will just assume what you mean. There's @PTetlock's famous Bay of Pigs anecdote, where an advisor told Kennedy there was a “fair chance,” meaning a 25% chance of success. Kennedy later reported he had assumed the advisor meant a 75% chance, and said he wouldn't have pursued the invasion if he had known the advisor only meant 25%! But this kind of miscommunication is ubiquitous. People assume different things about likelihood when speakers use qualitative language — it's an inherently less clear way to communicate what you are thinking. If you want your speaker to understand you, use numbers! Or at the very least, refer to the literature on perceptions of probability (see below) and pick your qualitative term very carefully so you communicate the right range! And don't use the extremely vague terms like "fair chance" or "improbable" that could mean literally anything to your listener. That is an extreme form of carelessness that we don't criticize often enough. 2. There haven't been many clear findings from the science of forecasting, but one of the clearest findings is that you make better predictions when you use precise numbers, even if these are completely made up. This is also true in group settings when aggregating the judgments of many people — which is essentially an idealized version of what we're doing pretty much any time we talk about probabilities. academic.oup.com/isq/article-ab… Here is an old thread I wrote on this topic some years ago: x.com/tyler_m_john/s… 3. Yes, people do perceive numbers as signaling more authority, and we shouldn't signal more authority than is appropriate. (How much is appropriate? Depends on the context. There isn't a universal answer in the context of existential risk from AI.) But you can do that without dropping numbers and losing the benefits of numbers I just set out. For example you can just use couching language, like "I would guess roughly 20%, but huge error bars, no one knows." 4. This can be studied!! It has already been studied a lot. I am finding it frustrating that no one in this debate is citing actual literature on perceptions of probabilities, especially in the age of LLMs where this information is readily available. We do know that percentages are viewed as more credible than qualitative language: papers.ssrn.com/sol3/papers.cf…. We do also know that hearing "61.87%" rather than "60%" triggers the inference that the speaker must have epistemic access that warrants the extra digits. frontiersin.org/journals/psych… How much higher-order confidence is it appropriate to convey when communicating the P(doom) of, say, a world expert on AI or an aggregate survey of every AI researcher publishing in NeurIPS? I don't know! If you want to make an argument that saying "20%" signals too much confidence, please cite some of this literature and explain why you think that the groundedness signaled to the audience is inappropriate. If you do want to advocate for a different communication style, it is not expensive to run a quick MTurk study to see what people's perceptions of it are and compare it to default rhetoric. Or even more cheaply you can run it on LLMs, which are a decent natural laboratory for testing hypotheses about human psychology in the absence of humans to test on. I hope to practice what I preach in the coming days and run some more LLM tests (I've ran one N = 7000 test yesterday) and set up a Mechanical Turk account so I can test my above claim about couching probabilities being just as good as using qualitative language, but with more clarity in communication and better epistemic practice.
Tyler John tweet media
English
9
9
83
14.8K
Scott Alexander
Scott Alexander@slatestarcodex·
What? You dare accuse us of not being situated within a lineage?! Our lineage is top notch! Most of us were initiated by @Raemon777 at one of his secret solstices. @Raemon777 was initiated by @ESYudkowsky at Benton-That-Was. @ESYudkowsky was initiated in an epileptic vision, St-Paul-style, by ET Jaynes. ET Jaynes was initiated at Cambridge by Harold Jeffreys. Harold Jeffreys was initiated in a storm off the Cape of Good Hope by Francis Galton. Francis Galton was initiated by Jeremy Bentham's Auto-Icon (it still counts! he formed the intention before death, and his body was almost perfectly preserved!) at University College London. Jeremy Bentham was initiated by Pierre-Simon Laplace in the Paris Catacombs. Pierre-Simon Laplace was initiated by The Reverend Thomas Bayes in Tunbridge Wells. The Reverend Thomas Bayes was initiated by either Isaac Newton or Gottfried Wilhelm Leibnitz, apparently the records are unclear and fiercely contested, but it doesn't matter because both of them were initiated in the same Amsterdam lens-grinding workshop on the same day by Christian Huygens. Christian Huygens was initiated by Rene Descartes, who fools and unbelievers claim "founded" the Rationalists, on the point (1, 1) on the coordinate plane. Rene Descartes was initiated by Francis Bacon during a secret meeting in Calais. Before Francis, the lineage apparently stayed in the Bacon family for seven generations, all the way back to Roger Bacon, whose brazen automaton was an early, though abortive, attempt at pre-empting bad actors and building a safe superintelligence. He was initiated by Leonardo Fibonnaci, who was initiated in Morocco by Nizami Aruzi, who was initiated in Nishapur by Omar Khayyam (the regular one, not Ravenhurst). Khayyam was initiated by a nameless philosopher who fled Baghdad during the persecutions of al-Ghazali, who was initiated by ibn al-Haytham, who was initiated by ibn Qurra (there may have been several ibn Qurras, the record is unclear), who was initiated by al-Khwarizmi (of "algorithm" fame!), who was initiated by Alhazred (falsely accused of being an early 20th century invention, but we know the truth), who was initiated by ibn Hayyan, by John of Damascus, by Jacob of Edessa, by Stephanus of Alexandria, and then something like thirty different Alexandrians I can never keep straight, mostly corresponding to the Secret Keepers of the Great Library. Origen was actually one of ours (the part with everyone being perfect spheres in Heaven gives it away), as was Heron (who aborted his incipient Industrial Revolution after realizing that it would increase existential risk too much if we got ultratechnology before Christianity had finished pacifying the West), as was Demetrius of Phalerum. Demetrius was initiated by Epicurus (so far nobody has noticed the sacred geometric similarities between Lighthaven and the Κήπος of Epicurus, probably because nobody except us retains the map of the latter). Epicurus was initiated by Nausiphanes, Nausiphanes by Diogenes of Smyrna (NOT the barrel-dweller, different guy with the same name!), and Diogenes by Democritus (Scott Aaronson's "Quantum Computing Since Democritus" is a nod to some of the content in their secret dialogues). Democritus was initiated by Leucippus in Miletus, Leucippus was initiated by Aristaeus in Croton, and Aristaeus was initiated by Pythagoras, who personally invented Reason in the year 536 BC after staring too hard at a right triangle. We still have it, by the way, although it no longer satisfies his famous Theorem since most the hypotenuse was lost during the Great Fire of London. It's too fragile for public view, but @ohabryka will let you see it if you donate more than $1K to Lightcone.
English
11
29
544
24.9K
Dean W. Ball
Dean W. Ball@deanwball·
It is so perfect that people who are obviously rationalists by any historical or objective standards would have as a major premise that We Are Not Rationalists, We Are Something New, Formed From The True Cognition. The premise is always denying history.
Eliezer Yudkowsky@allTheYud

@deanwball Sir. We are empiricists. That there was once an anti-empiricist movement called Rationalism is a sheer accident of history.

English
15
1
110
23K
Dean W. Ball
Dean W. Ball@deanwball·
Opus 4.7 just invented a new wife for me
Dean W. Ball tweet media
English
29
3
213
101.2K
Aryeh L. Englander
Aryeh L. Englander@AryehEnglander·
@tyler_m_john As I understand it though, there are different approaches to 2nd order uncertainty (if used at all) - mainly 2nd order probabilities and various types of imprecise probability.
English
1
0
0
101
Aryeh L. Englander
Aryeh L. Englander@AryehEnglander·
@tyler_m_john 2nd-order uncertainty can be decision relevant. Risk aversion is an example: If ample data shows flood risk 1%, but there's also a war looming, w/ est. chance of home destruction anywhere b/w 0.01%-2% (EV≈1%). For risk averse, war insurance is more valuable than flood insurance.
English
2
0
0
95
Aryeh L. Englander
Aryeh L. Englander@AryehEnglander·
@tyler_m_john So does that mean you don't actually use higher-order uncertainty in forecasting & decision making except to communicate lack of precision for those who might get hung up about that, plus maybe to help keep track of underlying reasons for updating purposes?
English
1
0
1
81
Tyler John
Tyler John@tyler_m_john·
@AryehEnglander 50% chance of 30% chance and 50% chance of 70% chance is equivalent to 50% chance unconditionally. Not that you shouldn't keep track of the underlying reasons so you can update appropriately.
English
1
0
1
82
Dean W. Ball
Dean W. Ball@deanwball·
Whenever I say “the doomers express way too much confidence” I inevitably get waves of responses saying “actually rationalists have highly calibrated probabilities.” Does this tweet convey a sense of carefully measured probability? Does it seem uncertain… at all?
Eliezer Yudkowsky@allTheYud

@deanwball This IS what we're seeing. It means that some stuff will go wrong when somebody actually builds ASI and it actually gets up there and then everyone dies because there was some gotcha that blew up the clever plan.

English
20
3
123
15.9K
Aryeh L. Englander retweetou
davidad 🎇
davidad 🎇@davidad·
have you seen this pattern before? - knows more STEM - knows less about celebrities and sports - worse at following instructions - better coding perf - worse performance at admin/ops - knows more literature - less engaged by pointless brainteasers and needle-in-haystack searches
English
61
81
1.9K
346.7K
Aryeh L. Englander retweetou
Peter Wildeford🇺🇸🚀
Peter Wildeford🇺🇸🚀@peterwildeford·
@sama April 2025: "we fix our model naming by this summer" OpenAI Summer 2025 to now: GPT5 -> GPT5.1 -> GPT5.2 -> GPT5.3 -> GPT5.4 -> GPT5.4-Cyber -> GPT-Rosalind ...So close. They almost had it.
OpenAI@OpenAI

GPT-Rosalind, our Life Sciences model series, is optimized for scientific workflows, with stronger performance in protein and chemical reasoning, genomics analysis, biochemistry knowledge, and scientific tool use.

English
5
2
37
3.1K
Aryeh L. Englander retweetou
Epoch AI
Epoch AI@EpochAIResearch·
Have AI capabilities accelerated? On 3 out of the 4 AI capability metrics we investigated, we found strong evidence of acceleration, around when reasoning models emerged.
Epoch AI tweet media
English
9
55
432
48.2K
Aryeh L. Englander retweetou
Kevin Roose
Kevin Roose@kevinroose·
New column: I went to visit @METR_Evals, the 30-person AI nonprofit that makes the Most Important Chart in the World. I learned a lot, but the most striking thing was how soon some of them think AI R&D could be fully automated. (This year!) nytimes.com/2026/04/17/tec…
English
11
56
530
90.2K
Drake Thomas
Drake Thomas@MaskedTorah·
@TheZvi Less juicy overall than last time, but I was happy we got to fit in section 6.1.3:
Drake Thomas tweet mediaDrake Thomas tweet media
English
4
0
32
6.8K
Aryeh L. Englander retweetou
Satya Nadella
Satya Nadella@satyanadella·
Our Fairwater datacenter in Wisconsin is going live, ahead of schedule. As the world’s most powerful AI datacenter, it will bring together hundreds of thousands of GB200s into a single seamless cluster. Congrats to all the teams who made this possible!
Satya Nadella@satyanadella

If intelligence is the log of compute… it starts with a lot of compute! And that’s why we’re scaling our GPU fleet faster than anyone else. Just last year, we added over 2 gigawatts of new capacity – roughly the output of 2 nuclear power plants. And today we’re going further, announcing the world's most powerful AI datacenter, located in southeastern Wisconsin. Fairwater is a seamless cluster of hundreds of thousands of NVIDIA GB200s, connected by enough fiber to circle the Earth 4.5 times. It will deliver 10x the performance of the world’s fastest supercomputer today, enabling AI training and inference workloads at a level never before seen. For AI training workloads, you need compute at exponential scale. That’s why we designed the datacenter, GPU fleet, and network together as one integrated system. This ensures a single job can run from day 1 at exponential scale across thousands of GPUs. Fairwater uses a liquid-cooled closed-loop system for cooling GPUs that requires zero water for operations after construction. And we’re matching all of the energy that is consumed with renewable sources. And of course, it is just one of several similar sites we’re lighting up across our 70+ regions. We have multiple identical Fairwater datacenters under construction in other locations across the US, in addition to our AI infrastructure already deployed in over 100 datacenters around the world, powering model training, test-time compute, RL tuning, and real-time inference at global scale. Too often during times like this, people go with the current and only later wonder, how did we get here? With Fairwater, we're charting a new path: doing the hard engineering work, bringing compute, network, and storage into one highly scaled cluster, and designing closed-loop energy systems to meet real-world computing needs. And partnering with local communities to ensure it's thoughtfully done in a way that is sustainable, creates new jobs, and expands opportunity. We are thrilled to see this take hold in Wisconsin, and we are just getting started.

English
324
456
3.9K
642.9K
Aryeh L. Englander
Aryeh L. Englander@AryehEnglander·
RT @emollick: AI keeps getting better but the last time the shape of the jagged frontier changed radically was o1 & the Reasoner. A good m…
English
0
2
0
1
Aryeh L. Englander retweetou
WarTranslated
WarTranslated@wartranslated·
"For the first time in the history of this war, Ukraine has captured an enemy position using only ground robots and drones. The occupiers surrendered. The operation was carried out without infantry participation and with zero losses on our side." - Zelenskyy.
WarTranslated tweet mediaWarTranslated tweet mediaWarTranslated tweet media
English
120
1.6K
8.6K
397.6K