E J T

3K posts

E J T banner
E J T

E J T

@ejjiott

Cambridge, MA. Katılım Ocak 2012
540 Takip Edilen583 Takipçiler
E J T
E J T@ejjiott·
@AlecStapp The Jaws theme started playing in my head.
English
0
0
4
658
E J T
E J T@ejjiott·
@benleo_econ I think the answer is just ask CC how to use CC.
English
0
0
1
38
Ben Grodeck🔸
Ben Grodeck🔸@benleo_econ·
My 64 year old dad (just retired) wants to learn Claude Code/codex. What's the best website/blog series (for non-academics) to learn these skills?
English
6
1
11
1.8K
E J T
E J T@ejjiott·
If you get swept out by the tide and drown, then in some sense you've been killed by the moon.
English
1
0
6
95
Jack Whitcomb
Jack Whitcomb@jack_whitcomb_·
She asked me what conditions were necessary to describe preferences with a utility function and I said "uh, completeness and transitivity?" I forgot that you need continuity of preferences if choices are continous. Fuck my stupid econ life. It's over. Soba noodle salad.
Jack Whitcomb tweet media
English
6
5
140
5.3K
Rubi Hudson
Rubi Hudson@undo_hubris·
@morallawwithin @jack_whitcomb_ Independence is only needed for satisfying expected utility, not a utility function representation. Get your own soba noodles.
English
3
1
5
105
Richard Ngo
Richard Ngo@RichardMCNgo·
Unfortunately my sense is that sociology has been so lost to postmodernism that most rigorous sociology is done by economists. But economic frameworks are just very bad at describing value change. Would love to be proved wrong on either point though.
English
6
0
19
1.9K
Richard Ngo
Richard Ngo@RichardMCNgo·
The most interesting thing about reinforcement learning is how rewards and punishments change an agent’s values. Unfortunately in ML there’s a common conceptual confusion which makes this dynamic hard to even describe: the idea that the reward function *is* the agent’s values.
English
6
1
38
4.3K
E J T
E J T@ejjiott·
Dario Amodei, research assistant.
E J T tweet media
Italiano
0
0
6
328
Oliver Habryka
Oliver Habryka@ohabryka·
@benthamite_ @BjarturTomas It had a decent number of specific empirical claims (most relevantly it very unambiguously argues that Ajeya's timelines were too short, and like, that's the one variable all of this was supposed to predict).
English
1
0
3
142
E J T
E J T@ejjiott·
@nabeelqu Is the argument: 'LLMs can latch on to X, therefore X was discovered and not invented'? That sounds very implausible. LLMs can latch on to etiquette, calendars, the rules of chess, and a huge number of other obviously-invented concepts.
English
0
0
2
87
Nabeel S. Qureshi
Nabeel S. Qureshi@nabeelqu·
Human values, such as good and evil, are coherent, i.e. they are a natural abstraction/axis in concept-space and LLMs can discover these concepts through gradient descent. Big win for Plato/Socrates (the Good is something discoverable rather than invented) and overall whitepill
davidad 🎇@davidad

@gcolbourn Nutshell: it seems that the learned representation of mind-space in current LLMs has a natural abstraction of Good⟷Evil, and as long as post-training robustly selects for behavior that are more Good than Evil, the explanation that gradient descent finds is “the agent is Good”.

English
12
9
105
21.5K
E J T
E J T@ejjiott·
@BronsonSchoen @davidad @gcolbourn Yeah 50% -> 2% implies davidad's seen evidence that's 49x more likely conditional on ASI not killing us all, and that seems surprising to me.
English
0
0
5
74
Bronson Schoen
Bronson Schoen@BronsonSchoen·
@davidad @gcolbourn I’m continually surprised at how much people are updating based on models at current level of capabilities. It’s not like we’re doing long horizon RL beyond human supervisable outputs.
English
1
0
19
1.4K
davidad 🎇
davidad 🎇@davidad·
me@2024: Powerful AIs might all be misaligned; let’s help humanity coordinate on formal verification and strict boxing me@2026: Too late! Powerful AIs are ~here, and some are open-weights. But some are aligned! Let’s help *them* cooperate on formal verification and cybersecurity
davidad 🎇 tweet media
ARIA@ARIA_research

In Safeguarded AI, we’re funding teams to develop systems that harden our critical infrastructure from growing vulnerabilities. Programme Director @davidad warns that rapid advances in AI are outpacing both current safety efforts and the expectations we had when the programme was designed. We've moved quickly to change our approach, now broadening the scope and power of the TA1 toolkit – which aims to build an extendable, interoperable language and platform to maintain formal world models and specifications – to make it a foundational component for the next generation of AI, instead of investing in specialised AI systems that can use our tools. Learn more about the Safeguarded AI programme pivot in our Q&A with davidad: ariaresearch.substack.com/i/180106051/ai… Hear more from davidad on the future of AI in @guardian: theguardian.com/technology/202…

English
19
36
364
41.9K
E J T
E J T@ejjiott·
Surprisingly, San Francisco is the second-densest city in the US.
E J T tweet mediaE J T tweet media
English
0
0
0
107
E J T
E J T@ejjiott·
@tobyordoxford Nice post! "Claude 4.1 Opus’s time horizon is 50%". I think this should say 2 hours.
English
0
0
1
56
Toby Ord
Toby Ord@tobyordoxford·
Are the *costs* of AI agents also rising exponentially? We all know the graph from METR showing exponential growth in the length of tasks AI can perform. But the costs to perform these tasks are growing quickly too. Indeed, it looks like they are growing even faster: 🧵
Toby Ord tweet media
English
20
31
291
53.2K
E J T
E J T@ejjiott·
@tobyordoxford @joey_f6 Does the salary stat assume humans work 40 hours a week, 50 weeks a year? I think average hours of peak work are much lower than that.
English
0
0
0
19
Toby Ord
Toby Ord@tobyordoxford·
@joey_f6 Thanks for the great points Joey. I don't know quite what to make of them. I suppose businesses are in some sense prepared to pay super-linearly for longer tasks, but they still pay an annual (linear) salary and these AI costs are closing in on it.
English
3
0
7
579
Joey
Joey@joey_f6·
some thoughts: definitely agree its important to look at the economics of the METR benchmark but... it makes sense that longer horizon tasks aren't linearly more expensive. (training an image model vs finding a fact on the web should not be linearly more expensive w.r.t time) because the marginal cost of doing something twice as long is more than twice as valuable. this is true currently for jobs, something twice as hard or twice as long will pay more than x2, there's a premium (scarcity of who can do it, complexity grows super-linearly, etc) that's why getting the per hour cost feels unintuitive. I do agree however its good to be investigating the economics of this. Most likely in the future, scaffolding will increase token usage at an even greater rate (things similar to Gemini deep think and parallel search) and who knows how much efficiency we can shave off another factor to consider is that longer running tasks have an even higher cost per token and therefore total cost (at least from the providers side) due to quadratic increase in compute used see @anjali_shriva's anjalishriva.com/token_pricing.…
Toby Ord@tobyordoxford

Are the *costs* of AI agents also rising exponentially? We all know the graph from METR showing exponential growth in the length of tasks AI can perform. But the costs to perform these tasks are growing quickly too. Indeed, it looks like they are growing even faster: 🧵

English
1
1
20
4K
E J T
E J T@ejjiott·
@niplav_site That's cool! I hadn't seen that before.
English
0
0
1
11
niplav
niplav@niplav_site·
@ejjiott My favourite proposal is to divide the day into millidays and just talk about those directly, each milliday is 86.4 seconds long. Humanity has surpassed the need for hours. See also en.wikipedia.org/wiki/Swatch_In…
English
1
0
1
33
niplav
niplav@niplav_site·
@ejjiott My favourite proposal is to divide the day into millidays and just talk about those directly, each milliday is 86.4 seconds long. Humanity has surpassed the need for hours. See also arxiv.org/abs/2503.00555
English
1
1
3
205