James

5.5K posts

James

James

@jmac_ai

Ask me about #ReinforcementLearning #AI research @SonyAI_global RL for games, robotics, and other real-world applications Views and tweets are my own.

Katılım Mart 2012
579 Takip Edilen783 Takipçiler
Sabitlenmiş Tweet
James
James@jmac_ai·
Reinforcement learning in #AI is hard, so I’ve made a website to collect answers I’ve given to common RL questions. It's named Decisions & Dragons. It’s launching with 8 questions and answers, but I will add to it in the future. A 🧵to give a preview with the link below.
James tweet media
English
5
33
251
24.2K
James retweetledi
James
James@jmac_ai·
@seohong_park Great post! I'll note that our work on GT is a real-world off-policy RL success. (It's in the game!) However, I share a similar conclusion. Off-policy Q-learning is brittle. I think MBRL where you learn from off-policy data but optimize on-policy in the model is more promising.
English
2
0
3
564
Seohong Park
Seohong Park@seohong_park·
Q-learning is not yet scalable seohong.me/blog/q-learnin… I wrote a blog post about my thoughts on scalable RL algorithms. To be clear, I'm still highly optimistic about off-policy RL and Q-learning! I just think we haven't found the right solution yet (the post discusses why).
Seohong Park tweet media
English
35
182
1.2K
168.2K
James
James@jmac_ai·
@bimald @coolkoon @GaryMarcus I’m not going to tell you LLMs will have no impact on coding. For simple things it can give non-programmers more flexibility. However, it’s not like chess, because the limitations if English in specifying behavior is an inherent bottleneck, where no such thing exists for chess
English
0
0
1
88
James
James@jmac_ai·
@bimald @coolkoon @GaryMarcus I think you misunderstand the nature of code. Code isn’t for computers, it’s for people. English is *bad* at specifying clear behavior. In fact, regularly in design discussions, we move to writing bits of code because it’s more clear than words.
English
2
0
1
159
Gary Marcus
Gary Marcus@GaryMarcus·
A computer scientist’s perspective on vibe coding:
Gary Marcus tweet media
English
625
1.3K
8.7K
831.5K
James
James@jmac_ai·
@kevinroose AI Expert here. First, a great many experts do not believe in the extreme hype/doom & the loudest hypers/doomers are usually not AI *scientists*. Second, unlike climate change, the hype/doom is not based on a scientific model. It is speculative rhetoric. They are not the same
English
0
0
1
68
Kevin Roose
Kevin Roose@kevinroose·
There is a weird amount of overlap between the people who insist that we take climate experts seriously when they warn of big changes on the horizon, and people who refuse to take AI experts seriously when they say the same thing.
English
232
54
798
637.5K
James
James@jmac_ai·
@chrisprucha @bradneuberg The bigger limiter has always been that you always need to do experiments in the real world to advance, and that means both resources and contending with the speed of reality. Science is necessary, not a kludge.
English
1
1
4
267
Chris Prucha
Chris Prucha@chrisprucha·
OpenAI's O3 model really makes the Doomer hard takeoff or "FOOM" theory look like a bunch of BS. As we start to enter the age of AGI, the massive amounts of required compute, interconnect bandwidth, and energy are real physical constraints that govern scale over the time dimension. FOOM can't happen when you need to build massive solar farms and nuclear power plants.
English
34
19
445
63.7K
James
James@jmac_ai·
@SenMastriano Resign. You are unfit for a booster seat.
English
0
0
0
21
Senator Doug Mastriano
Senator Doug Mastriano@SenMastriano·
It is inconceivable that the federal government has no answers nor has taken any action to get to the bottom of the unidentified drones. The fecklessness of this administration was on display last year when a Chinese surveillance balloon was allowed to fly over the entire continental United States before being shot down. Such should be viewed as a threat to our nation and citizens and action is long overdue. We have recourses and assets in our arsenal to get answers, but I suppose Ukraine is more important to the White House. January 20th can’t come soon enough.
Senator Doug Mastriano tweet media
English
3.9K
779
4.9K
5.5M
James
James@jmac_ai·
@agrimgupta92 Very impressive. Although it feels like this person is moments away from cutting off their finger and it’s giving me an anxiety attack :p
English
0
0
1
1.1K
Agrim Gupta
Agrim Gupta@agrimgupta92·
"A pair of hands skillfully slicing a ripe tomato on a wooden cutting board" #veo
English
140
211
2.7K
3.6M
James
James@jmac_ai·
@rao2z @natolambert I think this may be a misunderstanding of the bitter lesson. Rich Sutton is very focused on online learning and not pretraining, to a fault.
English
0
0
0
60
Subbarao Kambhampati (కంభంపాటి సుబ్బారావు)
long chain.. Bitter lesson, IMHO, ignores efficiency aspects. Go with compute rather than knowledge. This is fine when most of the compute is done behind the scenes at pre-training time. This was the case with original auto-regressive LLMs. But they have huge limitations anywhere reasoning is needed . Because their performance depends, to a large extent on compiling reasoning to retrieval). After the initial euphoria and wild claims about LLM reasoning abilities, it is clear even to the most upbeat LLM optimists that there just ain't enough data to just pretrain and do well at inference time. So you need synthetic data (a whole bunch of which comes from synthetic--existing--solvers--which, to the chagrin of bitter lesson, were made by people. Strike 1.) Even with this, pre-training won't be enough, and so you wind up doing inference time compute.. Which makes efficiency considerations of the whole workflow a whole lot more paramount. [My joke to my students is that deep RL is the best source of easy efficiency papers--since almost any second rate idea improves on its rank inefficiency..😅] YMMV.
English
2
2
6
1.4K
James
James@jmac_ai·
@Intrinsic29 I don't think I've ever seen a non-doomer say "Some things are impossible, therefore AI cannot be a threat." It's always been a reaction to magical thinking where a doomer will literally be equating a future AI with a "god."
English
1
0
2
35
James
James@jmac_ai·
@Intrinsic29 Often, if a non-doomer brings up limits, its because a doomer presented an argument assuming no limits. Doomers regularly hand wave "of course the AI finds a way to do x because it's superintelligent." You can't automatically conclude that without making magical assumptions.
English
1
0
2
38
Kevin
Kevin@Intrinsic29·
As an experiment, please do me a favor and just start paying attention to how many AI doomer arguments are clever refutations of arguments that basically nobody is making.
Kevin tweet media
English
2
0
3
137
Тsфdiиg
Тsфdiиg@tsoding·
Your language is just a wrapper around actually useful C code. That's why you hate C. You hate it 'cause you are nothing without it.
English
169
270
4.5K
244.9K
James
James@jmac_ai·
@Miles_Brundage I've been all aboard the "we can solve the AI problem" train my entire adult life. But claiming it's not hard shows a lack of respect for the challenging problem we've tackled and is falling victim to hype instead of science.
English
0
0
2
53
James
James@jmac_ai·
@Miles_Brundage What unwillingness? The timeless pattern is AI researchers thinking it will be solved soon and being wrong every time. The founders of the field thought they'd solve much of the problems in a few months with a small team. There is far too much willingness to believe it s easy.
English
1
1
4
197
Miles Brundage
Miles Brundage@Miles_Brundage·
The real wall is an unwillingness to believe that human intelligence isn’t that hard to replicate and surpass
English
44
58
475
91.7K
Erbun Ninja
Erbun Ninja@ErbunnNinja·
@jmac_ai @Intrinsic29 We’re talking about people interacting with, and even communicating with systems that mimic agency. I think it’s fair to say this is more than a scientific and technical issue.
English
2
0
0
34
Kevin
Kevin@Intrinsic29·
There's zero reason to believe that we're building a God or that intelligence alone can lead to anything remotely similar to a God. The real bottleneck for knowledge is scientific experiment not intelligence and AI will have extreme logistic limitations to conducting experiments.
Blokhaus@Blokhaus

"For better or worse, we are building a God." Tim Urban of @waitbutwhy fame describes how he started down the rabbit hole of AI and realized the importance of the topic.

English
2
1
10
371
James
James@jmac_ai·
@ErbunnNinja @Intrinsic29 There will be a world of difference. People who use the word god end up confusing themselves precisely because they're assuming the difference. This is a scientific and technical issue. The word "god" has no value in discourse of it and can -- and does -- confuse matters.
English
1
0
1
39
Erbun Ninja
Erbun Ninja@ErbunnNinja·
@jmac_ai @Intrinsic29 To the original part of people saying they are building gods, it’s because to the common user there will be little difference between the experience of interacting with them.
English
1
0
0
28