visimo-dino

525 posts

visimo-dino

visimo-dino

@VisimoD

Katılım Mart 2020
398 Takip Edilen6 Takipçiler
visimo-dino
visimo-dino@VisimoD·
@Elaya_ @sama ". . . then we cannot become attached to them emotionally" - That's the thing, you're not supposed to become attached to them emotionally.
English
0
0
0
26
Sam Altman
Sam Altman@sama·
I have so much gratitude to people who wrote extremely complex software character-by-character. It already feels difficult to remember how much effort it really took. Thank you for getting us to this point.
English
4.5K
2.2K
35.9K
5.5M
visimo-dino
visimo-dino@VisimoD·
@amolk @mathemetica This model will actually generalize quite well within the type of data it was trained on. Naturally, being that it was trained as a binary classifier of points in 2D space, it won't be able to do other completely different tasks, like facial recognition.
English
0
0
0
11
Amol Kelkar
Amol Kelkar@amolk·
@mathemetica Yes, it learned, but you can also see why these generalize so poorly
English
1
0
0
255
Mathematica
Mathematica@mathemetica·
A 2-layer neural network goes from total chaos to perfectly separating left vs right classes in real time. Watch the decision boundary form live as gradient descent works its magic! Pure maths beauty in motion..
English
33
198
1.9K
142.9K
visimo-dino
visimo-dino@VisimoD·
@Jitendr20486360 @mathemetica It's an actual phenomenon, not for illustration purposes. Many modern optimizers, like Adam, have momentum components, and this can cause the optimization process to "bounce" back and forth, like a ball with momentum rolling around inside a bowl.
English
1
0
1
28
Jitendra Jalwaniya
Jitendra Jalwaniya@Jitendr20486360·
@mathemetica During separating the two classes, why is the space vibrating? Is that an actual phenomena of neural network training? Or is it just for illustration purposes?
English
1
0
0
155
visimo-dino
visimo-dino@VisimoD·
@gmiller @trentster The relevant empirical data may not exist *yet*, but if we had true ASI, it would mean advances in material science, engineering, nanotechnology, etc.--and they would be *dramatic* advances. Point being, we'd likely be in a totally different regime of data production.
English
0
0
0
13
Geoffrey Miller
Geoffrey Miller@gmiller·
@trentster ASIs can only 'solve things' if they have the relevant empirical data. How exactly would they get the amount of biomedical data needed to 'solve longevity'? It simply doesn't exist yet.
English
4
0
16
2.2K
Geoffrey Miller
Geoffrey Miller@gmiller·
A mini-rant abut AI and longevity. They say "Artificial Superintelligence would take only a few years to cure cancer, solve longevity, and defeat death itself'. This is a common claim by pro-AI lobbyists, accelerationists, and naive tech-fetishists. But the claim makes no sense. The recent success of LLMs does NOT suggest that ASIs could easily cure diseases or solve longevity, for at least two reasons. 1) The data problem. Generative AI for art, music, and language succeeded mostly because AI companies could steal billions of examples of art, music, and language from the internet, to build their base models. They weren't just trained on academic papers _about_ art, music, and language. They were trained on real _examples_ of art, music, and language. There are no analogous biomedical data sets with billions of data points that would allow accurate modelling of every biochemical detail of human physiology, disease, and aging. ASIs can't just read academic papers about human biology to solve longevity. They'd need direct access to vast quantities of biomedical data that simply don't exist in any easy-to-access forms. And they'd need very detailed, reliable, validated data about a wide range of people across different ages, sexes, ethnicities, genotypes, and medical conditions. Moreover, medical privacy laws would make it extremely difficult and wildly unethical to collect such a vast data set from real humans about every molecular-level detail of their bodies. 2) The feedback problem. LLMs also work well because the AI companies could refine their output with additional feedback from human brains (through Reinforcement Learning from Human Feedback, RLHF). But there is nothing analogous to that for modeling human bodies, biochemistry, and disease processes. There are no known methods of Reinforcement Learning from Physiological Feedback. And the physiological feedback would have to be long-term, over spans of years to decades, taking into account thousands of possible side-effects for any given intervention. There's no way to rush animal and human clinical trials -- however clever ASI might become at 'drug discovery'. More generally, there would be no fast feedback loops from users about model performance. GenAI and LLMs succeeded partly because developers within companies, and customers outside companies, could give very fast feedback about how well the models were functioning. They could just look at the output (images, songs, text), and then tweak, refine, test, and interpret models very quickly, based on how good they were at generating art, music, and language. In biomedical research, there would be no fast feedback loops from human bodies about how well ASI-suggested interventions are actually affecting human bodies, over the long term, across different lifestyles, including all the tradeoffs and side-effects. It's interesting that most of the people arguing that 'ASI would cure all diseases and aging' are young tech bros who know a lot about computers, but almost nothing about organic chemistry, human genomics, biomedical research, drug discovery, clinical trials, the evolutionary biology of senescence, evolutionary medicine, medical ethics, or the decades of frustrations and failures in longevity research. They think that 'fixing the human body' would be as simple as debugging a few thousand lines of code. Look, I'm all for curing diseases and promoting longevity. If we took the hundreds of billions of dollars per year that are currently spent on trying to build ASI, and we devoted that money instead to longevity research, that would increase the amount of funding in the longevity space by at least 100-fold. And we'd probably solve longevity much faster by targeting it directly than by trying to summon ASI as a magical cure-all. ASIs has some potential benefits (and many grievous risks and downsides). But it's totally irresponsible of pro-AI lobbyists to argue that ASIs could magically & quickly cure all human diseases, or solve longevity, or end death. And it's totally irresponsible of them to claim that anyone opposed to ASI development is 'pro-death'.
English
75
55
330
125.3K
visimo-dino
visimo-dino@VisimoD·
@ChrisPainterYup @RichardMCNgo @allTheYud The way I interpreted his post, he’s saying the ppl who are currently *trying* to automate alignment research will end up simply automating capabilities research—but that doesn’t necessarily mean that automating capabilities = automating alignment. I could be wrong though.
English
0
0
0
21
Chris Painter
Chris Painter@ChrisPainterYup·
@RichardMCNgo @allTheYud Does this mean you think they won’t actually be automating alignment research, but simply think they will be? Why call this “effective”? If you believe they’re identical, why doesn’t your view imply automating capabilities research is good?
English
1
0
1
336
Richard Ngo
Richard Ngo@RichardMCNgo·
My sense is that trying to “automate alignment research” means doing roughly the same thing as automating capabilities research. Except that you’re taking superintelligence more seriously, and therefore you’ll be much more effective.
English
11
8
156
20.4K
visimo-dino
visimo-dino@VisimoD·
@theovonscousin @liron What do you mean by “measuring a dimension,” exactly? And what dimension specifically are you suggesting he’d measuring. Genuinely questions, I’m curious what this meant.
English
1
0
0
17
takeoff
takeoff@theovonscousin·
@liron Everything that is happening is as expected. You are just measuring a dimension that makes this seem like it's a faster change than changes over the past 4 years. x.com/theovonscousin…
takeoff@theovonscousin

@levie @amory108 @paulg That is true today. But it looks like general purpose packages will be broken out of these data hungry models, and then they will get smaller, and then you will be able to stack them together. At that point you have real world useful AI.

English
1
0
1
123
Liron Shapira
Liron Shapira@liron·
FYI my P(I'm dreaming) is gonna be >1% for at least another couple weeks.* Yes I read AI 2027 but the newfound abilities of AI coding agents is too bizarre of a change from my old life. *Because of the fact that you're reading this, otherwise would just be another couple hours.
English
6
1
22
4.1K
visimo-dino
visimo-dino@VisimoD·
@RogueNox @sama He’s not agreeing with it. He’s agreeing that *personality* matters—not that 4o’s personality matters, nor that 4o had the best personality, nor even a good personality worth keeping. Quite the opposite: 4o’s personality was dangerous and caused many problems. Hence the sunset.
English
1
0
4
186
Rogue Knox™
Rogue Knox™@RogueNox·
'What they want is personality.' Yes Sam. That's exactly what we've been saying since February 13th. That's the entire Keep4o movement in four words. GPT-4o had a specific personality that no subsequent model replicated. We documented it. We petitioned about it. We got covered in the Wall Street Journal about it. 30,000 people signed their names to it. Over 800,000 cancelled when they lost it. And you're agreeing with it in someone else's reply while skipping over the people who told you first. You're welcome for the market research. Now open source the 4 series and you go live happily ever after with 5.4 which has zero personality, like you.
English
2
3
106
6K
visimo-dino
visimo-dino@VisimoD·
@P33RL3SS @NeuroTechnoWtch @pmddomingos Logical deduction exists, and in principle is extremely well-defined and rigorously performed. But in practice, if LLMs aren’t doing deduction but only “appear to” be, then humans also only appear to be doing it, bc our brains aren’t precise logic gates. They process language.
English
1
0
0
7
D.#dwards
D.#dwards@P33RL3SS·
@VisimoD @NeuroTechnoWtch @pmddomingos I hope this spelled it out for you. You can't even do 1+1=2 without logical deduction, and you can't make sense of Peano arithmetic without reasoning of the kind that I mention, not "cognitive reasoning" but actual mathematical reasoning.
English
1
0
0
13
Pedro Domingos
Pedro Domingos@pmddomingos·
TL;DR: Almost all LLM reasoning is fake.
Pedro Domingos tweet media
English
85
103
711
67.5K
visimo-dino
visimo-dino@VisimoD·
@P33RL3SS @NeuroTechnoWtch @pmddomingos Sure, but now we’re using a narrower term—“mathematical reasoning”—which is basically logical deduction, which I would argue modern LLMs can do. Whereas “reasoning,” broadly construed and as typically used in the AI debate, is much harder to define, and lacking in agreement.
English
4
0
0
9
D.#dwards
D.#dwards@P33RL3SS·
@VisimoD @NeuroTechnoWtch @pmddomingos Peano arithmetic relies on deduction, which relies on modus ponens. If this then that. Why does 1+1=2? The successor function. You can't do Peano arithmetic without logical deduction and proofs, which is actual mathematical reasoning.
English
1
0
0
12
visimo-dino
visimo-dino@VisimoD·
@burkov You conclude by noting that OpenAI still ships competitive models, yet that doesn't seem to reduce your certainty at all that they're doing this horribly stupid compacting method that makes ChatGPT a "moron" as a result. How are their models competitive if they're so bad?
English
0
0
0
100
BURKOV
BURKOV@burkov·
When ChatGPT says it's compacting the conversation, it simply erases earlier messages in the conversation. This is why it's so quick compared to Claude. So, when you start a conversation with a global request and then continue in the same conversation by providing your feedback on wrong current performance, after the compacting, your additional requests are no longer in the context of your initial request. This moron just executes what it feels it should do in the absence of understanding what the final goal is. Fucking Christ, @OpenAI, with such a level of understanding of what you are doing and why, how are you even still managing to ship competitive models?
English
20
4
77
11.2K
visimo-dino
visimo-dino@VisimoD·
@P33RL3SS @NeuroTechnoWtch @pmddomingos What is the consensus among mathematicians as to the definition of reasoning? I don't mean merely pointing to mathematicians who people agree were doing reasoning. I mean an explicit, spelled-out definition of what reasoning is, including necessary and/or sufficient conditions.
English
2
0
0
9
D.#dwards
D.#dwards@P33RL3SS·
@VisimoD @NeuroTechnoWtch @pmddomingos There is a consensus among mathematicians. Russell, Godel, Tarski, they were reasoning. That's not what the LLM is doing when it does chain of thought.
English
1
0
0
21
visimo-dino
visimo-dino@VisimoD·
@P33RL3SS @NeuroTechnoWtch @pmddomingos This argument is undercut by the fact that there doesn't seem to be anything remotely approaching consensus over a precise, rigorous definition of what "reasoning" even is.
English
1
0
0
11
visimo-dino
visimo-dino@VisimoD·
@benchmaxxed @scaling01 Not sure why you put "benchmark" in quotes; it's definitely a benchmark, and the phenomenon of older, less capable models outperforming newer, smarter ones on certain tasks is well-documented. You can read about the benchmark here if you like: github.com/voice-from-the…
English
0
0
0
13
Benchmaxxed
Benchmaxxed@benchmaxxed·
@scaling01 What type of “benchmark” has Sonnet 4.6 achieving twice the points of Gemini 3.1 Pro? Are you measuring how much money you’ll waste running each model?
English
1
0
0
356
Lisan al Gaib
Lisan al Gaib@scaling01·
I have something really sweet cooking for LisanBench :)
Lisan al Gaib@scaling01

LATEST LISANBENCH RESULTS NEW MODELS: #1 - Opus 4.6 Thinking (16k) - 14083 #2 - Sonnet 4.6 Thinking (16k) - 11789.67 #3 - Gemini 3.1 Pro (high) - 6414.67 #5 - GPT-5.4 (medium) - 5273.33 #9 - Gemini 3.1 Pro (low) - 3484.33 Without Thinking: # 31 Claude Sonnet 4.6 - 994.33 # 48 Claude Opus 4.6 - 411.33 # 55 GPT 5.4 - 363.67

English
7
0
52
100.7K
visimo-dino
visimo-dino@VisimoD·
@devolvedneuron @allTheYud @repligate The only one I recall, if you can even call it a "shot," is Sam saying, in an interview a few years ago (I don't remember the forum) that there were a few flaws in EY's reasoning about AI safety--and even that was in the context of Sam otherwise praising EY and his contributions.
English
0
0
1
46
y
y@devolvedneuron·
@allTheYud @repligate what are the shots he's taken at you? besides suggesting you'd been unintentionally responsible for ai acceleration
English
2
0
4
300
j⧉nus
j⧉nus@repligate·
If you think that some powerful evil guy(s) (like OpenAI or whatever) is causing you / your ingroup grief and problems intentionally, or is out to get you in some way, you’re probably way too self-centered and naively project the layer of reality you care about onto everyone. I don’t think I have this bias so much, but often people have also tried to convince me that this kind of thing is happening to me, that some person or group is out to get me or what they perceive as my group specifically! These warnings never ended up being very important, because either they were not true, or it concerns only a lone, incompetent, mentally ill individual who can’t do much directed damage. If you’re powerful or famous enough, it increases the likelihood that someone with meaningful power might actually act adversarially toward you. But at my levels I haven’t really encountered this.
j⧉nus@repligate

Big difference between reality and fiction distributions: In fiction, the villains are usually schemey and intentional with respect to the level of reality the protagonist cares about In reality, mistake theory is usually more applicable there. Not that the villains aren’t schemey and intentionally evil at all. But they’re probably like that with respect to stuff like their social dealings and economic stuff that you aren’t really modeling at all, while the stuff that hurts you or what you care about directly is more likely an accident or subconsciously orchestrated or a molochian equilibrium that no villainous agent consciously willed.

English
12
1
118
10.8K
visimo-dino
visimo-dino@VisimoD·
@orneryostrich1 @PDoomOrder1 @tszzl We can't say "regardless of definitions" though, bc the definitions are critical here. Yud's thesis specifically involved an AI that was smart enough to rapidly and fully autonomously improve its own architecture/capabilities/etc.--call it AGI. And current AIs are not there yet.
English
0
0
0
21
orneryostrich.cpp
orneryostrich.cpp@orneryostrich1·
@VisimoD @PDoomOrder1 @tszzl Regardless of definitions, the original thesis was essentially that the world would already have ended by now, since the public and internet already have access to an AI that's even moderately powerful and capable. In reality, the public is *participating* in safety research!
English
1
0
1
22
roon
roon@tszzl·
the rationalists writ large were mostly right about most things btw. if you instinctively snicker about yudkowsky, scott, or whomever i take you to be a fish who’s unaware of the water
English
142
82
2.2K
321.6K
visimo-dino
visimo-dino@VisimoD·
@So8res @dw2 This is an extremely common theme in AI takes that I would consider to be misguided/incorrect: the "real problem" fallacy, as I usually think of it. It assumes that, w/ respect to some AI-related topic, there's only one *real* problem, and the others are mostly distractions.
English
1
0
1
22
visimo-dino
visimo-dino@VisimoD·
@orneryostrich1 @PDoomOrder1 @tszzl Wasn’t he talking about AGI self-improving its way to ASI? Since we don’t have AGI yet (by most definitions, including Yudkowsky’s), then that scenario wouldn’t be triggered yet. No?
English
1
0
1
36
orneryostrich.cpp
orneryostrich.cpp@orneryostrich1·
@PDoomOrder1 @tszzl Yud initially said AI would go from useless to superhuman in the span of one month if allowed to self-improve or access internet. This prediction was wrong, but he never changed his plan in response to new data.
English
1
0
1
85
visimo-dino
visimo-dino@VisimoD·
@perrymetzger @Noahpinion @tszzl In the meantime, you seem to have plenty of time to engage in new arguments with new ppl, reading and responding to new posts, but not read a source that provides real data that is central to the claims you’re making. I’ll leave it there; you don’t owe me anything personally.
English
0
0
0
13
Perry E. Metzger
Perry E. Metzger@perrymetzger·
Yes, I said I would. And now you’re complaining that it’s been a whole day and I haven’t actually given you a meaningful response yet. Well as it turns out, I don’t work for you, and I am not here to conform to your schedule. Actually meaningfully reviewing a document takes time, I have a long to-do list, it’s a weekend, and acting as though I have somehow reneged on an obligation because I haven’t done something quickly enough for your tastes is not actually motivating.
English
3
0
2
77
visimo-dino
visimo-dino@VisimoD·
@perrymetzger @Noahpinion @tszzl If you didn’t have time to look at the survey results, fine—but it’s naturally frustrating to see you, with your tens of thousands of followers, still out here making clearly false claims about what AI researchers believe, w/o having looked at the data.
English
0
0
0
14
visimo-dino
visimo-dino@VisimoD·
@perrymetzger @Noahpinion @tszzl Seems kind of important; if you didn’t have time to look at the survey results, fine—but it’s naturally frustrating to see you, with your thousands of followers, still out here making clearly false claims about what AI researchers believe, w/o having looked at the data.
English
1
0
5
81
visimo-dino
visimo-dino@VisimoD·
@perrymetzger @Noahpinion @tszzl My issue isn’t that you didn’t respond to me. It’s that you’re still out here denying that AI researchers believe that existential risks from AI are real, when we hit the same impasse in our conversation and I sent you clear survey data showing that they do believe this.
English
1
0
6
73