TBPN@tbpn
.@mcuban says he's not worried about AI doom scenarios related to robotics because AI "doesn't understand the consequences of its recommendations":
"Right now, LLMs are basically bimodal: it’s almost all text and pictures, with some video. You can’t model the world with that."
"AI, right now, doesn't understand the consequences of its recommendations. It has no idea what happens next. A 2-year-old in a highchair and a sippy cup knows if it pushes the sippy cup off the highchair, mom's coming running, and the kid's going to start laughing at mom. LLMs don't understand [that]."
"So we have to evolve to models that can capture the world and physics, and deal with the latency of not having access to video that you can't see — you have to try to model that. And not only does that take up a lot of processing power, but it takes up a lot of bandwidth."
"And so the Terminators taking over, I just don’t see how it’s going to happen."