To prominent AI experts such as Geoffrey Hinton, Ilya Sutskever and Chris Olah, it was obvious: Buried somewhere deep within an LLM¡¯s thicket of virtual neurons must lie ¡°a small-scale model of external reality,¡± just as Craik imagined.This reveals a deep disconnect between what AI people think about knowledge and what knowledge really is.
Google DeepMind and OpenAI are betting that with enough ¡°multimodal¡± training data ¡ª like video, 3D simulations, and other input beyond mere text ¡ª a world model will spontaneously congeal within a neural network¡¯s statistical soup.This is not completely impossible in principle, the way that deriving a world model from text is. But my guess is that it will not work without a way for the system being trained to actually interact with the real world as part of the training. Because the product is still going to be an ad-hoc pile of heuristics, just one that is more suited to tasks like deriving a model of New York City's streets that can reroute in the face of street closures.
Behold, this year's remarkable collection of visionaries who looked at the cutting edge of artificial intelligence and thought, "Hold my venture capital." Each nominee has demonstrated an extraordinary commitment to the principle that if something can go catastrophically wrong with AI, it probably will¡ªand they're here to prove it.On Working with Wizards:
The hard thing about this is that the results are good. Very good. I am an expert in the three tasks I gave AI in this post, and I did not see any factual errors in any of these outputs, though there were some minor formatting errors and choices I would have made differently. Of course, I can¡¯t actually tell you if the documents are error-free without checking every detail. Sometimes that takes far less time than doing the work yourself, sometimes it takes a lot more. Sometimes the AI¡¯s work is so sophisticated that you couldn¡¯t check it if you tried. And that suggests another risk we don't talk about enough: every time we hand work to a wizard, we lose a chance to develop our own expertise, to build the very judgment we need to evaluate the wizard's work.posted by TheophileEscargot at 12:02 PM on September 14 [4 favorites]
But I come back to the inescapable point that the results are good, at least in these cases. They are what I would expect from a graduate student working for a couple hours (or more, in the case of the re-analysis of my paper), except I got them in minutes.
This is the issue with wizards: We're getting something magical, but we're also becoming the audience rather than the magician, or even the magician's assistant. In the co-intelligence model, we guided, corrected, and collaborated. Increasingly, we prompt, wait, and verify¡ if we can.
« Older The Latest Roundup | How woodpeckers provide the heartbeat of forests... Newer »
The part that really caught my eye was this passage: 'To prominent AI experts such as Geoffrey Hinton, Ilya Sutskever and Chris Olah, it was obvious: Buried somewhere deep within an LLM¡¯s thicket of virtual neurons must lie ¡°a small-scale model of external reality,¡± just as Craik imagined. The truth, at least so far as we know, is less impressive. Instead of world models, today¡¯s generative AIs appear to learn ¡°bags of heuristics¡±: scores of disconnected rules of thumb that can approximate responses to specific scenarios, but don¡¯t cohere into a consistent whole. (Some may actually contradict each other.)'
I simply do not think there's any difference between what your brain is doing, and what the LLM is doing when it appears to have a model of Othello lurking inside it. It shouldn't be surprising at all, when the entire point is to develop a network of relationships around pieces of words--words that we have written down because they express meanings about our own world models. It shouldn't be surprising that it's there, shouldn't be surprising that it's incomplete and contradictory. These heuristics are very similar to our own, with the only difference being that we have (a) a lot more modalities around which to build a model, and (b) a large chunk of our brain devoted to coordinating those modalities so they fit together into a worldview that doesn't get us eaten by lions. (Obvs that's not all our brains are up to with modeling, don't get me wrong on that.)
I better leave it there, or else I'll start pulling up all the links you guys have provided on modeling and embodied cognition and then I'm gonna be here all day!
posted by mittens at 5:54 AM on September 14 [4 favorites]