[HN Gopher] 2022 Letter ___________________________________________________________________ 2022 Letter Author : oli5679 Score : 63 points Date : 2022-12-30 18:18 UTC (4 hours ago) (HTM) web link (zhengdongwang.com) (TXT) w3m dump (zhengdongwang.com) | mdorazio wrote: | Do any mainstream AGI researchers believe that GPT-style ML | methods will plausibly get us to AGI? I have only a very shallow | understanding of the state of the art, but from the outside | playing with the tools it seems much more likely that we'll get | to a local maximum far below AGI and require a different approach | altogether. I'd love to read some discussion on the topic if | anyone has good non-hype/biased fanboy links to share. | 0x008 wrote: | Actually they don't. You can listen to the podcasts by Lex | Fridman with Yan Le Cun or Andrej Karpathy regarding that | topic. But basically what Le Cun is saying is that the | information density in text is not high enough in order to | learn a realistic representation of the world from it. | [deleted] | readonlybarbie wrote: | As a professional programmer and a relatively optimistic AGI | enthusiast, why would the current ML methods not work, given | sufficient CPU/RAM/GPU/latency/bandwidth/storage? | | In theory, as long as you can translate your inputs and outputs | to an array of floats, a neural network can compute _anything_. | The required number of neurons might not fit into the world 's | best RAM, and the required number of weights and biases for | those neurons might not be quickly calculated by a CPU/GPU | however. | tmoertel wrote: | One big gap is causal learning. A true general intelligence | will have to learn how to intervene in the real world to | cause wanted outcomes in novel scenarios. Most current ML | models capture only stastical knowledge. They can tell you | what interventions have been associated with wanted outcomes | in the past. In some situations, replaying these associations | seems like genuine causal knowledge, but in novel scenarios | this falls short. Even in current day models designed to make | causal inferences, say for autonomous driving, the causal | structure is more likely to have been built into the models | by humans, rather than inferred from observations. | azinman2 wrote: | Yes but that doesn't mean you won't need new architectures or | training methods to get there, or data that doesn't currently | exist. We also don't know how many neurons / layers we'd | need, etc. | | The brain itself is infinitely more complex than artificial | neural networks. Maybe we don't need all of what nature does | to get there, but we are so many orders of magnitude off its | redonk. People talk about number of neurons of the brain as | if there's a 1:1 mapping with an ANN. Real neurons have | chemical, physical properties, along with other things | probably not yet discovered going on. | readonlybarbie wrote: | This is an interesting comment. I agree that I hear the | "all we need is 86 billion neurons and we will habe parity | with the human brain", and I feel it is dubious to think | this way because there is no reason why this arbitrary | number _must_ work. | | I also think it is a bit strange to use the human brain as | an analogy because biological neurons supposedly are | booleans and act in groups to achieve float level behavior. | For example I can have neurologic pain in my fingers that | isn't on off, but rather, has differences in magnitude. | | I think we should move away from the biology comparisons | and just seek to understand if "more neurons = more better" | is true, and if it is, how do we shove more into RAM and | handle the exploding compute complexity. | jiggawatts wrote: | The current AI approach is like a pure function in | programming: no side effects, and given the same input you | always get the same output. The "usage" and "training" steps | are seperate. There is no episodic memory, especially there | is no short term memory. | | Biological networks that result in conscious "minds" have a | ton of loops and are constantly learning. You can essentially | cut yourself off from the outside world in something like a | sensory deprivation bath and your mind will continue to | operate, _talking to itself_. | | No current popular and successful AI/ML approach can do | anything like this. | readonlybarbie wrote: | Agreed, but I also wonder if this is a "necessary" | requirement. A robot, perhaps pretrained in a highly | accurate 3d physics virtual simulation, which has an | understanding of how it can move itself and others in the | world, and how to accomplish text defined tasks, is already | extremely useful and much more general than an image | classificiation system. It is so general, in fact, that it | would begin reliably replacing jobs. | jimbokun wrote: | But it's not AGI. | readonlybarbie wrote: | Ok, so now we just have to define "AGI" then. A robot, | which knows its physical capabilities, which can see the | world around it through a frustrum and identifies objects | by position, velocity, rotation, which understands the | passage of time and can predict future positions for | example, which can take text input and translate that | into a list of steps it needs to execute, which is | functionally equivalent to an Amazon warehouse employee, | we are saying is not AGI. | | What is an AGI then? | phphphphp wrote: | An Amazon warehouse worker isn't a human, an Amazon | warehouse worker is a human engaged in an activity that | utilises a tiny portion of what that human is capable of. | | A Roomba is not AGI because it can do what a cleaner | does. | | "Artificial general intelligence (AGI) is the ability of | an intelligent agent to understand or learn any | intellectual task that a human being can." | readonlybarbie wrote: | I think the key word in that quote is "any" intellectual | task. I don't think we are far from solving all of the | mobility and vision-related tasks. | | I am more concerned though if the definition includes | things like philosophy and emotion. These things can be | quantified, like for example with AI that plays poker and | can calculate the aggressiveness (range of potential | hands) of the humans at the table rather than just the | pure isolated strength of their hand. But it seems like a | very hard thing to generally quantify, and as a result a | hard thing to measure and program for. | | It sounds like different people will just have different | definitions of AGI, which is different from "can this | thing do the task _i_ need it to do (for profit, for fun, | etc) " | dinkumthinkum wrote: | Well, I would put it back like why would it? When you | understand how these things work, does it sound anything like | what humans do? When prompted with a question, we do not | respond by predicting words that come next based on a | gigantic corpus of pre-trained text. As a professional | programmer, do you think Human intelligence works like a | Turing machine? | isthisthingon99 wrote: | Did mainstream AGI researchers predict something like GPT would | exist 15 years ago? I would listen to those who did on their | opinion. | [deleted] | abecedarius wrote: | There's a 2018 book of interviews of many well-known | researchers where they're asked about future prospects: | http://book.mfordfuture.com/ (list of interviewees on that | page). The actual interview dates weren't specified but don't | seem to be earlier than 2017, in my reading. Almost all of | them greatly underestimated progress up to now, or refused to | say much. (I'm hedging a little bit because it's a year since | I read it, and memory is fuzzy.) | | Shane Legg of DeepMind wrote a blog post at the opening of | the 2010s where he stuck his neck out to predict AGI with a | time distribution peaking around 2030. He thought the major | development would be in reinforcement learning, rather than | the self-supervised GPT stuff. | isthisthingon99 wrote: | Sounds about right. | hiddencost wrote: | I don't think anyone serious thinks or talks in terms of AGI. | The feverishly simplistic idea of the singularity is quite | silly. | | Most notably, neural networks alone will not reach any kind of | AGI. | | Start adding the capacity to read from massive knowledge | stores, and a place to keep long term information (i.e., | memory, probably also in a database), plus a feedback loop for | the model to learn and improve? Plus the ability to call APIs? | Now you're talking. I think all of those pieces are close to | doable right now, maybe with a latency of 5s. If one of the big | players puts that in place in a way that is well measured and | they can iterate on, I think we'll start to see some really | incredible advances. | localhost wrote: | The gpt_index project looks very promising in this area. | | "At its core, GPT Index is about: | | 1. loading in external data (@NotionHQ, @Slack, .txt, etc.) | 2. Building indices over that data 3. Inputting a prompt -> | getting an output!" | | https://twitter.com/jerryjliu0/status/1608632335695745024 | amelius wrote: | Interesting. How are these indexes stored and how are they | fed into the transformer model so that GPT can use them? | Does this require an additional training step? | hiddencost wrote: | One really exciting place things might improve is in data | cleaning. Right now preprocessing your data and putting it in | a format that can be learned efficiently and without bias is | a huge pain // risk. This next generation is allowing us to | largely ignore a lot of that work. | | Similarly, transfer learning is finally good. | | And the models are generalist, few shot learners. | | As a consequence, individuals with minimal expertise can set | up a world class system to solve niche problems. That's | really exciting and it's going to get easier. | hiddencost wrote: | Cross-language learning (what was referred to as an | "interlingua" in the 90s) means we're seeing some stunning | advances in low resource languages. It used to be that | everyone ignored languages other than English, and then | provides them with mediocre support. | | I think we're at a point where there's very little excuse | not to launch in many languages at once. | ripe wrote: | Sort of the opposite of what you want, but Gary Marcus says we | need to cross a few hurdles first: | | http://rebooting.ai/ | alsodumb wrote: | I am a PhD student working in learning and autonomy space and | every researcher I know thinks Gary Marcus is a joke. I'm not | saying he doesn't know things, but all I am saying is machine | learning at scale is not his area of expertise although he | pretends it is. Period. He passes on very generic, obvious | statements about the future without any details and when | someone does something in that direction he claims 'I told | you so!, you should have listened to me in the past!'. Look | at the entire chain of discussion between Gary Marcus and | Yann LeCun in this thread you'll get a sense of I am talking | about: https://twitter.com/ylecun/status/1523305857420824576 | | Gary Marcus is an academic grifter and to me he is no | different than crypto bros who grift non-experts. | hiddencost wrote: | Seconding reports that Gary Marcus is almost as big a waste | of your time as Jurgen Schmidhuber. | | Marcus has been writing some variant of exactly the same | article multiple times a year for the last 15 years. | Symmetry wrote: | We still seem to be missing an equivalent of explicit memory | formation, serializing digested perceptions into working then | short term and long term memory. The however many thousand | tokens in a GPT's buffer can span a much larger span of time | than the second's worth of sense impressions your brain can | hold without consciousness[1] and memory getting involved but | the principle seems to be the same. | | This isn't to say that there wouldn't be some simple hack to | allow memory formation in chat agents, just that there's at | least one advance we need besides simple scale. | | [1] As in not subliminal, not anything to do with philosophical | notions of qualia. | mikepurvis wrote: | I'm completely a bystander, but I feel like one flag for me | with current approaches is the ongoing separation between | training and runtime. Robotics has been through a similar thing | where you have one program that does SLAM while you teleop the | robot, and you use that to map your environment, then afterward | shut it down and pass the static map into a separate | localization + navigation stack. | | Just as robots have had to graduate to the world of continuous | SLAM, navigating while building and constantly updating a map, | I feel like there's a big missing piece in current AI for a | system that can simultaneously act and learn, that can reflect | on gaps in its own knowledge, and express curiosity in order to | facilitate learning-- that can ask a question out of a desire | to know rather than as a party trick. | ben_w wrote: | I think that depends on which definition of AGI you prefer. It | knows more than I do about most topics (though I still beat it | at the stuff I'm best at), so I'd say it's got the A and the G | covered, and it's "only" at the level of a university student | at the stuff I've seen it tested on, so it's "pretty | intelligent" in the way toy drone submarines are "pretty good | swimmers". | | It's not as fast a learner (efficient with samples) as humans | are; it doesn't continuously learn from interactions with | others like we do; and it's certainly not superhuman at | everything (or probably anything other than _breadth_ of | knowledge)... | | ...but Yudkowsky recently criticised Musk for taking that too | far and limiting the definition of "AGI" to normality meant by | "ASI" (Yudkowsky was also saying that no, GPT isn't AGI): | https://twitter.com/ESYudkowsky/status/1600362288149856256?c... | fmajid wrote: | Yes, Dan Wang's letters are outstanding. ___________________________________________________________________ (page generated 2022-12-30 23:00 UTC)