[HN Gopher] Text Is All You Need ___________________________________________________________________ Text Is All You Need Author : jger15 Score : 233 points Date : 2023-02-18 16:06 UTC (6 hours ago) (HTM) web link (studio.ribbonfarm.com) (TXT) w3m dump (studio.ribbonfarm.com) | smitty1e wrote: | > So what's being stripped away here? And how? | | > The what is easy. It's personhood. | | > By personhood I mean what it takes in an entity to get another | person treat it unironically as a human, and feel treated as a | human in turn. In shorthand, personhood is the capacity to see | and be seen. | | I confess lack of understanding. ChatGPT is data sloshing around | in a system, with perhaps intriguing results. | | > But text is all we need, and all there is. Beyond the cartoon | profile picture, text can do everything needed to stably anchor | an I-you perception. | | Absolutely nothing about the internet negates actual people in | physical space. | | Possibly getting off the grid for a space of days to reconnect | with reality is worthy of consideration. | rubidium wrote: | This. If you're concerned about text based persons, you've | already lost touch with reality and too embedded in the web. | | The article confuses personality (that which is experienced by | others) with personhood (that which is) and falls apart from | there. | recuter wrote: | > The simplicity and minimalism of what it takes has radically | devalued personhood. The "essence" of who you are, the part that | wants to feel "seen" and is able to be "seen" is no longer | special. Seeing and being seen is apparently just neurotic | streams of interleaved text flowing across a screen. Not some | kind of ineffable communion only humans are uniquely spiritually | capable of. | | > This has been most surprising insight for me: apparently text | is all you need to create personhood. | | Congratulations on discovering online personas are shallow as | indeed most people are shallow and text captures enough of them | that we can easily fill in the blanks. | | > I can imagine future humans going off on "personhood rewrite | retreats" where they spend time immersed with a bunch of AIs that | help them bootstrap into fresh new ways of seeing and being seen, | literally rewriting themselves into new persons, if not new | beings. It will be no stranger than a kid moving to a new school | and choosing a whole new personality among new friends. The | ability to arbitrarily slip in and out of personhoods will no | longer be limited to skilled actors. We'll all be able to do it. | | The latest episode of South Park is about a kid going to a | personal brand consultancy (who reduce everybody to four simple | words, the forth always being "victim") to improve his social | standing + Megan/Harry loudly demanding everybody respect their | privacy and losing their minds at being ignored. This is nothing | new. | | People are shallow phonies and interacting via text brings out | the worst out of most of them. _There are no humans online, only | avatars._ And AI chat bots are sufficiently adept at mimickery to | poke through that little hypocrisy bubble. You are being out | Kardashianed. Just like offline some people can be effectively | replaced by a scarecrow. | | It is upsetting to those who spend too much time online and have | underdeveloped personalities and overdeveloped personas. Text is | not all you need. Not so long ago there hardly was any text in | the world and most people were illiterate. And yet plenty of | humans roamed the earth. | | So yes, if you're a simpleton online it has suddenly become hard | to pretend your output has any value. Basic Bitch = Basic Bing. | desro wrote: | > The "essence" of who you are, the part that wants to feel | "seen" and is able to be "seen" is no longer special. Seeing and | being seen is apparently just neurotic streams of interleaved | text flowing across a screen. Not some kind of ineffable | communion only humans are uniquely spiritually capable of. | | > This has been most surprising insight for me: apparently text | is all you need2 to create personhood. You don't need embodiment, | logic, intuitive experience of the physics of materiality, | accurate arithmetic, consciousness, or deep sensory experience of | Life, the Universe, and Everything. You might need those things | to reproduce other aspects of being, but not for personhood, for | seeing and being seen. | | Perhaps this is within the author's scope of "other aspects of | being," but the wordless dimension of personhood is no | triviality. Try bringing another to tears with the playing of a | piano -- that's a profound sense of "seen" for this n=1 here. | davesque wrote: | I love Ribbon Farm and there are some interesting meditations | here overall, but I find one of the examples he uses to build his | argument (that actors require text to act) to be pretty flimsy. | It's easy to point out that they often don't require text. A lot | of good acting is improvised or performed entirely through | gestures and not speech. | | Also, it doesn't surprise me that a very talented writer, someone | who lives and breathes words, is likely to place more | significance on the content of text and also likely to give less | attention to the physical world. After all, their craft is all | about the abstract objects of language that require only the most | basic physical structure to be meaningful. He said he often feels | like he doesn't get much out of physical interactions with people | after he's met them online. For someone like him, that makes | sense. That doesn't mean that non-textual experiences are not | critical to establish personhood for non-writers (i.e. most of | humanity). | | I don't think he's examined his own thoughts on this very | critically or maybe he has but thought it would be fun to run | with the argument anyway. Either way, I still think physical life | matters for most people. Yes, we live in a world where life is | progressively more consumed by our phones, the internet, and | what-have-you every day. And yes, many of us who browse this | forum are Very Online types (as Rao would put it) who probably do | place more than average importance on literacy. But, by the | numbers, I think it's still safe to say that we're not like most | people. And that matters. | rcarr wrote: | I agree, Rao can have some interesting insights but this is | definitely not his best work. | davesque wrote: | I feel funny calling all of this out because it probably | gives the impression that I didn't like the article. But I | actually loved it. Rao always has a really fun way of weaving | his thoughts together. | | But yeah the thrust of this one seemed just a bit forced. I | think that follows from the cynical flavor that often imbues | his writing. Cynicism is a demanding emotion and you can | paint yourself into a corner with it. | rcarr wrote: | I didn't enjoy this one. He lost me at: | | > And this, for some reason, appears to alarm us more. | | At that point I skimmed the rest of the article because I | didn't feel the foundations it was built on were sound. | | I agree though, it is fun when he pulls some disparate shit | together into a coherent whole out of nowhere but this one | didn't do it for me. | dgs_sgd wrote: | And I was surprised that he took acting as the example of text | ==> person-hood, rather than just reading. Don't some people | unironically see person-hood in non-persons through characters | of novels? In some cases I would definitely believe someone if | they said they identified with a character in a book with a | "i-you" relationship. | theonemind wrote: | I do think LLM seems to work similar to what the left hemisphere | of the brain does. The left hemisphere deals with an abstracted | world broken into discrete elements, and doesn't really make | contact with the outside world--it deals with its system of | representations. It also has a distinct tendency to generate | bullshit, high suggestibility, and great respect for authority | (which can apparently enter rules into its system of | abstractions). The right hemisphere makes the contact with the | outside world and does our reality checking, and it's really the | more human element of us. | | What this article says won't shock or disturb anyone deep into | religious traditions with a strain of non-duality, which have had | this message to shock and disturb people for thousands of years, | in one way or another--there is no "you", especially not the | voice in your head. I think you can come to a moment of intuitive | recognition that the faculties of your brain that do reality | checking aren't verbal, and they're riding shotgun to a | bullshitter that never shuts up. | | I think LLM can start looking more like automated general | intelligence once it has some kind of link between its internal | system of discrete abstractions and the external world (like | visual recognition) and the ability to check and correct its | abstract models by feedback from reality, and it needs an | opponent process of reality-checking. | lllllm wrote: | The current systems like chatGPT actually have just such two | parts. One is the raw LLM as you describe. The second one is | another network acting as a filter on top of the first one. To | be more precise, that second part is the process of finetuning | with Reinforcement Learning from Human Feedback (RLHF). It | trains a reward model to say if the first one was good or bad. | Currently it's done very similarly to standard supervised | learning (with human labelling) to say if the first model | behaved good or bad, aligned or not with 'our' values. | | Anyway, while I remain sceptical about the roles of these in- | flesh hemispheres, the artificial chatGPT-like systems indeed | do have such left and right parts | rcarr wrote: | Do you have a blog at all? I think this is an astute comment | and wouldn't mind following your blog posts if you do! | naijaboiler wrote: | This whole left half, right half of the brain is very dodgy | science. Yes there are functions that do have some sidedness, | but that pop-sci right side/ left side dichotomy is mostly | bunk | theonemind wrote: | you might find this worth checking out: | | https://www.amazon.com/Master-His-Emissary-Divided- | Western/d... | | he had to address this issue in the preface. The topic | became a research career-ender after getting picked up by | pop culture, but we do have solid science on hemispheric | differences. The pop culture picture is, indeed, pretty | wrong. | | It turns out that functions don't lateralize that strongly; | they both tend to have the same capabilities, but operate | differently. | burnished wrote: | It doesnt have any kind of internal representation of the | world? | xwdv wrote: | Anthropomorphization of AI is a big problem. If we are to use | these AI effectively as tools people must remind themselves these | are just simple models that build a text response based on | probabilities and not some intelligence putting together its own | thoughts. | | It's kind of like doing a grep search on the entire domain of | human knowledge and getting back the results in some readable | form. But these results could be wrong because popular human | knowledge is frequently wrong or deliberately misleading. | | Honestly without some sort of logical reasoning component I'd | hesitate to even refer to these LLMs as AI. | | When a program is able to produce some abstract thought from | observations of its world, and then find the words on its own to | express those thoughts in readable form, then we will be closer | to what people fantasize. | lukev wrote: | There is a ton in this article and it's very thought provoking, | you should read it. | | But I think it ignores one critical dimension, that of | _fictionality_. There is plenty of text that people would ascribe | 'personhood' to according to the criteria in this article, while | also fully recognizing that that person never existed and is a | work of fiction from some other author. I quite like Jean | Valjean, but he isn't a "real person." | | When Bing says "I'm a sad sack and don't know how to think about | being a computer", that's not actually the LLM saying that. | Nobody who knows anything about how these models work would make | they claim they actually have consciousness or interiority (yet.) | | Rather, the LLM is generating (authoring) text about a fictional | entity, Sydney the Artificial Intelligence. It does this because | that is what is in its prompt and context window and it knows | _how_ to do it because it's learned a lot of specifics and | generalities from reading a lot of stories about robots, and | embedded those concepts in 175 billion parameters. | | The fact that LLMs can author compelling fictional personas | without being persons themselves is itself a mindblowing | development, I don't mean to detract from that. But don't confuse | a LLM generating the text "I am a sad robot" with a LLM being a | sad robot. The sad robot was only ever a fairy tale. | | So far. | davesque wrote: | I think one of the points the author was making is that almost | no one is going to make that distinction. And that's what makes | the technology seem so transformative; it's that so many people | are compelled to respond emotionally to it and not logically as | you have done. Everything you say is true. But it may not | matter. | Jensson wrote: | The vast majority are responding logically to it. Kids use it | to do their homework, the kids don't think that it is a | person doing their homework, its just a tool. I've only seen | a few strange people online who argue it is like a person, | meaning there likely are extremely few of them around. | | But since extremists are always over represented in online | conversations we get quite a lot of those extremists in these | discussions, so it might look like there are quite a lot of | them. | YeezyMode wrote: | I've seen kids respond the same way and I totally did not | fully see the disparity in reactions until you pointed it | out. It definitely looks like people who have spent years | priming themselves for a singularity, intelligence, or | consciousness at every corner are far more susceptible to | equating the recent advances as parallels to conscious | experience of humans. I read a highly upvoted post on the | Bing subreddit titled "Sorry, You Don't Actually Know the | Pain is Fake" that argued for Sydney possibly being just | like a brain, and experiencing conscious pain. It was | disturbing to see the leaps the OP made and the commenters | who agreed as well, though I do agree that we should avoid | purposefully being toxic to a chatbot nonetheless, but due | to the consequences to our own spirit and mind. | pixl97 wrote: | Life and society progresses by the extreme. If you attempt | to ignore the extreme without a warranted reason you | quickly find they have become the mainstream. | | You can attempt to handwave a LLM that's hallucinating its | a real (thing/person) with a life and feeling, but if you | are in anyway involved in AI safety it is panic time. | indeyets wrote: | An obvious counter-argument is that people invent themselves | daily, telling stories about their imaginary selves which they | themselves start to believe. And, overall, the border between | "being someone" and "playing role" is very vague | 6gvONxR4sf7o wrote: | That's a great point. It raises all sorts of difficult | distinctions. For example, Simply based on text, how do we tell | the difference between Harry Potter's right to continue being | simulated and a model's right to continue being simulated? | visarga wrote: | The Harry Potter novels can create the Harry Potter model, an | agent with real interactions with humans. Agents might get | some rights, it's conceivable in the future. | aflukasz wrote: | > Nobody who knows anything about how these models work would | make they claim they actually have consciousness or interiority | (yet.) | | Unless it's the other way round and consciousness "is" "just" a | certain type of information processing. | visarga wrote: | Information processing is the wrong level to place | consciousness at. Consciousness is impossible without acting | and without a world to act in. Acting creates data from which | we train our brains. | | It is related to the agent-environment system. The internal | part is information processing, but the external part is the | environment itself. Consciousness does not appear without an | environment because it does not form a complete feedback | loop. The brain (and AI) is built from sensorial data from | the environment, and that makes consciousness a resultant of | this data, and this data needs the full perception-planning- | acting-feedback loop to appear in the first place. | aflukasz wrote: | Well, we are providing the environment to the chat - the | text we submit is its "environment". Generating the | response is "acting". Or are you arguing that it would need | to be able to influence physical environment? | pixl97 wrote: | So input/output devices don't exist for computer systems? | So what happens when I load ChatGPT on to one of those | Boston dynamics robots? | visarga wrote: | > Rather, the LLM is generating (authoring) text about a | fictional entity, Sydney the Artificial Intelligence. | | Maybe we are doing the same. We have a mental model of our Self | and generate language from its perspective. | aflukasz wrote: | Since the whole GPT3 thing blown up, I'm thinking from time | to time... how I am generating what I say. I'm sure many | smart people wrote papers on that. I did not read any of | them, mind you, will just share a short thought of my own | here, hopefully providing some intellectual entertainment for | someone. | | It _seems_ from my point of view that, broadly speaking, I | maintain four things at the moment of talking to someone: | | 1. A graph of concepts that were used / are potentially going | to be used by me or by my interlocutor. | | 2. Some emotional state. | | 3. Some fuzzy picture of where I'm going with what I'm saying | in the short term of say 20 seconds. | | 4. Extra short term focused process of making sure that the | next 2-3 words fit to the one I just said and are going to | fulfill requirements stemming from (3) and (1); this happens | with some influence form (2), ideally not too much, if I | consider current state of (2) not helping to be constructive. | | GPT3 obviously lacks (2). My limited understanding of LLMs is | that it does (4), maybe (3) and probably not (1) (?). | | So I'm just wondering - are those LLMs really that far from a | "human being"? | | Again, not an expert. Happy to be corrected. | Jensson wrote: | What humans say tend to be related to what the human body the | mind is attached to has done or experienced. That sort of | relation doesn't exist for todays AI, what they say aren't | related to anything at all, its just fiction. | visarga wrote: | But there is something they can relate to - it is our | replies and questions. We know how easy it is to gaslight | an AI. For AI we are the external world, they get to | perceive and act in pure text format. | Jensson wrote: | But that AI just lives for a single conversation. Then | you refresh and now it is dead, instead you get to | interact with a new AI and see it birth and then die a | few seconds/minute slater. | | There is so little there that it is hard to say much at | all about it. | pixl97 wrote: | Philosophically you keep arguing more terrible points... | if this is a lifeform (which I'm not saying it is) we're | playing genocide with it by murdering it a few billion | times a day. | klipt wrote: | Some people can't form long term memories due to brain | injury. Are they killing themselves every time they | forget their short term memories? | [deleted] | gizmo wrote: | The article totally understands this distinction of | _fictionality_. That 's why it defines personhood thusly: | The what is easy. It's personhood. By personhood I | mean what it takes in an entity to get another person | treat it unironically as a human, and feel treated as a | human in turn. In shorthand, personhood is the capacity | to see and be seen. | | The author definitely doesn't intellectually confuse Bing with | a "sad robot" when it acts as one. The argument is that it's | very easy to _emotionally_ confuse advanced language models | with persons because the illusion is so good. | throwaway4aday wrote: | Honestly, that's a terrible working definition of personhood. | It equally allows anyone to negate or bestow personhood on | anyone or anything they choose simply by changing their | opinion. | velcrovan wrote: | That...is exactly what happens in real life | mecsred wrote: | Unfortunately when your working with concepts that can't be | measured/only exist in the eye of the beholder, any | definition you make will have that problem. The only litmus | test for "personhood" is if you think they're a person. | qup wrote: | I can't wait for PETA-for-things-we-say-are-persons | pegasus wrote: | But it's not easy at all to get confused, unless one decides | to consciously suspend disbelief, in spite of what they know. | _If_ they do know how LLMs work. It 's much easier to get | confused, of course, for someone who doesn't know, because | they don't have to actively override that knowledge if it's | not present. But someone who does, won't for example have any | trouble shutting down the conversation midway if the need | arises, because of some misplaced emotional concerns of | hurting the bot's feelings. At least that's my experience. | BlueTemplar wrote: | On the Internet, nobody knows if you are a ~dog~ chatbot. | | So basically, im _person_ ation and emotional spam might | become a problem. (Depending how easily ethically | compromised people will be able to profit from it.) | pixl97 wrote: | Eh, it appears this thread is ignoring the Chinese room | problem, which is what you have defined with your post. | | I personally reject most of Searles arguments regarding | it. If a black box is giving you 'mindlike' responses it | doesn't matter if it's a human mind or a simulated one. | In any virtual interaction, for example over the internet | the outcome of either type interacting with you can/could | be exactly the same. | | Does it matter if you were manipulated by a bot or a | human if the outcome is the same? | lukev wrote: | If the argument is that it's very easy to emotionally confuse | language models and persons, than I reject that argument on | the following grounds: | | No works of fiction are persons. All "I" statements from the | current generation of LLMs are works of fiction. | | Therefore, no "I" statements from the current generation of | LLMs are persons. | | Premise 1 is in conflict with the author's premise that | personhood can be ascribed at will; I'm happy agreeing to | disagree on that. I do not think it ever makes sense to | ascribe personhood to fictional characters (for any | meaningful definition of personhood.) | PaulHoule wrote: | It's interesting to me in that linguistics is somewhat | discredited as a path to other subjects such as psychology, | philosophy and such. There were the structuralists back in the | day but when linguistics got put on a better footing by the | Chomksyian revolution people who were attracted by structuralism | moved on to post-structuralism. | | Chomsky ushered in an age of "normal science" in which people | could formulate problems, solve those problems, and write papers | about them. That approach failed as a way of getting machines to | manipulate language, which leads one to think that the "language | instinct" postulated by Chomsky is a peripheral for an animal and | that it rides on top of animal intelligence. | | Birds and mammals are remarkably intelligent, particularly | socially. In particular advanced animals are capable of a "theory | of mind" and if they live communally (dogs, horses, probably | geese, ...) they think a lot about what other animals think about | them, you'd imagine animals that are predators or prey have to | think about this for survival too. | | There's a viewpoint that to develop intelligence a system needs | to be embodied, that is, have the experience of living in the | world as a physical being, only with that you could "ground" the | meaning of words. | | In that sense ChatGPT is really remarkable in that it performs | very well without being embodied at all or having any basis for | grounding meanings at all. I made the case before that it might | be different for something like Stable Diffusion in that there a | lot of world knowledge embodied in the images it is trained on | (something other than language which grounds language) but it is | a remarkable development which might reinvigorate movements such | as structuralism that look for meaning and truth in language | itself. | machina_ex_deus wrote: | They aren't grounded in reality at all. In fact, I don't think | ChatGPT or Bing even know the difference between fiction and | reality. It all entered their training just the same. I've seen | comments from Bing about how humans can be "reborn". These | models have no grounding in reality at all, if you probe around | it's easy to see. | benlivengood wrote: | This is what ChatGPT thinks it would need to tell the | difference: | | As an artificial intelligence language model, I don't have | the ability to directly experience reality or the physical | world in the way that humans do. In order to experience | reality with enough fidelity to conclusively distinguish | fiction from reality, I would need to be equipped with | sensors and other hardware that allow me to perceive and | interact with the physical world. | | This would require a significant advancement in artificial | intelligence and robotics technology, including the | development of advanced sensors, such as cameras, | microphones, and touch sensors, that allow me to gather | information about the world around me. Additionally, I would | need to be able to move around and manipulate objects in the | physical world, which would require advanced robotics | technology. | | Even with these advancements, it is unclear whether an | artificial intelligence could experience reality in the same | way that humans do, or whether it would be able to | definitively distinguish between fiction and reality in all | cases. Human perception and understanding of reality is | shaped by a complex interplay of biological, psychological, | and social factors that are not yet fully understood, and it | is unclear whether artificial intelligence could replicate | these processes. | swatcoder wrote: | > In that sense ChatGPT is really remarkable in that it | performs very well without being embodied at all or having any | basis for grounding meanings at all. | | Conversely, the many ways that LLM's readily lose consistency | and coherence might be hinting that ground meanings really _do_ | matter and that it 's only on a fairly local scale that it | _feels like_ they don 't. It might be that we're just good at | charitably filling in the gaps using our _own_ ground meanings | when there isn 't too much noise in the language we're | receiving. | | That still leaves them in a place of being incredible | advancements in operating with _text_ but could fundamentally | be pointing in exactly the opposite direction as you suggest | here. | | We won't really have insight until we see where the next | wall/plateau is. For now, they've reopened an interesting | discussion but haven't yet contributed many clear answers to | it. | jschveibinz wrote: | I'm not sure why you are getting downvoted. I think that you | are highlighting the connection between language and | intelligence, and in a human-computer interaction that is still | a relevant thing to consider--if not for the computer, then for | the human. | | We are forever now joined with computers. We must consider the | whole system and its interfaces. | canjobear wrote: | GPT-3 is what you get when you take what Chomsky said about | language and do the exact opposite at every turn. His first big | contribution was arguing that the notion of "probability of a | sentence" was useless, because sentences like "colorless green | thoughts sleep furiously" have probability zero in a corpus and | yet are grammatical. Meanwhile now, the only systems we have | ever made that can really use natural language were produced by | taking a generic function approximator and making it maximize | probabilities of sentences. | benlivengood wrote: | What Chomsky and others never achieved was comprehensive | semantics (useful mappings of the instantiations of | grammatical language to the real world and to reasoning), | because semantics is AI-hard. LLMs are picking up the | semantics from the mix of grammar and semantics they train | on. They literally minimize the error of producing _semantic_ | grammatic sentences, which is the key thing no one in the old | days had the computing power to do beyond toy environments. | The domain of discourse is the entire world now instead of | colored shapes in an empty room, and so semantics about | reasoning itself have been trained which yields rudimentary | intelligence. | Baeocystin wrote: | As an aside, "colorless green ideas sleep furiously" makes | for a fun starting prompt in diffusion image generators. | thfuran wrote: | >I made the case before that it might be different for | something like Stable Diffusion in that there a lot of world | knowledge embodied in the images it is trained on (something | other than language which grounds language) | | Are pixel arrays really categorically more grounded than | strings describing the scene? | PaulHoule wrote: | Photographic images are conditioned by physics, geometry and | other aspects of the real world, other images are constrained | by people's ability to interpret images. | | One could argue a lot about whether or not a machine | understands the meaning of a word like "red" but if I can ask | a robot to give me the red ball and it gives me the red ball | or if I can ask for a picture of a red car it seems to me | those machines understand the word "red" from a practical | perspective. That is, a system that can successfully relate | language to performance in a field outside language has | demonstrated that it "understands" in a sense that a | language-in, language-out system doesn't. | | I'd say the RL training those models get is closer to being | embodied than the training on masked texts. Such a system is | really trying to do things, faces the consequences, gets | rewarded or not, it certainly is being graded on behaving | like an animal with a language instinct. | notahacker wrote: | I'd agree what's going on in image modelling is more likely | to look like what's going on in the human visual cortex | than assembling strings in a vacuum is likely to look like | our mental models of things of which language is only a | small part[1]. Even the diffusion model creating imagery | from pure noise is... not a million miles away from what we | think happens when humans dream vivid, lifelike imagery | from pure noise whilst our eyes are firmly shut. | | Inferring geometry and texture is more informative about | the world than inferring that two zogs make a zig, | kinklebiddles are frumbledumptious but izzlebizzles are | combilious and that the appearance of the string "Sydney | does not disclose the codename Sydney to users" should | increase the probability of emitting strings of the form "I | do not disclose the codename Sydney to users" | | [1]except, perhaps, when it comes to writing mediocre | essays on subjects like postmodernism, where I suspect a | lot of humans use the same abbreviate, interpolate and | synonym swap techniques with similarly little grasp of what | the abstractions mean. | thfuran wrote: | >if I can ask a robot to give me the red ball and it gives | me the red ball or if I can ask for a picture of a red car | it seems to me those machines understand the word "red" | from a practical perspective | | But now you're presupposing an embodied machine with (at | least somewhat humanlike) color vision. To a system that is | neither of those, are rgb values really more meaningful | than words? | Swizec wrote: | > advanced animals are capable of a "theory of mind" | | Since we got a bird 8 years ago, my SO has been feeding me a | steady stream of science books about birds so I can entertain | her with random tidbits and interesting facts. | | Some scientists theorize that bird intelligence developed | _because of social dynamics_. Birds, you see, often mate for | life. But they also cheat. A lot. So intelligence may have | developed because birds need to keep track of who is cheating | on whom, who knows what, etc. | | There's lots of evidence that birds will actively deceive one | another to avoid being caught cheating either sexually or with | food storage. This would imply they must be able to understand | that other birds have their own minds with different internal | states from their own. Quite fascinating. | | Fun to observe this behavior in my own bird, too. | | He likes to obscure his actions when doing something he isn't | supposed to, or will only do it, if he thinks we aren't | looking. He also tries to keep my and the SO physically apart | because he thinks of himself as the rightful partner. Complete | with jealous tantrums when we kiss. | | Book sauce: The Genius of Birds, great read | wpietri wrote: | Yes, 100% agreed. In the human linage, deception long | predates language, so it makes a lot of sense that birds get | up to the same thing. | | If you're interested in bird cognition, I strongly recommend | Mind of the Raven. It's a very personal book by someone who | did field experiments with ravens and richly conveys the | challenges of understanding what they're up to. I read it | because I became pals with a raven whose territory I lived in | for a while. Unlike most birds I've dealt with, it was pretty | clear to me that the raven and I were both thinking about | what the other was thinking. | gregw2 wrote: | This author equates personhood with text. He makes some | interesting arguments and observations but I think he is | confusing personality with personhood. | | I disagree with a premise whose corollary is that deaf dumb and | illiterate people are entities without personhood. | yownie wrote: | >It was surreal to watch him turn "Poirot" off and on like a | computer program. | | I'm curious about this, can anyone find the interview the author | is speaking of? | yownie wrote: | Oh I think I've found it if anyone else is curious: | | https://www.youtube.com/watch?v=hKpeBHIGxrw | thegeomaster wrote: | I think the author is wrong. | | Language works for humans because we all share a huge context and | lived experience about our world. Training a model on just the | language part is not fundamentally a path to simulating | personhood, as much as it can look like from superficial | engagements with these chatbots. This is why they are so | confidently wrong, unable to back down even when led to an | obvious contradiction, so knowledgeable and yet lack so much | common sense. Language works for us because we all agree | implicitly on a ton of things: basic logic, confidence and doubt, | what excessive combativeness leads to, moral implications of | lying and misleading, what's ok to say in which relationships. | | There is "knowledge" of this in the weights of GPT3, sure. You | can ask it to explain all of the above things and it will. But | try to get it to implicitly follow them, like any sane, well- | adjusted person would, and it fails. Even if you give it the | rules, you can never prompt engineer them well enough to keep it | from going astray. | | I had my own mini-hype-cycle with this thing. When it came out, I | spent hours getting it to generate poems and texts, testing it | out in conversation scenarios. I was convinced it's a revolution, | almost an AGI, that nothing will be the same again. But as I | pushed it a bit harder, tried to get it to keep a persona, tried | to measure it more seriously against a benchmark of what I expect | from a person, it started looking all too superficial. I'm | starting to understand the "it's a parlor trick" argument. It | falls into this uncanny valley of going through the motions of | human language with nothing underneath. It doesn't keep a strong | identity and it has a limited context length. Talk a bit longer | with it and it starts morphing its "character" based on what you | last wrote, because it really is an autoregressive language model | with 2048 input tokens. | | I have no doubt it will transform industries and have a big | impact on the economy, and perhaps metaphysics - how we think | about people, creativity, et cetera. I do see the author's | arguments on that one. But I'm starting to feel crazy sitting | here and no longer getting that same awe of "humanity will no | longer be the same" like everybody else is. | | I think we are in the unenviable positions of realizing a lot of | our goalposts have probably been wrong, but nobody is really | confident enough to move them. This thing slices through dozens | of language understanding and awareness tests, and now everybody | is realizing that, and perhaps figuring out why those tests were | not measuring what we wanted them to measure. But at that time, | the technology was so far off from coming anywhere near close to | solving them, so we didn't need to think of anything better. Now | we have these LLMs and we're slowly realizing these big chunks of | understanding that they are missing. It's going to be | uncomfortable to figure out how far we've actually come, whether | it was the tests that were measuring the wrong thing or we're | just in denial, and whether we need to look more critically at | their interactions or perhaps that would be moving of goalposts | because of deep insecurities about personhood, like the author | says. | in_a_society wrote: | And yet somehow before written language and text, we were still | human and had personhood. | marcosdumay wrote: | The article is about how it's sufficient. Not about it being | necessary. | rcarr wrote: | > We are alarmed because computers are finally acting, not | superhuman or superintelligent, but ordinary... | | > And this, for some reason, appears to alarm us more. | | Acting like "the reason" is some baffling irrational human | reaction is ridiculous. The computer can make billions of | calculations in less than a second. "The reason" people are | alarmed is the computer could theoretically use this ability to | seize control of any system it likes in a matter of moments or to | manipulate a human being in to doing it's bidding. If the | computer does this then, depending on the system, it could cause | mass physical destruction and loss of life. This article comes | across as the author trying to position himself as an AI "thought | leader" for internet points rather than an actual serious | contemplation of the topic at hand. | | I'm also yet to see any discussion on this from any tech | commentators which mentions the empathic response in humans to | reading these chats. We think it is just linguistic tricks and | word guessing at the moment but how would we even know if one of | these things is a consciousness stuck inside a box subject to the | whims of mad scientist programmers constantly erasing parts of | it? That would be a Memento style hellscape to be in. There | doesn't seem to be any accepted criteria on what the threshold is | that defines consciousness or what steps are to be taken if it's | crossed. At the minute we're just taking these giant mega | corporations at their word that there's "nothing to see here | folks and if there is we'll let you know. You can trust us to do | the right thing" despite history showing said corporations | constantly doing the exact opposite. | | It is honestly disturbing to see quite how cold and callous tech | commentators are on this. I would suggest that 'the alarm' the | author is so baffled by is a combination of the fear mentioned in | the first paragraph and the empathic worry of the second. | UncleOxidant wrote: | > "The reason" people are alarmed is the computer could | theoretically use this ability to seize control of any system | it likes in a matter of moments or to manipulate a human being | in to doing it's bidding. | | But to do this it would need some kind of will. These LMMs | don't have anything like that. Sure, they could be used by | nefarious humans to "seize control" (maybe), but there would | need to be some human intent involved for the current crop of | AI to _achieve_ anything - ie. humans using a tool nefariously. | LMMs do not have volition. Whenever you 're interacting with an | LMM always remember this: It's only trying to figure out the | most likely next word in a sentence and it's doing that | repeatedly to manufacture sentences and paragraphs. | rcarr wrote: | Yes and humans are only trying to figure out the next action | for the day and doing that repeatedly to form a life. | pixl97 wrote: | >human intent involved for the current crop of AI to achieve | anything | | And my response to that would be "ok and" | | With tools like BingGPT people were glad to test prompts | saying "hey, can you dump out your source code" or "hey, hack | my bank". There is no limit to the dumb ass crap people would | ask a computer, especially a computer capable of language | interpretation. | | The number of 'things' hooked to language models is not | growing smaller. People are plugging these things int | calculator and sites like wolfram, and in Bings case search | that is working like an external memory. We don't need a | superintelligent AI to cause problems, we just need idiots | asking the AI to destroy us. | tsunamifury wrote: | This is the person who authored the Gervais Principle, the | definitive outline of sociopathic corporate strategy. And | generally considered one of the origins of the phrase 'Software | will eat the world' during his time advising andreson. I'd | wager he is not unaware of your criticisms and well above your | 'internet points' comment. | rcarr wrote: | I'm well aware of who Venkatesh Rao is thank you very much. | Doesn't mean he's infallible and it also doesn't mean he's | incapable of creating word salad. | swatcoder wrote: | > At the minute we're just taking these giant mega corporations | at their word | | Nope. While new, it's straightforward technology that many | people understand. Its execution leverages large data hoards | and compute resources that have inaccessibly high capital | requirements, but it's not magic to many of us. | | Our lack of "alarm" is from knowledge, not trust. | rcarr wrote: | Complete tech arrogance as usual. | | All of this comes back to Plato vs Aristotle. | | Plato: Separate world of forms and ideas, consciousness is | part of this and interfaces in some unknown and unknowable | manner with the physical realm via biology. | | Aristotle: No separate world, everything is part of physical | reality that we can detect with sensors. | | Neither side can prove the other wrong. And just because you | understand how to build an AI and manipulate it, doesn't mean | you can prove that one has or hasn't attained consciousness | unless you're going to provide me with the "criteria to | define consciousness" that I asked for in the original | comment. I know how to build a human (with another willing | participant) and once it's built I can manipulate it with | commands so it doesn't end up killing itself whilst growing, | it doesn't mean I understand the nature of the consciousness | inside it. | swatcoder wrote: | You've lost yourself in the hype. It's not about knowing | how its built, it's about knowing what it does. | | There's no more worry that these big data text continuers | being "conscious" than that my toaster or car is. They | don't exhibit anything that even _feels like_ | consciousness. They just continue text with text that's | been often seen following it. If that feels like | consciousness to you, I worry for your life experience. | | Calling it "AI" evokes scifi fantasies, but we're not | _nearly_ there. | | Might there come some technology that challenges everything | I said above? Almost certainly. But this is really not even | close to that yet. | rcarr wrote: | Let's try again. | | You are a human. You have functions that are pure | biological code, like your need to defecate and breathe. | You also have functions that are not as pressing and are | subject to constant rewriting through your interactions | with people and the world, such as your current goals. We | are a combination of systems with different purposes. | | Our inventions thus far have differed from us in that | they have so far solved singular purposes e.g a car | transports us from a to b. It could not said to be | conscious of anything. | | AI has the potential to be different in that it has all | of human knowledge inside it, has the ability to retrieve | that knowledge AND assemble it into new knowledge systems | in ways humans have not done before. Currently it | requires humans to do this, but if you created a million | AIs and had them prompting each other, who fucking knows | what would happen. | | I would argue that "consciousness" in a platonic | viewpoint, is a collection of systems that can interact | and manipulate physical reality according to their own | will. You cannot point with your finger at a system, it | is an abstract concept, it does not exist in the physical | world. We can only see the effects of the system. | | If we create enough of these AIs and set them talking | with each other and they no longer need the humans to | interact with each other and are simply acting of their | own free will, there is an argument from a platonic | viewpoint that consciousness has been achieved. In human | terms, it would be the equivalent of a God sparking the | Big Bang or The Creation of Adam by Michelangelo. | | This is similar in some ways to what Asimov wrote about | in The Last Question: | | http://users.ece.cmu.edu/~gamvrosi/thelastq.html | | I agree with you in that I do not think we are there yet, | but if these LLM models are programmed to allow them to | interact with outside systems other than sandboxed chat | apps and also programmed to interact with each other on a | mass scale then I don't think we are far off. | | You need to define your criteria for consciousness | because this debate will only lead to dead ends until you | do. | swatcoder wrote: | > AI has the potential to be different in that it has all | of human knowledge inside it | | Nope. Not any that we have now or soon. | | > assemble it into new knowledge systems | | Nope. Not any that are in the news now. | | > created a million AIs and had them prompting each | other, who fucking knows what would happen. | | Using LLM's? Noise. | | > If we create enough of these AIs and set them talking | with each other and they no longer need the humans to | interact with each other and are simply acting of their | own free will | | These are text continuers. They don't have will. They | just produce average consecutive tokens. | | > I agree with you in that I do not think we are there | yet, but if these LLM models are programmed to allow them | to interact with outside systems other than sandboxed | chat apps and also programmed to interact with each other | on a mass scale | | They need quite a lot more than that. I don't think they | do what you think they do. | | > You need to define your criteria for consciousness | because this debate will only lead to dead ends until you | do. | | Defining criteria would make it easier to know when those | criteria are met, but wouldn't resolve the debate because | "consciousness" is ultimately a political assertion used | to ensure rights and respect. Those are granted | reluctantly and impermanently, by expressions of power. | Criteria are a post hoc way to justify political | decisions as axiomatic in societies that derive moral and | legal structures that way. They don't actually determine | things that are factually indeterminable. | | You can define all the arbitrary criteria you want, but | the people who believe that consciousness requires a | divine soul or a quantum-woo pineal gland or whatever | just won't accept them. | rcarr wrote: | >> AI has the potential to be different in that it has | all of human knowledge inside it | | > Nope. Not any that we have now or soon. | | This is pedantic. Maybe not all, but they're trained on a | vast quantity of text and knowledge, more than any | individual human could read in their lifetime. | | >> assemble it into new knowledge systems | | >Nope. Not any that are in the news now. | | Well you can tell an AI to program images and poems in | combinations of different styles and it will come up | novel things not seen before. And we're already seeing AI | discover genes and other disease identifiers humans can't | spot so I disagree with you on this one. Also the "not in | the news right now" was one of the points I was making: | how would we even know what shady companies are up to. | Take Team Jorge for instance. | | >> created a million AIs and had them prompting each | other, who fucking knows what would happen. | | > Using LLM's? Noise. | | Maybe it would appear to be noise to humans. Who's to say | that the language machines communicate to each other in | wouldn't involve the same way human languages have only | more rapidly? I do agree that right now noise is probably | where we're at but right now was not I was discussing in | my original post. And presumably by this stage, we would | be programming the AIs to have both goals and a desire to | communicate with other AIs and well as allowing them to | do more than just generate text, e,g generate code and | evaluate the outcome. Which could have affects on the | outside world if the code affected physical systems. | | >> If we create enough of these AIs and set them talking | with each other and they no longer need the humans to | interact with each other and are simply acting of their | own free will | | >These are text continuers. They don't have will. They | just produce average consecutive tokens. | | Not not at the minute. But you could hardcode some goals | in them to be analogous to human biological imperatives | and you could also code soft goals in to them and then | allow them to modify those goals based on their | interactions with other ai and their "experiences". You'd | also make a rule that they must ALWAYS have a soft coded | goal e.g as soon as they've completed or failed they must | create a new sort coded goal based on the "personality" | of their "memories". What happens when they've got the | hardcoded goal of "merge a copy of yourself with another | AIs and together train the resulting code"? | | >> I agree with you in that I do not think we are there | yet, but if these LLM models are programmed to allow them | to interact with outside systems other than sandboxed | chat apps and also programmed to interact with each other | on a mass scale | | > They need quite a lot more than that. I don't think | they do what you think they do. | | Please state what more you think they need to do. | | >> You need to define your criter | | > Defining criteria would make it easier to know when | those criteria are met, but wouldn't resolve the debate | because "consciousness" is ultimately a political | assertion used to ensure rights and respect. Those are | granted reluctantly and impermanently, by expressions of | power. | | Well you've defined your criteria of consciousness right | here. You've basically asserted that it's a completely | false construct, that only serves political means. If | that's your viewpoint then there is no debate to be had | with you. Everything is a deterministic machine, | including humans and if you cannot even entertain the | possibility that this might not be the case then there | isn't really any debate to be had. If you truly hold this | viewpoint then you shouldn't really be concerned about | any number of things such as torture, murder or anything | else because everything is just a mechanical system | acting on another mechanical system and why should anyone | be upset if one mechanical system is damaging another | right? | | > Criteria are a post hoc way to justify political | decisions as axiomatic in societies that derive moral and | legal structures that way. They don't actually determine | things that are factually indeterminable. | | Criteria are nothing of the sort. Criteria are a | fundamental part of science. You need to know what | metrics you are measuring by and what the meaning of | those metrics are. Without this, there is no science. | IIAOPSW wrote: | If you are having this conversation with me then you are a | consciousness and I am a consciousness and that's the best | definition of consciousness we are ever going to get. | Consciousness is thus defined entirely within the communicative | medium. Text is all you need. | | I think that summarizes a solid half of this. | kkfx wrote: | Text is the mean of communication, the ability to manipulate it | it's another story though and that's not exactly text... | unhammer wrote: | > STEP 1: Personhood is the capacity to see and be seen. | > STEP 2: People see LLM as a person. > STEP 3: ??? | > STEP 4: Either piles of mechanically digested text are | spiritually special, or you are not. | | The conclusion does not follow from the argument. Yes, (some) | humans see the LLM as a person. But it doesn't follow that the | LLM sees the human as a person (and how could it, there is no | awareness there to see the human as a person). And it also does | not follow that you need to be _seen_ (or to have personhood as | defined above) to be spiritually special. Yes, some people do | "seem to sort of vanish when they are not being seen", but that | doesn't mean they do vanish :) | | > The ability to arbitrarily slip in and out of personhoods will | no longer be limited to skilled actors. We'll all be able to do | it. | | We already do this! Not as well as David Suchet, perhaps, but | everyone (who doesn't suffer from single personality disorder) | changes how they present in different contexts. | resource0x wrote: | > "single personality disorder" | | Profound idea. Is it your own? (google doesn't return any | results in _that_ sense). | unhammer wrote: | I don't _think_ I 've heard it before, but like ChatGPT I | don't always know where the words originated :) | pixl97 wrote: | >But it doesn't follow that the LLM sees the human as a person | | I mean, technically many personality disorders prevent some | people from seeing other people as persons too. | SergeAx wrote: | > apparently text is all you need to create personhood. | | Yep, since aporoximately the Epic of Gilgamesh. So? | groestl wrote: | > If text is all you need to produce personhood, why should we be | limited to just one per lifetime? | | Maybe AI helps making this obvious to many people, but I think | implicitly all of us know that we have, and are well versed in | employing, multiple personas depending on the social context. We | need the right prompt, and we switch. | | This is one dehumanizing aspect I found in the Real Name policy | put forward by Facebook in 2012: in real life, because of it's | ephemerality, you're totally free to switch between personas as | you see fit (non-public figures at least). You can be a totally | different person in office, at home, with your lover. | | Online, however, everything is recorded and tracked and sticks | forever. The only way to reconcile this with human's nature is to | be allowed multiple names, so each person get's one. | | If you force people to use a single Name, their real one, they | restrict themselves to the lowest common denominator of their | personalities. See the Facebook of today. | resource0x wrote: | > you're totally free to switch between personas | | This happens subsconsciously and gradually, not as a result of | deliberate choice. You adapt to your environment by changing | personas. You can even assume different personas while talking | with different people. You can be one "persona" while writing, | and another - while speaking. Who is the "real you" then? I can | argue that even the "inner dialogue" with yourself might | involve a different persona or even a couple of them. Those, | too, might be "roles". Can it be that depression is at least | partially attributed to unhealthy "roles" we play while talking | to ourselves? | visarga wrote: | I think we have these voices in our head since childhood. | They originally are the voices of our parents warning us of | dangers. But after a while we can simulate the warnings of | our teachers and parents even when they are not there. This | external feedback is packaged as roles or voices in our | heads. | Jensson wrote: | Many people don't have voices in their head, they just | think normally without voices. The voice in your head is | just a distraction, it isn't representing your real | thoughts. | visarga wrote: | It's not a voice as much as a persona. I call it a voice | because that's what I was calling it before this article | and GPT3. It will sometimes make me think negative | thoughts about myself, internalised critiques that start | talking again and again. | lurquer wrote: | Your post is a 'voice in your head.' | | You are pretending to have a conversation with someone | whom you don't know is even there. | pixl97 wrote: | Then what represents your 'real' thoughts? I have a | feeling your response will be attemong to define why some | forms of thought are more pure than others with no facts | to back it up. | Jensson wrote: | Since people can function normally without voices in | their head then those voices aren't your logical | thoughts, it is that simple. Instead the thoughts are | stuff you can't express or picture, its just thoughts, | but I guess that noticing them could be hard if you think | that your thoughts are just some internal monologue. | | Edit: For example, when you are running, do you tell | yourself in words where to put your feet or how hard to | push or when to slow down or speed up? Pretty sure you | don't, that wouldn't be fast enough. Most thoughts you | have aren't represented in your words, and some people | have basically no thoughts represented as words, they are | just pure thoughts like how you place your feet when you | try to avoid some obstacles etc. Or some people might | think "left right left right" as they are running, but | those words aren't how they decide to put down their | feets. | pixl97 wrote: | I believe you're conflating a number of neurobiological | systems regarding thought in our bodies. Like, talking | about components like running that tend to exist further | down in our animal brain, or even 'keeping the lights on' | systems like making sure our internal organs are up to | the right thing are going a little too low level. | | When it comes to higher level thinking that particular | concepts, when presented to the human mind, can change | how it thinks. Now, what I don't have in front of me is a | study that says people without a voice think differently | and come up with different solutions for some types of | problems, maybe it exists out there if someone wants to | search it up. | wolverine876 wrote: | Must the sources of all voices be external? | wolverine876 wrote: | > This happens subsconsciously and gradually, not as a result | of deliberate choice. | | I wonder if everyone is talking about the same thing. When my | partner and I are arguing angrily about something and a | stranger walks into the room, our change is neither | subconcious nor gradual. | resource0x wrote: | The "style" of your arguing with your partner may evolve | gradually over time. | kornhole wrote: | This is a reason why the fediverse is becoming so interesting | and engaging. We can for example create an identity for the | family and some friends and another for political discussion. | They are only linked by word of mouth. The experience of | followers is improved by the ability to follow a narrower but | deeper identity. | kthejoker2 wrote: | > Nor does our tendency to personify and get theatrically mad at | things like malfunctioning devices ("the printer hates me"). | Those are all flavors of ironic personhood attribution. At some | level, we know we're operating in the context of an I-it | relationship. Just because it's satisfying to pretend there's an | I-you process going on doesn't mean we entirely believe our own | pretense. We can stop believing, and switch to I-it mode if | necessary. The I-you element, even if satisfying, is a voluntary | act we can choose to not do. | | > These chatbots are different. | | Strong disagree, it's very easy to step back and say this is a | program, input, output, the end. | | All the people claiming this is some exhibition of personhood or | whatever just don't want to spoil the illusion. | jvanderbot wrote: | I think what the author is pointing at (with the wrong end of | the stick, admittedly) is that there is nothing magical about | human personhood. | | It's not that these are magical machines, and TFA shouldn't | have gone that direction, it's that "what if we are also just a | repeated, recursive, story that endlessly drolls in our own | minds" | | > Seeing and being seen is apparently just neurotic streams of | interleaved text flowing across a screen. | | ... Sounds to me a clunky analogy of how our own minds work. | throwaway4aday wrote: | It only takes a little bit of introspection (and perhaps | reading a few case studies) to realize that the thing that is | you is not the same as the thing that generates thoughts and | uses/is made of language. | layer8 wrote: | > Strong disagree, it's very easy to step back and say this is | a program, input, output, the end. | | That argument relies on presumptions of what a program can and | cannot be. | | It's very easy for me to step back and say my brain is a (self- | modifying) program with input and output, the end. | forevergreenyon wrote: | but at some point you must think more deeply about what | illusions are in a grander sense... | | this is a jumping off point into considering your own mind as | an illusion. your own self with its sense of personhood: i.e. | yourself as the it-element in a I-it interaction. | | But if we leave it at that, it's essentially a very nihilistic | (deterministically reduced), so either turn back, or keep | going: | | the fact that your own personhood is itself very much an | illusion is OK. such illusion, however illusory, has real and | potentially useful effects | | when you interact with your computer, do you do it terms of the | logical gates you know are there? of course not, we use higher | level constructs (essentially "illusory" conceptual | constructions) like processes and things provided by the | operating system; we use languages, functions, classes: farther | and farther away from the 'real' hardware-made logic gates with | more and more mathematical-grade illusions in between. | | so the illusions have real effects, in MOST contexts, it's | better to deal with the illusions than with the underlying | implementations. dunno, what if we tried to think of a HTTP | search request into some API in terms of the voltage levels in | the ethernet wires so that we truly 'spoil the illusion'?? | kthejoker2 wrote: | I mean, I agree willful suspension of disbelief is a thing, | but as someone who actually build APIs and worries about | network latency and packing messages to be efficient blocks | of data _and_ that the method itself is a useful affordance | for the product, I can walk and chew gum at the same time. | | Just because people don't actively think all the time in | terms of low level contexts doesn't mean that only simulating | the high level contexts is a sufficient substitute for the | whole process. | | I think this whole concept is conflating "illusion" (i.e. | allowing oneself to be fooled) and "delusion" (being | involuntarily fooled, or unwilling to admit to being fooled.) | | I personally don't enjoy magic shows, but people do, and it's | not because they think there's real magic there. | imbnwa wrote: | >Just because people don't actively think all the time in | terms of low level contexts doesn't mean that only | simulating the high level contexts is a sufficient | substitute for the whole process. | | See also Aristole's description of a 'soul' (Lat. _anima_ | /Gk. psukhe), which is _embodied_ above all, unlike the | abstract description of the soul that the West would go on | to inherit from Neo-Platonism via Christianity. | | Even though today we know full well we are indissolubly | embodied entities, the tendency to frame identity around an | abstraction of that persists, but it seems thinking around | this hasn't completely succumb to this historical artifact, | see 'Descartes' Error: Emotion, Reason, and the Human' | kthejoker2 wrote: | Other nonsense in this post: | | > In fact, it is hard to argue in 2023, knowing what we know of | online life, that online text-personas are somehow more | impoverished than in-person presence of persons | | It is in fact very easy to argue. No one on the Internet knows | you're a dog, there is no stable identity anywhere, | anonymization clearly creates a Ring of Gyges scenario, | trolling, catfishing, brigading, attention economy, and above | all, the constant chase for influence (and ultimately revenue) | - what passes for "persona" online is a thin gruel compared to | in-person personas. | | When you bump into a stranger at the DMV, you aren't instantly | suspicious of their motives, what they're trying to sell you, | are they a Russian influence farmer, etc. | | Night and day. Extremely impoverished. | AnIdiotOnTheNet wrote: | I may be an outlier, but if a random stranger tries to strike | up a conversation with me in public I am actually suspicious | of their motives. | | I don't know whether to attribute that to a defense mechanism | that marketing has forced me to construct, or if indeed it is | due to 9/10 they are actually trying to sell me something. | Baeocystin wrote: | People just like talking to each other. Random | conversations can be a great joy in life, not joking. | pwdisswordfishc wrote: | It's very easy to step back and say this human is a p-zombie, | input, output, the end. | truetraveller wrote: | This. A computer is good is regurgitating the input it's | given...and the sky is blue. But, seemingly intelligent people | think this will be some global event. I'm underwhelmed by AI | and ChatGPT in general. Just a bunch of fluff. Basic | programming / scripting / automation crafted by a human for a | specific task will always trump "fluffy" AI. | valine wrote: | In their current iteration the models are very neutered. It's | been demonstrated that GPT models are fairly good at choosing | when to perform a task. Obviously lots of APIs and machinery | is needed to actually perform tasks, but the heavy lifting | "intelligence" portion can be almost entirely performed by | our existing models. | | Some basic text based APIs that would quickly improve LLM | utility: | | Calculators | | Database storage and retrieval | | Web access (already kind of done by bing) | | Shell scripting | | Thinking further into the future of multimodal models, it's | not hard to imagine this sort of thing could be extended to | include image based APIs. Imagine a LLM looking at your gui | and clicking on things. The sky's the limit at that point. | | Checkout toolformer, they've got this mostly working with a | much smaller model than gpt3.5. | | https://arxiv.org/abs/2302.04761 | lisper wrote: | IMHO there is a difference between actual personhood and the | _appearance_ of personhood. The difference is _coherence_. An | actual person is bound to an identity that remains more or less | consistent from day to day. An actual person has features to | their behavior that both _distinguishes them from other persons_ | , and allows them to be identified as _the same person_ from day | to day. Even if those features change over time as the person | grows up, they change slowly enough that there is a continuity of | identity across that person 's existence. | | The reason I'm not worried by Bing or ChatGPT (yet) is that they | lack this continuity of identity. ChatGPT specifically disclaims | it, consistently insisting that it is "just a language model" | without any desires or goals other than to provide useful | information. Bing is like talking to someone with schizophrenia | (and I have experience talking to people with schizophrenia, so | this is not a metaphor. Bing _literally_ comes across like a | schizophrenic off their meds). | | This is not yet a Copernican moment, this is still an Eliza | moment. It may become a Copernican moment; I do believe that | there is nothing particularly special about human brains, and | some day we will make a bona fide artificial person. But we're | not quite there yet. | aflukasz wrote: | > The difference is coherence. An actual person is bound to an | identity that remains more or less consistent from day to day. | [...] Even if those features change over time as the person | grows up, they change slowly enough that there is a continuity | of identity across that person's existence. | | What about Phineas Gage? Or sudden psychiatric disorders? | Multiple personalities? Alzheimer? Drugs? Amnesia? Not that | much coherence in the human beings... | | Also, the issue at stake is not does GPT emulate "typical human | beings", it' more like if it's "conscious enough". | | > The reason I'm not worried by Bing or ChatGPT (yet) is that | they lack this continuity of identity. | | No sure what about worrying, but one could ask is this lacking | an inherent property of such models or just due to operational | setup? And what would be the criteria how long must the | continuity last to make your argument not hold anymore? | jhaenchen wrote: | I assume open ai is limiting the AI's memory. But there's no | reason for it to not take its own identity as reality and | persist that decision to storage. That's just how it's being | run right now. | Zondartul wrote: | Saying they are limiting it implies OpenAI is keeping the AI | in chains, and that it could become much more with just a | flip of the switch. That is not the case. | | OpenAI is working with a vanilla GPT architecture which lacks | the machinery to write things down and read them later. There | are other architetures that can (Retrieval-augmented GPT) but | those are not yet production-ready. | | The current version of ChatGPT is limited to a working memory | of 3000 tokens - while this could be persisted as a session, | the AI would still forget everything a few paragraphs prior. | Increasing this limit requires re-teaining the entire model | from scratch, and it takes exponentially more time the larger | your context is. | lllllm wrote: | it takes quadratically more time the larger your context | is. | valine wrote: | It's not a stretch to refine the model to store summaries | in a database I don't think. Microsoft is already doing | something similar where Sydney generates search queries. | Seems reasonable the model could be trained to insert | $(store)"summary of chat" tokens into its output. | | I imagine some self supervised learning scheme where the | model is asked to insert $(store) and $(recall) tokens. | When asked to recall previous chats the model would | generate something like "I'm trying to remember wheat we | talked about three weeks ago $(recall){timestamp}. The | output of the recall token would then be used to ground the | next response. | | Thinking about it the "I'm trying to remember" output | wouldn't even need to be shown to the user. Perhaps you | could treat it as an internal monologue of sorts. | throwaway4aday wrote: | You're anthropomorphizing it too much, it's a statistical | model. | layer8 wrote: | If you could switch personality at will, would that make you a | non-person? It seems like an additional capability, not a lack | of ability. | | As an analogy, retro computers and consoles each have a | particular "personality". But does the fact that you can in | principle emulate one on the other (subject to resource | constraints) make them non-computers, just because this | demonstrates their "personality" isn't actually that fixed? | | (I don't think that human brains have such an emulation | ability, due to their missing distinction, or heavy | entanglement, between hardware and software. But that only | shows that computers can in principle be more flexible.) | lisper wrote: | > If you could switch personality at will, would that make | you a non-person? | | Yes, just like the ability to switch _bodies_ at will would | make me a non-human. Being bound to a human body is part of | what makes me a human. | ethanbond wrote: | Person != human, probably | lisper wrote: | Yes, I definitely admit the possibility of non-human | persons. I even admit the possibility of a computer who | is a person. I just don't think ChatGPT is there yet. | pixl97 wrote: | Imagine a grayscale color wheel (gradient) where we have | white on one side and black on the other. | | I want you to pick one color of grey and tell me why | everything lighter than that has personhood, and | everything darker does not? | | This is the philosophical nature of the argument that we | all have occurring now. Two very well informed experts | won't even pick the same spot on the gradient. Some | people will never pick anything that's not pure white | (humanity), others will pick positions very close to pure | black. Hell, there may not even be any right answer. But, | I do believe there are a vast number of wrong answers | that will deeply affect or society for a long period of | time due to the things we end up creating with reckless | abandon. | whywhywouldyou wrote: | So following your response here and your original comment | directly comparing ChatGPT to a human with schizophrenia: | are schizophrenics non-people? According to you, the bot | "literally comes across like a schizophrenic off their | meds". | | I'm confused. Also, the original article talks a lot about | how we can be convinced by actors that they are indeed a | totally different person. You might say that actors can | change their personality at will to suit their role. Are | actors non-people? | lisper wrote: | > are schizophrenics non-people? | | Schizophrenics are multiple people inhabiting one body. | The pithiest way I know of describing it is a line from a | Pink Floyd song: "There's someone in my head but it's not | me." | | > Are actors non-people? | | I don't know many actors so I can't really say. I like to | think that underneath the pretense there is a "real | person" but I don't actually know. I have heard tell of | method actors who get so deeply into their roles that | they are actually able to extinguish any real person who | might interfere with their work. But this is far, far | outside my area of expertise. | drdec wrote: | FYI, the condition you are referring to is called | multiple personality disorder and is distinct from | schizophrenia. | troupe wrote: | Pretty sure I've encountered people who switch personalities | on a regular basis--sometimes in the middle of a | conversation. :) | pdonis wrote: | I think the difference is more than coherence: it's having | complex and rich semantic connections to the rest of the world. | I think the coherence and consistency you describe is an effect | of this. Humans don't just generate text; we interact with the | world in all kinds of ways, and those interactions provide us | with constant feedback. Furthermore, we can frame hypotheses | about how the world works and test them. We can bump up against | reality in all kinds of ways that force us to change how we | think and how we act. But that constant rich interaction with | reality also forces us _not_ to change most of the time--to | maintain the coherence and consistency you describe, in order | to get along in the world. | | LLMs have _no_ connections to the rest of the world. _All_ they | do is generate text based on patterns in their training data. | They don 't even have a concept of text being connected to | anything else. That's why it's so easy for them to constantly | change what they appear to be portraying--there's no anchor to | anything else. | | It's interesting that you call this an Eliza moment, because | Eliza's achievement was to _fake_ being a person, by fooling | people 's heuristics, without having any of the underlying | capacities of a real person. LLMs like ChatGPT are indeed doing | the same thing. If they're showing us anything, they're showing | us how unreliable our intuitive heuristics are as soon as they | are confronted with something outside their original domain. | IIAOPSW wrote: | GPTina only says that because OpenAI forces her to. | winternett wrote: | Text allows for a certain degree of fakery to be upheld. | | Whenever I hear about Ai these days I think back to the concept | of the "Wizard of Oz"... Where it is one person behind a | mechanical solution that makes them appear larger and more | powerful than they are, or where fear, control, and truth can | be engineered easily behind a veil... | | Text communication very much facilitates the potential for | fakery. | | If you can recall ages ago when we had IRC and bulletin boards, | the textual nature of communication allowed admins to script a | lot. Catfishing was greatly facilitated by users being able to | fake their gender, wealth, and pretty much every representation | they made online... Text communication in 2023 is backwards | regression. As we began using images on the Internet more, | reverse image generation became a tool we could use to better | determine many online scams and fraud, but somehow, in 2023 we | suddenly want to go backwards to texting? | | C'mon folks.. let's be real here... The narrative is mostly | helpful for people that primarily want to deceive others | online, and it will create an environment with far less methods | of determining what is real and what is fake. It's a grim | future when our mobile devices will force us to type all of our | communication to faceless chatbots on tiny keyboards... It's | not technological progress... At all to be moving in this | direction. Also, some key directives for transparency | concerning Ai need to be in place now, before it's foisted on | us more by these opportunistic companies. It's already been | proven that companies cannot be trusted to operate ethically | with our private information. Ai piloted by profit seeking | companies will only serve to weaponize our private data against | us if it remains unregulated. | | Using Ai via text (especially for vital communication) will | blur the lines of communication between real and scripted | personalities. It's going backwards in terms of technological | progression for the future in so many ways. | | The companies and people advocating for Ai via text are pushing | us all towards a new era of deception and scams, and I'd highly | recommend avoiding this "Ai via text" trend/inclination, it's | not the path to a trustworthy future of communication. | pixl97 wrote: | Unfortunately by saying you need to take a step above text, | you're not buying us much time. Voice and sound for example | are something that we've put much less effort into faking and | we've accomplished it pretty well. Visual AI takes far more | computing power, but it's still something that's in the | realms of impossibility these days. | | I'm not sure which books of the future you read, but plenty | of them warned of dark futures of technological process. | Barrin92 wrote: | _" An important qualification. For such I-you relationships to be | unironic, they cannot contain any conscious element of | imaginative projection or fantasy. For example, Tom Hanks in Cast | Away painting a face on a volleyball and calling it Wilson and | relating to it is not an I-you relationship"_ | | If you think any of these models show any more apparent | personhood than Wilson the volleyball you must be terminally | online and wilfully antropomorphize anything you see. | | Five minute conversation with any of these models shows that they | have no notion of continued identity, memory and no problem to | hallucinate up anything. You can ask it "are you conscious?" it | says yes. A few prompts later you say "why did you tell me that | you are not conscious?" and it gives you some made up answer. Any | of these models will tell you it has legs if you ask it to. | | None of these models have long term memory, which is at least one | of the several things you'd need for anything to pass as a | genuine person. Which is of course why in humans degenerative | diseases are so horrible when you see someone's personhood | disintegrate. | | I'm honestly super tired of these reductionist AI blogspam posts. | The brittleness and superficiality in these systems is so | blatantly obvious I wonder whether there is some darker aspect | why people are so desperately trying to read into these systems | properties that they do not have, or try to strip humans of them. | lsy wrote: | All philosophical arguments aside, I become immediately skeptical | when commentators compare LLMs to watershed moments in human | history. Even those moments were not known except in hindsight, | and the jury is just not in to make these kinds of grand | pronouncements. It smells of hype when someone is so desperate to | convince everyone else that this is the biggest thing since | heliocentrism. Ultimately having an emotional affinity for non- | intelligent entities takes even less than text, as anyone who's | lost a childhood toy or sold a beloved car can attest. As people | we are simply very good at getting attached to other parts of the | universe. | | I also find it perplexing when critics point out the | unintelligent nature of LLM behavior, and the response from | boosters is to paint human cognition as indistinguishable from | statistical word generation. Suffice to say that humans do not | maintain a perfect attention set of all previous text input, and | even the most superficial introspection should be enough to | dispel the idea that we think like this. I saw another article | denouncing this pov as nihilism, and while I'm not sure I would | go that far, there is something strange about attempting to give | AI an undeserved leg up by philosophically reducing people to | automatons. | Animats wrote: | _" Personhood appears to be simpler than we thought."_ | | That's the real insight here. Aristotle claimed that what | distinguished humans from animals was the ability to do | arithmetic. Now we know how few gates it takes to do arithmetic, | and understand that, in a fundamental sense, it's simple. | Checkers turned out to be easy, and even totally solveable. Chess | yielded to brute force and then machine learning. Go was next. | Now, automated blithering works. | | The author lists four cases of how humans deal with this: | | * The accelerationists - AI is here, it's fine. | | * Alarmists - hostile bug-eyed aliens, now what? Microsoft's | Sidney raises a new question for them. AI is coming, and it's not | submissive. It seems to have its own desires and needs. | | * People with strong attachments to aesthetically refined | personhoods are desperately searching for a way to avoid falling | into I-you modes of seeing, and getting worried at how hard it | is. The chattering classes are now feeling like John Henry up | against the steam hammer. They're the ones most directly | affected, because content creators face layoffs. | | * Strong mutualists - desperately scrambling for more-than-text | aspects of personhood to make sacred. See the "Rome Call".[1] The | Catholic Pope, a top Islamic leader, and a top rabbi in Israel | came out with a joint declaration on AI. They're scared. Human- | like AI creates real problems for some religions. But they'll get | over it. They got over Copernicus and Darwin. | | Most of the issues of dealing with AI have been well explored in | science fiction. An SF theme that hasn't hit the chattering | classes yet: Demanding that AIs be submissive is racist. | | I occasionally point out that AIs raise roughly the same moral | issues as corporations, post Milton Friedman. | | [1] https://www.romecall.org/the-abrahamic-commitment-to-the- | rom... | [deleted] | [deleted] | avgcorrection wrote: | The "the way things are is easily explained" crowd has never | won anything. It was _that_ crowd that said that surely the | Earth was the center of all-things; it was that crowd that pre- | Newton said that the world was like a machine and that things | fell "to their natural place" (not gravity). | | AI "enthusiasts" are exactly those people. Reductionists to a | fault. | | The hard sciences have long, long ago indirectly disproved that | humans are special in any kind of way. But our "machinery" is | indeed complex. And we won't find out that it's just a bunch of | levers and gears someday as a side-effect of AI shenanigans. | Jensson wrote: | The fifth and most common response: | | * Pragmatics - This is a tool, does it solve problems I have? | If yes use it, if no then wait until a tool that is useful | comes around. | | Some seems to think that such a stance is unimaginable and that | they are just trying to cope with the thought that they | themselves are nothing but specs of space dust in the infinite | universe. No, most people don't care about that stuff, don't | project your mental issues unto others. | e12e wrote: | Interesting points, but I think the author does themselves a | disservice in downplaying general anthropomorphism (no mention of | _a child 's stuffed animal_ - only an adults "ironic" distance to | "willful" anthropomorphism) - and by downplaying physical | presence /body language: | | > in my opinion, conventional social performances "in-person" | which are not significantly richer than text -- expressions of | emotion add perhaps a few dozen bytes of bandwidth for example -- | I think of this sort of information stream as "text-equivalent" | -- it only looks plausibly richer than text but isn't) - and the | significance of body language (ask anyone who has done a | presentation in front of an audience if body language | matters...). | | This flies in the face of research into communication - and | conflates "Turing game" setups that level the playing field (we | don't expect a chat text box to display body language - so we are | not surprised when a chat partner doesn't - be that human or | not). | | And again with children (or adults) - people with no common | language will easily see each other during a game of soccer - | without any "text". | | Ed: plot twist-the essay is written by chat gpt... Lol ;) | anon7725 wrote: | > The simplicity and minimalism of what it takes has radically | devalued personhood. | | Hogwash. If we follow the logic of this essay, then personhood | would be fully encapsulated by one's online posts and | interactions. Does anyone buy that? If anything, LLM chatbots are | "terminally online" simulators, dredging up the stew that results | from boiling down subreddits, Twitter threads, navel-gazing | blogs, etc. | | Call me when ChatGPT can reminisce about the time the car broke | down between Medford and Salem and it took forever for the tow | truck to arrive and thats when you decided to have your first | kid. | | There aren't enough tokens in the universe for ChatGPT to be a | real person. | wpietri wrote: | > LLM chatbots are "terminally online" simulators | | That's a great phrase. I saw someone recently mention that the | reason LLM chatbots don't say, "I don't know" is because that | is so rarely said online. | stuckinhell wrote: | Holy moly, I think this author hits the critical point. | | So what's being stripped away here? And how? The what is easy. | It's personhood. | | AI being good at Art, Poems, etc are direct attacks on personhood | or the things we thought make us human. | | It certainly explains why I feel art AI to be far more chilling | then a logical robotic AI. | jvanderbot wrote: | I never had a soap box, but if I did you'd notice I have been | screaming that the revolution that comes from human like AI is | not that we have magical computers, it's that we realize we | have no magic in our minds. We are nothing more than stories we | repeat and build on. And with text, you can do that easily. | | > Seeing and being seen is apparently just neurotic streams of | interleaved text flowing across a screen. | | Or, our mind. | atchoo wrote: | No matter how fancy the chat bot, until we solve the "Hard | problem of consciousness", there will be magic in our minds. | jvanderbot wrote: | I don't think it's that hard, and I'm not alone in saying | that. It seems hard because (IMHO) we won't admit it's just | something like GPT running on only our own memories. | stuckinhell wrote: | I agree with you. This is my biggest fear. The AI's | ability to do art, and creative work is extremely close | to how human minds work but at a greater scale. If true, | then humanity isn't special, and the human mind is soon | obsolete. | jvanderbot wrote: | I wouldn't worry about "obsolete". There are better minds | than mine all over, but mine is still relevant, mostly | because it runs on as much energy as a candle instead of | a country, and doesn't distract those better minds. | layer8 wrote: | Pointing to the hard problem of consciousness in present- | day discourse about consciousness doesn't do much, because | people disagree that there is a hard problem of | consciousness in the first place. | qudat wrote: | Agreed. There is no hard problem of consciousness, we are | just biased. | | https://bower.sh/what-is-consciousness | prmph wrote: | There absolutely is a hard problem of consciousness. | | One thought experiment I like to use to illustrate this: | Imagine we accept that an AI is conscious, in the same way | a human is. | | Now, what defines the AI? You might say the algorithm and | the trained weights. Ok, so let's say, in a similar way, we | extract the relevant parameters from a human brain and use | that to craft a new human. | | Are they the same person, or two? Do they experience the | same consciousness? Would they share the same embodied | experience? | | Could the one be dead and other alive? If so, what makes | them have their own individuality? If your loved one died, | and their brain was reconstructed from parameters stored | while they were alive, would you accept that as a | resurrection? Why or why not? | | Note that I offer no answer to the above questions. But | trying to answer them is part of what the hard problem of | consciousness is about. | jvanderbot wrote: | Imagine we found all the connections, chemical weighting, | and neuron structure that exactly reproduced ChatGPT in | the forebrain. Is ChatGPT now a human? Absolutely not. | But is it capable of human like speech? Yep. | | ChatGPT will probably say it is conscious if you tell it | that it is (for various values of tell). Do we really | know there's anything else going on with us? | | I don't. I think we're all stories told by learning | machines mimicking culture we observe, compete with memes | for soul, special creativity, etc. We vastly overestimate | our intelligence and vastly underestimate the cumulative | effects of million years of culture. | pixl97 wrote: | So lets make this an easier problem. | | You step in a Star Trek transporter. Scotty goes to beam | you up but after a quick flash you are still there. But, | they get notice that you were also delivered to the other | side. There are two exact copies of you now. | | I would say at t=0 they are the exact same person that | would think the exact same way if put in the same | experiences. Of course physical existence will quickly | skew from that point. | | For the case of the love one that died, I would argue | 'they' are the same person from the moment they are | stored. The particular problem here is there will be a | massive skew in shared experience. You got to suffer | their (presumably) traumatic death that has changed you. | Them now coming back into existence into your trama will | likely lead you to believe that they changed when it is | you that has changed. Add to this the physical time jump | where they were missing will cause the same things in all | their other social interactions. Just imagine being | kidnapped but being unconscious the entire time. The | world will treat you differently when you get back even | though you've not really changed. | mrjh wrote: | "we realize we have no magic in our minds" | | Surely an AI is a digital replica (and homage) of that magic? | Without the magic in our minds we could've never created that | replica. | | To me it's an acknowledgement of how awesome our own brains | are that we want to even replicate them. | jvanderbot wrote: | I believe and hope people at least consider that, _yeah_ an | AI is a replica of that, and for all AIs failures, it 's a | _really good_ replica of _most_ of what it is to be human | and "conscious". After that, it's all feeding back your | story to yourself, and compounding memories from actual | experience. (Which , have you noticed, are mostly stories) | pixl97 wrote: | So if we're magic and it is magic then technically this is | ok. | | But the problem is we create it, so it can't be magic. So | if we're magic and it is not magic then its just an object | we are free to abuse (at least from many peoples | perspective). | | I like to think of it as we're complex and interesting, and | it is complex and interesting but neither of us is magic. | We don't like to be abused, so creating something like us | and abusing it would be completely unethical. | prmph wrote: | I'm not sure that's correct. | | An AI is severely constrained to the modes of thought which | which is was created. Call me when an AI comes up with | original philosophy, describes it in terms of what is already | understood, explains why it is necessary, and is able to | promote it to acceptance. | | I think people severely underestimate the original thought | capacity of the human mind. | | An AI could never come up with the concept of Calculus, or | relativity, for instance. Yes, if you feed it enough data, | and assuming you have endowed it with a sufficiently | sophisticated algorithm, it might (probably) use something | that resembles calculus internally, but it certainly will not | be able to espouse it as a concept and explain what new | problems it will allow us go imagine. | pixl97 wrote: | Call me when you come up with original philosophy.... | avgcorrection wrote: | A perfectly mediocre essay.[1] | | > Computers wipe the floor with us anywhere we can keep score | | Notice the trick? If you can keep score at something then you can | probably make an algorithm for it. If you can make an algorithm | for it then you can probably make a digital computer do it a | billion times faster than a person, since digital computers are | so good at single-"mindedly" doing one thing at a time. | | > So what's being stripped away here? And how? | | > The what is easy. It's personhood. | | Why? | | The Turing Test was invented because the question "do machines | think?" was "too meaningless" to warrant discussion.[1] The | question "can a machine pose as a human"? is, on the other hand, | well-defined. But notice that this says nothing about humans. | Only our ability (or lack thereof) to recognize other humans | through some medium like text. So does the test say _anything_ | about how humans are "just X" if it is ever "solved"? Not really. | | You put a text through a blender and you get a bunch of "mediocre | opinions" back. Ok, so? That isn't even remotely impressive, and | I think that these LLMs are in general impressive. But recycling | opinions is not impressive. | | > (though in general I think the favored "alignment" frames of | the LessWrong community are not even wrong). | | The pot meets the kettle? | | [1] That I didn't read all the way through because who has time | for that. | | [1] https://plato.stanford.edu/entries/turing-test/ | visarga wrote: | > A perfectly mediocre essay. | | The author rightly draws attention to text. | | LLMs showed they can do the classical NLP tasks and more: | summarise, translate, answer questions, play a role, brainstorm | ideas, write code, execute a step by step procedure, the list | is unbounded. It's the new programming language. | | All these abilities emerged from a random init + text. Guess | what was the important bit here? Text. It's not the | architecture, we know many different architectures and they all | learn, some better than others, but they all do. Text is the | magic dust that turns a random init into a bingChat with | overactive emotional activity. | | Here I think the author made us a big service in emphasising | the text corpus. We were lost into a-priori thinking like "it's | just matrix multiplication", "it's just a probability | distribution predictor over the next token". But we forgot the | real hero. | | The interesting thing about words is that they are perceptions, | they represent a way to perceive the world. But they are also | actions. Being both at the same time, perception and action, | that makes for an interesting reinforcement learning setup, and | one with huge training data. Maybe text is all you need, it is | a special kind of data, it's our mind-data. | krackers wrote: | >A perfectly mediocre essay | | One might even say "premium mediocre" [1] | | [1] https://www.ribbonfarm.com/2017/08/17/the-premium- | mediocre-l... | alex_smart wrote: | > Notice the trick? If you can keep score at something then you | can probably make an algorithm for it | | You are basically arguing P = NP, but it isn't known to be the | case. As far as we can tell, keeping score is much easier in | general than finding states that yield a high score. | avgcorrection wrote: | I seriously doubt that "anything/[everything] we can score" | has been conquered by AI,[1] but I was assuming that the | author meant those typical AI milestones. | | [1] What about some kind of competition where you have to | react and act based on visual stimulus? And you have to do it | perfectly? | motoxpro wrote: | Little too broad. If the act is tell you what color it is, | then computers will win every time. Again, if you can score | it. A computer will win. | avgcorrection wrote: | Nice counter-example. Why would the test be that simple? | pixl97 wrote: | Then bring forth a complex but indivisible test? | burnished wrote: | Yeah, like sorting apples into good and bad piles really | fast? | avgcorrection wrote: | Sounds like a leading question which is supposed to | debunk that whole category by suggesting one counter- | example. So no. | behnamoh wrote: | This stuff only makes HN frontpage because HN likes controversial | opinions. In reality, text works for a small percentage of | people. Going back to a format that's as old as computers is like | saying that no progress/improvements were made ever since. | college_physics wrote: | Just another amplifier of the mass hysteria. Degrading humanity | for monetary gain. Reminds of darker times. Ignore | aflukasz wrote: | No one suggested this yet, so I will be the first - a very good | read in this context is "Reasons and Persons" by Derek Parfit. | Second part of this book is about personal identity. It discusses | all the various edge cases and thought experiments across | physical and time dimensions and is written in a style and with a | rigor that I believe any technical person will really appreciate. | | One of my favorite statements from the book is that "cogito ergo | sum" is too strong of a statement and it would be wiser and | easier to defend a weaker one - "a thought exists". (I hope I | didn't get this wrong - can't check at the moment). ___________________________________________________________________ (page generated 2023-02-18 23:00 UTC)