[HN Gopher] A Multi-Level View of LLM Intentionality
       ___________________________________________________________________
        
       A Multi-Level View of LLM Intentionality
        
       Author : zoltz
       Score  : 40 points
       Date   : 2023-09-11 16:59 UTC (5 hours ago)
        
 (HTM) web link (disagreeableme.blogspot.com)
 (TXT) w3m dump (disagreeableme.blogspot.com)
        
       | lukev wrote:
       | I'm not sure the definition of "intention" the article suggests
       | is a useful one. He tries to make it sound like he's being
       | conservative:
       | 
       | > That is, we should ascribe intentions to a system if and only
       | if it helps to predict and explain the behaviour of the system.
       | Whether it _really_ has intentions beyond this is not a question
       | I am attempting to answer (and I think that it is probably not
       | determinate in any case).
       | 
       | And yet, I think there's room to argue that LLMs (as currently
       | implemented) cannot have intentions. Not because of their
       | capabilities or behaviors, but because we know how they work
       | (mechanically at least) and it is incompatible with useful
       | definitions of the word "intent."
       | 
       | Primarily, they are pure functions that accept a sequence of
       | tokens and return the next token. The model itself is stateless,
       | and it doesn't seem right to me to ascribe "intent" to a
       | stateless function. Even if the function is capable of modeling
       | certain aspects of chess.
       | 
       | Otherwise, we are in the somewhat absurd position of needing to
       | argue that all mathematical functions "intend" to yield their
       | result. Maybe you could go there, but it seems to be torturing
       | language a bit, just like people who advocate definitions of
       | "consciousness" wherein even rocks are a "little bit conscious."
        
         | Icko wrote:
         | > Primarily, they are pure functions that accept a sequence of
         | tokens and return the next token. The model itself is
         | stateless, and it doesn't seem right to me to ascribe "intent"
         | to a stateless function. Even if the function is capable of
         | modeling certain aspects of chess.
         | 
         | I have two arguments against. One, you could argue that state
         | is transferred between the layers. It may be inelegant for each
         | chain of state transitions to be the same length, but it seems
         | to work. Two, it may not have "states", but if the end result
         | is the same, does it matter?
        
         | haltist wrote:
         | Ascribing human properties to computers and software has always
         | seemed very bizarre to me. I always assume people are confused
         | when they do that. There is no meaningful intersection between
         | biology, intelligence, and computers but people constantly keep
         | trying to imbue electromagnetic signal processors with
         | human/biological qualities very much like how children
         | attribute souls to teddy bears.
         | 
         | Computers are mechanical gadgets that work with electricity.
         | Humans (and other animals) die when exposed to the kinds of
         | currents flowing through computers. Similarly, I have never
         | seen a computer drink water (for obvious reasons). If
         | properties are reduced to behavioral outcomes then maybe
         | someone can explain to me why computers are so averse to water.
        
           | version_five wrote:
           | He was inspired by lesswrong which from my scan more
           | mysticism and philosophy (with a handful of self importance)
           | than anything about how computers work. Advanced technology
           | is magic to laypeople. It's like how some people believe in
           | homeopathy. If you don't understand how medicine works, it's
           | just a different kind of magic.
        
             | haltist wrote:
             | I guess that might be tied up with human biology, the need
             | to attribute agency to inanimate objects. That one is a
             | worthwhile puzzle to figure out but most people seem more
             | mesmerized by blinking lights and shiny gadgets than any
             | real philosophical problems.
        
         | version_five wrote:
         | Not because of their capabilities or behaviors, but because we
         | know how they work (mechanically at least) and it is
         | incompatible with useful definitions of the word "intent."
         | 
         | I've never seen this deter anyone. I can't understand how
         | people that know how they work can have such ridiculous ideas
         | about llms.
         | 
         | I'd add though that inference is clearly fixed but there is
         | some more subtlety about training. Gradient descent clearly
         | doesnt have intelligence, intent (in the sense meant),
         | consciousness either, but it's not stateless like inference and
         | you could argue has a rudimentary "intent" in minimizing loss.
        
           | og_kalu wrote:
           | The most useful definitions have predictive power.
           | 
           | When you say upsetting things to bing chat, you'll find the
           | conversation prematurely end.
           | 
           | You can cry all you want about how bing isn't _really_ upset.
           | How it doesn 't _really_ have intention to end the chat but
           | those are evidently useless defitions because the chat _did_
           | end.
           | 
           | A definition that treats Bing as an intentful system is more
           | accurate to what happens in reality.
        
             | version_five wrote:
             | That might be useful in helping a child learn to use it, it
             | has no value when actually studying neural networks. You
             | could equally pretend the sun sets every night because it's
             | upset from shining all day.
        
               | og_kalu wrote:
               | >That might be useful in helping a child learn to use it
               | 
               | It is useful for anyone looking to use such systems. A
               | LLM piloted robot could potentially pick up a knife and
               | stab you because you obstructed some goal or said mean
               | words and pretending it didn't have intent to do so won't
               | bring you back to life. Acting like it does could help
               | avoid such a scenario.
               | 
               | >You could equally pretend the sun sets every night
               | because it's upset from shining all day.
               | 
               | No you couldn't.
               | 
               | The conversation ended prematurely because of your input.
               | There is zero ambiguity on the relation.
               | 
               | But because you ascribe intent to bing, you can predict
               | (accurately) that saying nice things will not end the
               | conversation.
               | 
               | LLMs act like they have intent. This is a matter of
               | conceding they do or not. Conceding so is more useful
               | because it has more accurate predictive power than the
               | alternative. This becomes plain when LLMs are allowed
               | more actions than just conversation.
               | 
               | >it has no value when actually studying neural networks.
               | 
               | On the contrary. Now you know that certain things are
               | unnecessary to build a system that acts like it has
               | intent.
        
               | lawlessone wrote:
               | >You could equally pretend the sun sets every night
               | because it's upset from shining all day.
               | 
               | That would describe quite a lot of ancient religious
               | beliefs about the Sun
        
         | jedharris wrote:
         | The article provides a very clear reason for using the idea of
         | "intention": that framing helps us understand and predict the
         | behavior. In contrast framing a mathematical function as having
         | "intention" doesn't help. The underlying mechanism isn't
         | relevant to this criterion.
         | 
         | Clearly the system we're understanding as "intentional" has
         | state; we can engage in multi-round interactions. It doesn't
         | matter that we can separate the mutable state from the function
         | that updates that state.
        
           | lo_zamoyski wrote:
           | A mathematical function isn't a mechanism. It has no causal
           | power.
        
             | nomel wrote:
             | Sure, but we're talking mathematical functions that have
             | been given physical form, by attaching them to input and
             | output devices. Or, am I missing something? For other
             | examples, see anything automated with a motor, and some
             | decision tree, semi chaotic or not.
        
         | lo_zamoyski wrote:
         | Uh, hold on. That's not what's meant by intentionality. No one
         | is talking about what a machine _intends_ to do. In philosophy,
         | and specifically philosophy of mind,  "intentionality" is,
         | briefly, "the power of minds and mental states to be about, to
         | represent, or to stand for, things, properties and states of
         | affairs" [0].
         | 
         | So the problem with this guy's definition of intentionality is,
         | first, that it's a redefinition. If you're interested in
         | whether a machine can possess intentionality, you won't find
         | the answer in interpretivism, because that's no longer a
         | meaningful question.
         | 
         | Intentionality presupposes telos, so if you assume a
         | metaphysical position that rules out telos, such as
         | materialism, then, by definition, you cannot have "aboutness",
         | and therefore, no intentionality _of any sort_.
         | 
         | [0] https://plato.stanford.edu/entries/intentionality/
        
           | jedharris wrote:
           | Some philosophers take that position. Dennett, explicitly
           | cited in the article, wrote _The Intentional Stance_ (1987)
           | about exactly the approach to intentionality taken in the
           | article. His approach is accepted by many philosophers.
           | 
           | As you point out, the approach you cite can't be used in a
           | materialist metaphysical position. That's a pretty severe
           | problem for that definition! So Dennett's approach, or
           | something like it, has major advantages.
        
           | jedharris wrote:
           | Also, you are obviously wrong (or rhetorical?) when you say
           | "No one is talking about what a machine intends to do." We
           | certainly do! You can say "No one should" or other normative
           | condemnations but then we're arguing on different territory.
        
       | og_kalu wrote:
       | >Unless you think that there is some fundamental reason why LLMs
       | will never be able to play chess competently, and I doubt there
       | is, then it seems that we could with the right prompts implement
       | some sort of chess AI using an LLM.
       | 
       | You can play a good game of chess (or poker for that matter) with
       | GPT.
       | 
       | https://twitter.com/kenshinsamurai9/status/16625105325852917...
       | 
       | https://arxiv.org/abs/2308.12466
       | 
       | There's also some work going on in the eleuther ai discord
       | training LLMs specifically for chess to see how they shape up.
       | They're using the pythia models. so far:
       | 
       | Pythia 70M, est ELO 1050
       | 
       | Pythia 160M, est ELO 1370
        
         | lawlessone wrote:
         | I've found they fall apart after a couple of moves and lose
         | track of the game.
         | 
         | Edit: This might not be the case anymore it seems, my below
         | point doesn't actually contradict you, seems it matters a lot
         | how you tell the model your moves. Also saying things like
         | "move my rightmost pawn" completely confuses them.
        
           | og_kalu wrote:
           | Not had it lose track with the format in the first link
           | (GPT-4, not really tried 3.5)
        
             | lawlessone wrote:
             | Yeah i was wrong. I think it has gotten better since i
             | tried this.
        
           | pixl97 wrote:
           | The token model of LLMs doesn't map well into how human
           | experience the world of informational glyphs. Left and right
           | is a intrinsic quality of our vision system. An LLM has to
           | map the idea of left and right into symbols via text and line
           | breaks.
           | 
           | I do think it will be interesting as visual input and
           | internal graphical output is integrated with text based LLMs
           | as that should help correct their internal experience to be
           | based closer to what we as humans experience.
        
             | lawlessone wrote:
             | " An LLM has to map the idea of left and right into symbols
             | via text and line breaks."
             | 
             | Oh yeah that's i suggested it :)
             | 
             | I do wonder though if we give the LLMs enough examples of
             | texts with people describing their relative spatial
             | position to each other and things will it eventually
             | "learn" to work things these out a bit better
        
               | og_kalu wrote:
               | >I do wonder though if we give the LLMs enough examples
               | of texts with people describing their relative spatial
               | position to each other and things will it eventually
               | "learn" to work things these out a bit better
               | 
               | GPT-4's spatial position understanding is actually really
               | good all things considered. By the end, 4 was able to
               | construct an accurate maze just from feedback about the
               | current position and possible next moves after each move
               | by GPT-4.
               | 
               | https://ekzhu.medium.com/gpt-4s-maze-navigation-a-deep-
               | dive-...
               | 
               | I think we just don't write much about moving through
               | space and that is why reasoning about it is more limited.
        
       | [deleted]
        
       | labrador wrote:
       | The authors of the text the model was trained on certainly had
       | intentions. Many of those are going to be preserved in the
       | output.
        
         | passion__desire wrote:
         | Can we say ChatGPT or its future versions would be like an
         | instantiation of the Boltzmann Brain concept if it has internal
         | qualia? The "brain" comes alive with the rich structure only to
         | disappear after the chat session is over.
        
           | SanJoseEngineer wrote:
           | [dead]
        
           | jedharris wrote:
           | Cool way of putting it. Let's run with that. A good actor can
           | be seen as instantiating a Boltzmann Brain while on stage --
           | especially when improvising (as always may be needed). Maybe
           | each of us is instantiating some superposition of Boltzmann
           | Brains in everyday life as we wend our way through various
           | social roles...
           | 
           | From now on I'll listen for the subtle popping sounds as
           | these BBs get instantiated and de-instantiated all around
           | me...
           | 
           | Of course a philosopher can object that (1) these BBs are on
           | a substrate that's richer than they are so aren't "really"
           | BBs and (2) they often leave traces that are available to
           | them in later instantiations which again classical BBs can't.
           | So maybe make up another name -- but a great way to think.
        
       ___________________________________________________________________
       (page generated 2023-09-11 22:00 UTC)