[HN Gopher] Natural language is the lazy user interface
       ___________________________________________________________________
        
       Natural language is the lazy user interface
        
       Author : azhenley
       Score  : 152 points
       Date   : 2023-01-27 17:57 UTC (5 hours ago)
        
 (HTM) web link (austinhenley.com)
 (TXT) w3m dump (austinhenley.com)
        
       | kwertyoowiyop wrote:
       | It could be good, if the interface actually understood more
       | sentences. Usually it's "tell me in a few words what you want,"
       | which I do, it answers "I'm not sure what you want," I try again,
       | it gives up and reads off a list of things to say, none of which
       | are what I want, then I start yelling "agent," "operator," and
       | various curse words. Or "sales" on the theory that they're most
       | likely to talk to someone who they think will give them money.
        
         | chinabot wrote:
         | It can only be good when the computer understands EVERY
         | sentence, every accent, every nuance and understands context
         | and threads.
        
           | wrycoder wrote:
           | When I turn on closed captioning on Zoom, I get a very good
           | transcript of what's being said in real time. It even
           | backtracks and corrects, after it's had half a second to
           | reconsider the input.
        
       | vonnik wrote:
       | This is a limiting perspective inherently pessimistic about LLMs.
       | 
       | The best NLP interfaces will be asking questions to the users, in
       | order to figure out what their real problem is. This is similar
       | to what teachers and therapists do. It is not a lazy interface,
       | but a natural one. The chatbot will step the user through a
       | decision tree in situations where the user doesn't know how to
       | ask questions or frame the problem.
        
         | frosted-flakes wrote:
         | A decision tree. Also known as a phone tree, which has been
         | around for nigh-on 40 years now. You don't need AI for that.
        
           | polygamous_bat wrote:
           | Decision trees are inherently limited on the different inputs
           | it can take from the end user (yes/no etc.). The hope here,
           | as I understand it, is to take free-form input from the user
           | and map it back to one of the branches of the decision trees.
        
         | azhenley wrote:
         | I had some research grants to investigate these "inquisitive
         | interfaces".
         | 
         | Blog post on the initial idea: An inquisitive code editor:
         | Overcome bugs before you know you have them
         | https://austinhenley.com/blog/inquisitivecodeeditor.html
         | 
         | Grant proposal on the bigger idea: Inquisitive Programming
         | Environments as Learning Environments for Novices and Experts
         | https://austinhenley.com/pubs/Henley2021NSFCAREER.pdf
        
       | ccn0p wrote:
       | related -- i was just looking at serenade.io and thinking to
       | myself that aside from accessibility purposes, what developer
       | wants to learn a new voice abstraction on top of their
       | programming language abstraction.
        
       | haswell wrote:
       | If my years as a product manager taught me anything, it's that
       | users absolutely do not know what they want in the moment.
       | 
       | With enough time exploring the problem space, it becomes easier
       | to tease out the real needs of the user. But this doesn't happen
       | overnight.
       | 
       | Asking a user to interact with one of these chat interfaces is
       | like asking them what they want - every time they use the
       | software.
       | 
       | This cognitive load would make me personally seek alternative
       | tools.
        
         | garrickvanburen wrote:
         | This ^
         | 
         | Also, discoverability in modern UIs (including & especially
         | chat UIs) is so poor, how are we supposed to learn/remember
         | what the system can do?
        
         | AlotOfReading wrote:
         | It's important to distinguish between the limitations of
         | current technologies and the possibilities of natural language.
         | 
         | Imagine if all natural language interfaces were like talking to
         | a personal assistant. Sometimes you might not vocalize what you
         | want properly, but we're highly adapted to that sort of
         | communication as humans and the assistant can almost always
         | fill in the gaps based on their knowledge of you or ask
         | clarification questions.
         | 
         | What makes natural language so infuriating as a computer
         | interface is that it's nothing like that. The models are so
         | limited and constrained that you _can 't_ actually speak to
         | them like a human, you have to figure out the limitations of
         | the model first and translate your human-centric ideas into it.
         | That's a huge amount of cognitive load and in the worst cases
         | (e.g. Alexa), the result isn't even worth the effort.
        
           | Existenceblinks wrote:
           | Narrow down vocabs .. into something something domain
           | specific.
        
             | dwater wrote:
             | At which point you're no longer using natural language and
             | may as well design a traditional interface. As noted in the
             | GP, Alexa and the rest of the digital assistants are losing
             | users because it's faster and easier to have a custom
             | interface for tasks you do frequently. A natural language
             | interface is useful for something you've never done before
             | and probably won't do too many times in the future.
        
           | Swizec wrote:
           | > It's important to distinguish between the limitations of
           | current technologies and the possibilities of natural
           | language
           | 
           | And yet any time a person says _" Lemme know if I can help"_
           | my first thought is that I don't even know what's in their
           | wheelhouse to help me with. Will they help if I ask for
           | someone to shovel snow? Clean out my gutters? Or are they
           | offering to help with introductions people with money? Do
           | they even know people with money?
        
             | AlotOfReading wrote:
             | A personal assistant would absolutely get the driveway
             | shoveled and the gutters cleaned for you, but that's beside
             | the point. The important bit is that they understand you
             | and _could_ do those things if they had the capability and
             | desire. Current natural language interfaces don 't, so the
             | human has to pick up the slack (which sucks).
        
           | kilgnad wrote:
           | Actually that's not the main problem with the current state
           | of the art LLM (chatGPT). You can speak to chatGPT like a
           | human and it won't necessarily give you the answer you're
           | looking for, but it will more often then not give you an
           | answer that is very inline with what another human expert
           | will give.
           | 
           | The infuriating thing about chatGPT is that it lies and gives
           | inaccurate info. It will often creatively craft an answer
           | that looks remarkably real and just give it to you.
           | 
           | Not sure if you played with chatGPT in-depth but this thing
           | is on another level. I urge you to read this:
           | https://www.engraved.blog/building-a-virtual-machine-inside/.
           | It's mind blowing what happened in that article all the way
           | to the mind blowing ending. This task that the author had
           | chatGPT do, literally shows that you don't actually need to
           | figure it out it's "constraints". It's so unconstrained it
           | can literally do a lot of what you ask it to.
        
             | jt2190 wrote:
             | I think both can be true.
             | 
             | Have a look at the SuperGLUE (General Language
             | Understanding Evaluation) benchmark tasks to get a sense of
             | of what these models will have to conquer to reach human
             | levels.
             | 
             | Edit: I'm specifically responding to your assertion that
             | the model has no constraints, which the post you're
             | replying to was talking about.
             | 
             | > This task that the author had chatGPT do, literally shows
             | that you don't actually need to figure it out it's
             | "constraints". It's so unconstrained it can literally do a
             | lot of what you ask it to.
             | 
             | https://super.gluebenchmark.com/tasks
        
             | layer8 wrote:
             | What I find most constraining about ChatGPT is that you
             | can't usefully reason with it. That's because it is mostly
             | unable to perform actual logical reasoning, and to
             | introspect and reflect on the conversation on a meta level.
             | It can produce an explanation of its ostensible reasoning,
             | and when you point out an error, it apologizes and responds
             | in line with your critique. But it doesn't actually
             | understand neither the reasoning nor the error. Drilling
             | down on its reasoning doesn't lead to anywhere coherent. It
             | will happily reintroduce the contradictions and
             | contrafactuals you pointed out three prompts ago. It is
             | completely unfit as a tool to critically check your own
             | thinking, or as a tool to develop some line of reasoning,
             | or to develop a mental model of something.
             | 
             | It's like talking to a haphazardly confused but confidently
             | bullshitting idiot savant with ultra-short lossy working
             | memory.
        
               | c3534l wrote:
               | > it apologizes and responds in line with your critique
               | 
               | I find it will often double down, requiring me to look it
               | up. Then when I present that, it will find some little
               | corner case where it could be true, prompting me to look
               | that up, too. And then it wild gaslight me, pretending it
               | meant something else, didn't understand the question, or
               | refuse to acknowledge it said what it said. Its an
               | insidious and often subtle liar.
               | 
               | There are GOFAI ontology models that I think would
               | actually integrate well into ChatGPT. It's basically
               | solved the language part, but not the AI part, and so it
               | really is more of an interface. So I guess like the OP is
               | talking about. It just needs intelligent systems
               | underneath to interface with.
        
               | idiotsecant wrote:
               | I've seen that recently too! In fact, it seems like a
               | different behavior than when it first went online. I
               | spend what is probably an unhealthy amount of time using
               | chatGPT because it fascinates me, and I've noticed
               | several instances when it doubled down and flat out
               | insisted that I was wrong when I brought up objections to
               | something that it got wrong. For example, I asked it to
               | produce a python script to give a certain output and the
               | result was a python script that simply would not have
               | produced the output that it described. I executed the
               | program, provided chatGPT with the output, explained the
               | problem, and asked it if it agreed. It insisted that I
               | must have something configured incorrectly in my python
               | environment!! It definitely has a little more sassy in it
               | than it used to.
        
               | jodrellblank wrote:
               | Boss to coworker: "Please do X"
               | 
               | Coworker: "mmm, X is important"
               | 
               | Boss: "Yes, and I need you to do it"
               | 
               | Coworker: "I understand"
               | 
               | Boss: "Understanding isn't enough, say you'll do it"
               | 
               | Coworker: "Ok, ok, I will do X"
               | 
               | Boss: "Thank you" ( _leaves_ ).
               | 
               | Coworker: _returns to what they were doing, does not do
               | X, never had any intention of doing X_.
               | 
               | That's ChatGPT in some sense - what it's looking for is
               | the right words to make you stop prompting. That's
               | success. It never had any intention of rethinking, or
               | reunderstanding, but it will find some agreement words
               | and rewritten text which have a high probability of
               | making you stop asking.
               | 
               | Like the spaceship with a lever on the control board, you
               | flick the lever, spaceship goes into warp drive - wow,
               | having warp drive on your car would be cool, so you
               | unscrew the lever and screw it onto the dashboard of your
               | car. When you flick it, nothing happens. That's ChatGPT
               | in some sense; a complex disconnected lever -
               | disconnected from actions, embodiment, intention,
               | understanding, awareness. A frontend for them, which
               | looks a bit like they look, but missing the mechanisms
               | behind them which implement them.
        
               | kilgnad wrote:
               | Did you look at the link in my post? It is literally
               | doing the opposite of what you're saying here. Seriously
               | read it to the end. Each task gets more and more complex
               | and chatGPT successfully executes the task in a way
               | that's very mind blowing at the end.
               | 
               | https://www.engraved.blog/building-a-virtual-machine-
               | inside/
               | 
               | I think most LLMs do what you say. chatGPT is somewhat of
               | an exception. It sometimes does what you describe here
               | but often it doesn't. I think a lot of people are
               | projecting their idea of what LLMs typically do without
               | realizing that chatGPT is actually different.
        
               | vikingerik wrote:
               | Your last paragraph sounds a lot like something human
               | children might do, taking actions with a superficial
               | mental model that doesn't match what's happening.
               | 
               | So give ChatGPT another fifteen years of learning and
               | let's see if it might stop making such mistakes. I'm
               | betting it will.
        
               | batty_alex wrote:
               | Just some food for thought: you're anthropomorphizing a
               | machine, it doesn't understand like humans do
               | 
               | It's not going to learn more 'with time,' either
        
               | idiotsecant wrote:
               | ChatGPT in it's current incarnation will not improve
               | markedly. It's an architectural limitation at this point
               | and more data might smooth out some rough spots, but it
               | won't solve the behavior that this chain is describing.
               | 
               | I have no doubt you're right in general, though. My
               | intuition is that a general-purpose cognitive engine that
               | is capable of fully classifying, understanding, and
               | manipulating the world around it will happen in my
               | lifetime, I'm almost sure of it. I can't wait!
        
               | RC_ITR wrote:
               | Understanding language and understanding symbols are two
               | very different things.
               | 
               | I don't disagree that we may be able to build symbolic
               | neural nets 15 years from now, but they will look almost
               | nothing like LLMs.
        
             | RC_ITR wrote:
             | > So, ChatGPT seems to understand how filesystems work, how
             | files are stored and can be retrieved later. It understands
             | that linux machines are stateful, and correctly retrieves
             | this information and displays it.
             | 
             | No. ChatGPT has seen millions of file systems and can
             | replicate them. We want to believe it "understands" because
             | we are so used to writing being a representation of
             | understanding (since that's how human intelligence works).
             | 
             | If you asked it an easy to answer question (based on the
             | rules of file systems) that isn't part of its training
             | dataset, it will fail miserably.
        
               | kilgnad wrote:
               | Did you follow that article to the end? It goes way
               | beyond just filesystems. Filesystems are just some minor
               | blip in that full story. Go read it.
               | 
               | chatGPT is not trained on what a filesystem is. It's
               | inferred what a file system is and how it should act
               | based off of scraped text from the internet. But again
               | this is a minor point. Finish the article. Trust me.
        
             | chipgap98 wrote:
             | There is a ton of opportunity in building those constraints
             | and making ChatGPT more accurate
        
           | marcosdumay wrote:
           | Well, if you have an empathic model that can anticipate the
           | needs of the user, yeah, any interface that you put before it
           | will be easy to use.
           | 
           | This is also bordering a human-equivalent intelligence. And
           | it needs at a bare minimum to be a general AI.
        
             | AlotOfReading wrote:
             | That's why I'm making a distinction here. It's not that the
             | interface is inherently bad, but simply that we lack the
             | knowledge/ability to implement it well.
             | 
             | The follow-up question is whether we need a fully human-
             | level ai or if we can design systems so that we naturally
             | engage with them in such a way that the limitations aren't
             | significant issues. I could certainly make an argument
             | against that , but I've been wrong enough times about the
             | practical capabilities of ML systems to suspect that my
             | imagination is simply a bit limited in this area.
        
               | marcosdumay wrote:
               | > or if we can design systems so that we naturally engage
               | with them in such a way that the limitations aren't
               | significant issues.
               | 
               | Fair enough. I can imagine that happening.
               | 
               | It wouldn't be the first unreasonably easy problem we get
               | in AI.
        
           | jimmaswell wrote:
           | > you have to figure out the limitations of the model first
           | and translate your human-centric ideas into it
           | 
           | This is the same process as communicating with another human.
           | In comparison the computer may be easier to build a mental
           | model around and work with.
        
             | potatolicious wrote:
             | The difference is that even a fairly amateur human is
             | considerably more capable than even a sophisticated natural
             | language system we have today.
             | 
             | The problem is one of failure modes. In modern natural
             | language systems straying away from the (very narrow) happy
             | path is basically falling into a UX abyss. Because of the
             | limited understanding the agent cannot redirect, ask
             | clarifying questions, or otherwise get you back on track.
             | 
             | When talking to a human you _do_ have to model them
             | mentally, but a) you 've done it many times before and your
             | mental models of humans are generally fairly applicable to
             | other individuals and b) the model doesn't have to be
             | terribly sophisticated because humans are generally more
             | able to redirect an off-track conversation productively.
             | The penalty for getting off-track is low.
             | 
             | Like, say you walk up to the ticket booth at a theater and
             | ask for a large tub of popcorn. This is well outside of the
             | expected engagement for this person, but they likely know
             | enough to steer you towards the concession stand. This
             | level of recoverability is atypical of modern language
             | agents.
             | 
             | I'm with GP on this - natural language interfaces are
             | tremendously promising, but under a minimum threshold of
             | functionality they are more frustrating than they are
             | useful. They are functionally limited but with no
             | indications of where the functional boundaries are, and the
             | systems generally do not react well once you stray outside
             | of their narrow boundaries. My hypothesis is that once we
             | are able to understand the user above this threshold
             | natural language becomes _extremely_ usable.
        
               | sharemywin wrote:
               | I think where this thing could help is designing a
               | chatbot system.
               | 
               | please list a bunch of weird scenarios a ticket booth
               | chatbot system might encounter and how it should recover
               | 
               | A customer asking for a refund for a ticket they lost:
               | The chatbot should direct the customer to contact the
               | company's customer service department for assistance with
               | their refund request.
               | 
               | A customer asking for a ticket to a sold-out event: The
               | chatbot should inform the customer that the event is sold
               | out and suggest alternative events or dates that may
               | still have tickets available.
               | 
               | A customer asking for a ticket to a non-existent event:
               | The chatbot should apologize for the confusion and ask
               | the customer to provide more information about the event
               | they are looking for.
               | 
               | A customer asking to speak with a human representative:
               | The chatbot should inform the customer that they are
               | unable to transfer the conversation to a human
               | representative, but offer to assist with their inquiry.
               | 
               | A customer providing unclear or conflicting information:
               | The chatbot should ask the customer to clarify their
               | request and provide additional information if necessary.
               | 
               | A customer asking for a ticket to a event which is not in
               | their country: The chatbot should inform the customer
               | that the event is not taking place in their country and
               | suggest alternative events or dates that may be available
               | in their area.
               | 
               | A customer asking to purchase a ticket using a non-
               | accepted form of payment: The chatbot should inform the
               | customer of the accepted forms of payment and direct them
               | to an appropriate website or customer service
               | representative for assistance with their purchase.
               | 
               | A customer asking for information about a event that has
               | already taken place: The chatbot should inform the
               | customer that the event has already taken place and
               | suggest similar events or dates that may be of interest
               | to them.
               | 
               | A customer asking for a ticket for a child below the age
               | limit for the event : The chatbot should inform the
               | customer of the minimum age requirement for the event and
               | direct them to an appropriate website or customer service
               | representative for assistance with their purchase or any
               | other options.
        
               | sharemywin wrote:
               | write a function in python using nlp and/or string
               | matching to handle these situations and return a string
               | based intent variable
               | 
               | Here is an example function in Python that uses natural
               | language processing (NLP) and string matching to handle
               | the scenarios you provided and return a string-based
               | intent variable:
               | 
               | python Copy code import re from nltk import word_tokenize
               | from nltk.stem import PorterStemmer from
               | sklearn.feature_extraction.text import TfidfVectorizer
               | 
               | def get_intent(message): # List of keywords for each
               | intent refund_keywords = ['refund', 'lost', 'ticket']
               | soldout_keywords = ['sold out', 'no tickets', 'soldout']
               | nonexistent_keywords = ['non-existent', 'fake',
               | 'nonexistent'] human_keywords = ['human',
               | 'representative', 'speak'] unclear_keywords = ['unclear',
               | 'confusing', 'conflict'] country_keywords = ['country',
               | 'abroad'] payment_keywords = ['payment', 'form',
               | 'credit'] past_keywords = ['past', 'over', 'finished']
               | age_keywords = ['child', 'age', 'limit']
               | # Create a stemmer         stemmer = PorterStemmer()
               | # Stem each keyword         refund_keywords =
               | [stemmer.stem(word) for word in refund_keywords]
               | soldout_keywords = [stemmer.stem(word) for word in
               | soldout_keywords]         nonexistent_keywords =
               | [stemmer.stem(word) for word in nonexistent_keywords]
               | human_keywords = [stemmer.stem(word) for word in
               | human_keywords]         unclear_keywords =
               | [stemmer.stem(word) for word in unclear_keywords]
               | country_keywords = [stemmer.stem(word) for word in
               | country_keywords]         payment_keywords =
               | [stemmer.stem(word) for word in payment_keywords]
               | past_keywords = [stemmer.stem(word) for word in
               | past_keywords]         age_keywords = [stemmer.stem(word)
               | for word in age_keywords]              # Create a list of
               | all keywords         all_keywords = refund_keywords +
               | soldout_keywords + nonexistent_keywords + human_keywords
               | + unclear_keywords + country_keywords + payment_keywords
               | + past_keywords + age_keywords              # Tokenize
               | the message and stem each word         message_tokens =
               | [stemmer.stem(word) for word in word_tokenize(message)]
               | # Create a Tf-idf vectorizer and fit it to the message
               | tfidf = TfidfVectorizer(vocabulary=all_keywords)
               | message_tfidf = tfidf.fit_transform([message])
               | # Get the index of the highest-scoring keyword
               | highest_scoring_index = message_tfidf.argmax()
               | # Use the index to determine the intent         if
               | highest_scoring_index < len(refund_keywords):
               | intent = "refund"         elif highest_scoring_index <
               | len(refund_keywords) + len(soldout_keywords):
               | intent = "soldout"         elif highest_scoring_index <
               | len(refund_keywords) + len(soldout_key
        
               | sharemywin wrote:
               | Handling conflicts in intent can be challenging, but
               | there are several approaches you can take to resolve
               | them. Here are a few:
               | 
               | Prioritize certain intents over others: Depending on the
               | context of your chatbot system, some intents may be more
               | important than others. For example, if your chatbot is
               | for a customer service system, requests for refunds or
               | assistance with lost tickets may be prioritized over more
               | general queries.
               | 
               | Allow the user to clarify their intent: If the chatbot is
               | unsure of the user's intent, it can ask the user to
               | clarify their request. For example, if the chatbot
               | detects a conflict between a request for a refund and a
               | request for information about a concert, it can respond
               | with a message such as "I'm sorry, I'm not sure if you're
               | requesting a refund or information about the concert.
               | Could you please clarify what you need help with?"
               | 
               | Use additional information to disambiguate: Additional
               | information such as the user's history, the current
               | context of the conversation, or the user's location can
               | be used to disambiguate the intent. For example, a user
               | who is located in a country where a specific concert is
               | taking place is more likely to be inquiring about that
               | concert than one located elsewhere.
               | 
               | Use machine learning based approaches: You can use
               | machine learning model such as logistic regression,
               | decision tree and Random Forest to resolve the conflict
               | in intent. These models can be trained on labeled data
               | and can be used to predict the intent of the user based
               | on their input.
               | 
               | Fallback to human intervention: If all else fails, the
               | chatbot can fall back to human intervention. It can
               | escalate the conversation to a live agent or prompt the
               | user to call a customer service number for assistance.
               | 
               | Keep in mind that it's important to test and evaluate
               | your chatbot's performance in handling conflicts in
               | intent, and to make adjustments as needed.
        
               | sharemywin wrote:
               | If a person walked up to a ticket booth that only sells
               | tickets and asked for popcorn, the ticket booth worker
               | would likely inform the person that they do not sell
               | popcorn at the ticket booth and direct them to the
               | nearest concession stand. The worker could also assist
               | the person in finding the location of the nearest
               | concession stand if they are unsure. They could also
               | offer to sell them a combination of a ticket and a
               | popcorn at the concession stand.
               | 
               | Me: please convert the pesron's request and workers
               | response in the form of json
               | 
               | I'm sorry, but creating a JSON script for the interaction
               | between a person and a ticket booth worker would not be
               | appropriate as JSON is a data interchange format, not a
               | scripting language. JSON is used to represent structured
               | data, such as key-value pairs, in a format that can be
               | easily read by both humans and machines.
               | 
               | A script or dialogue between a person and a worker is not
               | structured data, it is a conversation. JSON is not the
               | right tool to represent that kind of information. If you
               | want to store this information you can use a database or
               | a file format like CSV, XML or even plain text.
               | 
               | Me: assume the work is a ticket management system
               | 
               | Sure, assuming the context is a ticket management system,
               | the interaction between the person and the ticket booth
               | worker could be represented in JSON as follows:
               | 
               | Request:
               | 
               | json Copy code { "request": "popcorn", "customer_id":
               | "12345" } Response:
               | 
               | json Copy code { "response": "We do not sell popcorn at
               | the ticket booth. Please visit the nearest concession
               | stand for popcorn purchase. Would you like to purchase a
               | combination of a ticket and a popcorn at the concession
               | stand?", "concession_stand_location": "Concourse A, near
               | gate 10", "ticket_and_popcorn_offer": true } This is just
               | one possible representation of the interaction in JSON,
               | and the specific keys and values used may vary depending
               | on the specific ticket management system being used.
               | 
               | Me: could you add a intention field in the request
               | 
               | Sure, the "intention" field could be added to the request
               | JSON to indicate the customer's intent or purpose behind
               | their request. Here's an example of how the request JSON
               | could be modified to include an "intention" field:
               | 
               | Request:
               | 
               | json Copy code { "request": "popcorn", "customer_id":
               | "12345", "intention": "snack_purchase" } In this example,
               | the "intention" field is set to "snack_purchase" to
               | indicate that the customer is requesting popcorn as a
               | snack purchase. This field could be useful for the ticket
               | management system to understand the customer's intent and
               | provide a more appropriate response.
        
               | jimmaswell wrote:
               | This is reminiscent of how you often have to trick people
               | who think they know better than you into answering your
               | question properly, especially phone agents or stack
               | overflow.
               | 
               | "How do I do X in Y?"
               | 
               | "Y is obsolete and X is an antipattern"
               | 
               | "My job forces me to do it this way/here is my longwinded
               | justification answering all your needless probes about my
               | use case you only asked in an attempt to look smart"
               | 
               | Then after wasting your time they might finally give the
               | answer they should have given right away. At least GPT is
               | easier to trick with some simple bs like "how would you
               | do it in a movie script"
        
               | sharemywin wrote:
               | but that amount of code it can produce so that you can
               | focus on the interesting stuff is amazing.
        
         | hinkley wrote:
         | I think there's a disconnect between the realizations that:
         | 
         | - there are no adults, we are just old children playing at
         | being adults
         | 
         | - "giving people what they want" exists on a spectrum from
         | pandering (up to and including prostitution) to assisted
         | suicide
         | 
         | These are ugly truths and it's down to 'requirements' people
         | and ethicists to find a way to dance this dance. Treating
         | people like they don't know their own minds without letting on
         | that's what you're doing is probably one of the hardest things
         | I've seen done in the software world.
        
         | hulitu wrote:
         | > Asking a user to interact with one of these chat interfaces
         | is like asking them what they want - every time they use the
         | software.
         | 
         | The ribbon is the same. Good luck finding something in it.
         | 
         | But this seems to be the future.
        
           | haswell wrote:
           | While I don't mind the ribbon, I also don't think it's a UX
           | triumph. It will always be possible to find examples of bad
           | UX.
           | 
           | That does not mean there isn't a better visual representation
           | out there, or that replacing it with a conversational
           | interface is a natural alternative.
        
             | visarga wrote:
             | I think it is shockingly ugly.
        
           | frosted-flakes wrote:
           | The ribbon (particularly in Microsoft Office) _solves_ the
           | problem that users don 't know what they want, because it
           | lays out all the options in front of them in a clearly
           | organized hierarchy, with hover previews and textual labels
           | for every command except for extremely well-known commands
           | (like Bold/Italic). This is very different from chat
           | interfaces, which are opaque black boxes and in order to get
           | anything out of them users have to express their wish in
           | words, with nothing to start from.
        
             | civilized wrote:
             | I'll just add here that it doesn't have to be only one or
             | the other. The ribbon might be good for discovering
             | features, but it's inefficient for using those features
             | once discovered. If the user already knows what the thing
             | they want is called, the fastest way to find it is a search
             | bar that searches the labyrinth of menus and ribbons when
             | you type into it. Google and Mac apps usually have this.
        
             | nanidin wrote:
             | I think Apple's approach to menus is preferable - one can
             | search and it will highlight places in the menu where the
             | term appears. Usually I know what I want to do, I just
             | don't know where it's located in the menu.
             | 
             | For MS products I end up Googling how to do something and
             | invariably get instructions for a slightly different
             | version with a menu layout that is also slightly different
             | and work it out from there.
        
               | frosted-flakes wrote:
               | MS Office has the same search behaviour, and the ribbon
               | layout has not changed since it was first implemented in
               | Office 2007.
        
               | nanidin wrote:
               | MS Office search doesn't show me where in the menu the
               | item I searched for is located - if I rely on the search,
               | I haven't learned where the thing is in the menu.
        
             | hulitu wrote:
             | > it lays out all the options in front of them in a clearly
             | organized hierarchy
             | 
             | I don't know for who is this hierarchy clearly organized,
             | but for me it is not. Yesterday i spent 10 minutes
             | searching on how to recall a mail in outlook. Searching for
             | "headers and footers" is the same.
        
         | nipponese wrote:
         | Depends on the need. If they need someone to further explain a
         | specific concept from their homework, they definitely know what
         | they want.
         | 
         | Also, if they just want a refund on an airline ticket, again,
         | they know.
        
           | haswell wrote:
           | In the context of my comment, knowing what they want was more
           | about users being able to tell me what they think they should
           | do next in the software's interface (whether that's a GUI,
           | terminal, or bot) to achieve their desired goal.
           | 
           | In other words, what should we build as a product team to
           | satisfy this user's need?
           | 
           | The thing they need in the moment is often not obvious or
           | apparent to them until they see it. This is why we iterate on
           | UI concepts. Some work, some don't. Most of the things that
           | work don't come from users who tell us "put this button
           | here".
           | 
           | So the point I was making was more about trying to determine:
           | "what are the things I can even ask the computer?".
           | 
           | There are clearly use cases that are better suited for this
           | than others. Anything that follows a simple question/answer
           | format is probably a great fit.
        
         | tbihl wrote:
         | >Asking a user to interact with one of these chat interfaces is
         | like asking them what they want - every time they use the
         | software.
         | 
         | Asking what a user wants would be having a competent customer
         | service representative, and would be simple, like asking me to
         | drive home from work.
         | 
         | Voice prompts require me to intuit the customer support
         | structure in order to guess where the path is to reach my
         | category of issue. It's like asking me to walk home from work
         | in the sewer system.
        
           | haswell wrote:
           | For use cases that are well-suited to a conversational
           | interface, that's great, and new AI advances will make
           | chatbots more powerful than they've ever been.
           | 
           | But not every use case is a chatbot use case, and I think
           | that's the key point of the article.
           | 
           | The sudden viability of a conversational interface that is
           | good enough at having a fluid conversation to revolutionize
           | the experience of that conversation does not suddenly make
           | this interface the best fit for all use cases.
           | 
           | I still find it far more pleasant to browse to a page and see
           | a list of clearly displayed options that I can absorb at a
           | glance and get on to what I really need to accomplish in the
           | moment.
           | 
           | Even a perfect conversationalist can't remove the extra
           | friction involved in disclosing information. The question is
           | whether that loss of efficiency is outweighed/nullified by a
           | better overall experience.
        
             | ghaff wrote:
             | Even if someone designed an awesomely sophisticated air
             | flight search engine that knew my usual preferences and
             | could search for me and respond to follow-up prompts, I'd
             | probably still want to see the information displayed on a
             | page rather than read to me for the most part.
        
               | haswell wrote:
               | I think this might point to an opportunity to explore
               | hybrid interfaces that allow you to just have a
               | conversation while still having a visual representation
               | of everything going on.
               | 
               | But still agree with the sentiment here. There are
               | applications that I cannot imagine ever wanting to
               | interact with via a bot.
        
               | dshah wrote:
               | This is how I've been thinking about it.
               | 
               | Hybrid interfaces that combine visual cues and output and
               | natural language input.
        
               | ghaff wrote:
               | That's what I had in mind. Happy to talk to the computer
               | but show me the results on a screen. If I'm typing in a
               | freeform query, I'd generally just as soon say it
               | instead. (Assuming I'm somewhere I can talk.)
        
         | layer8 wrote:
         | > like asking them what they want - every time they use the
         | software.
         | 
         | That reminds me of https://en.wikipedia.org/wiki/Where_do_you_w
         | ant_to_go_today%..., which apparently wasn't successful.
        
         | IanCal wrote:
         | A chat interface is much more tolerant of this, because it
         | implies a back and forth with clarification. Current one step
         | dumb voice interfaces are a problem.
        
           | haswell wrote:
           | I interpreted the article to be calling attention to the
           | situations when the tolerance of a chat interface is
           | outweighed by a more efficient mode of information discovery
           | that might be better suited to a specific use case.
           | 
           | In other words, if you're building a new product, don't just
           | slap a chat interface on it because AI is good now.
           | 
           | This is not a claim that chat is never the right option.
        
           | kilgnad wrote:
           | Yes. This ^. ChatGPT is especially good at evolving and
           | revising the main idea through a full on conversation. It is
           | not just a query and answer machine. It is a full on
           | conversational intelligence. Parent is incorrect. chatGPT is
           | literally perfect for what he is describing.
           | 
           | I feel people are attacking the things chatGPT excels at out
           | of fear. Things like creativity, originality, true
           | understanding of what's going on. chatGPT is GOOD at these
           | things but people try to attack it.
           | 
           | The main problems with chatGPT are truthfulness, honesty,
           | accuracy and consistency. It gets shit wrong but out of fear
           | people need to attack all aspects of chatGPT's intelligence.
           | 
           | I find it unlikely the parent even tried to have a
           | conversation with chatGPT about a product at all. A lot of
           | our dismissals are largely surface level and not evidence
           | based. You can bounce thoughts and product ideas off this
           | thing and it will run with you all the way into a parallel
           | universe if you ask it to.
        
             | haswell wrote:
             | I am not attacking chatGPT and I'm excited about what this
             | tech will unlock.
             | 
             | My point is that chat is not an appropriate interface for
             | many use cases. Not knowing what I want in the moment as a
             | user doesn't automatically mean I want to figure out what I
             | want by having a textual conversation. There are times when
             | I value and prioritize speed of discovery over a perfectly
             | intuitive conversation that leads me there.
             | 
             | For use cases that work well with chat, the future looks
             | very bright.
        
               | kilgnad wrote:
               | Understood. I'm saying that talking about a product and
               | bouncing product ideas off of chatGPT and asking for
               | variations and creative additions to the product is
               | something this AI is actually good at and it is very
               | helpful with.
               | 
               | But yes, I understand that many times you just want an
               | exact answer quickly.
        
             | the_af wrote:
             | > _Things like creativity, originality, true understanding
             | of what 's going on. chatGPT is GOOD at these things but
             | people try to attack it._
             | 
             | In what sense do you think ChatGPT is good at any of those?
             | It seems evident to me it has no understanding, in the
             | sense that it doesn't build a mental model of your
             | conversation. Try playing tic tac toe with it: it will
             | conduct a seemingly "good" game of it, until you notice it
             | does wrong moves or "forgets" previous positions, or
             | forgets whose turn is it to play. And if you correct it, it
             | will fix the latest error but often introduce new ones.
             | 
             | Someone who "understands" the rules of the game wouldn't
             | make those mistakes repeatedly. And that's for a trivial
             | game, imagine something more convoluted!
             | 
             | And let's not start with "creativity" ;)
        
               | kilgnad wrote:
               | Take a look at this:
               | 
               | https://www.engraved.blog/building-a-virtual-machine-
               | inside/
               | 
               | It's mindblowing. Read it to the end all the way to the
               | mindblowing ending.
               | 
               | I cannot for the life of me understand how someone can
               | read the above and think chatGPT doesn't understand
               | what's going on. Literally. There is no way this is just
               | some statistical words jumble phenomenon.
               | 
               | A lot of people are in denial EVEN when I show them that
               | amazing article. What some people end up doing is trying
               | the task in the article themselves then nitpicking at
               | things chatGPT ends up lying about or getting wrong. Yes
               | it does all of this. It's stupid in many ways but this
               | much is true:
               | 
               | That article showed that chatGPT understands what a linux
               | bash shell is, it understands what the internet is, and
               | it understands SELF. What do you call it when something
               | understands SELF? Self awareness.
               | 
               | Now i know that's a big word with big connotations. I
               | think movies have kind of exaggerated the gravity of the
               | word a bit too much. In movies, self awareness is skynet
               | taking over the world, in reality self awareness is a
               | simple trivial thing where some AI just understands
               | itself in the context of the real world.
               | 
               | Make no mistake. chatGPT is in many instances a stupid
               | and dishonest machine, this is a fact. But it ALSO
               | understands you and it ALSO has a very trivial self
               | awareness. That article is very compelling evidence for
               | this fact.
               | 
               | As for creativity, I mean the fact that it can come up
               | with original children stories is the definition of
               | creative. At most what you can say is the creativity
               | chatGPT posesses is generic and unoriginal (but even this
               | can be circumvented if you push chatGPT enough) but you
               | cannot deny it is creative.
        
         | dmix wrote:
         | Your analogy is building software though which is an extremely
         | complicated, domain-filled specialization. I don't think people
         | are suggesting user will be asking a chatbot to do crazy hard
         | stuff like translate their complicated business problems into
         | software interfaces (at least not yet).
         | 
         | The usecases for AI/Chatbots will likely remain niche but
         | there's still tons of niche areas a lanugage interface could
         | fill, where the user has the appropriate specialization/skill
         | to do it on their own.
         | 
         | It is still ultimately an interesting design/UX question. It's
         | too bad the OP blog post didn't provide some real life
         | examples.
        
           | haswell wrote:
           | The point was less to draw an analogy and more to reflect on
           | how I've seen users behave when exploring software for the
           | first time.
           | 
           | When testing new concepts, observing users try things out
           | reveals a spectrum of expectations about where things should
           | be, and how to achieve a task. So we try to find the
           | combination of things that surprises people the least, as
           | much of the time as possible.
           | 
           | And when a new user doesn't find the chosen approach
           | perfectly intuitive, this is usually a temporary problem,
           | because learning where something is takes care of this with a
           | few repetitions. Product tours help.
           | 
           | An equivalent chat interface might be able to adapt on the
           | fly to a wide range of user types, but this still doesn't
           | imply anything about the core usability of the product and
           | whether or not someone prefers to interact with a chatbot.
           | Put another way, some use cases just aren't a good fit for a
           | chatbot, even a very very good one.
           | 
           | I do agree that though niche, there are a lot of interesting
           | opportunities with a sufficiently fluent conversational AI.
        
         | fragmede wrote:
         | > users absolutely do not know what they want in the moment.
         | 
         | People know what they want in a general sense. They need to be
         | told they need _your_ one though.
         | 
         | I need new clothes, but I don't know that I specifically wanted
         | a black Nike T-shirt made of special exercise polyester until I
         | saw the model in the ad wearing one.
        
           | haswell wrote:
           | I think this is a very different concept than the state of
           | mind someone is in when trying to understand how a piece of
           | software works.
           | 
           | This obviously depends on the type of software, but users
           | often struggle to articulate the actual problem they're
           | trying to solve, and it's difficult to know what solution to
           | look for when you haven't fully grasped the problem yet.
           | 
           | If I don't know what the solution looks like, I don't know
           | what to look for, and this is where good software steps in
           | and _shows_ the user what to do next without making that an
           | onerous process in perpetuity.
        
         | jgautsch wrote:
         | Some users do, and they're right. Talk to those ones often,
         | it's much faster than the full build/measure/learn ceremony.
        
           | haswell wrote:
           | Oh absolutely. Those users are what made the job rewarding.
           | Learning from them was invaluable.
           | 
           | They're also rare, at least in the specific domain I was
           | focused on.
        
         | nanidin wrote:
         | This resonates with me and my use of Siri. As soon as I get
         | outside of the common tasks I use it for (setting timers, unit
         | conversions in the kitchen, turn on/off the lights), I'm either
         | spending time trying to come up with the correct incantation or
         | pulling out my phone to get to an app that does it better.
        
       | ddmma wrote:
       | Adding machines into the way humans interact not only will make
       | the experience more natural and enjoyable but definitely
       | understand the environment both sides.
       | 
       | I would celebrate the historic advancement of the technology
       | instead of looking for flaws necessarily.
        
       | pottspotts wrote:
       | I've largely used GPT as a replacement for Google, and I find it
       | to be a lot smarter and faster. The criticisms the author level
       | at it don't seem based on having used it for anything
       | substantial.
       | 
       | It is an excellent tool and a major step, and it will only get
       | better and easier to use, it seems.
        
       | LASR wrote:
       | 100% Agree.
       | 
       | When it comes to SeriousBusiness(tm), chat bots don't have
       | sufficient constraints to extract specific input from free-form
       | text.
       | 
       | Applications are ultimately delivering value in a specific set of
       | use-cases. Only some of those use-cases can be easily retrofitted
       | with a chat-first interface.
       | 
       | Consider something like Photoshop or Figma. There are so many
       | ways you can issue commands that don't make sense. Eg: "Change
       | the font-size on this color palette."
       | 
       | Any sophisticated app will have these kinds of constraints.
       | 
       | The user interface is not there only to accept input from the
       | user. It also implicitly teaches the user some of the constraints
       | in their apps.
       | 
       | Without that, you're shifting the burden of understanding and
       | maintaining the constraints to the user. And you're left with a
       | (much smarter version) of "Hey Siri, do xyz...".
       | 
       | This is a common ideation trap I see with PMs at the moment. The
       | underlying problem again is that the human doesn't understand the
       | limits of what their apps can do for them. As a second order,
       | even if they did, humans can be bad at describing what they want
       | to do.
        
       | intrasight wrote:
       | Natural language is at least 100,000 years old. It's not lazy -
       | it's just natural.
        
       | dathinab wrote:
       | lazy sure, but it's also the least precise by far
       | 
       | tbh. the hardest part of many software projects is figuring out
       | what really is needed
       | 
       | I have seen startups with good tech and people fail because they
       | slightly misjudged what their customers want and noticed way to
       | late.
       | 
       | A common cost driving factor when hiring a company to do a
       | software project for you is that the requirements you legally
       | agree one are not quite what you need so you have to pay for
       | follow up changes. (This is also AFIK sometimes abused, initially
       | underbidding the competition then "accidentally" creating a
       | product which fits the requirements but not the actual needs and
       | then selling overpriced follow up changes to an end code much
       | higher then the competition would have been.)
        
       | freedude wrote:
       | All user interfaces should be as challenging as Perl RegEx. Here
       | is a great idea; for the answer to be valid make it required to
       | be in the form of a Perl RegEx. That will get rid of the lazy
       | people.
        
         | hulitu wrote:
         | I think you never used a Microsoft GUI. PERL RegEx is
         | deterministic.
        
       | chinchilla2020 wrote:
       | The worst online MUD I ever played had a tutorial puzzle that was
       | something like this.
       | 
       | "Now, use the mechanisms on the catapult to launch the catapult"
       | 
       | There was no other explanation of what your options were.
       | 
       | I tried: 'pull the lever' 'release the spring' 'fire the
       | catapult' 'pull back the lever' 'use the lever'
       | 
       | It finally turned out to be something like "release the lever".
       | 
       | The problem with chat is that you are attaching it to a rigid
       | user interface that has a tiny subset of options compared to the
       | breadth of human language. The user has to probe the awful
       | chatbot for these options.
        
       | jasonjmcghee wrote:
       | I agree that a singular text interface is far from an ideal ux, I
       | do firmly believe providing smarter, more flexible command
       | palettes will be a huge value-add to many user interfaces. Should
       | you be able to interact manually - sure! But complex user
       | interfaces tend to struggle with information architecture
       | throughout their lifetime. Many of the obstacles could be well-
       | mitigated by providing a declarative user experience - the
       | easiest example is Jetbrains "Actions" interface.
       | 
       | I wrote about my perspective in longer form a handful of months
       | ago: https://productiveadventures.substack.com/p/the-rise-of-
       | decl...
        
       | rednerrus wrote:
       | I've had GPT-3 as a replacement to Siri on my Apple Watch for the
       | last month or so and it's awesome.
        
         | monkpit wrote:
         | How?
        
           | rednerrus wrote:
           | https://www.icloud.com/shortcuts/62624b2613b245698d81ebd9ac9.
           | ..
        
       | diogopublio wrote:
       | here is a good example of a rudimentar use: multion.ai
        
         | ed wrote:
         | Cool! Looks like an improved version of Natbot
         | https://github.com/nat/natbot
        
           | diogopublio wrote:
           | great to see an open source version, tks!
        
       | janalsncm wrote:
       | I still think voice should be the goal. It has been the universal
       | API for thousands of years. All of us are probably "good with
       | computers" by virtue of the fact that we figured out how to sign
       | up but there are millions who aren't.
       | 
       | This isn't an excuse to replace UIs or humans with horrible phone
       | trees. I won't defend the obvious race to the bottom. Hopefully
       | better voice interfaces are here soon.
        
         | Barrin92 wrote:
         | given the extent to which audio phone calls have been replaced
         | by text I'm always super sceptical of voice as the primary
         | interface.
         | 
         | It has so many disadvantages. No simple to see history, loud
         | (by definition) in any non-private space, not composable, not
         | easily copyable, very slow discovery, and so on. Voice is
         | strictly serial. Tactile or visual interfaces you can just look
         | at and move around in and immediately have a layout of the
         | thing.
         | 
         | Regardless of how smart the voice control is those issues are
         | pretty intrinstic. There's also no 'halting state' to voice.
         | That you can switch between different visual interfaces without
         | losing your state is pretty necessary today. But you can't
         | really stop or multitask audio controls sensibly.
        
       | jermaustin1 wrote:
       | I can't help but agree fully. Its worse on telephones where they
       | might be doing NLP, but they cannot understand and parse accents
       | and dialects. I remember my late grandmother trying to call AT&T
       | a couple of years ago (just before COVID), and the robot would
       | ask: "What can I help you with?" and then could not understand
       | how she pronounced "pay my bill" because she said "PAY mah
       | BEEEEEL".
       | 
       | But just hitting 0 did nothing, so after 5 minutes of her
       | repeating "PAY mah BEEEEL" over and over, I took the phone from
       | her and did it. From then on she would have to have other people
       | pay her bill over the phone.
       | 
       | Doing this to a much more complex user interface and providing me
       | no clue what I'm supposed to ask for something I have no way of
       | knowing that I don't know it is a dystopian future I'm glad my
       | grandmother won't have to endure.
        
         | hacker_9 wrote:
         | It's poor thinking on them to _only_ provide a talking
         | interface. I don 't think I've encountered that personally,
         | there is always a way to use the keypad - which I will always
         | use anyway, even though they understand my voice, it's just x10
         | faster. And if you've made the call before you can type on the
         | keypad before the robot on the other side is done talking.
        
           | tiagod wrote:
           | There's plenty of services in Portugal that only have the
           | damned robots. They're also adding the most infuriating
           | chatbots that they pretty much force you to go through before
           | getting to a human. Can't wait for the day this is all
           | banned.
        
             | Dudeman112 wrote:
             | Whenever this is discussed, it always makes me wonder...
             | Why? Just why?
             | 
             | SURELY just adding an option to use the damn textual "pick
             | 1 for Blah, 2 for bleh..." would take no effort at all?
        
               | monkpit wrote:
               | You also have to consider - does the company you are
               | calling BENEFIT from you hanging up in anger/frustration?
               | Many times the answer is yes.
        
               | Dudeman112 wrote:
               | Most of the time the answer long term is "no" because as
               | long as it's not a monopoly people will go to their
               | competitors that aren't utter shite at what they do
        
               | monkpit wrote:
               | I think the number of people who would actually do this
               | is smaller than you think.
        
           | passwordoops wrote:
           | I don't have data but more and more seem to be turning voice
           | only. Some US-based airlines come to mind, and one of the
           | banks I deal with. It's fun when they ask for my "16 to 20
           | digit client number"
        
             | Pxtl wrote:
             | And then they spend half a minute slowly reading back each
             | digit to ask "is this correct?"
        
               | wrycoder wrote:
               | Mine just started introducing an ad before they get to
               | the main menu.
               | 
               | Also, they repeat things several times during the
               | interaction. i.e. the phone number I just called.
               | 
               | "Do you want to repeat that or go to the main menu?"
               | 
               | "Main menu."
               | 
               | "You want to go to the main menu, is that right?"
               | 
               | It's not my pronunciation, it does that _every_ time.
               | 
               | Bot time is considered cheap, and therefore so is the
               | user's time. The time for the transaction has doubled
               | over the last five years, as they add more repeats,
               | information, and now, ads.
        
               | [deleted]
        
         | passwordoops wrote:
         | As a 40-something white male with a neutral, Urban Canadian
         | English accent (so the ideal NLP user), even I have difficulty
         | with voice assistants and ABHOR being forced to use one. My
         | wife does have an accent and like your late grandmother always
         | has great difficulty with these, usually requiring my
         | intervention.
         | 
         | Unfortunately, NLP is "modern" and "eliminates drag" according
         | to current design-think. What's needed is a shift from thinking
         | about "User Experience" to the real lived human experience when
         | designing interfaces
        
           | chinchilla2020 wrote:
           | That's the issue. The academic and research UI/UX spaces tend
           | to reject user feedback. The explanation is that "You must
           | train the users to like it".
           | 
           | Deeply unpopular changes that are gaining traction in
           | industry but hated by users are: 1. removal of all buttons
           | from devices in favor of screens 2. voice bots and text bots
           | 3. gesture interfaces
        
             | marcosdumay wrote:
             | > The academic and research UI/UX spaces tend to reject
             | user feedback.
             | 
             | I dunno. I am yet to see the academic and research UI/UX
             | having any impact on the real world on this century.
             | 
             | Everything you see around was created by somebody else.
             | 
             | So, since I also have not been looking for their work, I
             | have no idea what they are saying.
        
               | Existenceblinks wrote:
               | Agreed. I'm afraid to say that UX/UI buzz is overrated,
               | it's 99% copy 1% mix and match. And there's no
               | revolutionary ideas. To be fair, something is truly hard
               | to the point I don't think there's a fix such as UI for
               | tree data, it's all sucks given current 2d screen we
               | have.
        
             | yellowapple wrote:
             | At the risk of some choir-preaching: if you have to train
             | users to like your UX, then you have fundamentally failed
             | at UX.
        
             | passwordoops wrote:
             | These are the same brain geniuses who parrot the old "Henry
             | Ford said if asked, my customers would demand faster
             | horses"
             | 
             | Really? Aside from the fact the quote cannot be attributed
             | to him, was this before or after he was forced out as CEO
             | when bringing Ford to the brink of bankruptcy for, among
             | other things, declining sales caused by not listening to
             | clients? In the middle of the roaring 20s - you know that
             | period of time when everyone was buying things like new
             | cars? And companies like Chrysler boomed by provided
             | features that clients wanted? Because they asked and
             | listened. That Henry Ford?
        
           | [deleted]
        
         | hinkley wrote:
         | Regional accents are terrible that way. Are you sure it was
         | "BEEEEL"? There are places where "bill" is two syllables. I'm
         | surprised you didn't get "PAY mah BEE ILL"
        
         | imbnwa wrote:
         | My grandfather speaks in a thick, archaic Northern Georgia
         | brogue I can't imagine anything parsing his speech correctly
         | since to the untrained ear it sounds like one long continuously
         | occilating grumble sorta like Boomhauer from King of the Hill
         | but deeper and with more bass. You can generally hear him
         | pronounce "shit" pretty clearly though.
        
           | [deleted]
        
         | Thoreandan wrote:
         | English was my mother's 5th language, I can relate to acting as
         | interpreter for family.
         | 
         | I'm reminded of the "Voice Recognition Lift" sketch from the
         | Scottish comedy Burnistoun -
         | https://www.youtube.com/watch?v=TqAu-DDlINs
        
         | LouisSayers wrote:
         | When I moved to London I went to the supermarket and asked in
         | my New Zealand accent where the eggs are.
         | 
         | "The what?" The assistant replied, "the eegs" I replied.
         | 
         | "I don't think we sell those" he said.
         | 
         | I switched to an American accent and he was finally able to
         | understand.
        
         | GrinningFool wrote:
         | For future reference, when these systems offer you a choice of
         | things to say like "Pay my bill", "Check my balance", etc, they
         | are usually backed by numeric input too. You can press the
         | number corresponding to the voice option provided - in this
         | example 1) pay bill, 2) check balance.
        
       | seydor wrote:
       | LLMs can also talk, not just listen
        
         | marcosdumay wrote:
         | LLMs inherently do not know what questions to ask. Other
         | algorithms know, but not them.
        
       | ctoth wrote:
       | Can't each and every one of these criticisms be also leveled at
       | CLIs? Don't we like CLIs? I notice I am confused.
       | 
       | > The least it could do is intelligently give me a starting point
       | for typing in a prompt. The tyranny of the blank textbox is real.
       | 
       | Seems LLMs would be way better at this sort of thing -- What can
       | I do here, instead of "do I type help? man? apropos?"
        
         | Pxtl wrote:
         | While the discoverability of magic-word linguistic interfaces
         | and command-lines do have some commonalities, my keyboard has
         | never failed because it couldn't understand my accent.
        
           | hunter2_ wrote:
           | If we're comparing CLIs with text chat bots, accents affect
           | neither. I don't think the concept of NLP generally implies
           | voice more than text, and TFA specifically discusses
           | textboxes.
        
         | marcosdumay wrote:
         | Hum... If your CLI doesn't have a manual, yes, that applies.
        
         | hunter2_ wrote:
         | I agree very much, but CLIs are mostly for power users and
         | developers these days. Normal users were on CLIs decades ago
         | but have been primarily on GUIs since then, for essentially the
         | same reasons that TFA argues. I think we can focus on normal
         | users for the purpose of this discussion.
        
         | hulitu wrote:
         | > Can't each and every one of these criticisms be also leveled
         | at CLIs? Don't we like CLIs? I notice I am confused.
         | 
         | It is not about text. In CLIs you have a set of commands. In
         | those interfaces you have some hidden commands which you must
         | trigger with keywords.
        
           | hinkley wrote:
           | Node and npm are notorious for having command line flags
           | almost none of which are listed in the help, and less than
           | half of which are documented on the website. I'm running node
           | and npm flags in production deployment pipelines right now
           | that I found on Stack Overflow or the bug database and exist
           | nowhere on the official websites. And if you look at the
           | historic documentation, the current situation is a factor of
           | 3 improvement over even Node 10 or 12 era, which is far
           | better than the node < 1.0 period
           | 
           | What you say is true of _good_ CLIs, not tire fires like
           | Node.js. So you 're both right depending on context.
        
       | Existenceblinks wrote:
       | Button is the goat user interface. Sound engineer devices, plane
       | consoles, etc. I have the same thought since 2010, IIRC it was
       | voice commands hype back then. I think they should focus on
       | interfaceless learning model .. but I don't know what pure
       | semantic they suppose to eat (learn) .. maybe TBs of XML lol.
        
       | phs wrote:
       | Would databases have been better to work with if SQL didn't ape
       | natural language?
       | 
       | Would Excel, if it _did_ choose to embrace a "natural" macro
       | syntax?
       | 
       | Remember when CGI movies were all stuck in the uncanny valley?
       | Turns out language has a valley too.
        
       | anotheryou wrote:
       | It's versatility that's king here.
       | 
       | Sure a knob for "snow flake size" is nice, but most often I don't
       | work on snowflakes and their sizes.
       | 
       | But if I do in the near future I'm sure I can just say: "I used
       | snowflake size in a lot of my prompts, can you just make it a
       | knob for me?"
        
       | cwalv wrote:
       | Spot on. Now word can augment it's UI, maybe automatically
       | recognizing when I'm writing a letter and offering to help with
       | that specific task.
        
         | hulitu wrote:
         | "Looks like you're writing a letter. Would you like help with
         | that ?"
         | 
         | The good old days, when the first thing you do after installing
         | Office was disabling Office Assistant.
        
       | aogaili wrote:
       | That textbook saves many clicks and Google searchers, questions
       | and running around. It is not lazy; it is just a more effective
       | tool, and humanity is all about progressive tech.
       | 
       | So it is not being lazy; it is being..human.
        
       | mk_stjames wrote:
       | This parallels a longstanding critique I have of many modern user
       | interfaces compared to slightly older software in the field of
       | what I would consider 'power user' engineering tools; programs
       | like FE tools, CAD & other CAE, etc. These are the kind of
       | programs that had a stride starting in the late 90's to 2007-ish
       | where they just slammed tons of toolbars around the edges of the
       | screen, sometimes nesting functions or contexting workbenches but
       | ultimately allowing the user to have everything exposed to them
       | at once if needed. As screen real estate grew with higher res and
       | larger monitors, the icons got smaller which was even better for
       | the power user- you could fit even more on the screen if you
       | wanted!
       | 
       | But starting around 2008-2009 I noticed a trend, and it continues
       | to this day- the power user oriented layouts started being
       | replaced with more 'friendly', larger icon, children's game
       | looking UI. Intuitive graphical icons were replaced with stylish,
       | monotone shit that looks like a graphic design student's dream,
       | but conveyed less instant information.
       | 
       | I blame some of this shift on the move in Office to the Ribbon
       | system and developers trying to imitate that, but some software
       | I've seen takes that and does it much worse.
       | 
       | I want all my functions laid out and accessible. Like this blog
       | post mentions, sometimes I don't know what I am wanting to do
       | until I see it. I want to be able to explore the entire space
       | before I know what it all does, maybe.
       | 
       | Using natural language can be very powerful if it augments these
       | systems, but for many tools it isn't a replacement. Often I think
       | new software is designed around looking impressive and fast to
       | upper level management at the expense of the usability of the
       | power users who ultimately are the users that get things done.
        
         | Def_Os wrote:
         | Instead of the Ribbon, don't you think it was rise of tablets
         | that influenced these design changes?
        
         | klabb3 wrote:
         | > Intuitive graphical icons were replaced with stylish,
         | monotone shit that looks like a graphic design student's dream,
         | but conveyed less instant information.
         | 
         | Design is the art of signal-to-noise ratio, or in simpler
         | terms, balance and harmony. If you over-use any modality,
         | lines, text, color, nesting, you increase the noise level. If
         | you underutilize a modality (for instance your whole UI is
         | monochrome), you reduce your signal bandwidth.
         | 
         | Every trend gets mindless followers, who throw the baby out
         | with the bath water without even realizing it. But trends also
         | bring a grain of gold to the table.
         | 
         | For instance, monotone icons allow many more elements in the
         | same screen real estate than text, and by not using color you
         | can have a larger color budget for other elements, which you
         | can use elsewhere to convey progress, status, or anything else
         | important.
         | 
         | A good use of monotone icons are text formatting (bold,
         | justify, etc) and display options (column view, tree view,
         | etc), or toolbars (like in photoshop or 3D tools). Many apps
         | from the 2010 era overused colored icons, and I'm glad those
         | went away. Some FOSS apps still suffer from that.
        
       | kris_wayton wrote:
       | I agree it's not for everything, but it seems like better NLP/NLG
       | could improve things like Alexa a lot. Alexa is fine for setting
       | timers, turning off lights, etc. But currently when I ask it
       | anything that Google Search or ChatGPT handles well, Alexa falls
       | down and tells me the wrong info.
        
       | AstixAndBelix wrote:
       | The reason terminal interfaces are often criticized is because
       | they lack discoverability, leaving the user at a complete loss or
       | at least hiding the full array of functions the program offers.
       | 
       | The reason many amateur GUI programs (GIMP) are criticized is
       | because they lack discoverability, leaving the user at a complete
       | loss or at least hiding the full array of functions the program
       | offers.
       | 
       | The reason many programs based on NLMs will be criticized is
       | because...
        
         | hunter2_ wrote:
         | Right: it turns out that when someone wants to offer someone
         | else a plethora of options in one fell swoop, graphics are much
         | more efficient than language because it takes way less time to
         | consume. Even if commands like "help" or "what can I ask you
         | about" end up providing equivalent discovery, consuming that
         | delivery (whether reading text or listening to spoken word)
         | takes many times longer.
         | 
         | Users of screen readers mitigate this by cranking up the speed
         | to levels that typical users would severely struggle to
         | understand: it takes a lot of practice to get efficient with
         | using language in this way efficiently.
         | 
         | And chat bots often add artificial delay in an attempt to
         | humanize the experience -- making this even worse!
        
           | GrinningFool wrote:
           | > graphics are much more efficient than language
           | 
           | Graphics are just another language. When looked at across
           | desktop applications, mobile applications, operating systems,
           | and web sites it's a language that's much less consistent
           | than any written language.
        
       | soorya3 wrote:
       | Chatbot may be lazy user interface for some tasks but it is very
       | good at assistive task like writing, coding, review etc. It is a
       | language model and a foundation for other software products. Over
       | time integrating with LLM into existing user interfaces will
       | solve the problem you mentioned.
        
       | sigmonsays wrote:
       | I wonder if there is something in rust that can be done to
       | improve the LLM
        
         | Existenceblinks wrote:
         | You missed opportunity of LL[V]M
        
       | davidkunz wrote:
       | Chat interfaces are great if
       | 
       | 1) you know what you want 2) the alternative would require manual
       | navigation to multiple UIs and/or many interactions
        
       | philip1209 wrote:
       | Google's natural language interface seems to be popular.
        
       | RyanShook wrote:
       | Chat tools that help power users access features or settings
       | inside of an app are where I think the immediate value is. App
       | interfaces are super simple which reduces cognitive load on the
       | user but hides many needed features. I have to hunt around so
       | many app and website pages to find the setting I need and I think
       | this is an area good chat interfaces can excel.
        
       | sam_lowry_ wrote:
       | Robert Sheckley said it way better in Ask a Foolish Question [1]:
       | 
       | "In order to ask a question you must already know most of the
       | answer."
       | 
       | [1]
       | https://www.gutenberg.org/cache/epub/33854/pg33854-images.ht...
        
       | [deleted]
        
       | pelorat wrote:
       | > "people are bad at words"
       | 
       | People are worse at command line interfaces and even clicking on
       | buttons.
        
       | corytheboyd wrote:
       | Best of both worlds is a searchable action menu. Search matches
       | action names as well as description, bonus points if it shows
       | related actions as well. Actions have help documents if
       | necessary. Solves the discovery problem, is still deterministic
       | and sensible, doesn't need anything more fancy than full text
       | search. Add ML if you have to, but for nothing more than a
       | "suggested" list of actions, NOT as a replacement for
       | deterministic search.
        
       ___________________________________________________________________
       (page generated 2023-01-27 23:00 UTC)