[HN Gopher] Good Old Fashioned AI is dead, long live New-Fangled AI
       ___________________________________________________________________
        
       Good Old Fashioned AI is dead, long live New-Fangled AI
        
       Author : isomorphy
       Score  : 61 points
       Date   : 2022-11-15 19:36 UTC (3 hours ago)
        
 (HTM) web link (billwadge.com)
 (TXT) w3m dump (billwadge.com)
        
       | shrimpx wrote:
       | It could affect commercial artist deeply, like game artists and
       | commercial illustrators making logos and icons and whatnot.
       | 
       | But it won't affect studio artists at all. Studio art is not
       | about "the image", it's about the practice, physical qualities of
       | the artifacts, and an ongoing evolution of the artist.
        
         | tomrod wrote:
         | The subreddit for stablediffusion has several examples of high
         | quality stitching of SD-generated images. If I am interpreting
         | your vernacular of "studio artist" correctly, then yes, studio
         | artists will be affected.
         | 
         | Artists that produce a "real" medium like charcoal, sculpting,
         | etc. aren't directly affected yet, but could be in the future.
         | 
         | As always, there is a power law distribution when it comes to
         | perceived value. It will be interesting to see how this
         | evolves.
        
       | time_to_smile wrote:
       | People spend an awful lot of time talking about current successes
       | in AI without often reflecting on how much (or little actually)
       | AI impacts their lives. Despite all of the energy put into
       | current gen AI, as far as every day impacts the biggest things I
       | can think of are:
       | 
       | - Spam filtering/email sorting
       | 
       | - Web search
       | 
       | - GPS/Wayfinding
       | 
       | - Voice assistants
       | 
       | These are the only practical applications of "AI" that I use more
       | or less everyday (I'd be happy to reminded of others). Of these 4
       | I personally have found spam filtering to be getting _worse_
       | recently, as well as web search. The first 3 were all more or
       | less solved over a decade ago, and, while I find Siri convenient,
       | I wouldn 't mind much at all if voice assistants completely
       | disappeared tomorrow.
       | 
       | I'm not denying we've had an amazing decade of pushing the
       | research needle further. There have been tons of impressive AI
       | projects out there. However the practical, day-to-day
       | improvements we've seen with the existence of AI seem to be few
       | and far between, and this is even more true when you start asking
       | about any AI work done in the last decade. I was happier with the
       | state of "AI" in my life in 2006 than I am today.
       | 
       | I just find it a bit fascinating how much energy has gone into
       | both generic data science as well as more serious AI research and
       | yet how little the reach of AI has grown in the last 10 years.
       | All of the cool AI that I use existed _before_ data science was
       | declared the  "sexiest job".
        
         | flooo wrote:
         | Most novel AI serves the user perfectly well. And since this
         | type of AI requires lots of labelled data, these users are
         | typically large data harvesting organisations.
         | 
         | Some other applications that you may use daily are translate
         | and face unlock/recognition.
         | 
         | It's interesting that you mention search and spam filtering
         | which both include an adversarial component. It seems to me
         | that the adversarial AI has become better, in line of
         | expectation from the democratisation of AI tools and knowledge.
        
         | tomrod wrote:
         | You might be surprised where AI shows up.
         | 
         | Use a credit card? Fraud monitoring, KYC, and other financial
         | models run through (e.g. Early Warning service).
         | 
         | Log into a website? Application monitoring with anomaly
         | detection.
         | 
         | Own a 401k with shares in a financial vehicle like an ETF? AI
         | used to predict the market for in-the-money trades.
         | 
         | Gone to the ER? Risk levels of mortality, sepsis, etc. are
         | constantly pushed to your medical record (in many top-tech
         | hospitals, like Parkland Hospital in Dallas and similar).
        
           | woopwoop wrote:
           | Do any of those applications use neural nets in any non-
           | trivial way? I'm pretty sure that kind of stuff is all
           | classical statistical modeling.
        
         | vagrantJin wrote:
         | I agree with your take mostly. But also, some things are
         | getting better work-wise like video/image editing. Where
         | something can take an animator a week or two now takes all of 1
         | minute or less. Some startups in the motion capture space are
         | doing some wild things and in a few years even small indie game
         | studios will have mocap parity with AAAs.
        
       | falcolas wrote:
       | As always, they mis-spelled the acronym for "Machine Learning".
       | There's nothing "Artificial" or "Intelligent" here but a
       | mathematical algorithm operating on an algorithmically-encoded
       | dataset.
       | 
       | If anything, it's closer to an encryption algorithm where the
       | keys can decrypt deterministic parts of the plantext from the
       | cyphertext and soften the edges a bit.
        
         | IanCal wrote:
         | This is a long lost battle since AI has been a term used to
         | describe far simpler things than that for over 60 years.
        
         | drdeca wrote:
         | While I don't really endorse it, I understand the objection to
         | a name with the word "Intelligence" as part of it.
         | 
         | I don't understand the objection to the word "Artificial".
         | 
         | Why do you say that there's nothing "Artificial" about these
         | programs. Now, they may be contexts in which you could call a
         | program "natural" in the sense of, "the natural way to do
         | something", but, at the same time, are not all computer
         | programs, in a different sense, artificial?
        
         | short_sells_poo wrote:
         | I like to equate it to a lossy compression.
        
           | nl wrote:
           | Compression _is_ intelligence[1] - which goes to point out
           | exactly how misguided the OP 's attempted distinction between
           | ML and AI is.
           | 
           | [1] http://prize.hutter1.net/#motiv
        
           | mattnewton wrote:
           | Auto encoders really hit this home for me- you are trying to
           | find an efficient compressed notation of the dataset, and the
           | most efficient way to do that hopefully ends up learning
           | useful rules about the data.
        
       | wwwtyro wrote:
       | Why do the eyes in the generated images always look a little off?
       | Most facial features usually appear photorealistic to me, but the
       | eyes always have a little smudge or something in them that gives
       | them away.
        
         | tjnaylor wrote:
         | https://blog.oimo.io/2022/10/21/human-or-ai-walkthrough/
         | 
         | This is a blog that investigates ai eyes and particular and how
         | to distinguish them from human artist made eyes.
         | *However, in the case of AI painting, it will almost certainly
         | change the coloring of the left and right eyes and how to add
         | highlights . Humans can understand that ``the left and right
         | eyes have the same physical shape and are placed in the same
         | situation, so there is naturally a consistency there. '' I
         | don't understand the theory "I don't really know what an eye
         | is, but it's something like this that's placed around here,
         | isn't it?" Still, it looks like it, so humans can recognize it
         | as eyes, but there are still many defects in the details.
         | Among them, the most distinctive feature is the " highlight
         | that melts into the pupil and breaks the pupil ". Humans know
         | that ``first there is the eyeball, there is the pupil in it,
         | and then the surrounding light is reflected to form a gloss'',
         | so ``the highlight does not block part of the pupil. It can be
         | understood as a matter of course that the shape of the pupil
         | itself does not collapse, even if the AI     does the same, but
         | AI that learns only by looking at the final illustration can
         | understand the ``logical relationship between the whites of the
         | eyes, pupils, and highlights''. I don't recognize anything . Or
         | rather, I can't. I didn't give it as data.            The
         | unnatural deformation of the pupil is also one of the judgment
         | materials. Humans know that "the pupil is originally a perfect
         | circle", but AI trained by looking only at the final completed
         | illustration does not know "the original shape of the pupil" .
         | Therefore, such an error occurs.            Another feature of
         | AI drawings is that they often subtly change the color of the
         | left and right eyes . Of course, there are characters with
         | different eye colors on the left and right (heterochromia), but
         | in most cases , characters designed that way can be clearly
         | recognized as having different colors . It is one of the
         | criteria for judging that the colors are similar at first
         | glance, but if you take a closer look, they are different.
         | However, even if there is such a character, it is not strange,
         | so it is not an important basis. Also, it is natural for the
         | color of the left and right eyes to change depending on the
         | surrounding environment, so be careful not to make a mistake.*
        
         | KaoruAoiShiho wrote:
         | Maybe because the training data have a lot of bad photos with
         | the red dot in the eyes.
        
       | triska wrote:
       | The "new-fangled" AI, as the article calls it, is often useful
       | when the stakes are low, and you can accept mistakes in outcomes.
       | Examples of such applications are: trying to determine which of
       | your friends occur in a photo, which movies a subscriber _may_ be
       | interested in, or which action _could_ lead to victory in a
       | computer game. Getting a rough translation of a newspaper entry,
       | as mentioned in the article, is also a good example.
       | 
       | As soon as you need reliable outcomes, such as certainty
       | _whether_ an erroneous state can arise in a program, _whether_ a
       | proof for a mathematical conjecture exists, or _whether_ a
       | counterexample exists, exhaustive search is often necessary.
       | 
       | The question then soon becomes: How can we best delegate this
       | search to a computer, in such a way that we can focus on a clear
       | description of the _relations_ that hold between the concepts we
       | are reasoning about? Which symbolic _languages_ let us best
       | describe the situation so that we can reliably reason about it?
       | How can we be certain that the computed result is itself correct?
       | 
       | The article states: _" The heart of GOFAI is searching - of trees
       | and, more generally, graphs."_ I think one could with the same
       | conviction state: "The heart of GOFAI is reasoning - about
       | relations and, more generally, programs."
        
         | bade wrote:
         | Well said
        
         | skissane wrote:
         | > As soon as you need reliable outcomes, such as certainty
         | whether an erroneous state can arise in a program, whether a
         | proof for a mathematical conjecture exists, or whether a
         | counterexample exists, exhaustive search is often necessary.
         | 
         | Proof checking requires 100% reliability. But if you are
         | searching the space of all possible proofs for a valid one,
         | that process does not require 100% reliability. On the
         | contrary, automated theorem provers rely on heuristics to guide
         | their exploration of that space, none of which work 100% of the
         | time. "Exhaustive search" is an infeasible strategy, because
         | the search space is just too large. Finding proofs is the
         | really hard part (NP-hard), and the part which most stands to
         | benefit from "AI" techniques - checking their validity is a lot
         | easier (polynomial time).
         | 
         | "New AI" deep-learning techniques can be used to augment
         | automated theorem provers, by giving them guidance on which
         | areas of the search space to target - see for example
         | https://arxiv.org/abs/1701.06972 - that produced a seemingly
         | modest improvement (3 percentage points) - but keep in mind how
         | hard the problem is, a 3 percentage point improvement on a very
         | hard problem can actually be a big deal - plus I don't know if
         | any more recent research has improved on that.
        
           | _carbyau_ wrote:
           | The idea of chainlinking "AI to guide/choose AI for the next
           | step" is where I expect more impressive results in future. It
           | will be important to understand the limitations of AI to be
           | sure of proper placement.
        
         | falcolas wrote:
         | So, there's an area of research that's under way called "AI
         | Assurance" which seeks to answer many of these questions.
         | 
         | Some things they're attempting:
         | 
         | - Creating explainable outcomes by tracing the inner works of
         | ML models.
         | 
         | - Looking for biases in models using random inputs & looking
         | for biased outputs.
         | 
         | - Using training sets with differently weighted models to find
         | attacks and biases.
         | 
         | etc.
        
           | nonrandomstring wrote:
           | The tragedy is that GOFAI did all these things as built-ins.
           | Procedural expert systems have been doing introspection,
           | backtracing, declaring confidence intervals etc since the
           | 1960s. Layering "assurance" on top of inherently jittery
           | statistical/stochastic and neural systems seems to
           | misunderstand how these models evolved, where they come from
           | and why there are alternatives.
        
             | ShamelessC wrote:
             | Another tragedy, GOFAI hasn't made a dent on any of these
             | problems for a very long time.
        
         | thwayunion wrote:
         | _> As soon as you need reliable outcomes, such as certainty
         | whether an erroneous state can arise in a program, whether a
         | proof for a mathematical conjecture exists, or whether a
         | counterexample exists, exhaustive search is often necessary._
         | 
         | Checking proofs is easier than finding proofs.
         | 
         |  _> The question then soon becomes: How can we best delegate
         | this search to a computer, in such a way that we can focus on a
         | clear description of the relations that hold between the
         | concepts we are reasoning about? Which symbolic languages let
         | us best describe the situation so that we can reliably reason
         | about it?_
         | 
         | These questions are largely answered. Or, at least, the
         | methodology for investigating these types of questions is well-
         | developed.
         | 
         | I think the more interesting question is co-design. What do
         | languages and logics look like when they are designed for
         | incorporation into new-fangled AI systems (perhaps also with a
         | human), instead of for purely manual use?
        
           | [deleted]
        
           | [deleted]
        
         | shafoshaf wrote:
         | The only barrier for higher stakes applications is going to be
         | the frequency of errors. Flying an airplane or running a
         | factory has a lot less margin for error, but humans don't do
         | those things perfectly either (Chernobyl, Three Mile Island,
         | Union Carbide-Bhopal disaster). It doesn't have to be perfect,
         | just better than humans. And in fact, I'd argue that by having
         | no deterministic outcomes prevents systemic failure, like
         | having a single point of failure for all those drones in the
         | crappy episodes of Star Wars.
        
           | MonkeyMalarky wrote:
           | There's a weird bit of induced demand like widening a
           | freeway, makes errors less often than a human but is more
           | scalable so the absolute number of errors increases. I guess
           | in the case of self driving cars, it could be from hordes of
           | autonomous shipping trucks that outnumber existing truck
           | drivers.
        
           | triska wrote:
           | In addition to the _frequency_ , it is also about the
           | _magnitude_ of unintended consequences. As an example,
           | consider the Therac-25 accidents:
           | 
           | https://en.wikipedia.org/wiki/Therac-25
           | 
           | Not only was the software not perfect, it was so erroneous
           | that the patients were struck with approximately 100 times
           | the intended dose of radiation.
           | 
           | Such extreme outliers should be completely and reliably ruled
           | out in safety-critical applications, even if they occur only
           | very rarely.
        
           | setr wrote:
           | > The only barrier for higher stakes applications is going to
           | be the frequency of errors.
           | 
           | Frequency and strength. My issue with e.g. image classifiers
           | is that when they're wrong, they're _catastrophically_ wrong
           | -- they don't misidentify a housecat as a puma, they
           | misidentify a cat as an ostrich.
        
           | nradov wrote:
           | But there's the rub. It's impossible to determine through
           | testing whether a particular AI system will actually have a
           | lower frequency of errors than humans. You can program an AI
           | system to handle certain failure modes and test for those in
           | simulation. But complex systems tend to have hidden failure
           | modes which no one ever anticipated, so by definition it's
           | impossible to test how the AI will handle those. Whereas an
           | experienced human can often determine the correct course of
           | action based on first principles.
           | 
           | For example, see US Airways Flight 1549. Airbus had never
           | tested a double engine failure in those exact circumstances
           | so the flight crew disregarded some steps in the written
           | checklist and improvised a new procedure. Would an AI have
           | handled the emergency as well? Doubtful.
        
           | joe_the_user wrote:
           | _The only barrier for higher stakes applications is going to
           | be the frequency of errors._
           | 
           | IE, "The only barrier to the software working perfectly is
           | it's tendency to fail".
           | 
           | Which is to say this sort of argument effectively assumes,
           | without proof, that are no structural barriers to improving
           | neural network performance in the real world. The thing is,
           | the slow progress on self-driving cars shows that reducing
           | the "frequency of errors" can turn from a simple exercise in
           | optimizing and pumping in more data to a decades long debug
           | process.
        
         | blueyes wrote:
         | GOFAI was never more than a rules engine. If-then statements.
         | 
         | Agree with you about probabilistic AI being useful in low-
         | stakes situations, at least at first.
        
           | mistrial9 wrote:
           | If-then questions can lead to non-deterministic outputs with
           | some simple feedback systems
           | 
           | not disagreeing completely, but.. both "questions that are
           | reasoned about", and "the code that reasons about
           | questions".. need more careful classification in order to
           | make use of these new data methods..
           | 
           | personally, I see the hype on DeepLearning to solve "find
           | pattern in varying digital content" that is so clearly useful
           | to the FAANG content Feudal Lords, is engaging in an investor
           | shouting match that paves over simple use cases where
           | DeepLearning is really not appropriate.
        
           | thwayunion wrote:
           | _> GOFAI was never more than a rules engine._
           | 
           | GOFAI "aged out" of the AI label and became: compilers,
           | databases, search algorithms, planning algorithms,
           | programming languages, theorem proving, a million things that
           | are still commonly used in NLP/CV/robotics, comp arch, etc.
           | Aka, most of what makes computers actually useful.
           | 
           | If something's over 30 years old and is still called AI,
           | that's just shorthand for saying it's a failed idea that in
           | the best case hasn't had its moment yet.
           | 
           |  _> If-then statements._
           | 
           | 99.999% of the software I use is just if-then statements.
           | 
           | (Also, this is like saying that deep learning is Linear
           | Algebra.)
        
           | cscurmudgeon wrote:
           | Sorry, thats really wrong.
           | 
           | I wouldn't call theorem provers just "if-then" statements. By
           | that logic, everything, even large models, are if-then
           | statements.
        
       | mxwsn wrote:
       | The author undersells himself - AlphaGo basically searches the
       | game tree the same way he describes of GOFAI. Monte Carlo Tree
       | Search is old yet essential. The neural network mainly improves
       | the game evaluation heuristic, using function approximation which
       | I'm sure the author is familiar with. Modern AI abilities are
       | mind boggling but they're not that complicated to understand,
       | especially with a GOFAI background!
        
       | jamesgreenleaf wrote:
       | > The image generator seems to understand that you can't see
       | through opaque objects
       | 
       | I thought this isn't the case for Stable Diffusion. Wasn't it the
       | humans making the source images who understood things like that,
       | and their knowledge became encoded in the latent space of the
       | model? I'm not an expert. Please correct me here.
        
         | _carbyau_ wrote:
         | Hmm. Wonder what "astronaut riding a glass horse" would do
         | then?
        
       | Barrin92 wrote:
       | >The heart of GOFAI is searching - of trees and, more generally,
       | graphs.
       | 
       | and because of this GOFAI, unlike the title suggests, and
       | algorithmic solutions will continue to underpin a huge chunk of
       | applications that don't fall in any 'natural' or generative
       | domain. When you have an algorithmically optimal or closed form
       | mathematical solution to a problem trying to approximate that
       | with some new method from data doesn't make sense just because
       | it's cool.
        
       | lawrenceyan wrote:
       | If you can do good old fashioned AI, you most definitely can do
       | new fangled AI. In fact, I'm almost certain you'll have a better
       | understanding of the fundamentals.
       | 
       | At the heart of AI is mathematics, and that will never change.
        
       | synapticpaint wrote:
       | "Finally, a vital question is, how will this affect today's
       | working artists? Here the answer is not so optimistic."
       | 
       | I have a different take on this. I think this technology will
       | allow more people, not less, to make money as a living (so,
       | professionally) in a visual arts related industry. So I'm
       | broadening the field to include not just "artists" but
       | "commercial art" as well (designers, commercial illustrators,
       | video/film post-production, etc.).
       | 
       | The reason is that it changes and lowers the bar to entry for
       | these fields, automates away a lot of the labor intensive work,
       | thereby lowering the cost of production.
       | 
       | Whenever something becomes cheaper (in this case, labor for art),
       | its consumption increases. So in the future, because producing
       | commercial art is so much cheaper, it will be consumed a lot
       | more.
       | 
       | At the same time, we're not at the point where we can actually
       | remove humans entirely from the process. AI generated art is a
       | different process and requires a different skillset, but it still
       | requires skill and learning to do well.
       | 
       | The analogy would be something like a word processor reducing the
       | number of secretaries needed in the workforce, but increasing the
       | number of office workers. People no longer need someone to take
       | notes / dictation, but all kinds of new workflows emerged on top
       | of the technology, and almost all office workers need to know how
       | to use something like a word processor.
       | 
       | Therefore, the opportunity here to do is to build tooling that
       | make it easier and more accessible for more people to work with
       | AI image generation.
       | 
       | Disclaimer: I'm doing exactly that (building tooling to make
       | content generation easier and more accessible) with
       | https://synapticpaint.com/
        
         | dleslie wrote:
         | > I think this technology will allow more people, not less, to
         | make money as a living (so, professionally) in a visual arts
         | related industry. > ... > Whenever something becomes cheaper
         | (in this case, labor for art), its consumption increases.
         | 
         | But not its price, and definitely not the compensation for the
         | labour to produce it.
         | 
         | Making a living as a mediocre-to-good artist is already
         | incredibly difficult; increasing the supply of poor-to-good
         | artists through AI-assistance isn't going to make it any
         | easier.
         | 
         | > The analogy would be something like a word processor reducing
         | the number of secretaries needed in the workforce, but
         | increasing the number of office workers.
         | 
         | Only if the word processor wrote documents without the
         | assistance of a typist, or an author.
        
         | hbn wrote:
         | > Whenever something becomes cheaper (in this case, labor for
         | art), its consumption increases. So in the future, because
         | producing commercial art is so much cheaper, it will be
         | consumed a lot more.
         | 
         | I'm not sure how that would apply here. There's never been a
         | shortage of art. Art has always had more supply than demand,
         | and now we just added even more supply to saturate the market.
         | I was previously a more likely client for an artist than I am
         | now where I can get my computer to spit out any image I want in
         | like 30 seconds. But I have no more desire for art than I did
         | before.
        
       | [deleted]
        
       | [deleted]
        
       ___________________________________________________________________
       (page generated 2022-11-15 23:00 UTC)