[HN Gopher] GPT-3 can create both sides of an Interactive Fictio...
       ___________________________________________________________________
        
       GPT-3 can create both sides of an Interactive Fiction transcript
        
       Author : raldi
       Score  : 77 points
       Date   : 2022-10-24 16:19 UTC (1 days ago)
        
 (HTM) web link (old.reddit.com)
 (TXT) w3m dump (old.reddit.com)
        
       | joshuahedlund wrote:
       | Is there a Stable-Diffusion-esque open-source GPT yet? Given the
       | incredible pace of advances in the image space this year, and my
       | (perhaps naive) assumption that text generation is less complex
       | and less resource-intensive than image generation, I'm hoping
       | we'll get something similar and surprised that we haven't yet.
        
         | visarga wrote:
         | Google recently released the Flan-T5 models that are very
         | efficient, they rival GPT-3 and can run on a normal desktop.
         | You can try them out.
         | 
         | https://huggingface.co/google/flan-t5-xxl?text=Translate+to+...
        
         | davnn wrote:
         | There are multiple open-source GPTs, but GPT-3 is absolutely
         | massive - larger than the image models actually! So,
         | unfortunately, text generation is probably even more complex
         | and resource intensive than image generation (especially to
         | train). Additionally, in image generation, we appreciate the
         | creativity of solutions, but in text generation the creative
         | solutions seem like utter nonsense.
        
           | joshuahedlund wrote:
           | I guess my intuition is based on the file size of text being
           | so much smaller than images, but I guess that doesn't really
           | map to the complexity of generating it. Fascinating!
        
             | davnn wrote:
             | I think large language models are still in their infancy.
             | The models are extremely sparse, but we don't have the
             | tooling yet to deal with these kinds of structures
             | efficiently. Your intuition might be right in a future,
             | maybe.
        
             | drusepth wrote:
             | If you think about the space both models are covering from
             | a rate-of-failure perspective, it kind of makes sense that
             | images end up being a bit easier than text: both text- and
             | image-models can output results that look plausible at
             | first glance, but when you analyze both outputs further
             | there are a _lot_ more gotchas in parsing meaning within
             | language than there are in pixel placement within an image.
        
         | spijdar wrote:
         | There is, called "GPT-NeoX", or its TPU-based predecessor GPT-
         | Neo. However, even running inference on these models is much,
         | much harder than Stable Diffusion -- the GPT-NeoX-20B weights
         | for GPT-NeoX requires a minimum of two GPUs with 24 GB of VRAM
         | _each_ to simply run inference, never-mind training or fine-
         | tuning.
         | 
         | I believe there are some tricks for cutting down the VRAM
         | requirements a bit by dropping precision at different points,
         | but the gist is that these big text models are actually quite a
         | bit _more_ resource intensive than the image models.
        
           | tinco wrote:
           | Interesting that they're saying they have a target of
           | replicating GPT-3 with 175B parameters, but since the RETRO
           | paper it should (in theory?) be possible to replicate GPT-3
           | with a tenth of the parameters. Are they not planning to use
           | a retrieval system, is it difficult to adopt their system for
           | retrieval, or is their readme out of date?
        
         | jncraton wrote:
         | There are smaller models that you can freely play around with
         | that work in roughly the same way. If you're working on a
         | fairly regular computer some reasonable options are GPT2 or
         | GPT-Neo. These can both perform inference on your local CPU if
         | you have 8GB or more of RAM.
         | 
         | They are much less powerful that GPT-3, but they can still be
         | fun for simple text generation or NLP tasks. You can play
         | around with one of the smaller GPT-Neo models that should fit
         | in RAM if run locally here:
         | 
         | https://huggingface.co/EleutherAI/gpt-neo-1.3B
         | 
         | That page includes instructions to run this locally in Python.
         | 
         | As others mentioned, there are larger models available, but
         | they tend to be expensive to setup and use as an individual.
        
         | gamegoblin wrote:
         | Emad Mostaque, CEO of Stability AI (behind Stable Diffusion)
         | has said they are training language models that they plan on
         | releasing. IIRC, though, even using all the most up-to-date
         | techniques like 1-byte parameter quantization and Chinchilla
         | scaling learnings, it will still probably be on the order of
         | magnitude of 64GB or something, so not quite usable for most
         | people locally yet, unless you happen to have a very beefy
         | multi-GPU machine at home.
        
           | ctoth wrote:
           | GLM-130b is available and uses 8-bit quantization, was
           | trained on 400 billion tokens, and runs on 4x3090, which, of
           | course, is why all the cheap Vast.AI instances are gone :)
        
       | goodside wrote:
       | I recorded a demo of this same premise here:
       | https://twitter.com/goodside/status/1562613028927205377
       | 
       | Text completions of exotic forms of session/action logs are a
       | seriously under-explored area. Here's what happens if, instead of
       | a text game, you do text completion on an IPython REPL:
       | https://twitter.com/goodside/status/1581805503897735168
        
       | jcims wrote:
       | Isn't that the premise of AI Dungeon?
       | 
       | https://aidungeon.io/
        
         | raldi wrote:
         | As far as I can tell, that can only generate one side of the
         | transcript.
        
           | jtvjan wrote:
           | Type a lone "> You" in "Story" mode and it will generate an
           | action for you.
           | 
           | NovelAI[1] just gives you a big text box to type in, so you
           | don't have to do any input hacks.
           | 
           | [1]: https://novelai.net/
        
       | neaden wrote:
       | It's interesting that the prompt here pretty much immediately
       | became a Dr. Who game. It looks like there has been a Dr. Who
       | Text adventure game before but it was made before Jack Harkness
       | and Torchwood. I wonder if there is some fan made text adventure
       | that was part of the training data. Probably would be best for
       | Open AI that it doesn't just spit out copywrited characters
       | unprompted too.
        
         | visarga wrote:
         | Copyright infringement detection could be formulated as a ML
         | task. A model can copy the idea but not the expression, if it's
         | copyrighted. Fortunately a model can sample again and again
         | until it looks ok.
        
       | planetsprite wrote:
       | 99 times out of 100 whenever I see a new tech demo that
       | absolutely blows my mind and makes me optimistic about the future
       | of technology it's always some big model deep learning AI thing.
       | When are we all going to admit that ML/AI is the final and
       | ultimate paradigm shift of our time
        
         | kleiba wrote:
         | It's all smoke and mirrors.
        
         | dilap wrote:
         | I think it when it starts having a broad impact on the way we
         | live, like the internet and smartphones did.
        
           | raldi wrote:
           | Last night the AI and I teamed up to write bedtime stories on
           | demand and on the fly for whatever themes my daughter mused.
           | I think about a billion families would enjoy an app that did
           | that.
        
             | dilap wrote:
             | I do think there's a good chance this stuff _will_ have
             | that level of impact, I just don 't think it has yet.
             | 
             | (Though I'm not _convinced_ it will. If you wanted to be a
             | skeptic, you could argue that we 're _already_ in a huge
             | content glut; there 's basically infinite content available
             | for almost free. So does radically lower the cost matter
             | that much?
             | 
             | (Maybe in low-level ways, like it'll increase the abilities
             | of small indie studio to produce high-level content. But to
             | your average consumer, maybe it's not noticable.)
        
               | airstrike wrote:
               | > So does radically lower the cost matter that much?
               | 
               | It's not about lowering the cost, it's about generating
               | content that fits your specific preferences in real time.
               | Forget on demand streaming of off-the-shelf content,
               | people will want on demand _content generation_.
               | 
               | And then, as is tradition, the next step after that is
               | giving you content you did not even know you wanted. That
               | can be perverse marketing, feeding you sensationalized
               | stuff that keeps you hooked 24/71 but the better
               | alternative is something that is tailored to your general
               | tastes, genres, writing styles, etc. so that you receive
               | the content that is best received by you at that
               | particular moment (matching your mood, your goals, your
               | style)
               | 
               | ----------
               | 
               | 1. This image comes to mind...
               | https://www.wallpaperflare.com/dystopian-cyberpunk-sad-
               | virtu...
        
             | thih9 wrote:
             | And yet people aren't doing this so I guess something is
             | still missing. Awareness? UX? Something else? All of the
             | above? The company that figures it out might earn a lot.
        
               | MintsJohn wrote:
               | See what is happening with StableDiffusion, a model was
               | released opensource, performance in the same league as
               | closed source, usable on consumer hardware and (non AI)
               | techies start to modify it. The biggest
               | steps/modifications are by specialists no doubt, yet
               | still opensource, but others are happily glueing parts
               | together to make something else. The key really seems to
               | be access, an hosted API is rather hostile to innovation,
               | using and especially experimenting is expensive,
               | modification can only happen within whatever the API
               | allows. For the tech to get bigger and more noticed
               | faster more people need to be able to tinker with it.
        
               | nonasktell wrote:
               | Price. I have a dozen ideas that could be done using
               | GPT-3. Most of them aren't financially viable, I'm not
               | burning half my income to build an MVP.
        
             | actually_a_dog wrote:
             | Maybe so, but unless there are a billion families out there
             | who are already privileged enough to have internet access
             | and are willing to pay for such an app, I don't see it
             | happening.
        
       | davnn wrote:
       | I recently wrote a VS Code plugin to use GPT-3 for text
       | completions and editing of text using instructions [1]. The main
       | intention was scientific writing, but I didn't find it
       | particularly useful until now, unfortunately.
       | 
       | [1] https://github.com/davnn/assistant
        
       | aaron695 wrote:
        
       | airstrike wrote:
       | OK, this is absolutely mindbogglingly fun. Here's my run at it,
       | with italics representing what the model spit out:
       | 
       | > OPEN THE BACKDOOR WITH THE OLD KEY
       | 
       | You open the manor's backdoor with the old key. Your torch lights
       | the way into the damp service halls, the warm glow of the nuclear
       | war in the horizon lighting the window panes. You can almost feel
       | the centuries old dust through your boots as you make your way up
       | the stairs. You climb up to the main floor and pause at the
       | entrance to the parlor room. It is quiet, except for the
       | persistent low rumble of the bombs in the distance.
       | 
       |  _You enter the parlor room and are immediately hit with the rank
       | smell of death and decay. The room is draped in cobwebs and
       | littered with the bones of small animals. The only light comes
       | from the faint glow of the bombs outside. You see a small, dust
       | covered table in the center of the room, and on it is a dusty old
       | book. You approach the table and dust off the book. It is titled,
       | "The Necronomicon"._
       | 
       | > _OPEN THE BOOK_
       | 
       |  _You open the book and begin to read. The first few pages are
       | filled with gibberish, but as you flip through the pages, the
       | text becomes clearer and more comprehensible. The book speaks of
       | dark things, things that should not be spoken of, things that
       | should not be known. It speaks of powerful beings that sleep
       | beneath the earth, waiting to be awaken. It speaks of a time when
       | the world will be consumed by darkness, and all will bow before
       | the old ones. You close the book and back away from the table,
       | your heart pounding in your chest. You have to get out of here.
       | You have to warn somebody._
       | 
       | > _LEAVE THE PARLOR ROOM_
       | 
       |  _You leave the parlor room and make your way back down the
       | stairs. You can hear something moving around in the darkness, but
       | you don 't dare look back. You make it to the backdoor and slam
       | it shut behind you. You lean against the door, your heart still
       | racing, and listen to the thing moving around inside the manor.
       | You hear it coming down the stairs, but it doesn't reach the
       | door. You wait there for a few minutes, but nothing happens.
       | Finally, you muster up the courage to peek through the keyhole.
       | 
       | You see a large, black, slimy creature slithering around the
       | parlor room. It is the size of a man, but it has no eyes, only a
       | large mouth filled with sharp teeth. You watch in horror as the
       | thing moves around the room, destroying everything it comes
       | across. Then, without warning, it turns towards the door and
       | starts towards you.
       | 
       | > RUN AWAY
       | 
       | You turn and sprint away from the manor as fast as you can. You
       | can hear the thing crashing through the door behind you, but you
       | don't dare look back. You run through the woods, your heart
       | pounding in your chest, until_ you find an old cottage by a
       | creek. You force the old door in and see a 12 gauge shotgun lying
       | on the table. You confidently load the gun and sprint out of the
       | cottage.
       | 
       | > _AIM THE SHOTGUN
       | 
       | You raise the shotgun and take aim at the creature. You fire off
       | two rounds and the creature falls to the ground, writhing in
       | pain. You approach the creature and finish it off with a bullet
       | to the head. You stand there for a moment, panting, your heart
       | still racing. You can't believe you just killed something. You
       | can't believe this is real._
       | 
       | > INSPECT THE CREATURE'S BLOOD
       | 
       |  _You kneel down next to the creature and inspect its blood. It
       | is black and viscous, and it smells like rot. You have no idea
       | what this thing is, but you know it isn 't natural. You have to
       | find out what is going on. You have to find out what these things
       | are._ Suddenly you notice the viscous blood pooling around you,
       | engulfing your right foot. You try to move, but by now your whole
       | right leg is covered with it. It is too late. You can feel the
       | power coursing through you, and you know that you are not the
       | same person you were before. You are something else now.
       | Something dark. Something evil.
       | 
       | > RETURN TO THE CITY
       | 
       |  _You make your way back to the city, your mind consumed with
       | dark thoughts. You know that you are not the same person you were
       | before, and you are not sure what to do with your new found
       | power. all you know is that you want more. You want to learn more
       | about these things, and you want to find more of them. You want
       | to become the most powerful being in the world, and you will stop
       | at nothing to achieve your goal._
        
         | stuckinhell wrote:
         | wow thats really good!
        
         | raldi wrote:
         | And if you don't like the way the story is going, you can just
         | erase that part and everything that followed and type a couple
         | words nudging it in the right direction.
         | 
         | For some reason it's really drawn to rusty keys.
         | 
         | > LOOK IN HOLE
         | 
         |  _You look in the hole and see a rusty key_ - NO, TRY AGAIN,
         | ERASE ERASE ERASE
         | 
         |  _You look in the hole and see an old rusty key_ - NO,
         | BACKSPACE BACKSPACE
         | 
         |  _You look in the hole and see a small box_ - YES!!!!
         | 
         |  _> OPEN BOX_
         | 
         |  _You open the box, revealing a rusty key_
        
           | airstrike wrote:
           | I hear you--it took this somewhat unexpected turn into horror
           | so I just ran with it, but it's not where I was going
           | originally... I guess the bombs dropping in the horizon made
           | it eerie? It was supposed to be post-apocalyptic but it
           | decided to dial things up to 11 and bam!! "The Necronomicon"
        
       | selimnairb wrote:
       | Who cares? Does GPT-3 know suffering and joy? No. Will it produce
       | art? No. This is about as interesting as creating a self-
       | fellating android.
        
         | [deleted]
        
         | sodality2 wrote:
         | > Does GPT-3 know suffering and joy?
         | 
         | Do you? Prove it. Your textual output here will do nothing to
         | prove you know suffering and joy _any more_ than GPT-3 would be
         | able to.
         | 
         | > Will it produce art?
         | 
         | Is a poem art? I think most consider it to be.
         | 
         | For what it's worth, I don't like where AI generated content is
         | going either, but to disregard it as worthless (instead of
         | simply lacking conscience) is not helpful.
        
           | visarga wrote:
           | GPT-3 has learned many identities, it's not lacking a
           | perspective but quite the opposite, it has too many
           | perspectives and you never know who you're talking to. Unless
           | you prompt it with an identity. Recently it was used to run
           | "in silico sampling" [1], a virtual poll where the
           | respondents were GPT-3 prompted with personality backstories.
           | 
           | So it should not be seen as an agent, it's not like us in
           | that regard. It is the synthesis of all human culture and can
           | assume any hat.
           | 
           | [1] https://arxiv.org/abs/2209.06899
        
         | avarun wrote:
         | According to who? You?
         | 
         | This creation is art to me, and much more interesting than your
         | comparison.
        
         | smcameron wrote:
         | I don't know, Rockit was pretty popular back in the day.
         | https://www.youtube.com/watch?v=GHhD4PD75zY
        
       ___________________________________________________________________
       (page generated 2022-10-25 23:00 UTC)