[HN Gopher] This AI Does Not Exist ___________________________________________________________________ This AI Does Not Exist Author : thesephist Score : 126 points Date : 2022-04-23 19:04 UTC (3 hours ago) (HTM) web link (thisaidoesnotexist.com) (TXT) w3m dump (thisaidoesnotexist.com) | BbzzbB wrote: | Are these procedural or is there a list of pre-generated "AI"s | next goes thru? | | I got this as my third which seemed either prophetic or | deterministic. | | HackerNewsReplyGuy: | | >from hackernews_response_guy import HackerNewsReplyGuy | | >model = HackerNewsReplyGuy(1) | | >model.predict_comments(comments, [u'comment_id']) | sillysaurusx wrote: | I got that too. It's pregenerated. But what's particularly | impressive is that you can generate your own outputs on the | fly. Usually with sites like these, it's solely pregenerated. | | Quick, generate your own before the server goes down! I don't | think the model can withstand HN for too long unless they have | some beefy servers. | | Aaand it's dead. Fun while it lasted. | BbzzbB wrote: | Ohh thanks, I had not noticed that. This makes the site | fairly more interesting. | [deleted] | thesephist wrote: | There's a pre-generated set to (1) spare my server some work | and (2) showcase some output I liked. But as sibling comment | noted, you can (or could) generate your own -- I'm working on | bringing that side back up... | | The pre-generated set is hand-curated, but they are still 100% | generated by the GPT-J model behind the scenes. More info -> | https://github.com/thesephist/modelexicon | BbzzbB wrote: | Thanks for the reply! And the generator option, tho I keep | getting timed out of the code, the descriptions sound | promisingly good at times. | | Sorry, I didn't mean to imply these were not produced as | described, I was just curious. Tho think of it it was a silly | question as it would have otherwise implied they're generated | in a blink. | [deleted] | sillysaurusx wrote: | Wait, it's just vanilla gpt-j? No fine tuning? | | "Back in my day, we had to train our own models.." already | sounds anachronistic. | | Nicely polished. | | Looks like bmk (nabla theta) was right that arxiv was an | impactful addition to The Pile. I bet that's where J got its | knowledge in this case. | thesephist wrote: | Yep! No fine tuning. Here's the prompt I use for the | description (from source, https://github.com/thesephist/mod | elexicon/blob/main/src/main... ) | | --- | | Proceedings of Deep Learning Advancements Conference, list | of accepted deep learning models | | 1. [StyleGAN] StyleGAN is a generative adversarial network | for style transfer between artworks. It uses a traditional | GAN architecture and is trained on a dataset of 150,000 | traditional and modern art. StyleGAN shows improved style | transfer performance while reducing computational | complexity. 2. [GPT-2] GPT-2 is a decoder-only transformer | model trained on WebText, OpenAI\'s proprietary clean text | corpus based on Wikipedia, Google News, Reddit, and others | comprising a 2TB dataset for autoregressive training. GPT-2 | demonstrates state-of-the-art performance on several | language modeling and conversational tasks. 3. [$MODELNAME] | sillysaurusx wrote: | That's awesome! How'd you get such great code usage | examples out of J? | | It almost seems like the code is properly related to the | names. GAN code seems to look like GAN code. But I'm not | sure. | thorum wrote: | Very cool! | | https://github.com/thesephist/modelexicon | | Looks like it's powered by GPT-J. My understanding is that GPT-J | has comparable performance to OpenAI's Curie model on many tasks | (their second-best variant of GPT-3) but it's an openly available | model that you can run yourself if you have the resources. | thesephist wrote: | Yep, that's spot on. The overall performance is comparable to | Curie, but depending on the particular task GPT-J performs | better or worse (I believe empirically it's slightly better at | chat and code, worse at some others). | furyofantares wrote: | Skynet is an end-to-end speech recognition model. It is based on | the Inception-v3 architecture and the Speech Transformer (Sphin) | speech model. Its speech model was trained on a dataset of 30,000 | hours of human speech, as well as speech recordings from the | Switchboard corpus and the Fisher corpus. The model achieves | 99.34% WER on the Switchboard-1.1 test set. | Terry_Roll wrote: | It asked me for a model, so I naturally thought of female models | and cars, decided upon "911" and get: "911 is a dataset for 9/11 | related tasks, including predicting the location of the first | plane crash, the location of the second plane crash, and the | location of the towers." | | Thats not what I had in mind so it still needs a bit of work I | think or at least the questions do. ;-) | ImpressiveWebs wrote: | I got: | | > SpotifAI is a system that uses deep learning to automatically | create playlists from user-submitted playlists. Its algorithm has | been trained on millions of playlists from Spotify. | | Which is pretty cool sounding and has a cool name. | hunterb123 wrote: | Certainly so! I got some generic BeatlesAI one. | | Very nice accidental wordplay (it didn't mean have the same | pronunciation) and it's a cool premise. | | I'd like something like that, I currently use Pandora and Apple | Music since Apple radio is trash. | | AI generation serves best for cherry picking, certainly good | for coming up with ideas or searching for leads. | recuter wrote: | Garbage in Garbage out | luxuryballs wrote: | Jesus is a fast and scalable language model trained on the Jesus | dataset, which consists of over 4.7 billion words from the Bible. | Jesus demonstrates state-of-the-art performance on several | language modeling and conversational tasks. | hprotagonist wrote: | it would be hypermeta levels of satisfying if indeed these | results are maybe 500 or so human-written precanned responses. | vampiretooth1 wrote: | Clicked into it, didn't read the description, and got an AI-based | project that could perfectly hedge my fixed income portfolio. I | won't lie, got a bit excited and then I realized what site I'd | clicked on. | | Very nifty! Is this your site? | sillysaurusx wrote: | As someone who has trained around 60 GPT-2s, this is damn | impressive work. It's very hard to get consistent code quality | when the training corpus is so small (as this one undoubtedly | was). | | https://thisaidoesnotexist.com/model/MozartNet/JTdCJTIyZGVmb... | | The url scheme is interesting. I wonder what it base64 decodes | to. If I were at a computer I'd check. It might be a complete | representation of the inputs to the model, which is then cached. | Which implies you might be able to fiddle with it to get specific | outputs. | Ndymium wrote: | Looks like the URL path just contains the generated output and | not the inputs. | thesephist wrote: | Yep. I didn't want to have to host user-generated data (for | all the perils that carries), so the sharable links work by | embedding all the generated data in the link itself. | ALittleLight wrote: | The base64 in the URL decodes to a URL formatted JSON blob that | seems to describe the contents of the page. { | "defn":"MozartNet is a sequence to sequence deep neural network | trained on the music of Wolfgang Amadeus Mozart. It is used to | generate music for a piano transcription in a completely | unsupervised fashion. MozartNet is an instance of a more | general family of networks known as 'autoregressive networks', | and is trained on a synthetic dataset of about 1 million short | sequences of piano notes. The network is a two layer LSTM and | is trained with L2 regularization to minimize the total number | of parameters. MozartNet is one of the most widely used and | best-performing autoregressive networks, and is often cited as | an example of using a neural network for the purpose of | learning the structure of music.", "usage":"from nmt | import \*\nnet = | Model()\nnet.load_weights(\"/tmp/mozartnet.h5\")\n\n# get the | source text\nsequence = net.encode(\"GDAEADBBGEDC\", | output_chars=\"p\", max_length=5)\n\n# decode | it\nsource_sequence = net.decode(sequence)\n\n# print | it\nprint(source_sequence.as_list())" } | sillysaurusx wrote: | If you insert a url, or html tags, does the site properly | sanitize the output? | | It's remarkably difficult to suppress pentesting urges after | doing it for a year. | | And if you try to generate your own, the usage section | usually fails. I wonder if it elides the usage key. | | Modern websites are pretty fun. I like the simplicity here. | And also the meta: https://thisaidoesnotexist.com/model/Hacke | rNewsReplyGuy/JTdC... | thesephist wrote: | Hey HN! Author of the site here. I tried a few tricks to keep the | text-generation part of the site up, but even leaning hard on | Huggingface's API and bumping time-outs up, it looks like the | site is struggling a bit. I'm going to see if there's anything I | can do to keep the text-generation part available, but in the | meantime, the pre-generated set should stay pretty stable. Not | sure if there's much else I can do without burning a hole in my | cloud bills -- sorry for the troubles! | | I've put up a more detailed description of how this works on the | GitHub - https://github.com/thesephist/modelexicon | | PS - if anyone at Huggingface is reading this and wants to help | out with keeping the API up, that would be super :) | jcims wrote: | My favorite name of the dozen or so projects i saw: SpotifAI | [deleted] | [deleted] | TheCraiggers wrote: | On FF, I get a blank page. Given the domain name, I thought it | was a joke until I came here and read the comments. | thesephist wrote: | Sorry about that. I don't think I'm leaning on any super new | browser/JS features, but if you share your FF version string | (or an error in the console) I can try to troubleshoot what's | missing! | [deleted] | mordae wrote: | > AutoProfit is a reinforcement learning model that trains itself | on a simulated trading environment. It is able to trade on its | own and generate its own trading signals, outperforming a | portfolio of human traders and making the most out of available | information. AutoProfit is a model for trading stock, | cryptocurrencies, and commodities in real time, generating | trading strategies for itself. It uses an iterative training | process, and has been tested on over 50 trading strategies. | | Cool. | daniel-cussen wrote: | You know now that I made friends with a homeless beggar I have no | trouble making friends with a bot. Why not? Has some humanity | breathed into them, like a book for instance, a book can be your | friend. A kind old family friend who let me stay with her told me | a long time ago just that when talking about a chest full of | books, _these books are my friends_. ___________________________________________________________________ (page generated 2022-04-23 23:00 UTC)