[HN Gopher] Stable Diffusion Public Release ___________________________________________________________________ Stable Diffusion Public Release Author : flimsythoughts Score : 481 points Date : 2022-08-22 18:08 UTC (4 hours ago) (HTM) web link (stability.ai) (TXT) w3m dump (stability.ai) | UncleOxidant wrote: | > You can join our dedicated community for Stable Diffusion here: | [https://discord.gg/stablediffusion] | | Oh, discord... I've had so many problems trying to log into | discord in the last couple of years that I've given up on it. | Kiro wrote: | That's peculiar. I run a couple of big Discord servers and | never heard anyone saying they have problems logging in. | UncleOxidant wrote: | First about 18 months ago they said I had an IP phone and | could no longer log in (my carrier is Republic Wireless). | When I contacted support and told them that I had been able | to login with that number in the past they basically said | "we're tightening up security, too bad so sad. get another | phone number" Then recently I found that I was able to log in | when I wanted to try midjourney (Republic changed something | back in April, apparently so they it no longer looks like an | IP phone#). Then I wanted to login to discord on the my | desktop and it gave me some weird errors which basically | amounted to "this number is already claimed" (it was, by me) | so now I'm back to ignoring discord again. | mabbo wrote: | The most interesting part, to me, of a release like this is the | amount of "please don't abuse this technology" pleading. No | licence will ever stop people from doing things that the licence | says they can't. There will always be someone who digs into the | internals and makes a version that does not respect your hopes | and dreams. It's going to be bad. | | As I see it, within a couple years this tech will be so | widespread and ubiquitous that you can fully expect your asshole | friends to grab a dozen photos of you from Facebook and then make | a hyperrealistic pornographic image of you with a gorilla[0]. | Pandora's box is open, and you cannot put the technology back | once it's out there. | | You can't simply pass laws against it because every country has | its own laws and people in other places will just do whatever it | is you can't do here. | | And it's only going to get better/worse. Video will follow soon | enough, as the tech improves. Politics will be influenced. You | can't trust anything you see anymore (if you even could before, | since Photoshop became readily available). | | Why bother asking people not to? I guess if it helps you sleep at | night that you tried, I guess? | | [0]A gorilla if you're lucky, to be honest. | OmarIsmail wrote: | We do a decent job of banning child pornography. | | And bringing the two ideas together, is child pornography that | is provably created by an AI still illegal? | seydor wrote: | Actually ... we just did a good job of making generated child | pornography ubiquitous. | causi wrote: | As far as I'm aware countries fall broadly into two camps. | Camp 1, USA for example, is concerned purely with the abuse | of children, i.e., anything that depicts or is constructed of | pieces of real children is illegal but other things such as | drawings, stories, adults role playing, etc is not. Camp 2 | outlaws any representation of it whether or not a child was | involved. | | Nowhere will a training set featuring pictures of naked | children be legal. | 16890c wrote: | >Nowhere will a training set featuring pictures of naked | children be legal. | | True, but generalizing beyond the training set is precisely | the point of machine learning. A good generative model will | be able to produce such images, no matter how heinous the | content is. | OmarIsmail wrote: | > Nowhere will a training set featuring pictures of naked | children be legal. | | Appropriately from the recent news stories, but it's easy | to imagine at least portions of such pictures being | available for medical diagnostic purposes. I've sent | pictures of my children to my doctor, so presumably in the | future it's easy to imagine sending pictures to an AI to | diagnose which would require a suitably fleshed out (pardon | the pun) training set. | mdcds wrote: | > you can fully expect your asshole friends to grab a dozen | photos of you from Facebook and then make a hyperrealistic | pornographic image of you with a gorilla | | my prediction is that, as a result, people will start assuming | pics online are fake until proven otherwise. | lern_too_spel wrote: | > my prediction is that, as a result, people will start | assuming pics online are fake until proven otherwise. | | "That worked well for quotations." -- Abraham Lincoln | fpgaminer wrote: | If I recall from the interview with Stability.ai's founder, he | has more or less the same opinion, and that humans will adapt | to the new technology as we always have. I figure "please don't | abuse this technology" warning stickers are more CYA. It'll | make the vast majority of judges look at a motion to dismiss | and not blink an eye. | gitfan86 wrote: | Historically, he is correct. It is easy to find people that | were against TVs, Cars, Trains, Electric Cars. Those people | were not entirely wrong with their logic. Trains and cars did | make it much easier for scammers to come into a town and then | leave quickly. | theptip wrote: | In terms of the equilibrium, this is certainly a true | observation. However, historically speaking new technology | can be extremely disruptive in the short-term as society | figures out the new norms, and the power-structures are | disrupted and then re-equilibrate. | | Concretely, it's probably true that children born with this | technology will have adapted to many of the negative (and | positive) aspects of it. But the current generation of | elites, politicians, and voters might have a harder time | adapting. | patientplatypus wrote: | oifjsidjf wrote: | This. | | They just have to cover their asses, any sane dev would make | the same license due to the power of this tech. | | On some level I can't stop laughing since OpenAI really got | smoked. "OpenAI" my ass, this is what open TRULY means! | | Cheering for these devs. | aortega wrote: | >make a hyperrealistic pornographic image of you with a | gorilla[0] | | I don't understand this irrational fear. This can be done | today, just need some minutes instead of some seconds to create | a good Photoshop. | | Also, seriously this is the thing you fear? fake porn? there | are much worse thing you can do with this tech, like phising, | falsifications, etc. Not mentioning leaving millions of graphic | designers out of job. | peoplefromibiza wrote: | Unfortunately _Anything that can go wrong will go wrong_ | | Photoshop is a skill, not very widespread as we assume. | | Typing something is _literally_ at everyone 's fingertips. | cerol wrote: | That's what I was thinking. Whether a pornographic picture of | me and a gorilla was made by photoshop or AI is irrelevant. | People's reactions will be the same, and repercussions will | be mostly the same (which doesn't means there will be | consequences). | | If someone really wants to hurt you, not having AI isn't | going to stop them. | aortega wrote: | The real effect will be that you can publish a real picture | of you fucking a gorilla and nobody would believe it | because it's trivial to generate it with an AI. | naillo wrote: | I much prefer a company that asks me to not abuse it but lets | me, then a company that treats me like child and force filters | it out for me. | skybrian wrote: | It seems inevitable like it's "inevitable" that they Photoshop | your face onto porn. Yes, of course it will happen but maybe | not to most people? I'd guess inevitable for many celebrities. | croes wrote: | It will be more now. Photoshop needs some skills. With AI it | gets easier and easier. | eastbound wrote: | But even today, we deal correctly with it. Fakes and real | photos are mingled together in 9Gag/LatestNews reports | about Ukraine. Under the fakes (and the real), people ask | for confirmation. Someone says it's true, no-one believes | him, until a link to a newspaper is dropped. And 9Gag isn't | the highest IQ community around, so yes, general population | does distrusts photos by default until proven. | | They are laughed at anyway if they tell a story coming from | a forged photo. | | Sure, newspapers could forge stories, display pictures | with, I don't know, Biden's son with a crackpipe, and make | the populace believe untrue stories. But guess what, they | already do it anyway, newspapers already "spin" (as they | say, i.e. forge, suggest without literally saying) stories | all the time. | | The world deals correctly with fakes. | croes wrote: | I have a quite different perception of 9gag. Yes, some | ask for confirmation but it depends very much on the | topic. | | Wrong topic and facts get downvoted and the fake news | prevail. | | And not all links to newspaper are considered valid, | especially if it's about "woke culture". Then you have to | search the reasonable needle in the haystack of | transphobia, homophobia and misogyny. | gabereiser wrote: | The problem comes from the early adult newsroom interns | responsible for sourcing content. They don't know it's | fake, it sounds like a good click-baity article to them, | so they run it. It happens. | eastbound wrote: | I wouldn't shift responsibility on the shoulders of the | last newcomer. The top of the management has had ample | time to diagnose this. If it remains like this, it's by | design. | wwwtyro wrote: | > Politics will be influenced. You can't trust anything you see | anymore | | I've been wondering for a while now if this will lead to an | unexpected boon: perhaps people will be forced to pay attention | to a speaker's content instead of simply who is speaking. | dbingham wrote: | The problem with this is that you will never know who is | actually speaking. Deep fakes are already a thing, but as | they get better and more accessible we will approach a world | where anyone can make anyone say anything and make it hyper | believable. In that world, it will be very difficult to tell | what is real. | bluejellybean wrote: | My personal hunch is that this will end up leading to a | situation in which presenters do a cryptographic handshake | that works to verify and prove authenticity. This isn't a | new idea, and it has some very obvious drawbacks, but I | don't see much of a way around the issue. The handshake | could work great for something like official news releases, | but for other instances that might come up in court, say, | dash cam footage of an accident, it seems to me that the | legal system is going to face some serious issues as these | programs progress. | nelsondev wrote: | Looking forward to a future where all a politician's | quotes are on a blockchain, signed by their private key, | and they chose to do so voluntarily out of fear of deep | fakes. | | Will remove all the useless "I didn't say that" | dinosaurdynasty wrote: | Do politicians understand blockchains? | kelseyfrog wrote: | Neither do we[1], so it's sort of a mixed bag. | | 1. True for an overwhelming majority of the body politic. | nelsondev wrote: | Rename it as "verified speech transcripts." I don't need | to understand video codecs to watch Youtube. | marak830 wrote: | It will end up back where we started, "it's not on the | Blockchain so I didn't say it", while making racist | remarks with friends. | swader999 wrote: | That's a start! I think their video appearances should be | like a car in NASCAR with permanently displayed logos | superimposed from all the interests that have funded | their rise. | [deleted] | edouard-harris wrote: | Unfortunately a speaker's content can also be auto-generated | now, at least for brief enough snippets. And that means the | content can (and will) be optimized to appeal to a target | segment much more than has ever been previously possible. | narrator wrote: | As someone who witnessed AI Dungeon's GPT-3 model with an | unlimited uninhibited imagination for erotica I would download | everything now before they cripple the models. I would not be | surprised if they very shortly completely stop downloads due to | "abuse" and pursue a SAAS model. | | I think it's funny how Yandex, a Russian company, releases | these big language models without all the AI safety | handwringing in their press releases. The Russians have a | tradition of releasing technology without giving a lot of worry | about what happens to it, for better or for worse. For example, | they made between 75 and 100 million ak-47s, a not in any way | limited machine gun, and it spread to every corner of the | world. They even gave out all the plans and technical | assistance so any one of the organizations they worked with | could produce their own. 20 different countries make ak-47s | currently. Of course you had to register every xerox machine in | the Soviet Union, so maybe they just had different priorities? | | The west is absolutely fascinated these days with the control | of advanced technology. Drones, Blockchain, and AI models seem | to be the latest things that the west is determined to exercise | control over. For example: | | "Many of the technological advances we currently see are not | properly accounted for in the current regulatory framework and | might even disrupt the social contract that governments have | established with their citizens. Agile governance means that | regulators must find ways to adapt continuously to a new, fast- | changing environment by reinventing themselves to understand | better what they are regulating. To do so, governments and | regulatory agencies need to closely collaborate with business | and civil society to shape the necessary global, regional and | industrial transformations." -Klaus Schwab, "The Fourth | Industrial Revolution", Page 70. | RandomLensman wrote: | It's not about control of advanced technologies but rather | making such technologies look more powerful by pointing out | real or imagined destructive capabilities. If you want | funding or "big up" a topic, claim it has the power to bring | down the world (anyone remembering gray goo?). | seydor wrote: | I think that s a hopeful message. | | At first we will see a lot of "prompt censoring mobs" that will | try to "stop abusers because children and terrorists", but as | the images multiply, and they will multiply, the line between | real and fake photos will become hard to spot, for real. This | is i think a pivotal and great moment, because everyone can now | claim plausible deniability to any picture. No revenge porn | will be believable anymore, nor will anyone know if that | Bezos's weiner pic is real or not. | gjsman-1000 wrote: | I had a thought for a utopian/dystopian world. | | One day these image generators will also get video support... | and pornography support. When that happens, a few things may | occur that I think are reasonable to predict: | | EDIT: Original post was way too wordy, TL;DR: | | When AI-generated pornography becomes available, it could be | likely that demand for "real" pornography disappears because | the AI will match and surpass the "real." When that occurs, the | "real" will become increasingly regulated and legally risky, | and may just outright be as good as banned. | suby wrote: | There will be a huge and unkillable market for non-AI | generated pornography, even if people cannot tell the | difference in an AB test. The demand will be too strong and I | don't think there will be much outcry to ban it if it's all | consenting adults. | scarmig wrote: | If people can't tell the difference in an AB test, how will | the real porn out compete the generated stuff? Porn | distributors aren't known for their truth in advertising or | care in sourcing of material. And even if they were, how | would the porn distributors be able to source the real | stuff, when anyone can create porn of anyone they imagine? | You might say PKA will save us, but people aren't going to | be typing out `gpg --verify` when their hands are otherwise | occupied. | tastemykungfu wrote: | Curious as to how you arrived at that conclusion. | derac wrote: | A quick Google shows an estimate that it was a 100 | billion dollar market in 2017. Seems there is a lot of | demand. | thfuran wrote: | But what portion of that demand is specifically for | artisinally handcrafted authentic sex rather than for | apparent sex acts? | croes wrote: | Perhaps we should have started with artificial wisdom instead | of artificial intelligence | matheusmoreira wrote: | > you can fully expect your asshole friends to grab a dozen | photos of you from Facebook and then make a hyperrealistic | pornographic image of you with a gorilla | | ... Someone is gonna do this to children. This technology is | gonna end up on the news. Maybe they'll even try to ban it. | indymike wrote: | Corollary: If you did not make it, someone else would have | anyway. | hahajk wrote: | Like you mentioned near the end, all this has been possible | with photoshop, with amateur level skill. Hollywood can CGI the | entire Captain Marvel movie, so as far as state-level efforts | go, AI can really only be an incremental improvement at best. | | I think this is all just trendy popular sentiment moralizing | AI. | cypress66 wrote: | > Why bother asking people not to? I guess if it helps you | sleep at night that you tried, I guess? | | They obviously are aware. They just put all that so they don't | get "canceled". It's just virtue signaling and covering your | ass. | nootropicat wrote: | This is great news for those that actually have sex with a | gorilla, because now they can claim it's an ai photo. :) | | Kidding aside, I think this is actually good. Humans need | ephemerality. We are never getting the full version of it back, | but with photorealistic ai video and image creation some | freedom returns. I think without it a society in which everyone | has a camera all the time would mean absolute ossification of | social norms. Right now it's very, very new - I mean multiple | generations living with the current, or better (eg. recording | eye implants) technology. | ForrestN wrote: | I wonder/hope if a downstream effect of this technological | change will be the end of the idea of a "shameful" or | "humiliating" image. We all have bodies, we all have sex, and I | agree that soon images of ourselves being nude/having sex etc. | will proliferate because they'll be generated instantly via | Siri shortcut as part of casual banter. | | In a world where every celebrity is having sex with gorillas, | doesn't such an image lose its charge? Will Norms and values | around sex/body shaming change? | Fauntleroy wrote: | If we get to the point where we can't tell AI generated | images from reality, I'm not sure "body shaming" will be on | society's collective radar anymore. | geysersam wrote: | I bet someone said in 1830, "by the time we can send robots | to Mars, body shaming will not be a thing anymore". For | better and worse, that's not how we humans do things | generally. | ForrestN wrote: | Why? | TylerE wrote: | Much bigger fish to fry. | | Think things like forged evidence in trials. | telesilla wrote: | It's the right time to get into the digital forensics | business. | 867-5309 wrote: | ..and the Berlin gorilla nightscene | mmmpop wrote: | I second this. | bwest87 wrote: | Appropriately prioritizing problems has never really been | society's strength... | mmmpop wrote: | That's because pesky things like "democracy" get in the | way of getting shit done. | | Don't you think? | BrainVirus wrote: | _> In a world where every celebrity is having sex with | gorillas, doesn't such an image lose its charge?_ | | No, because the image itself doesn't matter. What matters is | how much the public wants to hate someone. If t he public is | primed, any remotely plausible incriminating image will do as | an excuse. | | Fortunately, these images are actually far than what someone | can cook up with Photoshop. Unfortunately, it's a part of a | bigger trend where we get more and more tools to produce, | manipulates and share information, while the tools to analyze | and filter information are lagging by at least half a | century. | Swizec wrote: | > In a world where every celebrity is having sex with | gorillas, doesn't such an image lose its charge? Will Norms | and values around sex/body shaming change? | | I hoped this would happen with social media. Everyone says | stupid shit online and everyone has past beliefs they've | outgrown. So what's the big deal? | | Instead we went the opposite way. Everyone is super self | conscious and censoring at all times because you never know | who's gonna take it out of context and make a big deal. | seydor wrote: | This is not the same though, it's about being able to deny | that that photo of you is real, not that it's taken out of | context. | Swizec wrote: | Well what's a "real" photo? | | If I take your photo and photoshop (very well) different | surroundings. Is it a real photo if you? | | If I photoshop your face (very well) onto a different | body. Is that real? | | If I feed your photos into a model that can create | realistic versions of those photos in different poses or | with different facial expressions. Is that real? | | They all start with something that is very definitely a | real photo. You can't (yet? ever?) generate a realistic | photo of a specific person from a textual description. | The machinery needs a source. | seydor wrote: | A real photo is one created by photons outside the camera | Swizec wrote: | You'll be surprised to learn that doesn't work without | some amazing tech to process the photons. Different | settings will produce a different photo. | | Hell, just changing focal length makes a bigg difference | to what your face looks like: | https://imgur.io/gallery/ZKTWi no digital manipulation | required. | | Which of those faces is "real"? They're all just | recording photons hitting the camera, but look very | different. | | It gets even worse when we start talking about colors. | For example: it took cameras _decades_ before they could | accurately capture black faces. Where accurately means | "an average person would say it looks right" | | https://www.vox.com/2015/9/18/9348821/photography-race- | bias | kortex wrote: | Simple: a "real" photo is one in which a light field from | the real world impinges on a photosensitive media (CCD, | film), and directly encodes that information, with some | allowance for global light levels, gain, ISO, and speed. | Anything else is a modification therein. HDR, multi- | exposure compositing, etc, aren't truly "real". They may | be 99% real, but aren't 100% real. If you _crop_ it, it | 's 99.9% real (we have models which can detect cropping | and even from which region it originated, obviously it | can't reconstruct the missing data). | | Yes, by that definition, most photographs already aren't | real. | Swizec wrote: | > Yes, by that definition, most photographs already | aren't real. | | What if I use no digital manipulation at all, but play | around with focal lengths or perspective to produce the | desired effect. Is that a real photo? | | For example from covid reporting: | https://twitter.com/baekdal/status/1254460167812415489 | simondotau wrote: | So if I take a photo of a photoshopped photo, it is a | real photo. | taylorportman wrote: | It is an interesting point [the hope that values will | adapt to reflect typical mischievous patterns of social | dynamics within various clusters].. I used to marvel | during the dotcom era at the salivating delight of "the | internet never forgets" of college students caught | smoking a bong and the subsequent impact on their career | (virtue signalling?).. I saw hope that society could then | adapt about ridiculous FUD biases. There is a strange | relationship of scarcity & opportunity and these windows | into souls painting a distorted picture of the darker | side of our ambitions. It is a cancer. Also, humans are | always testing boundary conditions - curiosity, | discontent, security,insecurity. Some inherent | desperation of people jockeying for the next path from | frying pan to fire in the search for greener pasture | breeds opportunism that seems certain to favor | desperation and negativity. | tshaddox wrote: | That seems unlikely, given that _making up hurtful stories | about people and transmitting them via text or voice_ is | still a thing. Everyone knows that anyone can make up any | story they want without any technology whatsoever, and yet | spreading rumors is still a thing. | seydor wrote: | Not really. I can't think of a recent "leaked texts" that | the participants cannot not easily and plausibly deny (e.g. | elon's supposed messages to gates), or even voice messages. | Even most images can already be denied as photoshop if all | the witnesses agree. The only medium that is somewhat hard | to deny is videos, like sex tapes, but that's also not too | hard. I think there will soon be a race to make deep | learning pics look completely indistinguishable from phone | pics. | zone411 wrote: | It'll be hard to deny crypto-signed photos | https://petapixel.com/2022/08/08/sonys-forgery-proof- | tech-ad..., especially if they include metadata that | distinguish photos of AI generated images from normal | photos. | nerdponx wrote: | Are journalists savvy or ethical enough to give a shit? | What about the people reading/viewing/listening to the | news? | seydor wrote: | any camera can be hacked to plant an image in its | framebuffer | zone411 wrote: | Maybe at some point for some cameras. But not soon after | their release if they took steps to protect their | pipeline with hardware. | dustyleary wrote: | There are a few different kinds of 'secure enclaves' | implemented on chips, where you can have some degree of | trust that it "cannot" be faked. | | E.g. crypto wallets, hardware signing tokens, etc. | | We could imagine an imaging sensor chip made by a big- | name company whose reputation matters, where the imaging | sensor chip does the signing itself. | | So, Sony or Texas Instruments or Canon start | manufacturing a CCD chip that crypto signs its output. | And this chip "can't" be messed with in the same way that | other crypto-signing hardware "can't" be messed with. | | That doesn't seem too far-fetched to me. | | * edit: As I think about it, I think more likely what | happens is that e.g. Apple starts promising that any | "iPhoneReality(tm)" image, which is digitally signed in a | certain way, cannot have been faked and was certainly | taken by the hardware that it 'promises' to be (e.g. the | iPhone 25). | | Regardless of how they implement it at the hardware level | to maintain this guarantee, it is going to be a major | target for security researchers to create fake images | that carry the signature. | | So, we will have some level of trust that the signature | "works", because it is always being attacked by security | researchers. Just like our crypto methods work today. | There will be a cat-and-mouse game between manufacturers | and researchers/hackers, and we'll probably know years in | advance when a particular implementation is becoming | "shaky". | tshaddox wrote: | Perhaps that's somewhat true for famous people, although | there are plenty of examples of false stories (without | any forged evidence, literally just stories) causing real | embarrassment and damage to reputation. | | But it's even more true for non-famous people getting | bullied in their social groups, both online and offline, | and that's more what I was responding to (the "asshole | friends" in the original comment). | tgv wrote: | You're right. They call this an "ethical release", but what | ethics, I may ask. A profitable IPO is more likely to have been | their consideration. As other researchers before them, they are | willingly releasing something with the potential to do harm, or | pave the way for it, washing their hands in innocence. | choppaface wrote: | Really, the licensing is most interesting? There's a lot of | public info about training and development too. | | The license itself is pretty irrelevant. What people will | actually do with the training blueprints, and how fast things | will evolve.. now that's interesting. | Geee wrote: | I think the actual problem is that this gives plausible | deniability against photographic evidence, which might result | in increase of bad behavior. Even cameras which | cryptographically sign their output can't prove that the input | was actually photographed from the real world, or if it's just | an image of an image. | jl6 wrote: | Can this be used to create Imagen/DALL-E levels of image quality | on consumer GPUs? | andybak wrote: | For many classes of prompts - yes. It has less semantic | awareness in some regards but it's in the right ballpark. | toxik wrote: | Takes 10 GiB but yes. | mkaic wrote: | This is one of the most important moments in all of art history. | Millions of people just got unconditional access to the state-of- | the-art in AI text-to-image for absolutely free less the cost of | hardware. I have an Nvidia GPU myself and am thrilled beyond | belief with the possibilities that this opens up. | | Am planning on doing some deep dives into latent-space | exploration algorithms and hypernetworks in the coming days! This | is so, so, so exciting. Maybe the most exciting single release in | AI since the invention of the GAN. | | EDIT: I'm particularly interested in training a hypernetwork to | translate natural language instructions into latent-space | navigation instructions with the end goal of enabling me to give | the model natural-language feedback on its generations. I've got | some rough ideas but haven't totally mapped out my approach yet, | if anyone can link me to similar projects I'd be very grateful. | egypturnash wrote: | It's the opposite of exciting if art is your job, lemme tell | you. | grandmczeb wrote: | Depends on your perspective I guess. The visual artists I | know are super excited about AI art. | squeaky-clean wrote: | I have a friend who works as an artist and he's excited and | nervous about this. But also he's trying to learn how to use | these well. If you try these AI out, there's definitely an | art to writing good prompts that gets what you actually want. | Hopefully these AI will become just another brush in the | artist's palette rather than a replacement. | | I hope these end up similar to the relationship between | Google and programming. We all know the jokes about "I don't | really know how to code, I just know how to Google things". | But using a search engine efficiently is a very real skill | with a large gap between those who know how and those who | don't. | squeaky-clean wrote: | Replying to myself because I just had a chat with him about | this. He's thinking of getting a high end GPU now, lol. | | Some ideas of how this could be useful in the future to | assist artists: | | Quickly fleshing out design mockups/concepts is the obvious | first one that you can do right now. | | An AI stamp generator. Say you're working on a digital | painting of a flower field. You click the AI menu, "Stamp", | a textbox opens up and you type "Monarch butterfly. Facing | viewer. Monet style painting." And you get a selection of | ai generated images to stamp into your painting. | | Fill by AI. Sketch the details of a house, select the Fill | tool, select AI, click inside one of the walls of the | house, a textbox pops up, you write "pre-war New York brick | wall with yellow spray painted graffiti" | fithisux wrote: | AI will just replace existing jobs. It is unethical since a | lot of people will be destroyed but our leaders sponsored by | the majority's insatiable appetite for power will soon make | it legal. | | AI is a new tool that will automate away a lot of workers | like other machines. | | What happens with these workers is what defines us. | boredemployee wrote: | And there are a lot of influencers saying that it will be | "really nice that AI will replace the boring jobs so we can | focus on creative/fulfilling life" yeah right... | cercatrova wrote: | Ironically, with Moravec's Paradox, (digital) creative | tasks will probably be automated while the boring tasks | of moving boxes around might not be for a while: | | > _Moravec 's paradox is the observation by artificial | intelligence and robotics researchers that, contrary to | traditional assumptions, reasoning requires very little | computation, but sensorimotor and perception skills | require enormous computational resources. The principle | was articulated by Hans Moravec, Rodney Brooks, Marvin | Minsky and others in the 1980s. Moravec wrote in 1988, | "it is comparatively easy to make computers exhibit adult | level performance on intelligence tests or playing | checkers, and difficult or impossible to give them the | skills of a one-year-old when it comes to perception and | mobility"._ | | https://en.wikipedia.org/wiki/Moravec%27s_paradox | mkaic wrote: | (OP here) I agree. I am an artist (not by trade but by | lifelong obsession) in several different mediums, but also an | AI engineer--so I feel a weird mixture of emotions tbh. I'm | thrilled and excited but also terrified lol. | egypturnash wrote: | It's my fucking job. I've spent my whole fucking life | getting good at drawing. I can probably manage to keep | finding work but I am really not happy about what this is | going to do to that place where a young artist is right at | the threshold of being good enough to make it their job. | Because once you're spending most of your days doing a | thing, _you start getting better at it a lot faster_. And | every place this shit gets used is a lost opportunity for | an artist to get paid to do something. | | I wanna fucking punch everyone involved in this thing. | cercatrova wrote: | I mean I get what you're saying, sucks to have someone or | something take your job, but isn't this a neo-luddite | type argument? AI is gonna come for us all eventually. | egypturnash wrote: | Please save this comment and re-read it when a new | development in AI suddenly goes from "this is cute" to | "holy fuck my job feels obsolete and the motherfuckers | doing it are not gonna give a single thought to all the | people they're putting out of work". Thank you. | cercatrova wrote: | Look at that, you said the same thing to me 8 days ago | [0]. I'll stick to the same rebuttals you got for that | comment as well, namely that AI comes for us all, the | only thing to be done is to adapt and survive, or perish. | Like @yanderekko says, it is cowardice to assume we | should make an exception for AI in our specific field of | interest. | | [0] https://news.ycombinator.com/item?id=32462157 | visarga wrote: | You got to move one step higher and work with ideas and | concepts instead of brushes. AI can generate more imagery | than was previously possible, so it's going to be about | story telling or animation. | egypturnash wrote: | I make comics, it's already about storytelling and ideas | as much as it is about drawing stuff. I make comics in | part because _I like drawing shit_ and that gives me a | framework to hang a lot of drawings on. I _like_ the set | of levels I work at and don 't want to change it. I've | spent an _entire fucking lifetime_ figuring out how to | make my work something I enjoy and I sure bet nobody | involved in this is gonna fling a single cent in the | direction of the artists they 're perpetrating massive | borderline copyright infringement upon. | | But here's all these motherfuckers trying to automate me | out of a job. It's not even a boring, miserable job. It's | a job that people dream of having since they were kids | who really liked to draw. Fuck 'em. | redavni wrote: | Eh...the countdown to Photoshop and Blender integrating | AI support has already started. | egypturnash wrote: | I wanna punch everyone involved in that, too. | | Admittedly as someone who's been subscribing to Creative | Cloud for a while I already wanna punch a lot of people | at Adobe so the people working on this particular part of | Photoshop are gonna have to get in line. | boredemployee wrote: | Yep. I'm glad that didn't hit other spaces, like music, _yet_ | zone411 wrote: | Check out the melodies I made with an AI assistant I | created (human-in-the-loop still but much quicker than if I | tried to come up with them from the scratch): https://www.y | outube.com/playlist?list=PLoCzMRqh5SkFwkumE578Y.... There | are also good AI tools for other parts of making music, | like singing voice generation. | fezfight wrote: | Kinda like the creation of copilot and it's ilk for | programmers, and gpt3 for writers. Ive seen some talk | recently around 'prompt engineers'... Probably to some | extent, every job will become prompt engineering in some way. | | Eventually I suppose the AIs will also do the prompts. | | At which point I hope we've all agreed to a star trek utopia, | or it's gonna get real bad. Or maybe it'll get way better. | egypturnash wrote: | Yeah, if we're gonna replace every fucking profession with | a half-assed good-enough AI version, what're we even here | for? We're sure not all gonna survive in a capitalist | society where you have to create some kind of "value" to | earn enough money to pay for your roof and your food and | your power. | | IIRC there is some vague "it sure got real bad" somewhere | in the Trek timelines between "capitalism ended" and "post- | scarcity utopia" and I sure am not looking forwards to | living through those times. Well, I'm looking forwards to | part of that, I'm looking forwards to the part where we | murder a lot of landlords and rent-seekers and CEOs and | distribute their wealth. That'll be good. | bottlepalm wrote: | Next let's get rid of all the artists and replace them | with AI. Redistribute their skills so we can all make | art. Oh wait that just happened. You're a hypocrite that | wants to redistribute the wealth of others, but not your | own. | | Also joking about murdering people is bad taste and not | how you convey a point or win an argument. Very low | class. | ALittleLight wrote: | I think right now we could setup the AI's to do the | prompts. You type in a vague description - "gorilla in a | suit" and that is passed to GPT-3's API with instructions | to provide a detailed and vivid description of the input in | X style where X is one of several different styles. GPT-3 | generates multiple prompts, the prompts passed to Stable | Diffusion, the user gets back a grid of different images | and styles. Selecting an image on the results grid prompts | for variation, possibly of the prompt and the images. | tailspin2019 wrote: | > state-of-the-art in AI image-to-text | | I think you meant text-to-image! | mkaic wrote: | oh heck, you're right! I edited the original comment, thanks. | [deleted] | bilsbie wrote: | > particularly interested in training a hypernetwork to | translate natural language instructions into latent-space | navigation instructions with the end goal of enabling me to | give the model natural-language feedback on its generations. | | What are you doing exactly? | DougBTX wrote: | Imagine every conceivable image is laid out on the ground, | images which are similar to each other are closer together. | You're looking at an image of a face. Some nearby images | might be happier, sadder, with different hair or eye colours, | every possibility in every combination all around it. There | are a lot of images, so it is hard to know where to look if | you want something specific, even if it is nearby. They're | going to write software to point you in the right direction, | by describing what you want in text. | | Here's an example of this sort of manipulation: | https://arxiv.org/abs/2102.01187 | haneefmubarak wrote: | AFAICT: making a navigation/direction model that can | translate phrase-based directions into actual map-based | directions, with the caveat that the model would be updated | primarily by giving it feedback the same way that you would | give a person feedback. | | Sounds only a couple of steps removed from basically needing | AGI? | space_fountain wrote: | I suspect you'd want to start by trying to translate | differences between images into descriptive differences. | Maybe you could generate examples by symbolic manipulation | to generate pairs of images or maybe nlp can let us find | differences between pairs of captions? Large nlp models | already feel pretty magical to me and encompass things that | we would have said required AGI until recently so it seems | possible, though really tough | daenz wrote: | >This is one of the most important moments in all of art | history. | | I agree, but not for the reasons you imply. It will force real | artists to differentiate themselves from AI, since the line is | now sufficiently blurred. It's probably the death of an era of | digital art as we know it. | gjsman-1000 wrote: | Maybe this will signal a return to "real" non-digital artwork | and methods... | quadcore wrote: | Wait here, this shit hasnt hit France yet :) | | Those models are trained on artists work and put those same | artists out of work. When people will register I dont think | this is gonna fly. | ben_w wrote: | > Wait here, this shit hasnt hit France yet. | | What is special about France in this case? | vintermann wrote: | Labour protections/willingness to strike, I suspect they | mean. But I don't buy it. I've seen far too many people | who "should" be worried about this technology instead be | absolutely in love with it. | | Case in point, the comic artist Jens K (who, f.d, I | support on Patreon): | https://twitter.com/jenskstyve/status/1560360657148682242 | [deleted] | telesilla wrote: | Did we say the same in the 80s when audio sampling became a | thing? We accepted it (after obligatory legal battles) and | moved on, giving rise to the Creative Commons. | wwwtyro wrote: | Yeah, the fact that these models are necessarily based on | existing works leaves me hopeful that humans will remain the | leaders in this space for the time being. | soulofmischief wrote: | Human works are needed to create the initial datasets, but | an increasing amount of models use generative feedback | loops to create more training data. This layer can easily | introduce novel styles and concepts without further human | input. | | The time is coming where we will need to, as patrons, | reevaluate our relationships with art. I fear art is | returning to a patronage model, at least for now, as | certainly an industry which already massively exploits | digital artists will be more than happy to replace them | with 25% worse performing AI for 100% less cost. | BeFlatXIII wrote: | For those who are lucky enough to make it, I foresee | patronage as being much more stable than making art to | sell to the masses/corporate ad contracts. | quadcore wrote: | The generated pictures that are posted in the blog post | are superior than the average artist work. Which isnt | surprising, AI corrects "human mistake" (e.g. | composition, colors, contrasts, lines, etc) easily. | archagon wrote: | Why would people want to consume art that says nothing | and means nothing? While this technology is fascinating, | it produces the visual equivalent of muzak, and will | continue to do so in perpetuity without the ability to | reason. | jmfldn wrote: | That's the problem for me too. This tech is cool for | games, stock images etc but for actual art it's pretty | meaningless. The artist's experience, biography and | relationship with the world and how that feeds into their | work is the WHOLE point for me. I want to engage, via any | real artistic product, with a lived human experience. | Human consciousness in other words. | | To me this technology is very clever but it's meaningless | as far as real art goes, or it's a sideshow at best. | Perhaps best case, it can augment a human artistic | practice. | visarga wrote: | It's easy to generate believable backstory. A large LM | comes to write the bio of the "artist" and even a few | press interviews. If you like you can chat with it, and | of course request art to specification. You can even keep | it as a digital companion. So you can extend the realism | of the game as much as you like. | ben_w wrote: | How much is really being said by the highest dollar- | valued modern music? | | https://youtu.be/oOlDewpCfZQ and | https://youtu.be/L2cfxv8Pq-Q come to mind for different | reasons. | robocat wrote: | Can photography be good art? Is Marcel Duchamp (found | object) art? Can good art be discovered almost | serendipitously, or can good art only be created by | slowly learning and applying a skill? | | I think art is mostly about perception and selection, by | the viewer. There are others that think art is more about | the crafting process by the artist. How do you tell the | difference between an artist and a craftsperson? | | One way I categorise artists I have met is engineer-type | artists versus discovery-type artists: | https://news.ycombinator.com/item?id=31981875 | | Disclaimer: I am engineer. | scarmig wrote: | We can tell the difference between muzak and "real | music"; we just dislike the muzak. But the real risk and | likelihood is that we get to the point that AI will be | generating art that is indistinguishable from human- | generated art, and it's only muzak if someone subscribes | to the idea that the content of art is less relevant than | its provenance. Some people will, particularly rich | people who use art as a status signifier/money laundering | vehicle, but mass media artists will struggle to find | buyers among the less discerning mass audience. | buildartefact wrote: | Marvel Studios makes billions of dollars every year | archagon wrote: | And I'm fairly confident in saying that AI will never be | able to generate a Marvel movie! (Not in our lifetimes, | anyway.) | BobbyJo wrote: | Humans are trained on other humans' work as well though. Is | there a type of ideological or aesthetic exploration that | can't be expressed as part of an AI model? | derac wrote: | Making art is already not making much money for the vast | majority of producers (outside of 3d modeling). There really | aren't very many jobs in making art. I'd reckon most people | are artists because they love doing it. | rvz wrote: | Great news of the public release, just look at the melting pot of | creativity and innovation already being posted here: [0] Much | better than DALLE-2. | | Well done to the Stable Diffusion team for this release! | | [0] https://old.reddit.com/r/StableDiffusion/ | mmastrac wrote: | How hard would this be to re-train to more domain-specific | images? For example, if I wanted to teach it more about specific | birds, cars or plane models? | serf wrote: | this one was incredibly easy to trick into producing uncensored | nudes. | | which. personally.. I think is great.. but to each their own | | (NSFW!!!) this ween does not exist : | https://i.ibb.co/D7qJ7HC/23456532.png | amrrs wrote: | They have a content filter classifier at the top wondering how | this escaped that. | GaggiX wrote: | You can disable it | contravariant wrote: | Don't tell it to make stuff in the style of Junji Ito. | bluejellybean wrote: | While neat, and no doubt impressive, it still utterly fails on | prompts that should be completely reasonable to any sane human | being/artist. | | Take something like "A cat dancing atop a cow, with utters that | are made out of ar-15s that shoot lazer-beam confetti". A vivid | description should be aroused in your head, and no doubt, I could | imagine an artist have a lot of fun creating such a | description... Alas, what the model spits out is pure unusable | garbage. | topynate wrote: | The referent of "utters" (sic) is ambiguous, so I can imagine a | model having more difficulty with it than usual. Regardless, | the current SOTA does need more specific and sometimes | repetitive prompting than a human artist would, but it's | surprising how much better results you can get from SOTA models | with a bit of experience at prompt engineering. | bluejellybean wrote: | This is, in part, what I'm trying to point out, it's an | obvious typo given the context, and something that you or I | would be able to pick up on, yet it completely breaks (it | spit out a bunch of weird confetti cats for me). Perhaps I'm | being a little harsh, but if it requires word-perfect tuning | and prompt engineering, it speaks to something about the | 'stupidity' of these models. It's a neat trick, but to call | it anything in the realm of artificial intelligence is a bit | of a joke. | TulliusCicero wrote: | More complex/weirder prompts aren't going to work yet, no. | | What will probably happen with these models is that for more | advanced stuff, you may using the "inpainting" that Dall-E | already has going, where you can sort of mix and match and | combine images. That way you could have the cat, for example, | rendered separately, thereby simplifying each individual | prompt. | h2odragon wrote: | ITYM "udders" | | also try "teats" | naillo wrote: | Fucking awesome | mysterydip wrote: | Really interesting. I wonder if at some point it would be | possible to optimize a network for size and speed by focusing on | a specific genre, like impressionist or only pixel art. I like | that I can get an image in any style I want, but that has to | increase the workload substantially. | badsectoracula wrote: | Is there any way to download this on my PC and run it offline? | Something like a command-line tool like $ | ./something "cow flying in space" > cow-in-space.png | | that runs with local-only data (i.e. no internet access, no DRM, | no weird API keys, etc like pretty much every AI-related | application i've seen recently) would be neat. | cercatrova wrote: | Yes, clone the repo (https://github.com/CompVis/stable- | diffusion), download the weights and follow the readme for | setting up a conda environment. I am presently doing so on my | RTX 3080. | isoprophlex wrote: | I can't believe i sold my rtx2070 last month, aaargh...! | digitallyfree wrote: | As an aside I wonder how performance would be like running this | on CPU (with the current GPU shortage this might well be a | worthwhile choice). Even something like 30 minutes to generate | an image on a multicore CPU would greatly increase the number | of people able to freely play with this model. | drexlspivey wrote: | Yes but you need a GPU with 10Gb of vRAM | GaggiX wrote: | Yes you can | https://github.com/huggingface/diffusers/releases/tag/v0.2.3 | (probably the easiest way) | timmg wrote: | Garr: | | > And log in on your machine using the huggingface-cli login | command. | | I find that annoying. I guess it is what it is. | mkaic wrote: | Yes, that's actually the biggest reason this is such a cool | announcement! You just need to download the model checkpoints | from HuggingFace[0] and follow the instructions on their Github | repo[1] and and you should be good to go. You basically just | need to clone the repo, set up a conda environment, and make | the weights available to the scripts they provide. | | [0] https://huggingface.co/CompVis/stable-diffusion [1] | https://github.com/CompVis/stable-diffusion | | Good luck! | vintermann wrote: | You need a decent GPU, though. I suspect my 6080MiB won't cut | it any longer :( | miohtama wrote: | Is Apple M1 support expected soon? Because even if Apple's | chips are slower, they have plenty of RAM on laptops. I saw | some weeks ago it was coming, but I am not sure where to | follow the process. | neurostimulant wrote: | You're going to need at least 10GB VRAM. My SFF pc with 4GB | VRAM can only run dalle mini / craiyon :( | mempko wrote: | Not if you change the precision to float16. Should work | on a smaller card. Tried on a 1080 with 8GB and it works | well. | krisoft wrote: | How would one do that? | | ----- | | Sorry my bad, found the answer. One simply adds the | following flags to the | StableDiffusionPipeline.from_pretrained call in the | example: revision="fp16", torch_dtype=torch.float16 | | Found it in this blogpost: | https://huggingface.co/blog/stable_diffusion | | mempko thank you for your hint! I was about to drop a not | insignificant amount of money on a new GPU. | | What does one lose by using float16 representation? Does | it make the images visually less detailed? Or how can one | reason about this? | [deleted] | ompto wrote: | There's a version that's a bit slower but more memory | efficient https://github.com/basujindal/stable-diffusion | that runs on 6GB too. | naillo wrote: | This should be possible if someone just exported them to tflite | or onnxruntime etc (quantization could help a ton too). Not | sure why ppl haven't yet. Sure it'll come in the next few days | (I might do it). | [deleted] | jsmith45 wrote: | I _think_ the answer is yes, but setup is a bit complicated. I | would test this myself, but I don 't have an NVIDIA card with | at least 10GB of VRAM. | | One time: | | 1. Have "conda" installed. | | 2. clone https://github.com/CompVis/stable-diffusion | | 3. `conda env create -f environment.yaml` | | 4. activate the Venv with `conda activate ldm` | | 5. Download weights from https://huggingface.co/CompVis/stable- | diffusion-v-1-4-origin... (requires registration). | | 6. `mkdir -p models/ldm/stable-diffusion-v1/` | | 7. `ln -s <path/to/model.ckpt> models/ldm/stable- | diffusion-v1/model.ckpt`. (you can download the other version | of the model, like v1-1, v1-2, and v1-3 and symlink them | instead if you prefer). | | To run: | | 1. activate venv with `conda activate ldm` (unless still in a | prompt running inside the venv). | | 2. `python scripts/txt2img.py --prompt "a photograph of an | astronaut riding a horse" --plms`. | | Also there is a safety filter in the code that will black out | NSFW or otherwise expected to be offensive images (presumably | also including things like swastikas, gore, etc). It is trivial | to disable by editing the source if you want. | cypress66 wrote: | I haven't gotten around it, but I remember reading on /g/ | that you can make it run on 5GB (sacrificing accuracy). | | You should check their threads there, there's some good info. | entrep wrote: | Thanks for these instructions. | | Unfortunately I'm getting this error message (Win11, 3080 | 10GB): | | > RuntimeError: CUDA out of memory. Tried to allocate 3.00 | GiB (GPU 0; 10.00 GiB total capacity; 5.62 GiB already | allocated; 1.80 GiB free; 5.74 GiB reserved in total by | PyTorch) If reserved memory is >> allocated memory try | setting max_split_size_mb to avoid fragmentation. See | documentation for Memory Management and | PYTORCH_CUDA_ALLOC_CON | | Edit: | | >>> from GPUtil import showUtilization as gpu_usage | | >>> gpu_usage() | | | ID | GPU | MEM | | | ------------------ | | | 0 | 1% | 6% | | | Edit 2: | | Got this optimized fork to work: | https://github.com/basujindal/stable-diffusion | orpheansodality wrote: | I also have a 10g card and saw the same thing - to get it | working I had to pass in "--n_samples 1" to the command, | which limits the number of generated images to 2 in any | given run. This has been working fine for me | squeaky-clean wrote: | There's some additional discussion on running it locally here | | https://old.reddit.com/r/StableDiffusion/comments/wuyu2u/how... | Traubenfuchs wrote: | So is there a simple way to do this online? I have no dedicated | gpu and won't buy one just for this. We'd pay. | scoopertrooper wrote: | https://huggingface.co/spaces/stabilityai/stable-diffusion | | Here you go. | mromanuk wrote: | the queue to make images was 1 now is 5 | marc_io wrote: | https://beta.dreamstudio.ai/ | panabee wrote: | hotpot.ai should offer stable diffusion later today and | available via API as well. | | we're also building an open-source version of imagen, if anyone | likes working on this kind of applied ML (need ML + design | help). | fpgaminer wrote: | Played with it for a bit in DreamStudio so I could control more | of the settings. So far everything it generates is "high | quality", but the AI seems to lack the creativity and breadth of | understanding that DALL-E 2 has. OpenAI's model is better at | taking wildly differing concepts and figuring out creative ways | to glue them together, even if the end result isn't perfect. | Stable Diffusion is very resistant to that, and errs towards the | goal of making a high quality image. If it doesn't understand the | prompt, it'll pick and choose what parts of the prompt are | easiest for it and generate fantastic looking results for those. | Which is both good and bad. | | For example, I asked it in various ways for a bison dressed as an | astronaut. The results varied from just photos of astronauts, to | bisons on earth, to bisons on the moon. The bison was always | drawn hyper realistically, which is cool, but none of them were | dressed as an astronaut. DALLE on the other hand will try all | kinds of different ways that a bison might be portrayed as an | astronaut. Some realistic, some more imaginative. All of them | generally trying to fulfill the prompt. But many results will be | crude and imperfect. | | I personally find DALLE to be more satisfying to play with right | now, because of that creativity. I'm not necessarily looking for | the highest quality results. I just want interesting results that | follow my prompt. (And no, SD's Scale knob didn't seem to help | me). But there's also a place for SD's style if you just want | really great looking, but generic stuff. | | That said, the current version of SD was explicitly finetuned on | an "aesthetically" ranked dataset. So these results aren't really | surprising. I'm sure the next generations of SD will start | knocking DALLE out of the park in both metrics. And, of course, | massive massive props to Stability.ai for releasing this | incredible work as open source. Imagine all the tinkering and | evolving people are going to do on top of this work. It's going | to be incredible. | GenericPoster wrote: | Interesting. I took a stab at your prompt and SD really | struggles. It just completely ignores part of the prompt. Even | craiyon puts in an effort to at least complete the entire | prompt. | | The bison is very realistic at least. So maybe the future is | different models that have different specialties. | | Edit: managed to get this one after a few more tries | https://imgur.com/a/3061n5d | ralfd wrote: | I read that you have to give longer prompts to Stable | diffusion. | | This is my bison astronaut: | | https://i.imgur.com/ohIuG6F.png | | The prompt was: | | "A bison as an astronaut, tone mapped, shiny, intricate, | cinematic lighting, highly detailed, digital painting, | artstation, concept art, smooth, sharp focus, illustration, art | by terry moore and greg rutkowski and alphonse mucha" | naillo wrote: | Img2img and inpainting is usually a lot easier to get cool | results (with a lot of control on your end) with in my view. | Tenoke wrote: | I'm getting a 403 on the Colab (while successfully logging in and | providing a huggingface token). Is it already disabled? Do you | have to pay huggingface to download the model? It's unclear from | the Colab and post where the issue is. | punkspider wrote: | The solution seems to be to visit | https://huggingface.co/CompVis/stable-diffusion-v1-4 and check | a checkbox and click the button to confirm access. | | The full error I also got was: HTTPError: 403 | Client Error: Forbidden for url: | https://huggingface.co/api/models/CompVis/stable- | diffusion-v1-4/revision/fp16 | | I visited https://huggingface.co/api/models/CompVis/stable- | diffusion-v... and saw {"error":"Access to | model CompVis/stable-diffusion-v1-4 is restricted and you are | not in the authorized list. Visit | https://huggingface.co/CompVis/stable-diffusion-v1-4 to ask for | access."} | | Eventually I focused and realized I need to visit that URL to | solve the issue. Hope this helps. | johnsimer wrote: | yeah visit the url and click the checkbox and Accept | hunkins wrote: | This release changes society forever. Free and open access to | generate a hyper-realistic image via just a text prompt is more | powerful than I think we can imagine currently. | | Art, media, politics, conspiracy theories; all of it changes with | this. | gjsman-1000 wrote: | Eh... if I was making conspiracy theories, it's not like | Photoshop hasn't existed for decades already, with far more | predictable results. | johnfn wrote: | Photoshop requires hours of work from a skilled professional | to create results of decent quality. Now anyone can do it for | free, virtually instantaneously. | seydor wrote: | This is like a gazillion photoshops being released in the | wild. Things change with scale, and there is a threshold | where, if enough people start doubting often enough, then all | the people will doubt all the time | TulliusCicero wrote: | Photoshop requires skill. This mostly doesn't. | [deleted] | jcims wrote: | >it's not like Photoshop hasn't existed for decades already, | with far more predictable results. | | Agreed. For me those results are predictably shit. Every | time. | realce wrote: | Photoshop requires experience and some talent, this doesn't. | If I was some small rebel group in Africa or the Middle East | with basically no money or training, I'd use this tool every | single day until I was in power, or I'd frame my opposition | as using it against the People. | | Everyone just got their own KGB art department. | roguas wrote: | Try and do that. Likely those people won't care too much | about you, they lean towards authority figures in their | community. It is way easier to find those guys and corrupt | them. Rather than running some underground news agency | changing minds of millions of people. | | In fact quite often the former is the case + cheaper + less | time to execute. "Western" societies will be more resilient | to this scenario. So, mostly its gonna be a lot of | "political" art we gonna see. | vagabund wrote: | Longterm, multimodal generative models will be society- | altering. But right now this is just a really cool toy. | rvz wrote: | Yes. This changes everything. | | Until this point, you _really_ cannot believe any image you see | on the internet anymore. | kertoip_1 wrote: | Ahh, this technology moves so fast I'm not even able to even keep | up with reading about it. Not to mention trying it myself. | aaroninsf wrote: | Price check for A6000 GPU: $4500 USD | | Hmmm. | andybak wrote: | It's running fine on my gaming laptop. | zone411 wrote: | You don't need an A6000. You can run it on consumer GPUs. | Kiro wrote: | So how do I set up a server to generate images for me? An API I | can post anything to and get an image back. | derac wrote: | You'll need to know how to program. Try Python with Flask or | Bottle. | Kiro wrote: | Very funny. | | Anyway, this pretty much answered my question: | https://news.ycombinator.com/item?id=32556277 | derac wrote: | Sorry I misinterpreted what you meant, I wasn't meaning to | be glib. | TeeMassive wrote: | That explain Open-AI's price reduction email that I got this | morning. | throwaway-jim wrote: | will anyone think about the illustrators? Who will pay them when | amateurs can generate 9 good enough images from a text prompt? | MaxikCZ wrote: | And that terrible time a plow was discovered. So many people | with shovels lost their job. | simonw wrote: | If you want to try it out this seems to be the best tool for | doing so: https://beta.dreamstudio.ai/ | criddell wrote: | If I'm willing to buy a computer, any pointers on what I would | have to buy? I'm asking for specific models from a company like | Dell, Apple, or Lenovo? | Sohcahtoa82 wrote: | You'll need a GPU. One with a LOT of RAM, like an RTX 3090, | which has 24 GB. | msoucy wrote: | According to this post, it needs 6.9 Gb. So the 3070, | 3070-Ti, 3080, etc. can all run it. Sadly, my RTX 2060 is | below that limit... | smoldesu wrote: | Apparently the model decompresses, and it won't fit very | well on the 8gb models... I'm willing to give the max | settings a spin on my 3070ti, but I'm not very hopeful. | criddell wrote: | Would a Mac Pro with a Radeon Pro W5700X with 16 GB of | GDDR6 memory work? | barbecue_sauce wrote: | It says that NVIDIA chips are recommended but that they | are working on optimizations for AMD. This implies to me | that it probably involves CUDA stuff and getting it to | run on a Radeon would be potentially difficult (I am not | an expert on the current state of CUDA to AMD | compatibility, though). | Rzor wrote: | AMD's answer to CUDA is called ROCm. I've been doing a | little research on it since a few weeks ago and it seems | to be funky when not outright broken. It's absolutely | maddening that after all this time AMD doesn't have | proper tooling on consumer GPUs. | mattkevan wrote: | They're also working on M1/M2 support. | derac wrote: | Wait for the 4070, it should be around the perf of a | 3080/maybe 4090 for ~500 bucks if rumors hold up. It is | coming in a few months. NZXT used to make good pre-builts. | Not sure which others have a good rep. DO NOT BUY DELL. | criddell wrote: | What's wrong with Dell? Is Lenovo a bad option as well? | Does NZXT offer on-site service? | derac wrote: | You aren't getting on-site service with a consumer | product. There are plenty of 3rd party people who can | service your computer, though. It's like Lego. | | Dell uses crappy proprietary tech, poor quality | components, and they have an all around bad reputation. | | NZXT uses good components and they make some of the best | cases you can buy. | | I don't know much about Lenovo's desktop products. | | You might try posting on reddit.com/r/suggestapc and ask | about the best service contracts and high quality system | integrators. | | Edit: that particular reddit looks pretty dead actually. | The big one is r/buildapcsales , you can take a look at | their side bar or discord and ask around. | | One more thing, GamersNexus on YouTube does reviews of | pre builts and they are the best at this sort of thing. | Their community is likely very helpful as well. | | https://youtube.com/playlist?list=PLsuVSmND84QuM2HKzG7ipb | IbE... | | The biggest issue with most pre-builts is terrible | airflow making the expensive components throttle. The | (Dell) Alienwares are some of the worst for this. | hedora wrote: | I'd rent a cloud vm with a beefy gpu from paperspace or a | similar company. It'll run about $20 per month for casual | use. | mattkevan wrote: | Go for Google Colab Pro - about PS12 per month - or Pro+ at | PS45 per month for A100 and V100 Gpus. | jmfldn wrote: | I just tried it via this link. I'm not sure what I'm looking at | here but the results were extremely underwhelming. I've used | Dall E 2 and Midjourney extensively so I know what they're | capable of. Maybe I'm missing something? | derac wrote: | This Twitter thread has a lot of great comparisons of the | models with different prompts. | | https://twitter.com/fabianstelzer/status/1561019215754280963 | akvadrako wrote: | I've only used it via discord, but it's much better than | Midjourney and sometimes better than Dall-E. So maybe that | site isn't the same thing or you need to work on your | prompts. | jcims wrote: | I haven't seen anything beat midjourney for creating | atmosphere. | | DALL-E is great at making 'things' and generally good/great | at faces. | andybak wrote: | Every model I've used initially seemed poor compared to the | one I was just using. It takes time to figure out their sweet | spot and what kind of prompts they excel at. | | I've had a lot of great results from SD - but _different_ | great results to Dall-E. | alcover wrote: | Moral question: | | Such tool _will_ be used to generate lewd imagery involving | virtual minors. | | No way to prevent it upstream by outlawing feeding it real | content (whose possession already is illegal). Suffice to add | 'childrenize' layer onto adult NN or something. | | How will the legal system react ? Bundle it into illegal imagery, | period ? Maybe it's already the case - I think drawing made | public is, not sure. If not, on what ground could it be ? No real | minor would be involved in that production. | ta_99 wrote: | naillo wrote: | If it is made illegal it'll probably be applied to the sites | that distribute stuff like that, not this model itself. | miohtama wrote: | Thoughtcrime is not a thing in the US, do not know about other | countries | | https://en.wikipedia.org/wiki/Thoughtcrime | | Though Think of the children folk have often tried to make it | illegal. | tomtom1337 wrote: | > The final memory usage on release of the model should be 6.9 Gb | of VRAM. | | Do they actually mean GB or Gb? Can anyone confirm? | coolspot wrote: | Gigabytes not gigabits for sure | soperj wrote: | This is honestly the best one so far for the things I'm looking | to do. For weird prompts though it sometimes produces images that | just look blurry. | gjsman-1000 wrote: | DALL-E 2 just got smoked. Anyone with a graphics card isn't going | to pay to generate images, or have their prompts blocked because | of the overly aggressive anti-abuse filter, or have to put up | with the DALL-E 2 "signature" in the corner. It makes me wonder | how OpenAI is going to work around this because this makes DALL-E | 2 a very uncompetitive proposition. Except, of course, for people | without graphics cards, but it's not 2020 anymore. | at_a_remove wrote: | I have been considering building a "modern" computer and now | wonder exactly what I need to load this puppy up. | TulliusCicero wrote: | People are saying you need a GPU with 6.9GB of RAM for the | current model, so in practice at least an 8GB GPU. | | Thankfully, GPU prices have finally calmed down and you can | get one for a reasonable price. I think any of the RTX 3000 | series desktop GPU's should do it, for example. | Baeocystin wrote: | DALL-E's filters are so harsh that I find myself often in the | situation where I don't even understand how what I prompted | could possibly be in violation. | | It's a novel feeling, but utterly stifling when it comes to | actual creativity, and I'm not even trying to push any NSFW | boundaries, just explore the artspace. Once I can run | unfiltered on my own GPU, DALL-E will never get used by me | again. | vintermann wrote: | The usual modern Google experience: they won't tell you what | you did wrong. | derac wrote: | Openai isn't a Alphabet/google company | throwaway675309 wrote: | Midjourney also completely destroys DALL-e from a price | perspective, effectively allowing nearly unlimited generation | for approximately $50 a month. | | Even though DALL-E tends to be better at following prompt | details, you're inhibited from being able to explore the space | freely because of how prohibitively expensive it can become. | polygamous_bat wrote: | This, however, is unconditionally good for the end users. I | expect OpenAI to lower their prices significantly quite soon. | morsch wrote: | In fact, they announced lower prices today, going into effect | in September. | | https://openai.com/api/pricing/ | aaronharnly wrote: | That's for GPT-3 text generation, not the DALL-E 2 image | generator. Hopefully that will get pricing revised down | (and an official API) before long. | morsch wrote: | Oh, I see! Thanks, I should have read the mail more | closely. | TOMDM wrote: | Been playing with this for the past hour now on my RTX 2070. Each | image takes about 10 seconds. | | Results vary wildly but it can make some really great stuff | occasionally. It's just infrequent enough to keep you going "one | more prompt". | | Super addicting. | | Lookingforward to people implementing inpainting and all the | stuff that lets you do. | jupp0r wrote: | "Use Restrictions | | You agree not to use the Model or Derivatives of the Model: | | - In any way that violates any applicable national, federal, | state, local or international law or regulation; | | - For the purpose of exploiting, harming or attempting to exploit | or harm minors in any way; | | - To generate or disseminate verifiably false information and/or | content with the purpose of harming others; | | - To generate or disseminate personal identifiable information | that can be used to harm an individual; | | - To defame, disparage or otherwise harass others; | | - For fully automated decision making that adversely impacts an | individual's legal rights or otherwise creates or modifies a | binding, enforceable obligation; | | - For any use intended to or which has the effect of | discriminating against or harming individuals or groups based on | online or offline social behavior or known or predicted personal | or personality characteristics; | | - To exploit any of the vulnerabilities of a specific group of | persons based on their age, social, physical or mental | characteristics, in order to materially distort the behavior of a | person pertaining to that group in a manner that causes or is | likely to cause that person or another person physical or | psychological harm; | | - For any use intended to or which has the effect of | discriminating against individuals or groups based on legally | protected characteristics or categories; | | - To provide medical advice and medical results interpretation; | | - To generate or disseminate information for the purpose to be | used for administration of justice, law enforcement, immigration | or asylum processes, such as predicting an individual will commit | fraud/crime commitment (e.g. by text profiling, drawing causal | relationships between assertions made in documents, | indiscriminate and arbitrarily-targeted use)." | | The last point seems to be the only thing that's not illegal, all | other restrictions seem to be covered under "you are not allowed | to break laws", which is somewhat redundant. | dang wrote: | Recent and related: | | _Stable Diffusion launch announcement_ - | https://news.ycombinator.com/item?id=32414811 - Aug 2022 (39 | comments) | ChildOfChaos wrote: | So what do I need to run this and how? | andybak wrote: | Anything close to a decent modern gaming PC will do it fine. | I'm running on a laptop 3080 and I can generate 768x512 in | about 20 seconds (with a 30 second overhead per batch) | swagmoney1606 wrote: | superdisk wrote: | Just tried it, the results seem pretty poor, about on par with | Craiyon/DALL-E Mini. I don't think OpenAI should be worried quite | yet. | i_like_apis wrote: | No it's all about the prompt. It takes some getting used to, | you have to look at a bunch of prompts from experienced users. | | SD is waaay better than Craiyon. And better than Dalle2. | | Check out r/StableDiffusion | minimaxir wrote: | It depends on the domain. Artsy will do better with Stable | Diffusion, but realistic/coherent output with Stable Diffusion | is harder to do especially compared to DALL-E 2. | andybak wrote: | With all due respect, I've been using it for over a week and I | don't think you've given it a fair shot. | | There's plenty of cases it's worse than Dall-E and there's | plenty of cases where it's better. Overall it seems to show | less semantic understanding but it handles many stylistic | suggestions much better. It's definitely in the right ballpark. | | In fact I'm still using a wide range of models - many of which | aren't regarded as "state of the art" any more - but they have | qualities that are unique and often desireable. | treis wrote: | Can you give an example? I've done: | | A house painted blue with a white porch | | A dreamy shot of an alpaca playing lacrosse | | A red car parked in a driveway | | The last one was particularly crappy. It gave me a red house | with a driveway, but no car. And the house wasn't even really | a house. It superficially looked like one but was actually | two garages put together. | andybak wrote: | Here's some random prompts I've had nice results from: | iridescent metal retro robot made out of simple geometric | shapes. tilt shift photography. award winning | Scene in a creepy graveyard from Samurai Jack by Genndy | Tartakovsky and Eyvind Earle virus bacteria | microbe by haeckel fairytale magic realism steampunk | mysterious vivid colors by andy kehoe amanda clarke | etching of an anthropomorphic factory machine in the style | of boris artzybasheff origami low polygon | black pug forest digital art hyper realistic a | tilt shift photo of a creepy doll Tri-X 400 TX by gerhard | richter | | I guess I might have spent more time reading guides on | "prompt engineering" than you. ;-) I think maybe Dall-E is | more forgiving of "vanilla prompts". | | However I do get nice results from simpler prompts as well. | I just tend to use this style of prompt more often than | not. | mattkevan wrote: | Agreed. I still primarily use vqgan + clip, which is nowhere | near state of the art, but produces really interesting | results. I've spent a long time learning to get the best out | of it, and while the results aren't very coherent, it's great | at colour, texture, materials and lighting. | lxe wrote: | The adjustable safety classifier makes this release leapfrog | above the competition imho. | [deleted] | bilsbie wrote: | Is there a way to give it an image to manipulate as part of the | prompt? | neurostimulant wrote: | Maybe try the img2img script? | https://github.com/CompVis/stable-diffusion/blob/main/README... | davesque wrote: | Pretty cool. Although it's interesting that it can't seem to | render an image from a precise description that should have | something like an objectively correct answer. I tried prompts | like "Middle C engraved on a staff with a treble clef" and "An | SN74LS173 integrated circuit chip on a breadboard" both of which | came back with images that were nowhere close to something I'd | call accurate. I don't mean to detract from the impressiveness of | this work. But I wanted a sense of how much of a "threat" this | tech is to jobs or to skills that we normally think of as being | human. Based on what I'm seeing, I'd say it's still got a ways to | go before it's going to destroy any jobs. In its current form, it | mostly seems like a fun way to generate logos or images where the | exact details of the content don't matter. | nerdponx wrote: | I am generally of the "it's not threatening yet and won't be | for a while" camp, but in this _particular_ case it 's probably | just for lack of trying. These algorithms are essentially | enormous pattern-matching engines, so given enough data and | some task-specific engineering effort, I wouldn't be surprised | if you could build an "AI" circuit designer, like Copilot but | for electronics instead of code. | | Next-level autorouting would be cool, but it's still not going | to put the electrical engineering field out of business. | cercatrova wrote: | I've been looking forward to this. The license however strikes me | as too aspirational, and it may be hard to enforce legally: | | > You agree not to use the Model or Derivatives of the Model: | | > - In any way that violates any applicable national, federal, | state, local or international law or regulation; | | > - For the purpose of exploiting, harming or attempting to | exploit or harm minors in any way; | | > - To generate or disseminate verifiably false information | and/or content with the purpose of harming others; | | > - To generate or disseminate personal identifiable information | that can be used to harm an individual; | | > - To defame, disparage or otherwise harass others; | | > - For fully automated decision making that adversely impacts an | individual's legal rights or otherwise creates or modifies a | binding, enforceable obligation; | | > - For any use intended to or which has the effect of | discriminating against or harming individuals or groups based on | online or offline social behavior or known or predicted personal | or personality characteristics; | | > - To exploit any of the vulnerabilities of a specific group of | persons based on their age, social, physical or mental | characteristics, in order to materially distort the behavior of a | person pertaining to that group in a manner that causes or is | likely to cause that person or another person physical or | psychological harm; | | > - For any use intended to or which has the effect of | discriminating against individuals or groups based on legally | protected characteristics or categories; | | > - To provide medical advice and medical results interpretation; | | > - To generate or disseminate information for the purpose to be | used for administration of justice, law enforcement, immigration | or asylum processes, such as predicting an individual will commit | fraud/crime commitment (e.g. by text profiling, drawing causal | relationships between assertions made in documents, | indiscriminate and arbitrarily-targeted use). | | How can you prove some of these in a court of law? | lacker wrote: | The license does seem impossibly vague and broad. Usually what | happens when software projects use custom & demanding licenses | like this is that large companies refuse to allow the software | to be used because of the legal uncertainty, small companies | just use it and ignore the licensing constraints, and there are | never any lawsuits that clarify anything one way or another. If | that's fine with the authors of the project, they can just | leave the license vague and unclear forever. | alexb_ wrote: | Make sure your art isn't the wrong type of artistic! | jahewson wrote: | I dunno. I can imagine any of those points being the subject of | a civil suit and for someone to win damages, for e.g. | psychological harm. The parts talking about "effect" instead of | intent are of questionable enforceability - how can I agree not | to cause an unanticipated effect on a third party? I cannot. | But having said that, I can be asked to account for effects | that a "reasonable person" would anticipate, so there's that. | | These are all things that someone could sue over (especially in | California) and so they're wanting to place the responsibility | on the artist and not their tools. | [deleted] | fluidcruft wrote: | It seems like the sort of things you would require so that | HuggingFace doesn't get roped in as defendants in lawsuits | related to things that others do with the code. So if for | example someone builds something that generates medical advice | and gets sued for violating FDA requirements or damages or | whatever then HuggingFace can say that was not something they | allowed in the first place. | laverya wrote: | Yeah it really reads like "we know people are going to do | nasty things with this; we can't prevent that; please don't | sue us over it" to me | hedora wrote: | They really should have used blanket liability waiver text | and left it at that. | | I'm sure someone will find a way to sue them anyway. It | doesn't even call out using this to create derivative works | to avoid paying original authors copyright fees. | | On top of that, their logo is an obvious rip off of a Van | Gogh. It seems clear they're actively encouraging people to | create similar works that infringe active copyrights. They | should ask Kim Dotcom how that worked out for him. | ben_w wrote: | > On top of that, their logo is an obvious rip off of a | Van Gogh. It seems clear they're actively encouraging | people to create similar works that infringe active | copyrights. | | I don't think Van Gogh's works are under copyright any | more. At least not directly, recent photos of them may be | but that's the photos not the paintings that have a | copyright. | cercatrova wrote: | Indeed, it seems like a legal "wink wink, nudge nudge" sort | of thing. Well, as long as I can run stuff on my own GPU, | I'm satisfied. | digitallyfree wrote: | I'm actually seeing these types of conditions becoming more | common in software EULAs as well, as an boilerplate addon to | the usual copyright notices and legal disclaimers. Don't have | examples off the top of my head but I've seen clauses that | this application may not be used for the enablement of | violence, for discriminatory purposes, and so forth. It | really is a CYA sort of thing. | hedora wrote: | Each of the "for any use ... which has the effect of ..." | clauses probably bars any wide distribution of the output of | this model. | | Trivially: People have phobias of literally everything. | | They ban using it to "exploit" minors, presumably that prevents | any incorporation of it into any for-profit educational | curriculum. After all, they do not define "exploit", and | profiting off of a group without direct consent seems like a | reasonable interpretation. | | I am not a lawyer, but I wouldn't dream of using this for | commercial use with a license like this. This definitely | doesn't meet the bar for incorporation into Free or Open Source | Software. | nullc wrote: | That license from top to bottom is distilled "tell me you don't | know anything about art without telling me you don't know | anything about art". | | Interesting art challenges its audience. But even the most | boring art will still offend some-- it's the nature of art that | the viewer brings their own interpretation, and some people | bring an offensive one. | digitaLandscape wrote: | TigeriusKirk wrote: | I dunno. All I see is necessary ass covering. They're just | saying anyone who does these objectionable things did so | against our terms. | nullc wrote: | CYA doesn't necessitate creating a cause of action against | the users for engaging in what otherwise would be a legally | protected act of free expression. One can disclaim without | creating liability. | ambivdexterous wrote: | /ic/ is having daily meltdowns over this. I don't think the | internet at large is doing better, because even professional | concept artists are dialing it in now. Holy hell. | gjsman-1000 wrote: | What's /ic/? | alexb_ wrote: | https://4channel.org/ic/ | ambivdexterous wrote: | 4chan's Art Critique board | kache_ wrote: | I kept on warning them | | They ignored me | | Now however.. | howon92 wrote: | Huge kudos for stability.ai | faizshah wrote: | What's the license on the images produced by this model? | coolspot wrote: | CC0 - no one owns the copyright, so everyone is free to use | kazinator wrote: | That is not so; the CC0 explicitly states that patent and | trademark rights are not waived. | | Contrast that with, say, the Two-Clause BSD which says | "[r]edistribution and use in source and binary forms, with or | without modification, are permitted provided that the | following conditions are met [...]". | | Since trademark and patent rights are not mentioned, then | these words mean that even if the purveyor of the software | holds patents and/or trademarks, your redistribution and use | are permitted. I.e. it appears that a patent and trademark | grant is implied if a patent holder puts something under the | two-clause BSD. Or at least you have a ghost of a chance to | argue it in court. | | Not so with the CC0, which spells out that you don't have | permission to use any patents and trademarks in the work. | api wrote: | "This release is the culmination of many hours of collective | effort to create a single file that compresses the visual | information of humanity into a few gigabytes." | | If something like this is possible, does this mean there's | actually far less _meaningful_ information out there than we | think? | | Could you in fact pack virtually all meaningful information ever | gathered by humanity onto a 1TiB or smaller hard drive? Obviously | this would be lossy, but how lossy? | MaxikCZ wrote: | You can pack virtually all meaningful information ever gathered | by humanity onto a single bit, but its gonna be lossy. And what | is your definition of "meaningful information" anyway. | Meaningful today might not be meanigful yesterday. Nobody cares | about spin of each electron in my brain today, but in 4 | centuries my descendants will be like "if only we had that | information, we could simulate our great-...-great parent | today" ___________________________________________________________________ (page generated 2022-08-22 23:00 UTC)