2024-05-09 Hoping for the next AI winter ======================================== @taichara@chirp.enworld.org pointed me to a blog post that Creative Commons (CC) wrote in 2023 about the court cases against GitHub, Stable Diffusion, Midjourney and the claim that their use of copyrighted works to train the machines is copyright infringement. CC does not agree. > As Creative Commons has argued elsewhere, and others agree, I > believe that this type of use should be protected by copyright’s > fair use doctrine. In this blog post I will discuss what fair use > is, what purpose it serves, and why I believe that using copyrighted > works to train generative AI models should be permitted under this > law. – Fair Use: Training Generative AI I guess I don't mind that blog post. For one, Fair Use is a very USA thing and I don't live there. Furthermore, the CC licenses don't say "this material is fine for AI training" – so whether this is fine or not will be decided by the courts. I would only worry if CC turns out to be working on a version 5 of their licenses with an explicit exception for AI training. And even then, we can keep on using version 4 of the license. There is no automatic upgrade option. This is why I feel fine. Imagine the Dude from The Big Lebowski saying, "Yeah, well, that's just, like, their opinion…" It doesn't look as if CC will be changing the license. The courts still get to decide. If the courts decide in favour of free-for-all AI training, then rich people win for the moment. But even then, we can still nationalise their businesses, or prohibit AI training, or raise prices on electricity and water. The fight isn't over. Not by a long shot. Maybe this is a thing with lawyers. We trust people who share our values and we’d trust somebody when they write something along the lines of “AI training is terrible and AI use is dubious but what can you do, it’s in the hand of the courts, we certainly don’t support it.” But lawyers trust other lawyers who say “we’ll see what the courts say but our understanding is that nobody should stop AI training because that’s the trade-off copyright makes and we support people and corporations exercising their rights.” The apparent moral support on copyright grounds is revolting, but this approach seems typical to me. They say: The point of the law is this or that. We say: The result of these actions is bad for society and bad for the climate. They are externalising all the costs. The artists and authors who want to produce new works are losing their jobs. The people double-checking the computers and their training are burning out. The communities where these corporations build their computing centres suffer from water scarcity because the water is used to cool the computers. The countries where these corporations build their computing centres need to build more power plants because so much electricity is required to train their models. The move to renewable energies is delayed. Phasing out nuclear power is delayed and more radioactive waste ends up on our lands. > Sending this message was important to us. We considered ourselves to > be a powerful culture. This place is not a place of honor... no > highly esteemed deed is commemorated here... nothing valued is here. > – Long-term nuclear waste warning messages, Wikipedia Or we can believe in nuclear fusion: > AI models made up of billions of parameters require huge amounts of > energy to train. OpenAI's old GPT-3 system reportedly consumed 936 > megawatt hours … the average household consumes about 10.5 MWh per > year. That means training GPT-3 consumed as much energy as about 90 > households consume in a year. – Energy breakthrough needed to build > AGI, says OpenAI boss Altman, The Register Sadly, nuclear fusion only seems to be working inside the sun: > 1955: At the first Atoms for Peace meeting in Geneva, Homi J. Bhabha > predicts that fusion will be in commercial use within two decades. – > Timeline of nuclear fusion, Wikipedia And what for? The people who are looking for answers get served bullshit slop that's wrong and hurtful and racist and misogynist. We see this in customer service degrading, in search engines degrading, in websites degrading. The corporations don't mind that they're feeding us slop as long as they are making money. We are enabling this because these corporations are powerful and rich. They buy our politicians with money and promises of jobs, they don't pay the taxes they ought to pay and then they don't deliver on the jobs, and the jobs they offer are menial and subservient when we'd prefer to be creative and free. We thought we'd be writing and painting all day while the machines serve us but we end up serving them while they do the writing and painting, badly! It's hard to defend against these corporations because the knowledge required to defeat them is not known to us all. We don't read about it in the media. Instead, they volunteer for all the posts, offer their experts, pay for their own research, write their own glowing press releases and commission promising reports. This is how regulatory capture works. So yes, we should nationalise them, regulate them, force them to pay for the true cost of resources they use, outlaw their energy waste, force them to label their output, shun them. We should kick them from our company. Like a creeping disease their rot spreads, destroying the technology we built, the search engines, databases, websites, turning it all into AI slop. It is revolting. And maybe copyright is not the right tool to stop corporations from training their machines on our works. But in the big picture, they are still wrong. @eloquence@social.coop says: > The folks who are calling such training "theft" might regret what > they seem to be implicitly asking for, i.e. much stricter copyright. > Copyright law won't prevent Microsoft, Google, OpenAI or Adobe from > making shady licensing deals, but they'll prevent the free/open > community from keeping up. It's true, I don't want to ask for stricter copyright. But I still oppose all the things that are wrong with the current AI bubble. @mcc@mastodon.social recently wrote about a similar thing: We use free software licenses in order to distribute our software for the benefit of our fellow humans but not necessarily to train the machines of corporations that are responsible for all the slop. > Like, heck, how am I supposed to rely on my code getting preserved > after I lose interest, I die, BitBucket deletes every bit of > Mercurial-hosted content it ever hosted, etc? Am I supposed to rely > on Microsoft to responsibly preserve my work? Holy crud no. We want > people to want their code widely mirrored and distributed. That was > the reason for the licenses. That was the social contract. But if > machine learning means the social contract is dead, why would people > want their code mirrored? – mmc We're using copyright to keep the software free. If AI regurgitating our software from its impenetrable memories circumvents copyright it also circumvents our licenses and therefore pulls the claws and teeth free software has to fight for its freedom. This is why opposing the current AI bubble is important. No matter how I look at it, the big picture shows that this kind of AI is our enemy. It's just that copyright is the wrong weapon. This is why we continue fighting all the negatives these corporations bring to the table. And if it turns out that their business model doesn't work unless they bring all these negatives to the table, then I guess that's just that. We are under no obligation to support their business model. In fact, it is our right and our duty to regulate the world in order to improve it for all of us. We've been doing it for a very long time. Here's to hoping for the next AI winter. > In the history of artificial intelligence, an AI winter is a period > of reduced funding and interest in artificial intelligence research. > The field has experienced several hype cycles, followed by > disappointment and criticism, followed by funding cuts, followed by > renewed interest years or even decades later. – AI winter, > Wikipedia ​#Copyright #Artificial_Intelligence ​#Artificial_Intelligence 2024-05-16. Cory Doctorow also thinks that copyright is the wrong weapon: > Creative workers who hope that they can use copyright law to prevent > AI from changing the creative labor market are likely to be very > disappointed in court … Individual creative workers rarely have any > bargaining leverage over the corporations that license our > copyrights. … Rather than going to Congress alongside our bosses > from the studios and labels and publishers to demand more copyright, > we could go to Congress alongside every kind of worker, from > fast-food cashiers to publishing assistants to truck drivers to > demand the right to sectoral bargaining. – AI "art" and uncanniness