[HN Gopher] Update on KDP Title Creation Limits ___________________________________________________________________ Update on KDP Title Creation Limits Author : ilamont Score : 199 points Date : 2023-09-18 20:37 UTC (2 hours ago) (HTM) web link (www.kdpcommunity.com) (TXT) w3m dump (www.kdpcommunity.com) | bern4444 wrote: | It seem that as a society we are coming to realize that enabling | anyone to do anything on their own and at anytime isn't the best | of ideas. | | Verifiability and authenticity matter and are valuable. Amazon | has long had a problem of fake reviews. This issue with kindle | books seems an extension of that. Massive centralized platforms | like Amazon makes fraud more likely and is bad for the consumer. | | The "decentralization" that we need as a society is not in the | form of any crypto based technical capability but simply for the | size of the massive players to be reduced so competition can | reemerge and give consumers more options on where and how to | spend their dollars. Other E-book stores may just pop up that | develop relationships with publishers and disallow independent | publishing if amazon were forced to be broken up. | | I hope the FTC can begin finding a strategy to force some of | these massive corporations to split making it more likely for | there to be more competition. | willio58 wrote: | It seems Amazon cares more about polluting search results in | Kindle than polluting the search results in their own e-commerce | business. I think low-effort books generated by AI are much less | detrimental than sketchy physical products being shipped to your | door in 2 days or less. | crooked-v wrote: | It's probably about volume rather than quality. Sketchy copycat | product lines are still hard limited by the number of factories | and shipping operations in existence, while sketchy AI- | generated books can easily keep growing exponentially in number | for a while. | japhyr wrote: | It'll be interesting to see where this goes. Amazon has had ML- | generated garbage books for years now, and I assume they haven't | taken them down because they make money even when they sell | garbage. | | _Maybe_ there 's so much garbage coming in now that they finally | have to do something about it? I feel for people trying to learn | about technical topics, who aren't aware enough of this issue to | avoid buying ML-generated books with high ratings from fake | reviews. The intro programming market is full of these scam | books. | ehsankia wrote: | I thought it was more so filled with low quality mechanical | turk garbage books. | m463 wrote: | I was thinking about buying an air fryer. My search came up | with cookbooks specific to that air fryer, and I was intrigued. | I found a good 5-star book, but then I found that ALL the | 5-star reviews were submitted the same day. | | I complained, but Amazon defended the book as legitimate, and | since I hadn't purchased it, they would not take any action. | (to be honest, I assume frontline customer service reps don't | have much experience or power) | | So I purchased it, complained, got a refund and then they were | able to accept my complaint (after passing the complaint higher | in the food chain). | | Seriously, how hard was it amazon? I guess they're starting to | notice. | | Take a look at air fryer cookbooks - there are books specific | to most makes and models. But everything is ML copypasta all | the way up and down - the title, the recipes and the reviews | all seem to be generated garbage. | tiew9Vii wrote: | It's the same across all big tech. The size/volume for | complaint handling doesn't scale. It's either filtered out by | some machine learning algorithm or some poor person in a 3rd | world country getting paid next to nothing who reviews the | complaints so quality isn't of importance. | | There been a recent influx of scammers on Facebook local | groups. Air con cleaning, car valeting, everyone's calling | out the scammers in the comments yet when you click report to | FB the response is we have reviewed the post and it has not | breached our guidelines, would you like to block the user. | japhyr wrote: | I'm the author of _Python Crash Course_ , the best selling | introductory Python book for some time now. Years ago, | someone put out a book listing two authors: Mark Matthes and | Eric Lutz. That's just a simple juxtaposition of my name and | Mark Lutz, the author of O'Reilly's _Learning Python_. The | subtitle is obviously taken from my book 's subtitle as well. | I assume the text is an ML-generated mess, but I haven't | bought a copy to verify that. | | I used to comment on reviews for books like these explaining | what was happening, but Amazon turned off the ability to | comment on reviews a long time ago. | | I've spoken with other tech authors, and almost all of us get | emails from people new to programming who have bought these | kinds of books. If you're an experienced programmer, you | probably know how to recognize a legitimate technical book. | But people who are just starting to learn their first | language don't always know what to look for. This is squarely | on Amazon; they have blocked most or all of the channels for | people to directly call out bad products, and they have | allowed fake reviews to flourish and drown out authentic | reviews. | nektro wrote: | i stopped frequenting the dev.to community because the | average quality of articles just got so low it stopped | being worth my time | quickthrower2 wrote: | Conversely if you post something sophisticated there it | will likely bomb. A bunch of emojis and explaining JS | closures for the hundredth time. Does well! | arrowsmith wrote: | dev.to is blocked on HN for this reason (try submitting a | dev.to link; it won't appear under New.) | | There's an old thread where dang explains that it's | blacklisted (along with many many other sites) due to the | consistently poor article quality. | ocdtrekkie wrote: | I think the best way to recognize a legitimate tech book | is... visit a Barnes and Noble. If it's a publisher or | series you can find printed on the shelf, books are legit. | | Unfortunately online market "platforms" are pretty much | widely untrustworthy for any sort of informational | purposes. | failTide wrote: | also, just doing your research on any platform other than | Amazon helps. | rmbyrro wrote: | Why don't beginners start at Python.org, though? It's such | a great resource to learn the language. | | - it's free, unlike books | | - always up-to-date, unlike even the best book after a few | months | | - easy to choose: heck, there's only one official | documentation! No chance of making a mistake here! | Armisael16 wrote: | Are you suggesting people just go read the documentation | like an encyclopedia? I don't know a single person who | got their start programming by doing that - just about | everyone wants some sort of guide to help lead them in | good directions. | rmbyrro wrote: | I guess book authors don't like my perspective... | nanidin wrote: | If I don't get where I want to be with the front door | customer service within a decent amount of time, I have | always had good success contacting jeff@amazon.com. Their | executive support team gets back quickly via email or phone | and they really seem to care. | hinkley wrote: | Ugh. I hate the, "You're not a customer yet so our CRM system | won't let me talk to you." | | And what happens when my problem is that your system won't | let me place an order? | me_again wrote: | I think that's a different issue. Amazon has thorny | problems with takedowns. Company A trying to get rival | company B's listing taken down probably happens 100's of | times a day. I believe Amazon uses "proof of purchase" | kinda like a CAPTCHA or proof of work - an extra hoop to | jump through to reduce the volume of these things they have | to adjudicate. | indymike wrote: | CRM should never mean Sales Prevention as a Service. | hinkley wrote: | The great thing about filtering is that you don't have to | hear the screams. | | These accidents play out in slow motion until someone | corners you at a family reunion and asks why their | friends can't create accounts and when you ask them how | long they say "months". | blululu wrote: | False Negatives and False Positives are always connected. | On the other side of the equation, there are plenty of bad | actors who will casually flag their competitors to score a | quick win. Crime doesn't like to go uphill - raising the | stakes for feedback lowers the prevalence of bad actors. | onlyrealcuzzo wrote: | > Seriously, how hard was it amazon? I guess they're starting | to notice. | | It's not hard. It's a cost center, and they're in the | business of making money - not providing the best service. | kristopolous wrote: | They're biggest risk has always been the perception they | peddle fraudulent simulacrums of worthy products. It's a | hard to shake perception but dangerously easy to acquire | given how they've set up their marketplace. | tetrep wrote: | > Maybe there's so much garbage coming in now that they finally | have to do something about it? | | It seems like this is preventative action rather than | reactionary, as they say that there hasn't been an increase in | publishing volume, "While we have not seen a spike in our | publishing numbers..." | harles wrote: | > ... I assume they haven't taken them down because they make | money even when they sell garbage. | | I'd be surprised if this is the case. The money they make is | probably a rounding error compared even just to other Kindle | sales. Much more likely is that they haven't seen it as a big | enough problem - and I'm willing to bet it's increased multiple | orders of magnitude recently. | throe37848 wrote: | I knew guy who made "generated" text books in 2010. He would | absorb several articles, and loosely stitch them into chapters | with some computer scripts and from memory. In a week he would | produce 400 pages on new subject. It was mostly coherent and | factual (it kept references). Usually it was the only book on | market about given subject (like rare disease). | | Current auto generated garbage is very different. | plagiarist wrote: | I wouldn't even consider that generated. That's like where | useful content and copyright infringement overlap on a Venn | diagram. | Karellen wrote: | > That's like where useful content and copyright | infringement overlap on a Venn diagram. | | That sounds like a description of LLM-generated content to | me ;-) | delecti wrote: | LLMs only ever accidentally generate useful content. They | fundamentally can't know whether the things they're | outputting are true, they just _tend_ to be, because the | training data also tends to be. | franze wrote: | explains the CouchDB Book from OReily from that time. | quickthrower2 wrote: | Do people still use CouchDB? Blast from the past! | miohtama wrote: | Garbage books are used for money laundering. | | You buy books using stolen credit cards and such. | | https://www.theguardian.com/books/2018/apr/27/fake-books-sol... | hinkley wrote: | I wonder if that means the Feds made a phone call to Jeff on | his private line and said we need to have a little chat. | | We can track money laundering when there are X fake books. We | can't when there are 10X fake books. | mortureb wrote: | In my opinion, all we learn over time is that we need | gatekeepers (publishing houses in this case). The general | public is a mess. | barrysteve wrote: | The standards for filtering internet data have dropped badly. | | Amazon and Google both abuse their filtering systems on a | daily basis to effect social change. | | We need new companies built with policies to keep the | filtering systems rigid, effective and unchanging. We need | filterkeepers. | mortureb wrote: | I'm good with Amazon and Google over some unknown. I don't | want some right wing shit to be my gatekeepers. | barrysteve wrote: | Yay, politics in my business soup. That'll generate a | quality outcome for my customers! | | /s | | The politics are ephemeral, the results matter. | chongli wrote: | I think what we're seeing here is a symptom of the broader | and more fundamental problem of trust in society. We've gone | from a very high trust society to a very low trust society in | just a few decades. We, as technology people, keep searching | (desperately) for technical solutions to social problems. | It's not working. | pixl97 wrote: | Because technology never was the solution for social | problems, it's a solution to the few people getting very | rich problem. | VancouverMan wrote: | Such systems just result in content that is terribly bland, | or worse, intentionally limited to push specific political | narratives. | | I'd rather have a much more diverse and interesting set of | content to choose from, even if some of it might not be to my | liking, and even if I'd have to put some effort into | previewing or filtering before I find something I want to | consume. | ozfive wrote: | Some people value their time, energy, and money more. I can | appreciate that you do not as we all have choices but I | imagine that most people would disagree. | dzink wrote: | Strategically, AI generated content is a boon for platforms like | Amazon. | | 1. The more content there is, the more you can't reliably get | good stuff without reviews, the more centralized distribution | platforms with reviews and rankings are needed. 2. Even if people | are making fake books for money laundering, Amazon gets a cut of | all sales, laundered or not. | | Just like Yahoo's directory once upon a time though, and Movie | theaters, the party gets ruined when most people learn they can | use AI to generate custom stories at home and/or converse with | the characters and interact in far more ways than currently | possible. Content is going from king to commodity. | blibble wrote: | amazon's reviews and rating are completely garbage and have | been for some time | Supply5411 wrote: | While not exactly the same, the invention of the printing press | caused a lot of controversy with the Catholic Church. With the | printing press, people could mass produce and spread information | relatively easily. I'm sure a lot of it was considered "low | quality" (also heretical)[1]. Seems like we're going through | similar growing pains now. Yes I know it's different, but it | rhymes. | | 1. https://en.wikipedia.org/wiki/Index_Librorum_Prohibitorum | BarryMilo wrote: | You're implying that what is being produced has actual value, | the problem is they're acting in patently bad faith. Weep not | for the spammers. | Barrin92 wrote: | >similar growing pains | | For what it's worth, these 'growing pains' took the form of the | wars of religion in Europe, which in Germany killed up to 30% | of the population, that's in relative terms significantly worse | than the casualties of World War I and II. So maybe the | Catholic Church had a point | lovemenot wrote: | >> So maybe the Catholic Church had a point | | Is that really the take-away? If the Catholic Church had not | been so belligerent, those wars would not have been needed. | Now that we are past that time, we should surely be thanking | those combatants who helped disseminate knowledge in spite of | the Church whose interest was in hoarding it. | dougmwne wrote: | I really dislike the comparison. The printing press | democratized knowledge. The LLM destroys it. LLM output is | perfect white noise. Enough of it will drown out all signal. | And the worst part is that it's impossible to distinguish it | from real human output. | | I mean think about it. Amazon had to stop publishing BOOKS | because it can no longer separate the signal from the noise. | The printing press was the birth of knowledge for the people | and the LLM is the death. | jedberg wrote: | If you asked the Church back then, they would tell you that | the printing press was the death of truth, because to them | only the word of god was truth, and only the church could | produce it. | | It's all just a matter of perspective. | | Yes, right now it looks like white noise, just like back then | it looked like white noise which could drown out the | religious texts. But we managed to get past it then and I'm | sure we'll manage now. | SketchySeaBeast wrote: | I'd argue that giving a group with unique thoughts and | ideas a voice is different than creating a noise machine. | jedberg wrote: | I think the jury is still out on whether an LLM produces | ideas any more or less unique than most humans. :) | duskwuff wrote: | This is an _astoundingly_ bad take. Surely you aren 't | trying to suggest that original, factual, human-authored | content has no more inherent value than randomly generated | nonsense? | rileyphone wrote: | That's Wittgenstein's argument. | jedberg wrote: | No not at all, I'm not sure why you would even think | that. | duskwuff wrote: | As I read it, your parent comment suggests that the | distinction in quality and utility between human-authored | and AI-generated content is merely "a matter of | perspective", i.e. that there is no real distinction, and | that they're both equally valuable. | | If you actually meant something else, you should probably | clarify. | lovemenot wrote: | I am not the person to whom you replied. I understood | their comment to be about paradigms shifting through | social awareness of the limits and opportunities of new | technology. | | It can be both true that _right now_ predominantly low | quality content emanates from LLMs and at some future | time the highest quality material will come from those | sources. Or perhaps even right now (the future is already | here, just unevenly distributed). | | If that was their reasoning, I tend agree. The equivalent | of the Catholic Church in this metaphor is the | presumption human-generated content's _inherent_ | superiority. | [deleted] | dambi0 wrote: | Suggesting clarification to suit your imaginary | inferences seems puzzling. The parent post pointed out | that perspectives on authorship have a historical | precedent, I didn't see the value judgement your reading | suggested. | rmbyrro wrote: | The discussion here is that we're not able to distinguish | them. | | If we cannot distinguish, I'd argue they have similar | value. | | They must have. Otherwise, how can we demonstrate | objectively the higher value in the human output? | SketchySeaBeast wrote: | If they were of similar value would there be a problem | with the deluge? | rmbyrro wrote: | Can't the deluge be delusional or an overreaction at | best? | snailmailman wrote: | They _can_ be distinguished. They are just becoming more | difficult to. Its slightly-more difficult, but also the | amount of garbage is _overwhelming_. AI can spit out | _entire books_ in moments that would take an individual | _months_ or _years_ to write. | | There are lots of fake recipe books on amazon for | instance. But how can you _really_ be sure without trying | the recipes? It might _look_ like a recipe at first | glance, but if its telling you to use the right | ingredients in a subtly-wrong way, its hard to tell at | first glance that you won 't actually end up with edible | food. Some examples are easy to point at, like the case | of the recipe book that lists Zelda food items as | ingredients, but they aren't always that obvious. | | I saw someone giving programming advice on discord a few | weeks ago. Advice that was blatantly copy/pasted from | chat GPT in response to a very specific technical | question. It _looked_ like an answer at first glance, but | the file type of the config file chat GPT provided wasn | 't correct, and on top of that it was just making up | config options in attempt to solve the problem. I told | the user this, they deleted their response and admitted | it was from chatGPT. However, the user asking the | question didn't know the intricacies of "what config | options are available" and "what file types are valid | configuration files". This could have wasted so much of | their time, dealing with further errors about invalid | config files, or options that did not exist. | rmbyrro wrote: | They can indeed distinguish them, I agree. So why the | fuss? | | I think the concern is that bad authors would game the | reviews and lure audiences into bad books. | | But aren't they already able to do so? Is it sustainable | long term? If you spit out programming books with code | that doesn't even run, people will post bad reviews, ask | for refunds. These authors will burn their names. | | It's not sustainable. | geraldwhen wrote: | It's sustainable if you can automate the creation of | amazon seller accounts. Based on the number of fraudulent | Chinese seller accounts, I'd say it's very likely | automated or otherwise near 0 cost. | snailmailman wrote: | It doesn't need to be sustainable as one author or one | book. These aren't _real_ authors. Its people using AI to | make a quick buck. By the time the fraud is found out, | they 've already made a profit. | | They make up an authors name. Publish a bunch of books on | a subject. Publish a bunch of fake reviews. Dominate the | search results for a specific popular search. They get | people to buy their book. | | Its not even book specific, its been happening with | actual products all over amazon for years. People make up | a company, sell cheap garbage, and make a profit. But | with books, they can now make the cheap garbage look | slightly convincing. And the cheap garbage is so cheap to | produce in mass amounts that nobody can really sort | through and easily figure out "which of these 10k books | published today are real and which are made up by ai". | | It takes time and money to produce cheap products at a | factory. But once these scammers have the AI generation | setup, they can just publish books on loop until someone | ends up buying one. They might get found out eventually, | and they will have to pretend to be a different author, | and they just repeat the process. | failuser wrote: | What's the fuss about spam? You can distinguish it from | useful mail? What's the fuss about traffic jams? You'll | get there eventually. | | The LLM allow DDoS attack by increasing the threshold | needed to check the books for gibberish. | | It's not like this stream of low quality did not exist | before, but the topic is hot and many grifters try LLMs | to get a quick buck at the same time. | emodendroket wrote: | I can't distinguish between pills that contain the | medicine that I was prescribed and those than contain | something else entirely. Therefore taking either should | be just as good. | rmbyrro wrote: | Really. Are you comparing a complex chemical analysis | required to attest the contents of a pill to reading | text? | emodendroket wrote: | It depends, is the text of a technical nature? How | exactly is one to know they're being deceived if, to take | one of the examples that has been linked in this | discussion, they receive a mushroom foraging guide but | the information is actually AI-generated? | rmbyrro wrote: | You first check who published it. Is the author an expert | in the matter with years, perhaps decades in the | industry? | | Heck, we always did that since before GPT. | | Good authors will continue to publish good content | because they have a reputation to protect. They might use | ChatGPT to increase productivity, but will surely and | carefully review it before signing off. | skydhash wrote: | I'm not a native English speaker, but ChatGPT answers in | each interaction I had with it sound bland. And I dislike | the bite-sized format of it. I'm reading "Amusing | Ourselves to Death" by Neil Postman and while you may | agree or disagree with his take, he developed it in a | very coherent way, exploring several aspects. ChatGPT's | output falls into the same uncanny valley as the robotic | voice from text to speech software, understandable, but | no human does write that way. | | ChatGPT as an autocompletion tool is fine, IMO. As well | as generating alternative sentences. But anything longer | than a paragraph falls back to the uncanny valley. | rmbyrro wrote: | I totally agree. So why are people so worried about books | being written by ChatGPT? | | These pseudo-authors will get bad reviews, will lose | money in refunds, burn their names. | | It's not sustainable. Some will try, for sure, but they | won't last long. | failuser wrote: | If you ask LLM something you know you can distinguish | noise from good output. If you ask LLM something you | don't know then how do you know if the output is correct? | There are cases where checking is easier than producing | the result, e.g. when you ask for a reference. | rmbyrro wrote: | Book buyers should give themselves primarily by who's the | author, I think. | | Choose a book from someone that has a hard earned | reputation to protect. | mostlylurks wrote: | A piece of human-written content and a piece of AI- | written content may have similar value if we cannot | distinguish between them. But if you can add the | information that the human-written content was written by | a human to the comparison, the human-written content | becomes significantly more valuable, because it allows | for a much deeper reading of the text, since the reader | can trust that there has been an actual intent to convey | some specific set of ideas through the text. This allows | the reader to take a leap of faith and put in the work | required to examine the author's point of view, knowing | that it is based on the desires and hopes of an actual | living person with a lifetime of experience behind them | instead of being essentially random noise in the | distribution. | courseofaction wrote: | The LLM does democratize knowledge, but you have to be the | user of the LLM, not the target of the user of the LLM. | | The LLM is the most powerful knowledge tool ever to exist. It | is both a librarian in your pocket. It is an expert in | everything, it has read everything, and can answer your | specific questions on any conceivable topic. | | Yes it has no concept of human value and the current | generation hallucinates and/or is often wrong, but the | responsibility for the output should be the user's, not the | LLM's. | | Do not let these tools be owned, crushed and controlled by | the same people who are driving us towards WW3 and cooking | the planet for cash. This is the most powerful knowledge tool | ever. Democratize it. | shitloadofbooks wrote: | Asking a statistics engine for knowledge is so unfathomable | to me that it makes me physically uncomfortable. Your | hyperbolic and relentless praise for a stochastic parrot or | a "sentence written like a choose your own adventure by an | RNG" seems unbelievably misplaced. | | LLMs (Current-generation and UI/UX ones at least) will tell | you all sorts of incorrect "facts" just because "these | words go next to each other lots" with a great amount of | gusto and implied authority. | __loam wrote: | This happened to me looking up am obscure c library. It | just confidently made up a function that didn't actually | exist in the library. It got me unstuck but you can | really fuck yourself if you trust it blindly. | Supply5411 wrote: | My mind is blown that someone gets so little value out of | an LLM. I get over software engineering stumbling blocks | _much_ faster by interrogating an LLM 's knowledge about | the subject. How do you explain that added value? Are you | skeptical that I am actually moving and producing things | faster? | lxgr wrote: | My mind is also blown by how much people seemingly get | out of them. | | Maybe they're just orders of magnitude more useful at the | beginning of a career, when it's more important to digest | and distill readily-available information than to come up | with original solutions to edge cases or solve gnarly | puzzles? | | Maybe I also simply don't write enough code anymore :) | Supply5411 wrote: | I'm very far from the beginning of my career, but maybe I | see a point in your comment, because I frequently try | technologies that I am not an expert in. | | Just yesterday, I asked if Typescript has the concept of | a "late" type, similar to Dart, because I didn't want to | annotate a type with "| null" when I knew it would be | bound before it was used. Searching for info would have | taken me much longer than asking the LLM, and the LLM was | able to frame the answer from a Dart perspective. | | I would say that that information neither "important to | digest" nor "readily available." | pests wrote: | I agree with you but at what point does it change? Aren't | we all just stochastic parrots? How do we ourselves | choose the next word in a sentence? | __loam wrote: | God dammit please stop comparing these things to brains. | Stop it. It's not even close. | barrysteve wrote: | If you wish to make an apple pie, first you must make the | universe from scratch. (carl sagan) | | We can generate thoughts that are spatially coherent, | time aware, validated for correctness and a whole bunch | of other qualities that LLMs cannot do. | | Why would LLMs be the model for human thought, when it | does not come close to the thoughts humans can do every | minute of every day? | | Aren't we all just stochastic parrots, is the kind of | question that requires answering an awful lot about the | universe before you get to an answer. | skydhash wrote: | We use languages to express ideas. Sentences are always | subordinate to the ideas. It's very obvious when you try | to communicate in another language you're not fluent in. | You have the thought, but you can't find the words. The | same thing happens when writing code, taking ideas from | the business domain and translating it into code. | [deleted] | ForHackernews wrote: | > and can answer your specific questions on any conceivable | topic | | Yeah, I mean, so can I, as long as you don't care whether | the answers you receive are accurate or not. The LLM is | just better at pretending it knows quantum mechanics than I | am. | scarmig wrote: | Even if a human expert responds about something in their | domain of expertise, you have to think critically about | the answer. Something that fails 1% of the time is often | more dangerous than something that fails 10% of the time. | | The best way to use an LLM for learning is to ask a | question, assume it's getting things wrong, and use that | to probe your knowledge which you can iteratively use to | prove the LLM's knowledge. Human experts don't put up | with that and are a much more limited resource. | halfmatthalfcat wrote: | > but the responsibility for the output is the user's, not | the LLM's. | | The current iteration of the internet (more specifically | social media) has used the same rationality for its | existence but at a level, society has proven itself too | irresponsible and/or lazy to think for itself but be fed by | the machine. What makes you think LLMs are going to do | anything but make the situation worse? If anything, they're | going to reenforce whatever biases were baked into the | training material, of which is now legally dubious. | [deleted] | lxgr wrote: | For a librarian, they're confidently asserting factual | statements suspiciously often, and refer me to primary | literature shockingly rarely. | arrowsmith wrote: | In other words they behave like a human? | Supply5411 wrote: | You could argue that speech is literally noise that drowns | out the signals of your environment. If you just babbled, it | would be useless, but instead you use it intelligently to | communicate ideas. LLM output is a new palette with which | humans can compose new signals. We just have to use it | intelligently. | | Prompt engineering is an example of this. A clever prompt by | a domain expert can prime an LLM interaction to yield better | information to the recipient in a way that the recipient | themselves could not have produced on their own. | woah wrote: | It used to be that a scribe would painstakingly copy a | manuscript, through the process absorbing the text at a deep | level. This same scribe could then apply this knowledge to | his own writing, or just understand and curate existing work. | The manual labor required to copy at scale employed many | scribes, who formed the next generation of thinkers. | | With the press, a greasy workman can churn out hundreds of | copies an hour, for whichever charlatan or heretic palms him | enough coin. The people are flooded with falsehoods by men | whose only interest in writing is how many words they can fit | on a page, and where to buy the cheapest ink. | | The worst part is that it is impossible to distinguish the | work of a real thinker from that of a cheap sophist, since | they are all printed on the same rough paper, and serve | equally well as tomorrow's kindling. | emodendroket wrote: | Where are the good AI-generated books that serve as the | positive side of this development? | __loam wrote: | People comparing the AI bullshit spigot to the printing press | are clowns. | [deleted] | rockemsockem wrote: | I really feel like you can't have used any advanced LLMs if | you legitimately think the out put "perfect white noise". The | results that you can get from an LLM like GPT-4 are | incredibly useful and are providing an enormous amount of | value to lots of people. It isn't just for generating phony | information to spread or having it do your work for you. | | I get the most value out of asking for examples of things or | asking for basic explanations or intuitions about things. And | I get so much value from this that I really think the | printing press is the most apt comparison. | mostlylurks wrote: | What you say is not in conflict with AI-generated content | being white noise. Even if you find some piece of AI- | generated content useful, it is still white noise if it is | merely combining pieces of information found in its dataset | and the result is posted online or published elsewhere. | There is no signal being added in that process, and it | pollutes the space of content. Humans are also prone to | doing this, but with the help of AI, it becomes a much | larger issue. | | "Signal" would mean new data, which is by definition not | possible via LLMs trained on publicly available content, | since that means the data is already out there, or new and | meaningful ideas or innovations beyond just combining | existing material. I have not seen LLMs accomplish the | latter. I consider it at least possible that they are | capable of such a feat, but even then the relevant question | would be how often they produce such things compared to | just rearranging existing content. Is the proportion high | enough that unleashing floods of AI-generated content | everywhere would not lower the signal-to-noise ratio from | the pre-AI situation? | softg wrote: | The problem is advanced LLMs are controlled by large | corporations. Powerful local models exist (in part thanks | to Meta's generosity oddly enough) and they're close to | GPT-3.5, but GPT-4 is far ahead of them and by the time | other models reach to that point whatever OpenAI or | Antropic, Meta etc. have developed behind closed doors | could be significantly better. In that case open models | will be restricted to niche uses and most people will use | the latest model from a giant corp. | | So it is possible that LLMs will centralize the production | and dissemination of knowledge, which is the opposite of | what people think the printing press did. I hope I'm wrong | and open models can challenge/overtake state of the art | models developed by tech giants, that would be amazing. | courseofaction wrote: | Precisely. I spent weeks learning about cybersecurity | when GPT-4 first came out, as I could finally ask as many | stupid questions as I liked, get detailed examples and | use-cases for different attacks and defenses, and | generally actually learn how the internet around me | works. | | Now it refuses, because OpenAI's morals apparently don't | include spreading openly available knowledge about how to | defend yourself. | | Scary. I have also been using it to generate useful | political critiques (given a particular theoretical | tradition, some style notes, and specific articles to | critique, it's actually excitingly good). What if OpenAI | decides that's a threat? What reason do we have to think | that a powerful institution would not take this course of | action, in the cold light of history? | blibble wrote: | how do you know what you learnt wasn't completely made up | gibberish? | Supply5411 wrote: | The same way you know that the things you learn from a | person isn't made up gibberish: You see how well it | explains a scenario, how well it lines up with your | knowledge and experience, and you sample parts to verify. | [deleted] | mrighele wrote: | > Amazon had to stop publishing BOOKS because it can no | longer separate the signal from the noise. | | That's because they are trying very hard not to check what | they are selling, hoping that their own users and a few ML | algorithms can separate the signal from the noise for them. | It seems to me that the approach is no longer working, and | they should start doing it by themselves. | pacman2 wrote: | I was told publishers dont promote a good book anymore these | days. They ask how many instagram followers do you have? | | Maybe the self-publishing and BoD will decline in the long | term due to ML white noise and publishers are a sign of | quality again. | rmbyrro wrote: | > the worst part is that it's impossible to distinguish it | from real human output | | Doesn't that make human content look bad in the first place? | | If we can't distinguish a Python book written by a human | engineer or by ChatGPT, how can we demonstrate objectively | that the machine-generated one is so much worse? | mostlylurks wrote: | That argument might work for content which serves a purely | informational purpose, such as books teaching the basics of | programming languages, for instance, but it doesn't work | for art (e.g. works of fiction) because most of the | potential for a non-superficial reading of a work relies on | being able to trust that there is an author that has made a | conscious effort to convey something through that work, and | that that something can be a non-obvious perspective on the | world that differs from that of the reader. AI-generated | content does not have any such intent behind it, and thus | you are effectively limited to a superficial reading, or if | were to instist on assigning such intent to AI, then at | most you would have one "author" per AI model, which | additionally has no interesting perspectives to offer, | simply those perspectives deemed acceptible in the culture | of whatever group of people developed the model, no | perspective that could truly surprise or offend the reader | with something they had not yet considered and force them | to re-evaluate their world view, just a bland average of | their dataset with some fine tuning for PR etc. reasons. | Nashooo wrote: | The problem is not that no one can distinguish it. It's | that the intended audience (beginners in Python in your | example) can't distinguish it and are not able to easily | find and learn from trusted sources. | rmbyrro wrote: | Aren't there already bad Python books written by humans? | | I bet ChatGPT can come up with above-average content to | teach Python. | | We should teach beginners how to prompt engineer in the | context of tech learning. I bet it's going to yield | better results than gate-keeping book publishing. | rmbyrro wrote: | Another great contribution would be fine-tuning open | source LLMs on less popular tech. I've seen ChatGPT | struggling with htmx, for example (I presume the training | dataset was small?), whereas it performs really well | teaching React (huge training set, I presume) | nneonneo wrote: | There are, but it used to take actual time and effort to | produce a book (good or bad), meaning that the small pool | of experts in the world could help distinguish good from | bad. | | Now that it's possible to produce mediocrity at scale, | that process breaks down. How is a beginner supposed to | know whether the tutorial they're reading is a legitimate | tutorial that uses best practices, or an AI-generated | tutorial that mashes together various bits of advice from | whatever's on the internet? | rmbyrro wrote: | Personally I don't subscribe to the "best practices" | expression. It implies an absolute best choice, which, in | my experience, is rarely sensible in tech. | | There are almost always trade-offs and choosing one | option usually involves non-tech aspects as well. | | Online tutorials freely available very rarely follow, | let's say, "good practices". | | They usually omit the most instructive parts, either | because they're wrapped in a contrived example or | simplify for accessibility purposes. | | I don't think AI-generated tutorials will be particularly | worse at this to be honest... | emporas wrote: | If beginners in Python programming are not capable of | visiting python.org, assuming they are genuinely | interested in learning Python, it would be very | questionable how good their knowledge on the subject can | really be. | rmbyrro wrote: | 100% agreed. | | I've seen many developers using technologies without | reading the official documentation. It's insane. They | make mistakes and always blame the tech. It's | ludicrous... | vosper wrote: | > The printing press democratized knowledge | | That's true, but it also allowed protestant "heretics" to | propagate an idea that caused a permanent schism with the | Catholic church, which led to centuries of wars that killed | who-knows-how-many people, up to recent times with Northern | Ireland. | | (Or something like that, my history's fuzzy, but I think | that's generally right?) | [deleted] | bbarnett wrote: | I thought it was a king wanting a divorce, and as he | couldn't get it from the catholic church, created his own. | mmcdermott wrote: | Henry VIII created the Church of England in 1534 for the | purposes of granting himself an annulment. Most histories | count Martin Luther's 95 Theses as beginning of the | Reformation in 1517 (a crisp date for a less-than-crisp | event; Luther did not originally see himself as | protesting the Roman Catholic Church). The Protestant | Reformation was a heterogeneous movement from the | beginning. | TRiG_Ireland wrote: | Not really, no. It was Luther who kick-started | Protestantism. Henry VIII attempted to supplant the Pope, | and kind of slid into Protestantism by accident. | vladms wrote: | That was the case just for the anglican church, which is | only one "part" of the reformation. | verve_rat wrote: | Protestantism started in Germany with Martin Luther | nailing his theses to a church door. Henry's reproductive | problems came later and where only sort of related. | peab wrote: | I'm reminded of the Library of Babel | OfSanguineFire wrote: | > The printing press democratized knowledge. | | Not for centuries. Due to the expense of the technology and | the requirement in some locations for a royal patent to print | books, the printing press just opened up knowledge a bit more | from the Church and aristocracy to the bourgeoisie, but it | did little for the masses until as late as the 1800s. | ls612 wrote: | A big part of this is that literacy didn't come to the | masses until the 1800s. But in England and the Netherlands | you had (somewhat) free press by the late 1600s and early | 1700s. | ben_w wrote: | > LLM output is perfect white noise. | | Not even close to white noise. White noise, in the context of | the token space, looks like this: | | auceverts exceptionthreat."<ablytypedicensYYY DominicGT | portaelight\\- titular Sebast | Yellowstone.currentThreadrition-zoneocalyptic | | which is literally the result of "I downloaded the list of | tokens and asked ChatGPT to make a python script to | concatenate 20 random ones". | | No, the biggest problem with LLMs is that the best of them | are simultaneously better than untrained humans and yet also | nowhere near as good as trained humans -- someone, don't | remember who, described them as "mansplaining as a service", | which I like, especially as it (sometimes) reminds me to be | humble when expressing an opinion outside my domain of | expertise, as it knows more than I do about everything I'm | not already an expert at. | | Specific example: I'm currently trying to use ChatGPT-3.5 to | help me understand group theory, because the brilliant.org | lessons on that are insufficient; unfortunately, while it | knows infinitely more than I do about the subject, it is | still so bad it might as well be guessing the multiple choice | answers (if I let it, which I don't because that would be | missing the point of using a MOOC like brilliant.org in the | first place). | gamepsys wrote: | The rhyme has a lot to do with how existing power structures | handle a sudden increase in the amount of written text | generated. In this comparison, they both try to apply the | breaks. Banned books didn't work well for the Catholic Church. | I think increasing QA for Amazon might actually help their book | business. Of course, a book seller has a greater responsibility | to society than to make money. | neilv wrote: | This sounds like a commendable move by Amazon. I especially like | the idea of requiring disclosure of use of "AI". | corethree wrote: | How do we even know this entire comment thread isn't polluted | with AI? | | Maybe it doesn't matter. The quality of the work matters more | than the process of actualization. | quickthrower2 wrote: | In a practical sense: AI generated stuff is crappy and often | subtly wrong and it can be generated faster than human | generated content. So it becomes untenable to even search for | good information. | DookieBiscuit wrote: | [dead] | fuddle wrote: | About time, YouTube is full of videos about making eBook's with | ChatGPT. e.g "Free Course: How I Made $200,000 With ChatGPT eBook | Automation at 20 Years Old" | https://www.youtube.com/watch?v=Annsf5QgFF8 | hinkley wrote: | How do we... | | I'm not entirely sure how to word this question. | | How do we make sure that most of the people we talk to are at | least humans if not necessarily the person we expect them to be? | And I'm not saying that like a cartoonish bad guy in a movie who | hates artificial intelligence and augmented humans. | | How do I not get inundated by AI that's good at trolling. How do | I keep the social groups I belong to from being trolled? | | These questions keep drawing me back to the concept of Web of | Trust we tried to build with PGP for privacy reasons. Unless I've | solicited it, I really only want to talk to entities that pass a | Turing Test. I'd also like it if someone actively breaking the | law online were actually affected by the deterrence of law | enforcement, instead of being labeled a glitch or a bug in | software that can't be arrested, or even detained. | | It feels like I want to talk to people I know to be human | (friends, famous people - who might actually be interns posing as | their boss online), and people they know to be human, and people | those people _suspect_ to be human. | | I have long term plans to set up a Wiki for a hobby of mine, and | I keep getting wrapped around the axle trying to figure out how | to keep signup from being oppressive and keep bots from turning | me into an SEO farm. | timeagain wrote: | This is only a problem for someone terminally online. The vast | majority of people talk to their friends and coworkers in | person. | mostlylurks wrote: | > This is only a problem for someone terminally online. | | Is it? Even those whose social life is entirely IRL, they | still have to increasingly interact with various businesses, | banks, healthcare providers, the government, and often more | distant collegues through online services. Do I want these to | go through LLM chatbots? No. Can I ensure that I'm speaking | to an actual human if the communication is text based? Not | really. | munificent wrote: | That was the solution that came to mind to me too, but it | doesn't work either. | | Even if you're never online and only talk to people in | person... over time those people will be increasingly | informed by LLM-generate pseudo-knowledge. We aren't just | training the AIs. They're training us back. | | If you want to live in a society where the people you | interact with have brains mostly free of AI-generated | pollution, then I'm sorry but that world isn't going to be | around much longer. We are entering the London fog era of the | Information Age. | invalidptr wrote: | This is a problem for anyone who is not actively vigilant | about the information they consume. A family member (who I | would not describe as "terminally online") came to me today | in a panic talking about how some major event had just | occurred and how social order was beginning to collapse. I | quickly glanced at the headlines on a few major news outlets | and realized that they just saw some incendiary content | designed to elicit that reaction. I calmed them down and | walked them through a process they could use to evaluate | information like that in the future, and they were a little | embarrassed. | | The concern isn't necessarily for _you_. It 's for the large | swaths of people who are less equipped to filter through | noise like this. | hinkley wrote: | I don't trust my friends for medical advice. Some of them | trust me for plant advice, and they really probably | shouldn't. I am very stove-piped. | | We have two and a half generations of people right now most | of whom think "I did the research" means "I did half as much | reading as the average C student does for a term paper, and | all of that reading was in Google." | | And Alphabet fiddles while Google burns. This is going to end | in chaos. | vorpalhex wrote: | Meet people in real life. This problem is trivially solved by | just using meatspace. | | Alternatively for sign ups, tell them to contact you and ask. | Chat with them a moment. Ask them about their hobbies and | family. | ethanbond wrote: | Using meatspace doesn't solve the problem, using meatspace | _exclusively_ solves the problem. And it 's not a great one | given, you know, how much of the world "happens" online now. | romseb wrote: | There is some irony in Sam Altman bringing us the cause (AI) | and purported solution (Worldcoin) for your problem at the same | time. | hinkley wrote: | It's what ad men do. Point out there's a problem, offer you | the solution. | Almondsetat wrote: | Livestreams where artists show their creative process and use the | streaming platform to immediately sell the thing they produced, | just to prove it had human origins. | | This is the future | SketchySeaBeast wrote: | You're only kicking the can down the road. | edgarvaldes wrote: | We have realtime filters, avatars, translators, TTS, etc. Soon, | all of this will be "good enough" to mimic the proposed | solution. | gz5 wrote: | I think we will see tidal waves of 'not-so-good' AI-generated | content. Not that AI can't generate or help generate 'good' | content, but it will be faster and cheaper to generate 'not-so- | good'. | | These waves will mainly be in places in which we are the product. | And those waves could make those places close to uninhabitable | for folks who don't want to slosh through the waves of noise to | find the signal. | | And in turn that perhaps enables a stronger business model for | high quality content islands (regardless of how the content is | generated) - e.g. we will be more willing to pay directly for | high quality content with dollars instead of time. | | In that scenario, AI could be a_good_thing in helping to spin a | flywheel for high quality content. | omnicognate wrote: | Except they shouldn't be islands. Unify/standardise the payment | mechanism, make it frictionless and only for content consumed. | There's no technical reason you shouldn't see an article on hn | or wherever, follow the link and read it _and pay for it_ | without having set up and pay for a subscription for the entire | publication or jump through hoops. It should be a click at | most. | | There will always be a place for subscriptions, but people want | the hypertext model of just following a link from somewhere and | there is absolutely no technical reason for that to be | incompatible with paying for content. The idea that ads are the | only way to fund the web needs to be challenged, and generative | AI might just provide the push for that to finally happen. | | Or maybe there will be no such crisis and it'll just make the | whole thing even more exploitative and garbage-filled. | munificent wrote: | _> There 's no technical reason you shouldn't see an article | on hn or wherever, follow the link and read it and pay for it | without having set up and pay for a subscription for the | entire publication or jump through hoops. It should be a | click at most._ | | People have been saying this and building startups on this | and having those startups crash and burn for _decades_. | | It's not a technical problem. It's a psychology problem. | | Paying after you've read an article doesn't provide the | immediate post purchase gratification to make it an inpulse | purchase [0]. The upside of paying for an article you've | already read is more like a considered purchase [1]. But the | amount of cognitive effort worth putting into deciding | whether or not to pay for the article is often less than the | value you got from the article itself. So it's very hard for | people to force themselves to decide to commit to these kinds | of microtransactions. See also [2]. | | It's just a sort of cognitive dead zone where our primate | heuristics don't work well for the technically and | economically optimal solution. It's sort of like why you | can't go into a store and buy a stick of gum. | | [0]: https://en.wikipedia.org/wiki/Impulse_purchase | | [1]: https://en.wikipedia.org/wiki/Considered_purchase | | [2]: https://en.wikipedia.org/wiki/Bounded_rationality | pixl97 wrote: | >It should be a click at most. | | Welcome to new and interesting ways to defraud people over | the internet for money school of thought. | | At least with Amazon it's a "one and done shop" of who I | spent my money with when I bought something. | | Imagine tomorrow with your click to pay for random links on | the internet you suddenly have 60,000 1 cent charges. They | all appear to go different places and to get a refund you | need to challenge each one. | bobthepanda wrote: | It sounds like the digital version of the CD scam. | https://viewing.nyc/nyc-scams-101-dont-get-fooled-by-the- | cd-... | rwmj wrote: | Assuming not too many people die eating mushrooms while we're | waiting: | https://www.theguardian.com/technology/2023/sep/01/mushroom-... | hinkley wrote: | Common foraging rhetoric is that you need two independent | sources asserting that a wild food is edible. Ones that cite | neither each other or the same chain of citations. And | preferably a human who says, "I've been eating these for | years and no problems." or scientists who did recent blood | work to make sure you aren't destroying your organs by eating | [1]. | | In a world with fake books, it would be quite easy for two | books to contain the same misinformation or mis- | identification (how many times have I found the wrong plant | in a google image search? More times than I care to count). | Two fake books putting the wrong mushroom picture next to a | mushroom because they were contiguous on some other page and | you have dead people. | | [1] In the ten years since I started working with indigenous | plants, wild ginger (asarum caudatum), has gone from quasi- | edible to medicinal to don't eat. More studies show subtler | wear and tear on the organs (wikipedia lists it as | carcinogenic!) and it is recommended now that you don't eat | them at all, even for medicinal purposes. I'm not sure I own | a foraging or native species book younger than 5 years, and | many are older. | cellu wrote: | Why do people read contemporary books is something I can't really | get my head around. There're so many classics to keep people busy | for life - and are 100% guaranteed to be insightful and | pleasurable. | bwb wrote: | Contemporary books are just new classics. It is like asking why | read :) | rustymonday wrote: | Should people stop telling new stories? A century from now the | best books of today will be classics. Books can act as a time | capsule of a certain time and place and mode of life. And that | has value. | gamepsys wrote: | I think the risk of reading a suboptimal book is not greater | than the risk of not allowing myself to be exposed to different | voices. | OfSanguineFire wrote: | There's a distinct demographic in the contemporary-fiction- | reading community, as can be seen in corners of Goodreads or | Instagram, that demands new fiction to tell the stories of | groups not covered, or supposedly unfairly covered, in that | classic literature: LGBT, BIPOC, the working class, etc. In | fact, they might even deny that the classics are "insightful | and pleasurable" due to these social concerns. | timeagain wrote: | That's really weird. People are making all kinds of books and | stories. And stories are relevant to their time. The matrix | wouldn't be written in 1900, a tale of two cities wouldn't be | written in 1200, ... | | It is true though that if you have a culturally diverse set | of friends and are open to their experiences and opinions, a | lot of "the classics" start to smell bad. Imagine being black | and reading Grapes of Wrath. You might think the situation of | the main characters as humorous or infantile, considering how | relatively fortunate they are. | Baeocystin wrote: | What's the name of the law where the longer something has | already been around, the longer it will likely stay around in | the future? | | I've found that it definitely applies to books. Starting at a | ~20 year horizon is a surprisingly good filter for quality. | savoyard wrote: | > What's the name of the law where the longer something has | already been around, the longer it will likely stay around in | the future? | | The Lindy effect. | carlosjobim wrote: | One of the best books I read last year was the story of the | rescue of the football team that was trapped in a flooded cave | in 2018 - written by cave diver Rick Stanton, who found the | team and led the rescue. How would that account have been | written into a book before it happened? | harles wrote: | The title of this story doesn't seem to match the content. This | seems like a proactive move to prevent individual publishers from | spamming many many submissions - and even then, they're willing | to make exceptions. | | > While we have not seen a spike in our publishing numbers, in | order to help protect against abuse, we are lowering the volume | limits we have in place on new title creations. Very few | publishers will be impacted by this change and those who are will | be notified and have the option to seek an exception. | freediver wrote: | This is just a tip of the iceberg, compared to what we are | heading into with the web. Very concerning. | | I would go long the value of genuine human writing, aka the | 'small web'. | adamredwoods wrote: | >> We require you to inform us of AI-generated content (text, | images, or translations) when you publish a new book or make | edits to and republish an existing book through KDP. AI-generated | images include cover and interior images and artwork. You are not | required to disclose AI-assisted content. | pcl wrote: | This is really interesting. I imagine that AI-generated art / | illustrations for books mostly-text is a pretty compelling | thing for authors, for all the same reasons that AI-generated | text is of value for non-authors. I wonder how this line will | work out in practice. | hiidrew wrote: | Their distinction: | | >AI-generated: We define AI-generated content as text, images, | or translations created by an AI-based tool. If you used an AI- | based tool to create the actual content (whether text, images, | or translations), it is considered "AI-generated," even if you | applied substantial edits afterwards. AI-assisted: If you | created the content yourself, and used AI-based tools to edit, | refine, error-check, or otherwise improve that content (whether | text or images), then it is considered "AI-assisted" and not | "AI-generated." Similarly, if you used an AI-based tool to | brainstorm and generate ideas, but ultimately created the text | or images yourself, this is also considered "AI-assisted" and | not "AI-generated." It is not necessary to inform us of the use | of such tools or processes. | | https://kdp.amazon.com/en_US/help/topic/G200672390#aicontent... | . | prvc wrote: | Allowing the use of tools to modify the contents erases any | clear distinction between the categories. | campbel wrote: | Gee, I sure hope people don't just lie about it... | skepticATX wrote: | It doesn't matter. It's garbage content and immediately | recognizable as being AI generated. | | It is absolutely possible to write a good article or even a | good book with AI, but at least for now it's just as hard, if | not harder, than doing it without AI. | | But of course people trying to make a quick buck won't put in | the required effort, and they likely don't even have the | ability to create great or even good content. | duskwuff wrote: | > It's garbage content and immediately recognizable as being | AI generated. | | It's also recognizable by its sheer volume. An "author" who | submits several new books every day is clearly not doing | their own writing. The AI publishing scam relies on volume -- | they can't possibly win on quality, but they're hoping to | make up for that by putting so many garbage books on the | market that buyers can't find anything else. | atrus wrote: | I'm not sure. Ghostwriting exists, and a person (or | organization) with enough money could easily pay enough | ghostwriters to output at a more than human pace. | duskwuff wrote: | Even at their most prolific, a ghostwritten author still | probably wouldn't publish more than one or two books a | month. Beyond that point, you're just competing with | yourself. (For instance, young adult series like | _Goosebumps_ , _The Baby-Sitters Club_ , or _Animorphs_ | typically published a book every month or two.) | | Publishing multiple books _per day_ is out of the | question. That 's beyond even what's reasonable for an | editor to skim through and rubber-stamp. | gamepsys wrote: | > It's garbage content and immediately recognizable as being | AI generated. | | Yea, but the Turning Test is actively being assaulted. Soon | we won't know the difference between an uninspired book | written by an AI and an uninspired book written by a human. | mostlylurks wrote: | > It doesn't matter. It's garbage content and immediately | recognizable as being AI generated. | | Is it? How do you immediately recognize a book as AI | generated before buying it, if the author isn't doing | something silly like releasing several books per day/month? | And even after you buy a book, how can you distinguish | between the book just being terrible and the book being | written with extensive use of AI? I don't believe AI can | write good books, but I would still like to distinguish those | two cases, since the former is just a terrible book, which is | perfectly fine, while the latter I would like to avoid. I | don't want to waste my limited time reading AI content. | tyingq wrote: | >It is absolutely possible to write a good article or even a | good book with AI, but at least for now it's just as hard, if | not harder, than doing it without AI. | | How hard is it though, to create a shitty book with AI, that | Amazon can't detect was written with AI? | ilamont wrote: | See also: "Tom Lesley has published 40 books in 2023, all with | 100% positive reviews" | | https://news.ycombinator.com/item?id=35687868 | ritzaco wrote: | I remember that one - interestingly the amazon link it goes to | shows only 3 books now, all that look real, not the 40 that I | remember seeing before. | | So I guess Amazon is doing _something_ even though I regularly | hear complaints from authors that they allow blatant piracy all | the time | phh wrote: | Possibly it's the author removing them at the first one star | rating to keep their author score high? | kmeisthax wrote: | Amazon has no reason to give a shit about piracy on KDP: they | make money either way. But having a load of AI generated | garbage on your platform makes it far less valuable. You want | your stolen books to actually be good. :P | bragr wrote: | >shows only 3 books now | | Those appear to be by different authors with similar names: | https://www.amazon.com/s?k=%22tom+lesley%22 | cogman10 wrote: | Here's a pretty good article about the problem with AI generated | books. "AI Is Coming For Your Children" [1] | | [1] https://shatterzone.substack.com/p/ai-is-coming-for-your- | chi... | el_benhameen wrote: | This doesn't seem surprising. Half of my YouTube ads these days | are for some kind of AI+Kindle-based get rich quick scheme. | NotYourLawyer wrote: | [flagged] ___________________________________________________________________ (page generated 2023-09-18 23:00 UTC)