[HN Gopher] Lawyer cited 6 fake cases made up by ChatGPT; judge ... ___________________________________________________________________ Lawyer cited 6 fake cases made up by ChatGPT; judge calls it "unprecedented" Author : umilegenio Score : 136 points Date : 2023-05-30 20:33 UTC (2 hours ago) (HTM) web link (arstechnica.com) (TXT) w3m dump (arstechnica.com) | ss108 wrote: | It would have taken about 1 minute to put each of them into a | tool like Casetext, Lexis, WL, or Bloomberg Law to determine they | didn't exist. | rafark wrote: | Exactly. This is how AI should be used. Maybe Chatgpt | hallucinates a lot, but that doesn't mean it's useless. It can | still provide a lot of good value, you just have to filter out | the crap, just like you'd do with any other tool. Especially if | you're using it in this context, you should double check its | facts. I'd say it's more the fault of the lazy lawyer that | didn't bother to verify than the fault of ChatGPT. | tombert wrote: | 100% agree. | | I get a little annoyed at people seeing this AI, seeing how | it's not _absolutely perfect_ , and then acting like it's | horrible. | | I think the expression "All models are wrong but some are | useful" applies very much to ChatGPT. It's a useful tool, | even if it's not perfect. | onion2k wrote: | To do that you first need to distrust AI, and a lot of people | don't. They think of GPT like Google-but-written-in-English. | That is a large part of the problem. | anigbrowl wrote: | That's not a valid excuse, though. Lawyers are paid big bucks | to _think_ , not to _assume_. Otherwise you could do your | litigation for free by just asking interested people on | Twitter. I went to law school and had to drop out due to an | injury & attendant medical costs; it's a crime (as in going | to jail) for me to practice law without being licensed, no | matter how good my work product might be. | ss108 wrote: | Well, leaving aside whether a lot of people don't distrust | and whether a lawyer would or should be among those people, | lawyers typically don't cite things without checking them | first. I work at a biglaw firm, can't imagine it never | getting caught that the cases don't exist. This sounds like a | midsize firm, should be fairly similar. | | There are a few small law firms/sole prop type guys, however, | who I have crossed paths with for whom this kind of stupidity | and carelessness would be on brand though. | | Guess he was just in a rush and figured this would be one of | the 2/10 times he files something without at least taking a | look at the opinions first, and it ended up being a massive | error. | MiddleEndian wrote: | A lot of Google's top suggested results / autogenerated | answers tell you nonsense or misunderstandings as well. | morkalork wrote: | Imagine being that clown's client. | lynx23 wrote: | Wait, this is handled as "ChatGPT made up these cases" and not as | "Layers deliberately used ChatGPT to fabricate stuff"? Is anyone | really believing a lawyer is that stupid? I know, adssume good | intentions and all, but in this case, really? | droopyEyelids wrote: | A very intelligent person can simultaneously be very stupid, | and stupid behavior can be increased with tight deadlines or a | high workload. | Projectiboga wrote: | It was his first try with the tech. He should have popped the | cases into West Law and at least skimmed the briefs, he is at | a big firm with that subscription. I'm not a lawyer and I at | least have a clue for the legal research shortcuts. This guy | must have been drinking and or having an affair to have pit | so little effort into his first try using a new tech. I'm | guessing there is a risk of humorous legal parody getting | scooped up by the chat-bot. | lamontcg wrote: | My bet would be that he cheated his way through life and | the bar somehow and landed the job on the basis of his | looks and presentation or who he knew or something like | that, so asking ChatGPT and cheating on it was more natural | than using actual legal research tools. | jeroenhd wrote: | I believe some lawyers are that stupid. | | I also believe lawyers who are stupid enough not to verify | ChatGPT's responses should be treated as if they willfully lied | to the court. "Oops, I didn't know" is a good defence when | you're caught accidentally walking on the grass, not when | you're in court. | wintogreen74 wrote: | "Oops, I didn't know" has never been a valid defense for | anyone; should definitely not be for lawyers, aka "people who | are paid to know". | pclmulqdq wrote: | This appears to be an "individual vs company" case, so I | think expecting some cost-cutting on the individual's side is | normal. Just not done this poorly. | anigbrowl wrote: | Especially after they replied to the first queries with the | assurance that the citations were real and one that couldn't be | found was taken from an unpublished opinion. Negligent _at | best_. Should be a career-ending move for the guy; he can | always pivot into the burgeoning legal AI field. | pkilgore wrote: | In an adversarial system, you know that at the very least your | opponent is checking your cases (you should also assume the | Judge/clerks are too, but I never practiced enough in State | courts to know how well that holds. In Federal District, it's | absolutely true). Usually it's for incredibly small things, | like unquoted/cited distinguishing remarks, later cases | reversing the decision, and misquotes. So a whole case not | existing is going to stand out like crazy. | | Occam's Razor here is that this person was lazy, ignorant, | careless, stupid, or any combination of those. To be | intentionally fraudulent in this circumstance is the equivalent | of trying to steal a gun from a cop. You're fucking with the | one person in society who definitely has the training, | motivation, and willingness to stop you. | lamontcg wrote: | > Is anyone really believing a lawyer is that stupid? | | There's over a million lawyers in the United States. | | You'd expect at least one of them to be a 1-in-a-million level | of bad, or 4.7 standard deviations below the mean assuming a | Gaussian distribution of competency. | | An average person would normally never come across that lawyer | in their lifetime, but media will find that lawyer and amplify | their mistakes to everyone in the population. | drc500free wrote: | Is there a term for this "winning the lottery jackpot is | unbelievably unlikely, but every week we hear about someone | winning it" effect? People get it when the media reports on | the lottery, but somehow miss the amplifying effect on pretty | much any other topic. | pavon wrote: | I can totally see a lawyer who hasn't been following closely | read headlines about ChatGPT acing the Bar Exam, and start | using it without being too skeptical. | | What gets me is that they doubled down when asked to provide | copies. Seriously, when that happens, you don't ask ChatGPT if | the cases are real, you do your own damn search, and apologize | profusely for your mistake. That really makes me question | whether they were trying to pull a fast one, and then play dumb | when caught, or if they really are that stupid. | Fauntleroy wrote: | I'd imagine the lawyer didn't understand what ChatGPT really | is. Many people seem to think that it "knows" things, and they | took that at face value. | [deleted] | thefourthchime wrote: | Sigh, I'm getting very sick of hearing about how "ChatGPT" makes | stuff up. Yes, 3.5 made a lot of stuff up, 4.0 still does, but | it's much rarer. | | I wish people would mention this, it's all treated as the same | thing. It's like talking about how unreliable these "Airplanes" | are when they are talking about prop planes, even though jets are | out. | inopinatus wrote: | The difference is vast and somewhat irritating to those with | GPT-4 access today, but it's also besides the point. The only | people who deserve to cringe at the conflation are OpenAI's | marketing team, but even they may take solace in the sheer cut- | through of column inches their cloven brand obtains, no doubt | anticipating a future retirement of GPT-3. | [deleted] | throwway120385 wrote: | If I had an airplane that usually flew just fine as long as I | didn't use it for business, but would sometimes crash directly | into the ground for no apparent reason, I would be talking | about how unreliable it is too. Especially if the maker of the | airplane touted its unusual benefits for searching for | information in a business setting. | dimgl wrote: | I'm so sick of seeing this strawman. "Well, ChatGPT 4 is | better!" Well, how should I know? I'm not paying for ChatGPT 4. | At least a product like Midjourney is consistent both in its | paid and unpaid form. | tracerbulletx wrote: | None of this reporting or line of criticism is good and I | completely agree, but it is worthwhile to remember that most | users are using the free version and so to most people that's | what ChatGPT is. | ok123456 wrote: | Most people are using 3.5 since that's the free model. | favorited wrote: | These are generative conversation bots. Their #1 priority is to | make things up. | macmac wrote: | That is a hilarious pun. | mabbo wrote: | I think the judge should not take into consideration anything | about where the lawyer said they got the case information from. | | When you go to court and cite previous cases, you are responsible | for ensuring they are real cases. If you can't do that, what | exactly is your job as a lawyer? | kevinmchugh wrote: | If someone hacked westlaw and maliciously replaced opinions, I | think it would not warrant censure of lawyers who cited | westlaw. Maybe a bad day with the judge, and they'd have to | amend the filings. This is malpractice in a way that relying on | an accepted if technically fallible resource isn't | dylan604 wrote: | >If you can't do that, what exactly is your job as a lawyer? | | to convince some (judge/jury) that your client is the one to | win the trial. ethics are meant to be the definer of how far to | go in that cause with licensing boards being the ultimate | decider if you've crossed the line and are allowed to continue | in the legal practice. | [deleted] | inopinatus wrote: | Grinding against the safety rails on a mountain hairpin is | not indicative of a competent driver. | dylan604 wrote: | no, but that driver might just be that much more successful | if they get away with it. in fact, while not a mountain | hairpin as your example, there was a recent race car driver | that did a similar thing by intentionally using the wall as | a push back to allow a maneuver that allowed for success. | | so a clever comment attempting to prove a point is not | always indicative of a proven point ;-) | Ankaios wrote: | It turns out there were precedents: | | Case: Thompson v. Horizon Insurance Company, Filing: Plaintiff's | Motion for Class Certification. Citation: The plaintiff's | attorney cites the influential case of Johnson v. Horizon | Insurance Company, 836 F.2d 123 (9th Cir. 1994), which | established the standards for certifying class actions in | insurance disputes. However, it has recently come to light that | Johnson v. Horizon Insurance Company is a fabricated case that | does not exist in legal records. | | Case: Rodriguez v. Metro City Hospital, Filing: Defendant's | Motion to Exclude Expert Testimony. Citation: The defense counsel | references the landmark case of Sanchez v. Metro City Hospital, | 521 U.S. 987 (2001), which set the criteria for admitting expert | witness testimony in medical malpractice cases. However, it has | now been discovered that Sanchez v. Metro City Hospital is a | fictitious case and does not form part of legal precedent. | | Case: Barnes v. National Pharmaceuticals Inc., Filing: | Plaintiff's Response to Defendant's Motion for Summary Judgment. | Citation: The plaintiff's lawyer cites the well-known case of | Anderson v. National Pharmaceuticals Inc., 550 F.3d 789 (2d Cir. | 2010), which recognized the duty of pharmaceutical companies to | provide adequate warnings for potential side effects. However, | further investigation has revealed that Anderson v. National | Pharmaceuticals Inc. is a fabricated case and does not exist in | legal jurisprudence. | ftxbro wrote: | > "Lawyer: ChatGPT said the cases were real" | glenngillen wrote: | Oh boy, and this one! | | > The plaintiff's lawyer continued to insist that the cases | were real. LoDuca filed an affidavit on April 25 in which he | swore to the authenticity of the fake cases | ftxbro wrote: | maybe they are trolling and their long game is to sue openai | banana_giraffe wrote: | > In Shaboon v. Egyptair, 2013 IL App (1st) 111279-U. | | I'm sure a case involving Egyptair is complicated, still .. | I'd love to see the 111,279 page volume this citation claims | to come from. | noman-land wrote: | That's assuming it's citing the very last page of the | document... | ketralnis wrote: | Doubling down on your bullshit which is itself fuelled by a | robot's bullshit is a move | sharemywin wrote: | we're really at this point aren't we. | sidewndr46 wrote: | I've always felt like Al Gore v. the Internet was a landmark | case | noman-land wrote: | AI Gore | [deleted] | mensetmanusman wrote: | "He was a bad lawyer. I am a good Bing. I will send him to jail." | xbar wrote: | I asked ChatGPT to disbar an attorney. | rayiner wrote: | Check out one of the fake opinions: | https://storage.courtlistener.com/recap/gov.uscourts.nysd.57.... | It even makes up a panel comprising real federal appellate judges | (although one is from the fifth circuit while the fake case is | from the eleventh circuit). I can see how someone unfamiliar with | what GPT can do could get fooled. | greenyoda wrote: | This is interesting. At the beginning of the document, we read | that Susan Varghese, representing the estate of George | Varghese, "appeals the district court's dismissal of her | wrongful death claim". However, in the "Factual background" | section, we read about someone named Anish Varghese who is | alleging a breach of contract due to being denied boarding on a | flight and having to purchase another ticket. | cj wrote: | I asked ChatGPT to tell me a riddle. | | It was "What is always hungry, needs to be fed, and makes your | hands red?" (Or something like that) | | I asked for a hint about 5 times and it kept giving more | legitimate sounding hints. | | Finally I gave up and asked for the answer to the riddle, and it | spit out a random fruit which made no sense as the answer to the | riddle. | | I then repeated the riddle and asked ChatGPT what the answer was, | and it gave me the answer ("Fire") which makes sense as the | answer to the riddle. | | But it was giving extremely bad hints, like "it starts with the | letter P" and "it's a fruit". | | That was a great way to show my non-tech family members the | limitations of AI and why they shouldn't trust it. | | Playing "20 questions" with ChatGPT is another great way to | expose its limitations. It knows the game and tries to play, but | is terrible at asking questions to narrow down possible answers. | | There really needs to be some confidence or accuracy | score/estimation displayed alongside its output. | | Or, learn how to say "I don't know" | marcus0x62 wrote: | > Or, learn how to say "I don't know" | | That would require semantic understanding the LLM doesn't | possess. | tanbog5 wrote: | Or, learn how to say "I don't know" | | It can't. Correct me if I'm wrong - I probably am - but as a | generative language model I was under the impression it has no | actual understanding of the meaning of it's words, only the | probability they work well together. | | You wouldnt expect dice to understand why you didn't want to | roll snake eyes | xahrepap wrote: | I've had a few moments with ChatGPT that are great anecdotes | similar to your own: | | - Asked it to generate a MadLib for me to play that was no more | than a paragraph long. It produced something that was several | paragraphs wrong. I told it "no. That's X paragraphs. I asked | for one that is only 1 paragraph long" and it would respond | "I'm sorry for the misunderstanding. Let me try again" and then | would make the same mistake. It never got it right | | - Asked it, "Can you DM a game of Dungeons and Dragons?" and it | said something like, "Yes! I'd love to DM a game of Dungeons | and Dragons for you". Dumped some text to the screen about how | we'd have to adapt it some. I asked it to begin, and it asked a | few questions about the character I would want to play. I | answered the few questions it asked. Then it finally dumped a | page of text to the screen as "background" to my character and | the quest I was going to embark on. Then it said something | like, "You win. Good job! Hope you enjoyed your quest!" | | I showed these to my family and they were all a little deflated | about AI. Like they realized how willing it was to pretend like | what you wanted and just make up its own answers. | Vvector wrote: | Did you really expect ChatGPT to be a full blown dungeon | master? | felipemnoa wrote: | With all the hype about it, the average person does. | mabbo wrote: | There's definitely a potential for a D&D DM with an LLM, but | you'd need a lot of careful prompting and processing to | handle the token limits today's models have. Simply put: a | d&d game has more story and state than the 30,000-ish words | an LLM can think about at once. | | I think there's a lot of interesting opportunities there. | bentcorner wrote: | Aetolia tried this in a 24 hr experiment: | https://nwn.blogs.com/nwn/2023/04/aetolia-mud-virtual- | world-... | | Sounds like it was a success! I suppose it comes down to | cost - I think it'd be fun to try a single player game | authored like this and would be willing to use my own API | token to try it out. | kbenson wrote: | I've also heard (here) that after you get 20-ish questions | into an instance you start getting the really weird output. | Some of the conjecture was because that's about how deep | they trained. | | In any case, if that's true, that's a very short role | playing session, unless there's a good way to retain info | but reset the state that accrues and causes problems (if | indeed that happens). | lisasays wrote: | I find those example quite encouraging, actually. In that it | shows that the current SOA is still pretty far off from | creating a planet-killing and/or species-ending AI. | pshc wrote: | I think part of this is because GPT doesn't have any "hidden | variable" storage and doesn't get any prep time up front to | come up with something coherent. Just completes the next token | based on the previous context. | agildehaus wrote: | I played 20q three times with GPT4. Aluminum can, plastic | straw, and tire pressure gauge were my items. It had no | problems getting to the answers with sensible questions | throughout. | travisjungroth wrote: | > That was a great way to show my non-tech family members the | limitations of AI and why they shouldn't trust it. | | These are the limitations of the version of ChatGPT you were | using at that moment. They are not categorical limitations of | AI or even LLMs. | | It's amazing to me how many people are sleeping on AI, mixing | up the failing cases of a freemium chatbot for the full | capability of the tech, even on HN. LLMs can say "I don't | know". Even ChatGPT can do it. Ask some super niche historical | questions of any version and see what you get. Is it perfect | every time? No. But that's something that can be reduced. | | Over the next year, you'll see more instances of lawyers citing | hallucinated cases. There will also be a handful of startups | that hook up LLMs to document stores, and they'll be able to | check for this sort of thing and do an even better job. | lettergram wrote: | I made https://AskHN.ai | | What it does is not try to answer, but collect previous | topics discussed by experts. Then answer the question based | on the text, a far more reliable approach. | bunabhucan wrote: | How does it qualify experts? I love the discussion here but | if it turns to international nuclear strategy or the | minutae of electrical networks (or presumably anything | outside the regular wheelhouse) I notice that the quality | goes down but the confidence stays the same. | Bilal_io wrote: | > LLMs can say "I don't know". Even ChatGPT can do it. | | That's the problem in my opinion. When you know something is | capable of saying "I don't know" but confidently spits out | some hallucinated BS is when the average person eats it up. | kbenson wrote: | I don't know exactly why, but for some reason this made me | think of qAnon, and now I'm thinking of an AI trained on | qAnon theories that people can form a community around like | they did qAnon, and frankly that's one of the most | terrifying things I've thought in quite a while. | travisjungroth wrote: | It is definitely a problem. OpenAI does a lot to warn | people, but I'm not really sure it's enough. | anigbrowl wrote: | _Playing "20 questions" with ChatGPT is another great way to | expose its limitations. It knows the game and tries to play, | but is terrible at asking questions to narrow down possible | answers._ | | I began doing this last winter, and while it tends to be a bit | slow I'm quite impressed that it can manage at all. | Our_Benefactors wrote: | Chat GPT versus Akinator! | seba_dos1 wrote: | GPT does nothing but predict the next word (token, to be more | exact). If you ask it for hints, it's going to come up with | some hints that sound like they could be plausible in context | of the original riddle. If you _then_ ask it what the answer | was, it 's going to try to take all the earlier hints into | account, because that's now the text it's trying to complete. | | There's no entity sitting on the other side of the screen | thinking about riddle's solution from the start. There's just a | model that completes the given text one token at a time. You | should expect similar results as if it was bunch of people | passing a single piece of paper to each other, each writing a | single word on it, not communicating with each other in any | other way and trying their best to come up with coherent | result. | jimnotgym wrote: | > Or, learn how to say "I don't know" | | This is the correct answer. It is like a sad salesman who is | out of his depth, but decides to keep bullshiting! | dylan604 wrote: | or a politician. even when presented with incontrovertible | evidence, double down | jiggawatts wrote: | ChatGPT 3.5 or GPT 4? | | Almost every negative comment about LLMs is by someone using an | older, weaker model and making generalisations. | | Here's GPT 4 giving me a riddle: | https://chat.openai.com/share/1753ce5a-d44d-44ac-bc97-599a26... | hospitalhusband wrote: | > But was it GPT4 | | I keep seeing this cop-out, which ignores that it's | fundamentally the same architecture, and has the same flaws. | More wallpaper to hide the cracks better makes it an even | worse tool for these use cases because all it does is fool | more people into thinking it has capabilities that it | fundamentally doesn't. | gerdesj wrote: | > Or, learn how to say "I don't know" | | It doesn't know that it doesn't know! It is, very roughly | speaking, a model that is designed to print out the most likely | word given its current input and training, and then the next | word etc. | | Whereas you or I might be mistaken about some of our faculties, | memories and skills, ChatGPT cannot possibly "know" what its | limitations are. It was never taught what it was not taught | (obvs) nor can it conceive of what it was not taught about. It | might come back with some useful chat in this regard but it | cannot reason. | okasaki wrote: | It seems that you don't know what you don't know, really. | | There's no way to definitively know what properties ChatGPT | has. It does seem to reason to some extent and it does often | say that some information isn't known/there's no data. And it | almost obnoxiously often tells you that it's simplifying a | complex and multifaceted situation. | [deleted] | ravenstine wrote: | Maybe you didn't get the riddle. | [deleted] | chrisco255 wrote: | Maybe the real definition of intelligence isn't being able to | answer questions effectively, but being able to know what you | know and know what you don't know. | waserwill wrote: | There's truth to this, but in a time and place. There are | times being very specific about what you know and don't know | is important, but most of the time we are learning little by | little, and we benefit from saying things we aren't entirely | comfortable with, if only to figure out whether they are true | and/or socially acceptable. | screamingninja wrote: | > I asked ChatGPT to tell me a riddle. | | Which version of ChatGPT, if you don't mind me asking? | cj wrote: | It was an iOS app (ChatOn) not the official app. So probably | GPT-3 (edit: the app claims to use GPT-4, so not sure) | | After the riddle, I bought the $20/mo subscription via the | official OpenAI app to try it on GPT-4. I started by trying | to play "20 questions" but we couldn't get past 10 questions | before getting an error message "rate limit exceeded, try | again in an hour" | macjohnmcc wrote: | I have asked it to write code. The code looked reasonable but | did not work (using a non-existent API or library). I said the | library XXX doesn't exist. And it would respond with an apology | and a new solution that maybe worked maybe didn't. Sometimes it | could pull off a save eventually. | runsWphotons wrote: | i keep getting it to write code that works but then its | example usage makes no sense. i point out the code is write | but the example it gives is wrong and it can almost never fix | it. version 4. | shahar2k wrote: | the interesting part is I've interacted with some people going | through manic phases and it's REALLY similar to chatgpt, the | ability to string words and thoughts together without any large | scale plans. | threeseed wrote: | ChatGPT doesn't know what is right or wrong. | | It simply knows what the highest probability next word should | be. | dabluecaboose wrote: | >Playing "20 questions" with ChatGPT is another great way to | expose its limitations. It knows the game and tries to play, | but is terrible at asking questions to narrow down possible | answers. | | The Rise of the Machines will be staved off as long as ChatGPT | doesn't absorb the powers of the mighty Akinator | BugsJustFindMe wrote: | I need to know what version of ChatGPT you were using, because | this is a critical piece of information that everyone just | blatantly ignores, and I can only imagine that it's out of | ignorance of the significance of the difference. | | This is what happened when I asked ChatGPT 4... | | ME | | Give me hints without outright telling me the answer to the | riddle: "What is always hungry, needs to be fed, and makes your | hands red?" | | ChatGPT 4 | | Hint 1: This object can be found in many homes, usually in a | specific room. | | Hint 2: It is not a living creature, but it does "consume" | things. | | Hint 3: The part about hands turning red relates to a common | physical response to a certain temperature. | | Hint 4: It's a crucial tool for a common household task. | | Hint 5: While it can't eat in the traditional sense, you do put | things into it quite frequently. | | ME | | What's the answer? | | ChatGPT 4 | | The answer to the riddle is a "fire" or "fireplace." It's | always hungry for more fuel (like wood), needs to be fed to | keep burning, and the heat from the fire can make your hands | red. | SparkyMcUnicorn wrote: | Just tried 20 questions on each version of Claude and it | wasn't bad! Some games it didn't even ask for hints. | | gpt-3.5-turbo struggled. | sn0wf1re wrote: | That's GPT-4, not ChatGPT (3.5-turbo I think). | | Also, yes you can get correct information by tailoring your | prompts, but that isn't the issue. The issue is that some | prompts lead to bad results and confusing/incorrect answers. | You changed what OP queried by providing the riddle and | asking for hints to that riddle, whereas OP asked for a | random riddle and then hints to that riddle. | BugsJustFindMe wrote: | > _That 's GPT-4, not ChatGPT_ | | It absolutely is ChatGPT, the paid monthly "Plus" version, | using the GPT4 model instead of the 3.5 model. | [deleted] | selcuka wrote: | I guess the point is GPT-4 hallucinates, too. Maybe it did | well for this example but still a lawyer should not trust its | output. | BugsJustFindMe wrote: | Maybe, but it's surprisingly good in the face of all the | non-version-indicating complaints about how terrible people | think it is. Mostly I doubt that the lawyer was using GPT4, | because the lawyer sounds like the kind of person who would | be ignorant of the significance of the difference. | jiggawatts wrote: | The kind of person too lazy to check the output of a | computer program before submitting it to a court of law | is the type of person too cheap to pay $20 for the good | version of the program. | | Think: Lionel Hutz. | dools wrote: | Is that 3.5 or 4? I don't play word games with GPT but for | coding tasks it went from unusable most of the time to | absolutely indispensible team member between 3.5 and 4. | boringg wrote: | Disbarred? | toyg wrote: | Actually, I think he shouldn't be - if suitably scared and | scarred, the guy will likely stay away from anything resembling | AI/ML for the rest of his life. | | Unlike language models, humans really do learn. | StrangeATractor wrote: | If he isn't he'll need to find new work anyway. Imagine | searching that guy's name while looking for an attorney. | Antoniocl wrote: | I would hope so, at a minimum! I'm shocked that signing an | affidavit in a court of law certifying the existence of made-up | courses isn't simply criminal. | anamexis wrote: | Intent (mens rea) is a core part of what makes a crime a | crime. | anigbrowl wrote: | Look up 'strict liability'. | kube-system wrote: | That's an exception, not the general rule. | occamrazor wrote: | Strict liability is extremely rare in criminal matters. | favorited wrote: | Almost certainly not. It's astoundingly difficult to get | disbarred. I heard a lawyer say the only reliable way to get | disbarred is to steal money from your clients. Judges rarely | even use their inherent power to sanction, though he | certainly will in this case. | ftxbro wrote: | I almost wonder if it could be something like one who | deliberately slips on a grocery store floor to do some legal | shenanigans. Like I'm sure there are some people who if they | saw some spilled milk on a store floor they would be like 'it's | my lucky day' and maybe similarly this lawyer saw a service | that tells lies and he could have been like 'this is too good | to be true'. | latexr wrote: | Prior discussion on the same matter from a different link: | https://news.ycombinator.com/item?id=36095352 | tromp wrote: | > The other five bogus cases were called Shaboon v. Egyptair, | Petersen v. Iran Air, Martinez v. Delta Airlines, Estate of | Durden v. KLM Royal Dutch Airlines, and Miller v. United | Airlines. | | Tyler Durden no doubt... | rvba wrote: | > The lawyer's affidavit said he had "never utilized ChatGPT as a | source for conducting legal research prior to this occurrence | (...)" | | I wonder if the court tried to verify that. | unsignedint wrote: | Simply put... | | ChatGPT is not a search engine. | | Probably OpenAI needs to prominently display this in their site. | | If you are using it to find any factual information, you are | using it wrong. You can perhaps find some hint on angle of how | you would search on things elsewhere, but that's about it... | seydor wrote: | this is funny. However GPTs are great linear interpolators | between stories and probably also between judgements. Maybe they | will be useful to replace judges | kristianbrigman wrote: | How do these usually get checked? I mean, maybe he's already done | this successfully a few times. | ajb wrote: | They get checked by the lawyer for the other side. | pkilgore wrote: | Usually the first thing a Judge's clerk will do before a | judge reads a brief is pull all of the authority cited for | the judge to have ready-to-review along side the brief itself | (if not analyzed, highlighted, summarized, etc. etc. | depending on the judge). | | Now, whether the judge _actually reads them_ is debatable (I | had my doubts sometimes). But you bet your ass that if the | Clerk simply cannot find a case, the Judge will be informed | of that. | | YMMV in State courts, which can be all over the place in | terms of professionalism. But you should _at least_ assume | your opponent is going to read your cases because the easiest | way to beat someone in court is to point out the law you rely | on is bad. | 0xbadcafebee wrote: | Were they any worse than the normal junk cases that get filed | every day in this country? | onionisafruit wrote: | "unprecedented"? ChatGPT says there is precedent and gave me | several citations. | hluska wrote: | For reference, here is the judge's order. Endnote #2 alone is | worth the price of admission: | | https://s3.documentcloud.org/documents/23826753/judgeaskingt... | StrangeATractor wrote: | That guy has some work to do on his resume. | josu wrote: | He'll just outsource it to chatGPT. | bitwize wrote: | So turds turn up in AI law as well... | Brendinooo wrote: | Yeah, AI-generated fake cases would definitely be unprecedented, | unless SmarterChild did a stint as a paralegal in the early | aughts... | not2b wrote: | At minimum the sanctions should include a fine sufficient to | cover the costs of everyone who had to deal with this: the court, | opposing attorneys. Maybe punitive damages too. But let the guy | keep his law license (unless there's a second offense). | CSMastermind wrote: | It feels like there definitely should be punitive damages? | | It's like they didn't even check the text that ChatGPT | generated for correctness. ___________________________________________________________________ (page generated 2023-05-30 23:00 UTC)