[HN Gopher] Don't believe ChatGPT - we do not offer a "phone loo...
       ___________________________________________________________________
        
       Don't believe ChatGPT - we do not offer a "phone lookup" service
        
       Author : freyfogle
       Score  : 158 points
       Date   : 2023-02-23 21:24 UTC (1 hours ago)
        
 (HTM) web link (blog.opencagedata.com)
 (TXT) w3m dump (blog.opencagedata.com)
        
       | hayksaakian wrote:
       | This marks the new age of "AI Optimization" where companies will
       | strive to get their business featured into answers in ChatGPT.
       | 
       | The OP's example is Unwanted demand, but it clearly shows that
       | ChatGPT can funnel potential customers towards a product or
       | service.
        
       | jefftk wrote:
       | _This is not a service we provide. It is not a service we have
       | ever provided, nor a service we have any plans to provide.
       | Indeed, it is a not a service we are technically capable of
       | providing._
       | 
       | I'm curious: why not? It seems like a lot of people would be
       | interested in this if you could figure out how to provide it.
        
         | insane_dreamer wrote:
         | > a lot of people would be interested in this
         | 
         | you mean like scammers and stalkers? (ok, and probably Meta)
        
         | iamflimflam1 wrote:
         | The service is possible:
         | 
         | If you are a mobile network operator.
         | 
         | Or, you can convince people to install something on their phone
         | that sends you their location along with their phone number.
        
         | ceejayoz wrote:
         | How would _you_ go about reliably providing the location of
         | someone 's _mobile_ phone without being their cell phone
         | carrier?
        
           | jraph wrote:
           | By partnering with said cell phone carriers.
           | 
           | But I hope it would be illegal.
        
         | simonw wrote:
         | How would this work?
         | 
         | If a phone number is for a mobile phone then looking up the
         | location doesn't make sense at all: mobile phones are mobile.
         | 
         | I guess you could try and crawl an index of business phone
         | numbers and associate those with the listed address for
         | businesses, but that's a completely different business from
         | running a geocoder.
         | 
         | You could provide a bit of geographical information about the
         | first three digits of a US phone number. I imagine that's not
         | what users are actually looking for though.
        
         | cactusplant7374 wrote:
         | You mean if they could figure out how to illegally track
         | millions of people?
        
       | mort96 wrote:
       | That's quite the predicament. I hope OpenAI will listen, to this
       | and to anyone else in a similar situation. I'm reminded of the
       | cases of ChatGPT recommends random people's personal phone
       | numbers for various services.
       | 
       | But yeah, don't trust ChatGPT for anything. Just earlier today I
       | tried my darnedest to convince it that 2 pounds of feathers
       | doesn't weigh the same as 1 pound of bricks, and it just would
       | not listen, presumably because it just regurgitated stuff related
       | to the common "1 pound of feathers and 1 pound of bricks"
       | question.
       | 
       | By the way, the last paragraph has some typos:
       | 
       | > _I wrote this post to have a place to send our new ChatGPT
       | users when they ask why it isn't work, but hopefully also it
       | serves as a warning to othrs - you absolutely can not trust the
       | output of ChatGPT to be truthful,_
        
         | insane_dreamer wrote:
         | > don't trust ChatGPT for anything
         | 
         | Agreed. But then it begs the question: what purpose does
         | ChatGPT serve (other than for entertainment purposes or
         | cheating on your HS/college exam)? If you have to verify its
         | information by other means, then you're not really saving much
         | effort.
        
           | shagie wrote:
           | It works really well for translating one "language" to
           | another "language".
           | 
           | Give it some structured data and ask it to summarize it (e.g.
           | hourly weather data and it gives a better summarization than
           | a template based one).
           | 
           | Give it HN titles and the categories and it does a passable
           | zero shot tagging of them (
           | https://news.ycombinator.com/item?id=34156626 ).
           | 
           | I'm toying around with making a "guided bedtime story
           | generator". A friend of mine uses it to create a "day in the
           | life of a dinosaur" stories for a child (a different story
           | each day!)
           | 
           | The key is to play to its strengths rather than testing its
           | bounds and complaining that they break in weird ways when
           | they will inevitably break in weird ways.
        
           | worldsayshi wrote:
           | This was my initial thought as well. But I've noticed that my
           | brain has started to find tasks that it would be quite useful
           | for. Too bad it's almost always seem to be at capacity when I
           | think of those cases. Guess I will have to pay up to figure
           | out if it's actually worth it.
        
         | geoelectric wrote:
         | I thought for sure that must be glitching, so just tried on my
         | ChatGPT Plus account with a fresh thread, but nope:
         | 
         | > _does 2 pounds of feathers weigh more than 1 pound of
         | bricks?_
         | 
         | > _No, two pounds of feathers weigh the same as one pound of
         | bricks. The weight of an object is determined by its mass, and
         | one pound of any object weighs the same as one pound of any
         | other object. However, feathers are less dense than bricks, so
         | two pounds of feathers will take up more space or have a larger
         | volume than one pound of bricks._
         | 
         | Huh. It's interesting that some of these "wrong paths" are
         | quite so repeatable. Wonder if that means there's also a way to
         | band-aid them when found, the way a human might be corrected
         | for a misconception.
        
           | jxy wrote:
           | This is hilarious. ChatGPT even gave me a more bizarre
           | example.
           | 
           | > does 2 pounds of bricks weigh more than 1 pound of bricks?
           | 
           | > No, 2 pounds of bricks do not weigh less or more than 1
           | pound of bricks. 2 pounds of bricks and 1 pound of bricks
           | refer to the same unit of weight, which is a pound.
           | Therefore, they weigh the same, which is one pound. The
           | difference between them is only the quantity, not the weight.
           | 
           | > It's similar to saying that two one-dollar bills do not
           | have more value than one five-dollar bill. Even though you
           | have two bills in one case and one bill in the other case,
           | the total value is the same.
        
             | mort96 wrote:
             | It even managed to give me this gem:
             | 
             | > 2 pounds of feathers do not weigh less than 1 pound of
             | bricks. 1 pound of bricks weighs more than 2 pounds of
             | feathers, but they can have the same mass.
        
       | b800h wrote:
       | Is this not defamation, at least in some jurisdictions?
        
       | 99_00 wrote:
       | I remember a time when "I saw it on the internet" was a punchline
       | for a joke about someone who's gullible or misinformed.
        
       | ninjakeyboard wrote:
       | ChatGPT gets the rules to the pokemon trading card game wrong. It
       | will tell you you can use 4 energy a turn. Convincingly. Not sure
       | how it hallucinates this. The rule is 1 per turn.
        
         | DoktorDelta wrote:
         | Could it have been referencing Blastoise's Deluge ability?
         | Jacob Van Wagner used it in the 2015 championship to use 4
         | water energy in one turn.
        
         | bigmattystyles wrote:
         | Isn't it just garbage went in, got weighed as a more reliable
         | source than it should have been and thus garbage came out. Good
         | old GIGO... It's just here, ChatGpt, as much as I love it, is
         | amazing at imparting the impression that its shit don't stink.
        
         | codetrotter wrote:
         | A few days ago I asked ChatGPT if "pannekake" and "kannepake"
         | are anagrams of each other.
         | 
         | It correctly stated that they are, but when it went on to prove
         | that this was the case, it generated a table of the frequencies
         | of the individual letters in these two words, and the table
         | looked like this.                   Letter | Frequency in |
         | Frequency in                | "pannekake"  | "kannepake"
         | - - - - - - - - - - - - - - - - - - -         a      | 2
         | | 2         e      | 2            | 2         k      | 2
         | | 2         n      | 2            | 2         p      | 2
         | | 2
         | 
         | This reminded me that yes indeed, AI just isn't quite there
         | yet. It got it right, but then it didn't. It hallucinated the
         | frequency count of the letter "p", which occurs only once, not
         | twice in each of those words.
        
           | int_19h wrote:
           | Anything that has to do with individual words doesn't work
           | well, but as I understand, this is an artifact of the
           | tokenization process. E.g. pannekake is internally 4 tokens:
           | pan-ne-k-ake. And I don't think that knowing which tokens
           | correspond to which letter sequences is a part of the
           | training data, so it has to infer that.
        
       | kelseyfrog wrote:
       | > All suggestions are welcome.
       | 
       | Monetize it!
       | 
       | Evil answer: Partner with an advertiser and sell
       | https://api.opencagedata.com/geocode/v1/json as an ad space. This
       | may be the first opportunity for an application/json-encoded
       | advertisement.
       | 
       | Nice answer: Partner with an actual phone lookup platform and
       | respond with a 301 Moved Permanently at the endpoint.
        
         | insane_dreamer wrote:
         | > actual phone lookup platform
         | 
         | uh, you mean stalker / scammer platform? This would be a major
         | privacy violation.
        
         | rosywoozlechan wrote:
         | there's no "actual phone lookup platform" you can't get a
         | person's location by knowing their phone number, that's a huge
         | privacy violation. You can get the location of your own phone
         | via icloud or google's system for android. You could also
         | install an app on your phone to track your phone's location.
         | You cannot find people based on knowing their phone number,
         | that would be a serious safety issue for you know people trying
         | to not, for example, get murdered by their ex-boyfriends.
        
           | throwaway29495 wrote:
           | What about phone numbers corresponding to a specific
           | location?
        
           | Hello71 wrote:
           | it's been reported numerous times that you can buy real-time
           | cell phone location data:
           | https://news.ycombinator.com/item?id=17081684,
           | https://news.ycombinator.com/item?id=20506624,
           | https://news.ycombinator.com/item?id=32143256. you might need
           | a little more info than just a phone number, but (allegedly)
           | not that much more.
        
       | ninjakeyboard wrote:
       | It hallucinates that you can use 4 energy per turn in Pokemon TCG
       | and confidently tells you so. No idea where that would come from.
        
       | aaron695 wrote:
       | [dead]
        
       | ntonozzi wrote:
       | Including the word 'phone' six times in a popular blog post is
       | not going to help their predicament.
        
         | elicash wrote:
         | Wouldn't they want this post to be at the top when people
         | search 'phone' and 'open cage data'? Seems like SEO towards
         | correcting this is only helpful. And maybe when GPT updates
         | data, this post gets pulled in, too. The more popular, the
         | better, I'd guess.
        
         | KomoD wrote:
         | Not gonna hurt either, ChatGPT data is not up to date
        
       | freyfogle wrote:
       | ChatGPT very convincingly recommends us for a service we don't
       | provide.
       | 
       | Dozens of people are signing up to our site every day, then
       | getting frustrated when "it doesn't work".
       | 
       | Please do NOT trust the nonsense ChatGPT spits out.
        
         | seedless-sensat wrote:
         | A new market opportunity for your company?
        
           | theWreckluse wrote:
           | > It is not a service we have ever provided, nor a service we
           | have any plans to provide. Indeed, it is a not a service we
           | are technically capable of providing.
        
           | hackernewds wrote:
           | this seems like a game-changing opportunity actually. I'd be
           | down to buy the domain
        
           | anaganisk wrote:
           | So, based on the BS these LLMs spout and companies start
           | pivoting. The govts should start writing laws?
        
           | input_sh wrote:
           | > This is not a service we provide. It is not a service we
           | have ever provided, nor a service we have any plans to
           | provide. Indeed, it is a not a service we are technically
           | capable of providing.
        
         | fire wrote:
         | have you been able to contact OpenAI about this? It sounds like
         | they're actively adding load to your CS ops with this
        
           | hackernewds wrote:
           | what are they going to do? add custom logic? where does it
           | stop?
           | 
           | the malady is that LLMs cannot do operational adhoc changes
           | such as these kinds of errors at scale
        
       | assdontgot wrote:
       | [dead]
        
       | coldtea wrote:
       | ChatGPT doesn't "recommended" anything. It just recombines text
       | based on statistical inferences that appear like a
       | recommendation.
       | 
       | It could just as well state that humans have 3 legs depending on
       | its training set and/or time of day. In fact it has said similar
       | BS.
        
         | mort96 wrote:
         | What would you call it instead?
        
           | qwertox wrote:
           | "Makes stuff up." And it's us, the users, who have to realize
           | this. I mean, I wouldn't blame OpenAI for this, at least not
           | at this point, and the company will have to live with it,
           | look how it can turn it into something useful instead, since
           | there's no one to complain to.
        
             | vlunkr wrote:
             | > I wouldn't blame OpenAI for this
             | 
             | They're offering the tool, it's at least partially their
             | responsibility to tell people how it should and should not
             | be used.
        
             | rodgerd wrote:
             | Why wouldn't you blame OpenAI for creating a harassment
             | campaign against the business based on nonsense?
        
             | [deleted]
        
           | coldtea wrote:
           | A glorified Markov chain generator.
           | 
           | Now, humans could very well also be statistical inference
           | machines. But they have way more tricks up their semantic-
           | level understanding sleeves than ChatGPT circa 2023.
        
         | circuit10 wrote:
         | > ChatGPT doesn't "recommended" anything. It just recombines
         | text based on statistical inferences that appear like a
         | recommendation.
         | 
         | I think that's a bit pedantic and not very helpful... I'm not
         | typing this comment, my brain is just sending signals to my
         | hands which causes them into input data into a device that
         | displays pixels that look like a comment
        
           | coldtea wrote:
           | > _I think that's a bit pedantic and not very helpful... I'm
           | not typing this comment, my brain is just sending signals to
           | my hands which causes them into input data into a device that
           | displays pixels that look like a comment_
           | 
           | Well, if you're just fed a corpus, with no real-time first-
           | person strem of experience that you control, no feedback
           | mechanism, no higher level facilities, and you're not a
           | member of a species with a proven track record of state-of-
           | the-art in nature semantic understanding, then maybe...
        
       | crazygringo wrote:
       | I'm curious -- does anyone know of ML directions that could add
       | any kind of factual confidence level to ChatGPT and similar?
       | 
       | We all know now that ChatGPT is just autocomplete on steroids. It
       | produces plausibly convincing _patterns_ of speech.
       | 
       | But from the way it's built and trained, it's not like there's
       | even any kind of factual confidence level you could threshold, or
       | anything. The concept of factuality doesn't exist in the model at
       | all.
       | 
       | So, is any progress being made towards internet-scale ML "fact
       | engines" that also have the flexibility and linguistic
       | expressiveness of ChatGPT? Or are these just two totally
       | different paths that nobody knows how to marry?
       | 
       | Because I know there's plenty of work done with knowledge graphs
       | et al., but those are very brittle things that generally need
       | plenty of human curation and verification, and can't provide any
       | of the (good) "fuzzy thinking" that ChatGPT can. They can't
       | summarize essays or write poems.
        
         | csours wrote:
         | I'm curious about falsifiable models.
        
         | alfalfasprout wrote:
         | By definition, an LLM doesn't have a semantic world model or
         | ontology. Even the most "dumb" (and I use that in quotes
         | because they really aren't) animal is able to reason about
         | uncertain concepts and understands risk and uncertainty.
         | 
         | Yann Lecun has posted a lot recently about this but basically
         | LLMs are a "useful offramp on the road to AGI".
        
         | BoorishBears wrote:
         | There's research being done on this:
         | https://arxiv.org/abs/2302.04761
         | 
         | At its core using an LM _alone_ to solve factual problems seems
         | silly: It 's not unlike asking Dall-E to draw DOT compliant
         | road signs.
         | 
         | I've gone at length at how unfortunate it would be if LMs start
         | to get a bad rap because they're being shoehorned into being
         | "Ask Jeeves 2.0" when they could be so much more.
        
         | irrational wrote:
         | Remember the guy a few weeks ago that was being gaslighted by
         | ChatGPT that this is the year 2022? Not only is it giving out
         | potentially false info, but it will double down that it is
         | right and you are wrong. Though, to be honest, that sounds like
         | a lot of real people. The difference is, people are smart
         | enough to not double down on try to say it is a different year
         | and your phone is probably reporting the year wrong.
        
           | amscanne wrote:
           | That was the Bing preview, which is supposed to be an actual
           | information product.
        
         | snowstormsun wrote:
         | I think "Explainable AI" is a related research direction, but
         | perhaps not popular for language models.
        
           | behnamoh wrote:
           | Impossible to explain the inner workings of GPT-3 without
           | having access to the model and its weights. Does anyone know
           | if any methods exist for this?
        
             | IncRnd wrote:
             | I asked ChatGPT for some in-depth source code that
             | realistically mimics chatgpt. ChatGPT replied with various
             | answers in python. I'm not sure any of them are correct,
             | though.
        
           | shawntan wrote:
           | I think part of the issue is what level of explanation is
           | satisfactory. We can explain how every linear transformation
           | computes its output, but the sum of it is in many ways more
           | than its parts.
           | 
           | Then there are efforts that look like this one:
           | https://news.ycombinator.com/item?id=34821414 They go probing
           | for specific capabilities of Transformers to figure out which
           | cell fires under some specific stimulus. But think a little
           | bit more about what people might want from explainability and
           | you quickly find that something like this is insufficient.
           | 
           | There may be a tradeoff we're looking at where explainability
           | (for some definition of it) will have to be exchanged for
           | performance (under some set of tasks). You can build more
           | interpretable models these days, but you usually pay for it
           | in terms of how well you do on benchmarks.
        
         | mochomocha wrote:
         | > But from the way it's built and trained, it's not like
         | there's even any kind of factual confidence level you could
         | threshold, or anything. The concept of factuality doesn't exist
         | in the model at all.
         | 
         | I'm not super familiar with ChatGPT internals, but there are
         | plenty of ways to tack on uncertainty estimates to predictions
         | of typical "large scale ML models" without touching Bayesian
         | stuff (which only work for small scale academics problems). You
         | can do simple parametric posteriors estimation or if all you
         | have is infinite compute and don't even want to bother with
         | anything "mathy", bootstrapping is the "scalable / easy"
         | solution.
        
           | pavon wrote:
           | Sure, but would that uncertainty estimate measure the
           | accuracy of the data or the accuracy of it being a reasonably
           | sounding sentence.
        
         | ericlewis wrote:
         | its super duper easy, prob not perfect and I don't have any
         | sort of proper "test": 1. I ask the model first if it seems
         | like a question that benefits from an external answer 2. I talk
         | to Wolfram alpha with some abstraction of the question 3. I
         | wait for a response 4. I "incept" it into the final response,
         | essentially a prompt that mixes in a context of sorts that
         | contains the factual information.
         | 
         | you could cross check this stuff too with yet more models.
        
         | simonw wrote:
         | That's basically what the new Bing is. It's a large language
         | model that can run searches, and then use what comes back from
         | those searches to generate answers to questions.
         | 
         | Whether or not the information that comes back from those
         | searches is reliable is a whole other question.
         | 
         | I would love to learn what the latest research is into "factual
         | correctness" detection. Presumably there are teams out there
         | trying to solve that one?
        
           | behnamoh wrote:
           | AFAIK, Bing AI is not itself an LLM, but rather a wrapper
           | around ChatGPT, which itself is based on GPT-3, which is
           | based on the GPT architecture, which is (roughly speaking)
           | half of a transformer architecture, which is based on
           | encoder/decoder neural nets which are based on ...
        
             | nl wrote:
             | It's a newer, different GPT model than chatGPT.
        
               | simonw wrote:
               | To quote the Bing announcement post:
               | https://blogs.microsoft.com/blog/2023/02/07/reinventing-
               | sear...
               | 
               | > Next-generation OpenAI model. We're excited to announce
               | the new Bing is running on a new, next-generation OpenAI
               | large language model that is more powerful than ChatGPT
               | and customized specifically for search. It takes key
               | learnings and advancements from ChatGPT and GPT-3.5 - and
               | it is even faster, more accurate and more capable.
        
         | nl wrote:
         | > does anyone know of ML directions that could add any kind of
         | factual confidence level to ChatGPT and similar?
         | 
         | Yes. It's a very active area of research. For example:
         | 
         | Discovering Latent Knowledge in Language Models Without
         | Supervision (https://arxiv.org/abs/2212.03827) shows an
         | unsupervised approach for probing a LLM to discover things it
         | thinks are facts
         | 
         | Locating and Editing Factual Associations in GPT
         | (https://arxiv.org/pdf/2202.05262.pdf) shows an approach to
         | editing a LLM to edit facts.
         | 
         | Language Models as Knowledge Bases?
         | (https://aclanthology.org/D19-1250.pdf) is some slightly older
         | work exploring how well LLMs store factual information itself.
        
       | singlow wrote:
       | Its not like ChatGPT made this up. There were pre-existing
       | YouTube tutorials and python scripts available that used OpenCage
       | an purported to do this. OpenCage even blogged about this problem
       | almost a year ago[1].
       | 
       | Honestly it looks more like OpenCage is trying to rehash the same
       | issue for more clicks by spinning it off the hugely popular
       | ChatGPT keywords. Wouldn't be too surprised if they created the
       | original python utilities themselves just to get some publicity
       | by denouncing them.
       | 
       | 1. https://blog.opencagedata.com/post/we-can-not-convert-a-
       | phon...
        
         | freyfogle wrote:
         | Hi, Ed from OpenCage here, author of the post.
         | 
         | We do have python tutorials and SDKs showing how to use our
         | service for ... geocoding, the actual service we provide.
         | 
         | I wrote the post mainly to have a page I can point people to
         | when they ask why "it isn't working". Rather than take the user
         | through a tour of past posts I need something simple they will
         | hopefully read. But fair point, I can add a link to last year's
         | post about the erronious youtube tutorials as well.
         | 
         | What I think you can't appeciate is the difference of scale. A
         | faulty youtube video drives a few users. In the last weeks
         | ChatGPT is sending us several orders of magnitude more
         | frustrated sign-ups.
        
           | singlow wrote:
           | I get frustrated at the number of things ChatGPT gets blamed
           | for that aren't its fault. It is completely understandable
           | that if there are repos out on GitHub like the one for
           | Phomber[1] thant ChatGPT would find that code and have no
           | idea that it was phoney. Suggesting that ChatGPT just made
           | this up out of thin air when you know it didn't is not very
           | responsible.
           | 
           | 1. https://github.com/s41r4j/phomber
        
             | jraph wrote:
             | You are blaming the victim. OpenAI is to be blamed.
             | 
             | They know what they are doing. They provide something that
             | sounds over-confident for anything it says, knowing full
             | well that it can't actually know if what it generated is
             | accurate because it is designed to generate plausible
             | sentences using statistics and probabilities, not verified
             | facts from a database. On top of it, they trained it on an
             | uncontrolled set of texts (though IIUC even a set of
             | verified text would not be enough, nothing guarantees that
             | a LM would produce correct answers). And they provide it to
             | the general population, which doesn't always understand
             | very well how it works and, above all, its limitations.
             | _Including developers_. Few people actually understand this
             | technology, including myself.
             | 
             | Inevitably, it was going to end up causing issues.
             | 
             | This post factually presents a problematic situation for
             | the authors of this post. How ChatGPT works or how it can
             | end up producing wrong results is irrelevant to the post's
             | authors problem. It just does, and it causes troubles
             | because of the way OpenAI decided to handle things.
             | 
             | And it's not "fair enough, because this false stuff can be
             | found on the internet".
        
             | mtmail wrote:
             | Phomber is not the best example. Ed contacted the developer
             | of that tool over a year ago about the issue and to remove
             | mentions of OpenCage and as far as I see the author removed
             | it https://github.com/s41r4j/phomber/issues/4
        
         | gus_massa wrote:
         | That explains why ChatGPT is confused.
         | 
         | It may be an old problem, but I guess users are more use to a
         | random YouTube video with wrong information. But the computer
         | is always right so ChatGPT is always right, so users may be
         | more annoyed to discover that the recommendation is wrong and
         | blame them instead of ChatGPT.
        
         | ceejayoz wrote:
         | That seems like a pretty nasty assertion to bandy around with
         | zero evidence.
        
           | singlow wrote:
           | I cannot think of any other reason why the new blog post
           | wouldn't have mentioned the obvious connection to the earlier
           | issues that they had. They want to make it seem like ChatGPT
           | invented this use case but they know that the sample code
           | that ChatGTP learned from was mentioned in their previous
           | blog post.
        
             | ceejayoz wrote:
             | There's a vast chasm between "whoever wrote this article
             | didn't think to link to a similar issue a year ago" and
             | "the first incident was a malicious hoax".
        
               | singlow wrote:
               | The author of both posts is purportedly the same person.
               | But he probably didn't write either of them. It was
               | probably his social media personal assistant. .
        
               | freyfogle wrote:
               | Just re-checked the org chart. There's no social media
               | personal assistant.
        
               | ceejayoz wrote:
               | That's another apparently evidence-free accusation.
               | 
               | Is there some undisclosed bad blood here?
        
               | singlow wrote:
               | [flagged]
        
               | mtmail wrote:
               | Ed is my co-founder, he writes all our blog posts because
               | I suck I writing. He also does more than half of our
               | podcast episodes https://thegeomob.com/podcast (the guy
               | on the left). Last I saw him (yesterday) he was real.
        
               | luckylion wrote:
               | I don't understand the original comment to suggest that.
               | Rather: it's a known issue. ChatGPT does nothing new, and
               | certainly doesn't do it by itself -- it just rehashes
               | what others have already written. Like Google might send
               | you visitors for something that's not even present on
               | your website because others link to you mentioning it.
               | 
               | What the comment suggested was that they're now bringing
               | this up again to get attention (and links) since it's
               | combined with ChatGPT. That's not "malicious", but it's
               | also not exactly "wow, we just realized this happens".
        
               | seszett wrote:
               | What the comment suggested is that the company
               | _deliberately created tools using their own API in a
               | wrong way in order to write a blog post about it_.
               | 
               | If that's not an accusation of being malicious I don't
               | know what could be.
        
           | vlunkr wrote:
           | There's also no clear motive. They want to attract users to a
           | fake feature their free tier?
        
       | VectorLock wrote:
       | This is the biggest problem I encounter when trying to use
       | ChatGPT on a daily basis for computer programming tasks. It
       | "hallucinates" plausible looking code that never existed or would
       | never work, especially confusing whats in one module or API for
       | something in another. This is where ChatGPT breaks when pushed a
       | bit further than "make customized StackOverflow snippets."
       | 
       | For example I asked ChatGPT to show me how to use an AWS SDK
       | "waiter" to wait on a notification on an SNS topic. It showed me
       | code that looked right, but was confusing functions in the SQS
       | library for those that would do the thing with SNS (but SNS
       | doesn't support what I wanted)
        
         | shagie wrote:
         | Have you tried using the code-davinci-002 model instead of
         | ChatGPT?
         | 
         | For example - https://platform.openai.com/playground/p/default-
         | translate-c...
         | 
         | The codex models are intended for doing work with code rather
         | than language and may give better results in that context.
         | https://help.openai.com/en/articles/6195637-getting-started-...
        
         | IncRnd wrote:
         | It does indeed sound problematic to use ChatGPT daily for
         | computer programming tasks. ChatGPT is not a snippets manager
         | but text completion.
         | 
         | It may be more helpful to look for better answers on Amazon's
         | help pages for SNS and AWS SDK.
        
         | wvenable wrote:
         | The problem is compounded by the fact that sometimes it
         | produces really good results. One task, good results. Next
         | task, totally hallucinated result.
        
       | gumballindie wrote:
       | ChatGPT is hilariously buggy - I asked "it" how to use an open
       | source library i made. The output was wrong ranging from a broken
       | github url to outright broken or nonexistent code. I suspect it
       | may even have used private code from other libs - couldnt find
       | some of the output it generated anywhere public.
        
       | IshKebab wrote:
       | Well for a start you could make it more obvious what your service
       | _does_ do. I don 't know what "geocoding" is. Converting things
       | to/from "text" is meaningless. You have to get all the way down
       | ... way down, past authentication to the details of the `q` query
       | parameter before it actually tells you.
       | 
       | At the top you should have a diagram like this:
       | 
       | Lat, lon <- opencage -> address
       | 
       | With a few examples underneath.
        
         | mtmail wrote:
         | "Past authentication", so you're looking at the
         | https://opencagedata.com/api page. Most people go to the
         | homepage first. Great feedback, we should make it clearer on
         | that page and add examples earlier. Thanks!
        
       | yieldcrv wrote:
       | lol it recommended their api and gave python code for using it
       | 
       | but the real api doesnt give results that the user asked ChatGPT
       | for
       | 
       | that is amusingly alarming
        
         | CabSauce wrote:
         | Not quite as alarming as these people most likely trying to
         | stalk someone without their permission.
        
           | hk__2 wrote:
           | > Not quite as alarming as these people most likely trying to
           | stalk someone without their permission.
           | 
           | It's so common to want to know where does a incoming call
           | come from that it's built-in in iOS. It has nothing to do
           | with stalking, just with guessing if who's calling you is a
           | scammer or a company trying to sell you stuff.
        
             | cjbgkagh wrote:
             | It's pretty simple to look up the location of a phone
             | number issuance, you can get a map or table that does this.
             | I guess these people want the current physical location of
             | the mobile phone. Either way these are not customers you'd
             | want.
             | 
             | Edit: reading the blog post from the same company listed
             | above, it is indeed people using an external API for what
             | is an incredibly simple country code. It is a shame that
             | programming has come to this and that ChatGPT continues to
             | propagate it. One way they could solve the problem would be
             | to provide sample code that does the same thing using a
             | built in table without using their API service. Sure it's
             | work but not much will get ppl off your back asap.
        
             | simonw wrote:
             | I'm willing to bet people asking ChatGPT to help them
             | resolve a phone number to a location are much more likely
             | to be stalkers than people who are trying to identify spam
             | calls.
        
         | goguy wrote:
         | Our jobs are safe! For now...
        
         | int_19h wrote:
         | The obvious follow-up is to create the non-existing API
         | endpoint but hook it into GPT so that it can hallucinate a
         | convincing address based on the phone number. Take GPT API key
         | as input so that the caller is paying for this.
         | 
         | Bonus points for using ChatGPT to implement this end-to-end.
        
       | CactusOnFire wrote:
       | Because ChatGPT is so new, we are in this weird period where
       | people haven't learned that is just as incorrect as the rest of
       | us.
       | 
       | I am hoping that in a year from now people will be more skeptical
       | of what they hear from conversational AI. But perhaps that is
       | optimistic of me.
        
         | ravenstine wrote:
         | AI will never be totally correct. If it ever is, then we've
         | found God.
        
         | austinshea wrote:
         | It's not incorrect like the rest of us. It's incorrect in a
         | very different way.
         | 
         | Providing detailed information on the usage of a service that
         | has never existed is a brand new kind of incorrect that is
         | carelessly causing the rest of us grief.
        
         | Xylakant wrote:
         | > Because ChatGPT is so new, we are in this weird period where
         | people haven't learned that is just as incorrect as the rest of
         | us.
         | 
         | It's worse than that. It's wrong, you cannot correct it and it
         | makes up supporting citations on the fly. Very few humans
         | behave like that.
        
           | renewiltord wrote:
           | I think very many humans behave like that, actually. A recent
           | example is people claiming that Flint, MI still has leaded
           | water.
           | 
           | But in the past, HN users "corroborated" that Apple is spying
           | on them etc. Fabrication is well and alive among us.
        
             | nl wrote:
             | > A recent example is people claiming that Flint, MI still
             | has leaded water.
             | 
             | Doesn't it?
             | 
             | According to [1]:
             | 
             |  _The residential lead service line replacement was
             | initially set to be finished in 2019, according to a
             | settlement agreement with the city. That deadline was
             | eventually pushed back to the fall of 2022 and has most
             | recently been set for completion in August 2023, according
             | to city officials._
             | 
             | and
             | 
             |  _" More than 95% of lead pipes in Flint have been
             | replaced, and we will continue the work until the job is
             | done," Flint Mayor Sheldon Neeley said in a recent
             | statement on the water filters._
             | 
             | It sounds to me a lot like Flint, MI still has leaded
             | water?
             | 
             | [1] https://abcnews.go.com/US/flint-residents-urged-filter-
             | water...
        
           | TehCorwiz wrote:
           | I can think of more than a few that regularly appear on TV.
        
             | Xylakant wrote:
             | So can I, but luckily TV is not representative of the world
             | at large.
        
         | annoyingnoob wrote:
         | > just as incorrect as the rest of us
         | 
         | Even worse because it has no clue when it might be completely
         | wrong and yet it will be confident in its answer.
        
           | DoktorDelta wrote:
           | That might be the most human thing it's ever done
        
             | mdp2021 wrote:
             | Dunning-Kruger, provisionality and delirating are different
             | things.
        
         | none_to_remain wrote:
         | Humans are capable of not bullshitting
         | 
         | ChatGPT can only bullshit
        
         | avgDev wrote:
         | It is quite interesting really. I took AI in school but I have
         | not dived deep at all in ChatGPT but isn't chatGpt just
         | learning from the internet?
         | 
         | Could someone push "wrong" opinion heavily online to sway the
         | opinion of AI?
         | 
         | I can only imagine a bot that learned from 4chan.
        
       ___________________________________________________________________
       (page generated 2023-02-23 23:00 UTC)