[HN Gopher] ChatGPT's API is so good and cheap, it makes most te...
       ___________________________________________________________________
        
       ChatGPT's API is so good and cheap, it makes most text generating
       AI obsolete
        
       Author : minimaxir
       Score  : 286 points
       Date   : 2023-03-11 18:20 UTC (4 hours ago)
        
 (HTM) web link (minimaxir.com)
 (TXT) w3m dump (minimaxir.com)
        
       | d23 wrote:
       | Off topic, but how'd you get that amazing header banner Max? I
       | tried a web-based ControlNet + Stable Diffusion combo[1], but the
       | quality is nothing near what you have there.
       | 
       | [1] - https://stablediffusionweb.com/ControlNet
        
         | minimaxir wrote:
         | https://news.ycombinator.com/item?id=35112641
        
       | zibzob wrote:
       | Does anyone know if there's a way to use this technology to help
       | understand a large codebase? I want a way to ask questions about
       | how a big unfamiliar codebase works. It seems like ChatGPT isn't
       | trained on open source code, so it can't answer questions there.
       | When I asked it how something works in the Krita source, it just
       | hallucinated some gibberish. But if there's a way to train this
       | AI on a specific codebase, maybe it could be really useful. Or is
       | that not possible with this type of AI?
        
         | freezed88 wrote:
         | This is what we've designed LlamaIndex for!
         | https://github.com/jerryjliu/gpt_index. Designed to help you
         | "index" over a large doc corpus in different ways for use with
         | LLM prompts.
        
         | roflyear wrote:
         | No, not large understanding. But if you are unfamiliar with
         | specific language features, or there is confusing code it can
         | help you figure things out. But no it is not good for any large
         | corpus of text, and you can't give it new stuff and teach it
         | anything.
        
         | scarface74 wrote:
         | It "knows" the AWS API, CloudFormation and from what others
         | have told me the CDK pretty well. I've asked it to write plenty
         | of 20-30 line utility scripts and with the proper prompts, it
         | gets me 95% of the way there.
         | 
         | I assume it would "understand" more popular open source
         | frameworks.
        
         | vorticalbox wrote:
         | In this case you can feed it bits of code you're interested in
         | and ask it to explain, the API has a limit of 4096 tokens
         | (which is a good chunk of text).
         | 
         | I actually built a slack bot for work and daily ask it to
         | refactor code or "write jsdocs for this function"
        
           | zibzob wrote:
           | Yeah, and this is pretty useful for small bits of code, but
           | what I want is a way to ask questions about large projects.
           | It would be nice to ask something like "which classes are
           | responsible for doing X", or "describe on a high level how Y
           | works in this code". But I'm not sure if that is actually
           | possible with the current technology.
        
             | roseway4 wrote:
             | It's possible to do this either by fine-tuning an existing
             | model or using an existing chat model prompts enriched by a
             | vector search for relevant code. See my comment elsewhere.
        
         | roseway4 wrote:
         | OpenAI Codex understands code. Though it's primary use case is
         | code completion, it might be to do Q&A well given a prompt with
         | context.
         | 
         | https://platform.openai.com/docs/guides/code
         | 
         | I'd you're interested in trying the very cheap models behind
         | ChatGPT, you may want to have a look at langchain and
         | langchain-chat for an example of how to build a chatbot that
         | uses vectorized source code to build context-aware prompts.
        
           | zibzob wrote:
           | Thanks for the links, I'll take a look at this and see if
           | it's something I could reasonably achieve.
        
         | Jiocus wrote:
         | Have you checked out Copilot Labs, the experimental version of
         | Copilot? It's bundled with ability to explain and document
         | source code, among other things.
         | 
         | https://githubnext.com/projects/copilot-labs/
        
           | zibzob wrote:
           | That looks promising! But I think it only works on small
           | snippets of code and doesn't have an overview of the whole
           | codebase...still, maybe it's coming down the line as they
           | improve it.
        
         | rocauc wrote:
         | There's good work happening in this area, e.g. Sourcegraph is
         | working on "Cody" to understand and search your code base
         | https://twitter.com/beyang/status/1614895568949764096
        
         | bcrosby95 wrote:
         | ChatGPT does not understand your code, does not have the same
         | mental model as you do of your code, and from my experiments
         | does not have the ability to connect related but spatially
         | disconnected concepts across even small codebases which will
         | cause it to introduce bugs.
         | 
         | Asking it about these things sounds like it would result in
         | questionable, at best, responses.
        
           | zibzob wrote:
           | I see, that's what I was worried about. It would be really
           | helpful if it could answer high-level questions about a big
           | confusing codebase, but maybe it's not just a matter of
           | showing it the code and having it work.
        
             | skissane wrote:
             | ChatGPT has a published context window of 4096 tokens.
             | Although, I saw someone on Twitter saying the real figure,
             | based on experiments, was closer to 8192 tokens. [0] Still,
             | that's an obvious roadblock to "understanding" large code
             | bases - large code bases are too big to fit in its "short-
             | term memory", and at runtime its "long-term memory" is
             | effectively read-only. Some possible approaches:
             | 
             | (A) wait for future models that are planned to have much
             | longer contexts
             | 
             | (B) fine tune a model on this specific code base, so the
             | code base is part of the training data not the prompt
             | 
             | (C) Break the problem up into multiple invocations of the
             | model. Feed each source file in separately and ask it to
             | give a brief plain text summary of each. Then concatenate
             | those summaries and ask it questions about it. Still
             | probably not going to perform that well, but likely better
             | than just giving it a large code base directly
             | 
             | Another issue is that, even the best of us make mistakes
             | sometimes, but then we try the answer and see it doesn't
             | work (compilation error, we remembered the name of the
             | class wrong because there is no class by that name in the
             | source code, etc). OOTB, ChatGPT has no access to
             | compilers/etc so it can't validate its answers. If one gave
             | it access to an external system for doing that, it would
             | likely perform better.
             | 
             | [0] https://mobile.twitter.com/goodside/status/159887467420
             | 46187...
        
           | panarky wrote:
           | Saying a machine cannot understand the way humans understand
           | is like saying airplanes cannot fly the way birds fly.
        
             | Dudeman112 wrote:
             | Which is correct and a big reason for why early flight
             | machines had no chance at all of working
             | 
             | Of course, that doesn't tell you whether the machine
             | understanding will be useful or not
        
         | sandkoan wrote:
         | I'm actually building this very thing--shoot me an email at
         | govind <dot> gnanakumar <at> outlook <dot> com if you'd like to
         | be a beta tester.
        
       | xyz_ielh wrote:
       | [dead]
        
       | recuter wrote:
       | [flagged]
        
       | superkuh wrote:
       | This might be true for the type of business and institutional
       | uses that can operate under the extremely puritanical filters
       | that are bolted onto gpt3.5-turbo. But for most human person uses
       | the earlier text completion models like gtp3 davinci are
       | incomparibly better and more responsive. But also 10x as pricey.
       | Still, it's worth it compared to the lackluster and recalcitrant
       | non-output of gpt3.5-turbo.
       | 
       | I think over the next couple months most human people will switch
       | away from gpt3.5-turbo in openai's cloud to self-hosted LLM
       | weights quantized to run on consumer GPU (and even CPU), even if
       | they're not quite as smart.
        
         | abraxas wrote:
         | I have a hard time imagining anything that comes even close to
         | ChatGPT being able to run on consumer hardware in the next
         | couple of years
        
           | yieldcrv wrote:
           | I could perceive something like in 1 or less.
           | 
           | M3 Macbook with eGPU functionality restored in conjunction
           | with more efficient programming would mean having enough
           | memory available to all the processors. This would definitely
           | count as consumer hardware.
           | 
           | Custom built GPU-like devices with tons of RAM could become
           | vogue. Kind of like the Nvidia A100 but even more purpose
           | built for running LLMs or whatever models come next.
        
           | superkuh wrote:
           | That's what I thought 2 weeks ago. I figured it'd be ~5 years
           | before I could do anything at home. But already people have
           | the leaked facebook llama weights running on CPU w/under 32
           | GB of system ram doing a token a second or so.
        
             | riku_iki wrote:
             | that llama is likely much smaller than chatgtp
        
           | mattnewton wrote:
           | I can certainly imagine it after seeing
           | https://github.com/ggerganov/llama.cpp
           | 
           | Still a couple years out but moving way faster than I would
           | have expected.
        
             | abraxas wrote:
             | ChatGPT has 175B weights if I'm not mistaken. Llama 7B
             | would not be in any way comparable.
        
               | v64 wrote:
               | One finding in the LLaMA paper [1] is that our current
               | large models are undertrained. LLaMA with 13B params
               | outperforms GPT-3 175B (not ChatGPT), but an "instruct"
               | version of LLaMA was finetuned over the 65B model and did
               | quite well.
               | 
               | [1] https://arxiv.org/pdf/2302.13971.pdf
        
               | superkuh wrote:
               | For people who think the number of parameters determines
               | LLM coherence, well, that's a good rule of thumb. But
               | there's an optimal training set data size to parameters
               | count and gpt3 was trained on too little data. The LLM
               | coming out now trained on more data with fewer parameters
               | and achieve something close.
               | 
               | Sure, the 7 billion parameter can't do long outputs. But
               | the 13 billion one is not too bad. They're not a full
               | replacement by any means but for many use cases a local
               | service that is stupider is far preferable to a paid
               | cloud service.
        
           | KeplerBoy wrote:
           | It's crazy, but it seems to be happening already. Granted,
           | that's probably still a far-cry from Chat-GPT, but it seems
           | inevitable a few years down the line.
           | 
           | https://news.ycombinator.com/item?id=35100086
        
           | zamnos wrote:
           | Moore's law isn't quite beaten yet, so the (hypothetical
           | future) RTX 5090 and 6090 is gonna be _insane_. Combined with
           | software optimization and refinement of the techniques, along
           | with training != inference, means I think we 'll see
           | something better, runnable locally, in a couple of years. The
           | leaps and bounds Stable Diffusion has gone is insane.
           | Facebook's LLaMA is also seeing a similar growth from just
           | having the model available.
        
         | echelon wrote:
         | llm-nasty will find a way.
         | 
         | Stable Diffusion broke free of the shackles and was pushed
         | further than DALL-E could have ever hoped for.
         | 
         | Just wait. People's desires for LLMs to say spicy things and
         | not be controlled by a single party will make this happen yet
         | again. And they'll be more efficient and powerful. Half the
         | research happening is from "waifu" groups anyway, and they'll
         | stop at nothing.
        
           | isatty wrote:
           | Exactly, there is so much money to be made by generating porn
           | that it'll be done by this year.
        
             | yieldcrv wrote:
             | seriously. entertain the humans.
        
             | [deleted]
        
             | echelon wrote:
             | VCR, cable, internet, usenet, web, streaming, VR...
             | 
             | There are so many technologies that were propelled forward
             | because of it, not in spite of it.
             | 
             | Twitter, Reddit, Tumblr...
             | 
             | Tumblr learned a hard lesson when they tried to walk away.
        
           | zirgs wrote:
           | Feeding the AI lots of porn is one way to fix broken fingers.
        
         | Spivak wrote:
         | Or just keep using davinci because is it's also really cheap
         | all things considered. I was excited about getting 1/10th the
         | cost but also came to the same conclusion as you as turbo can't
         | actually _do_ anything. I could care less about getting it to
         | write porn or nazi propaganda but good lord it can't even write
         | code, do line edits or follow instructions more complicated
         | than simple call /response.
        
           | superkuh wrote:
           | My use case is IRC bots. If you just have the bot responding
           | to only a single line and not knowing any of the chat
           | history, yeah, it can be fairly cheap. But once you try to
           | start giving it short term memory by feeding in the prior
           | ~handful of lines you blow through that $18 of free credit in
           | a couple weeks. Something that costs $25/mo is not cheap for
           | a human person.
           | 
           | I am not happy with your implication that gpt3.5-turbo only
           | doesn't respond to "nazi" stuff and that my users are such
           | people. But I guess getting Godwin'd online isn't new. It
           | literally won't even respond to innocuous questions.
        
             | Spivak wrote:
             | What kinda volume are you pushing though because I also do
             | that and even have it ingest whole pdfs/word docs as
             | conversation context and I get charged like $3/mo on
             | average.
             | 
             | Edit: I'm literally agreeing with you and describing
             | innocuous questions that it doesn't respond to. I'm saying
             | that if all it refused to do was write hate and erotica it
             | would be fine and I would use it but the filter catches
             | things like code.
        
               | superkuh wrote:
               | With gpt3 davinci we were doing about ~30 requests for
               | text completion per hour (at peak activity) each having
               | ~11 lines of chat history (including up to 190 token
               | responses from davinci) which added up to about ~1000 to
               | 5000 tokens each. So 30*3000 at $0.0200/1000 tokens
               | equals a few dollars per day.
        
           | tracyhenry wrote:
           | what is an example that you can do with Davinci, but not
           | chatgpt? In my limited experience with prompting you can ask
           | chatgpt to do a lot of things
        
             | superkuh wrote:
             | gpt3.5-turbo fails the turing test due to it's constant
             | butt covering. Davinci can pass for a human. I am speaking
             | only of the API responses. The "chatgpt" web interface is
             | something different.
        
         | LeoPanthera wrote:
         | It doesn't matter that the older model will happily generate
         | text to make your grandmother blush. The usage policy
         | specifically says you can't do that. They even provide an API
         | endpoint for checking whether the input, and output, is allowed
         | or not.
         | 
         | There's nothing stopping you from ignoring it, except for the
         | certainty that OpenAI will simply block you.
        
         | faizshah wrote:
         | Its actually good for most human person uses too like writing
         | or learning. I've never encountered it refusing to do a task in
         | my actual work.
         | 
         | I would guess the risk to their brand vs the number of actual
         | applications of the unfiltered ai makes it an obvious trade
         | off.
         | 
         | I mean who turns off google safe search when writing an essay
         | or lyrics?
        
           | zamnos wrote:
           | Wait, there are people that aren't children that actually
           | have SafeSearch turned on as more than an accident? Not
           | trying to be insulting, I just genuinely have it turned off
           | in my settings and haven't noticed any of my search results
           | being particularly NSFW and assumed everyone else did too.
        
             | [deleted]
        
             | bilbo0s wrote:
             | If the default is off, most will have it off.
             | 
             | If the default is on, most will have it on.
             | 
             | All of which to say, no one cares, and google very likely
             | knows that. Google will only care if enough of their users
             | care. And they will probably operate in a fashion that
             | keeps the maximum number of their users in the "don't care"
             | camp. It's just business.
        
             | kayodelycaon wrote:
             | If the default is on, most people would have it on.
        
         | tracyhenry wrote:
         | A couple months might be too soon imho. But I hope that in 2-3
         | years there will be a model with similar performance but much
         | smaller size, small enough to run incredibly fast inference +
         | training on my laptop. OpenAI might need to rethink their moat
         | in case that happens.
         | 
         | Think about all the smart ML researchers in academia. They
         | can't afford training large models on large datasets, and their
         | decades of work is made obsolete by OpenAI's bruteforce
         | approach. They've got all the motivation in the world to work
         | on smaller models.
        
           | soulofmischief wrote:
           | I actually don't think that we will make significant
           | advancements in reducing model size before we make
           | significant advances in increasing available power and
           | compute.
           | 
           | One reason is that the pressure is still on for models to be
           | bigger and more power hungry, as many believe compute will
           | continue to be the deciding factor in model performance for
           | some time. It's not a coincidence that OpenAI's CEO, Sam
           | Altman, also runs a fusion energy r&d company.
        
             | flangola7 wrote:
             | But processing hardware has been seeing diminishing returns
             | for years. My CPU from 2013 is still doing what I need; a
             | 1993 processor in 2003 would have been useless.
             | 
             | Where do you see hardware improvements coming from?
        
       | alpark3 wrote:
       | Everyone loves to hate on OpenAI and talk about how they're
       | really ClosedAI and an evil corporation vying for power, but the
       | opposite way is also interesting to think about. I think it's
       | fair to say that majority of scientists at OpenAI wouldn't be
       | working there if they knew they were working for an evil
       | corporation. These are some of the brightest people on the
       | planet, yet I've only heard good things about OpenAI leadership,
       | especially Sam Altman, and their commitment to actually guiding
       | AI for the better.
       | 
       | I'm not saying that OpenAI is benevolent, but let's assume so for
       | the sake of argument. They definitely would need real-world
       | experience running commercial AI products, for the organizational
       | expertise as well as even more control over production of safe
       | and aligned AI technologies. A hypothetical strategy, then, would
       | be to a) get as much investment/cash as needed to continue
       | research productively (Microsoft investment?) b) with this cash,
       | do research but turn that research into real-world product as
       | fast as possible c) and price these products at a loss so that
       | not only are they the #1 product to use, other potentially
       | malevolent parties can't achieve liftoff to dig their own niche
       | into the market
       | 
       | I guess my point is that a company who truly believes that AI is
       | potentially a species-ending technology and requires incredible
       | levels of guidance may aim for the same market control and
       | dominance as a party that's just aiming for evil profit. Of
       | course, the road to hell is paved with good intentions and I'm on
       | the side of open source(yay Open Assistant), but it's
       | nevertheless interesting to think about.
        
         | Sol- wrote:
         | > and their commitment to actually guiding AI for the better
         | 
         | I think the Silicon Valley elite's definition of "for the
         | better" means "for the better for people like us". The
         | popularity of the longtermism and transhumanism cult among them
         | also suggests that they'd probably be fine with AI wiping out
         | much of humanity1, as long as it doesn't happen to them - after
         | all, they are the elite and the future of humanity, with the
         | billions of (AI-assisted) humans of that will exist!
         | 
         | And they'll think it's morally right too, because there's so
         | many utility units to be gained from their (and their
         | descendants') blessed existence.
         | 
         | (1 setting aside whether that's a realistic risk or not, we'll
         | see)
        
         | recuter wrote:
         | > These are some of the brightest people on the planet, yet
         | I've only heard good things about OpenAI leadership, especially
         | Sam Altman, and their commitment to actually guiding AI for the
         | better.
         | 
         | Hear hear. It ought to be remembered that there is nothing more
         | difficult to take in hand, more perilous to conduct, or more
         | uncertain in its success than to take the lead in the
         | introduction of a new order of things.
        
         | wpietri wrote:
         | > These are some of the brightest people on the planet, yet
         | I've only heard good things about OpenAI leadership
         | 
         | This is a deeply ahistorical take. Lots of technically bright
         | people have been party to all sorts of terrible things.
         | Don't say that he's hypocritical         Rather say that he's
         | apolitical         "Vunce ze rockets are up, who cares vere zey
         | come down         "Zats not mein department!" says Werner von
         | Braun
        
           | ben_w wrote:
           | While "smart people do terrible things" is an absolutely fair
           | point, it's also the kind of thing I hear AI researchers say,
           | even with similar references.
           | 
           | Sometimes they even say this example in the context of "why
           | human-level AI might doom us all".
        
           | masfuerte wrote:
           | I think you're agreeing. The "yet" implied a contrast.
        
         | edgyquant wrote:
         | I think it's safe to say people wouldn't be working for any
         | company if they thought it was evil, so your whole point is
         | moot.
        
           | soulofmischief wrote:
           | The conclusion derived from this argument, that there are no
           | evil companies, doesn't seem to match up with empirical data
        
           | scotty79 wrote:
           | > I think it's safe to say people wouldn't be working for any
           | company if they thought it was evil
           | 
           | Did you read what you wrote?
        
           | ben_w wrote:
           | Hah no.
           | 
           | Lots of people work for organisations they actively think are
           | evil because it's the best gig going; plenty of other people
           | find ways to justify how their particular organisation isn't
           | evil despite all it does so they can avoid the pain of
           | cognitive dissonance and keep getting paid.
           | 
           | My _current_ approval of OpenAI is conditional, not certain.
           | (I don 't work there, and I at least _hope_ I will be  "team-
           | think-carefully" rather than "team OpenAI can't possibly be
           | wrong because I like them").
        
         | al2o3cr wrote:
         | Similarly, if you feel the need to fart it COULD be a monkey
         | trying to escape - sure, it's been eggy gases every single time
         | before but THIS TIME COULD BE DIFFERENT!
         | 
         | Don't hold it in, the monkeys need your help!
        
         | croes wrote:
         | Ever read The Physicists from Durrenmatt?
         | 
         | Or let me quote Dr. Ian Malcolm:
         | 
         | "Your scientists were so preoccupied with whether they could,
         | they didn't stop to think if they should."
        
       | [deleted]
        
       | avereveard wrote:
       | yeah remember when a lot companies based themselves on the bing
       | search api and then the price increase 3x-10x depending on usage?
       | thanks, but no thanks.
        
       | [deleted]
        
       | vintermann wrote:
       | One kind of text generation AI it already makes obsolete, is
       | specialized translation models. It's no surprise it outdoes
       | Google Translate, that feels like it hasn't been updated in a
       | while. But it also outdoes Deepl now, and Deepl is good.
       | 
       | And it seems to handle translating from low-resource languages
       | extremely well. Into them, it's a bit harder to judge.
       | 
       | It handles translation between closely related languages such as
       | Swedish and Norwegian extremely well. Google Translate goes via
       | English and accumulates pointless errors.
        
       | DuckFeathers wrote:
       | The biggest problem with ChatGPT (and alternatives) is the risk
       | of being coopted for generating the content someone gets in
       | trouble for. Someone very important will get in BIG BIG trouble
       | and try to blame OpenAI for it... and the series of lawsuits that
       | will follow will kill them.
       | 
       | While other such models will be impacted, hopefully, there will
       | be significant variations in alternatives so that we don't lose
       | this technology over giant corporations trying to get out of
       | their trouble by suing their service providers.
       | 
       | There will also be companies that will use modified versions of
       | open source alternatives... to make them much more conservative
       | and cautious, so that they don't get in trouble. There will be
       | these variations that will be shared by certain industries.
       | 
       | So, while the generative AI is here to stay, there will be a LOT
       | of variations... and ChatGPT will have to change a lot if they
       | want to stay alive and relevant over time.
        
         | freedomben wrote:
         | would you consider OpenAI (in its current iteration) to be
         | conservative?
        
         | sebzim4500 wrote:
         | You may be right that some of the smaller AI players could be
         | overwhelmed by lawsuits but OpenAI has a nearly $2 trillion
         | company bankrolling them so they can hire every lawyer in the
         | US if necessary.
        
       | FuckShadowBans wrote:
       | [dead]
        
       | margorczynski wrote:
       | What's the catch? How do they plan to make money out of it? Or
       | maybe the plan is to use the massive amount of data gathered to
       | make it better for e.g. Bing search? Cut out the competition
       | before it has a chance to flourish?
       | 
       | Companies, especially giant publicly traded ones like MS (the de
       | facto owner of OpenAI) don't give out freebies.
        
         | m3kw9 wrote:
         | OpenAI wins by innovating faster than everyone because a lot of
         | these models inner workings are known and can be trained to
         | meet ChatGPTs metrics. so all they have to do is hire the best
         | and move faster, as long as they have on par or better, people
         | won't be switching
        
         | dragonwriter wrote:
         | > What's the catch?
         | 
         | The catch is its a tactic to discourage investment in competing
         | technologies, enabling OpenAI to build their lead to the point
         | it is insurmountable.
         | 
         | > How do they plan to make money out of it?
         | 
         | Altman's publicly-stated plan for making money from OpenAI is
         | (I'm completely serious) [0]:
         | 
         | (1) Develop Artificial General Intelligence under the control
         | of OpenAI.
         | 
         | (2) Direct the AGI to find a way to make a return for
         | investors.
         | 
         | [0] https://techcrunch.com/2019/05/18/sam-altmans-leap-of-
         | faith/
        
           | guiriduro wrote:
           | That idea might actually work. If a startup is a build-
           | measure-learn loop, then coming up with ballpark viable
           | ideas, devising experiments to test them and optimising for
           | traction/profit should be a cinch for AGI. So just train it
           | to build a business for itself.
        
         | overcast wrote:
         | The API is a paid service, like all the other APIs.
        
           | alpark3 wrote:
           | I believe his underlying assumption is that the API is so
           | cheap that there's no way they're making money off of it. Yes
           | it's paid, but doesn't matter if they're losing money on
           | every API call.
        
           | ianmcgowan wrote:
           | If they're selling below cost, it doesn't matter. When you're
           | selling a dollar for 90 cents, it's hard to make up for that
           | in volume.
        
             | scottLobster wrote:
             | It's not like there are individual units of ChatGPT. With
             | enough subscribers they could sell it for 1 cent per month
             | and profit.
        
               | supermatt wrote:
               | Not sure what you mean by "individual units" but the
               | suggestion is that it costs more than they charge. i.e
               | it's not profitable, and the more they sell the more they
               | lose.
        
               | scottLobster wrote:
               | My point was "making it up on volume" is largely
               | irrelevant when it comes to mass market web-apps.
               | 
               | Costs are relatively fixed outside of infrastructure, and
               | potential customers are any number up to and including
               | the internet-connected population of the world.
               | 
               | The marginal cost of a new subscription is way less than
               | they charge. The more they sell the less they lose, even
               | if they're still losing overall to gain market-share.
        
               | pixl97 wrote:
               | This depends on the compute power quantum stepping....
               | 
               | That is what is the upgrade cost to expand capacity as
               | new customers are added. If for example adding 1 million
               | new users requires $200,000k in hardware expenditure and
               | $20k in yearly power expenditure, but your first year
               | return on those customers is only going to be $50k,
               | you're in a massive money losing endeavor.
               | 
               | The point here is we really don't know the running and
               | upkeep costs of these models at this point.
        
         | chessgecko wrote:
         | People are speculating that gpt3.5 turbo is actually much
         | smaller and that they are very likely currently making a profit
         | on it. It seems likely just given how quickly some of the 3.5
         | turbo responses are from the api, and how much they push users
         | to it. I haven't seen any really compelling theories of how
         | they did it though, just the results...
        
           | ethbr0 wrote:
           | They wouldn't be the first business to have showroom halo
           | products to attract customers, who instead but more
           | profitable mass-market products. Auto industry 101.
        
         | deeviant wrote:
         | I don't know why everybody is asking themselves why they are
         | offering it so cheaply, it seems rather obvious:
         | 
         | 1. Get near every company to jump on the hype train and
         | integrate openai api into their processes.
         | 
         | 2. Get overwhelming market share.
         | 
         | 3. Slowly reduce costs by increasing model and computation
         | efficiency and raise prices.
         | 
         | 4. Profit.
        
           | sacred_numbers wrote:
           | Alternatively:
           | 
           | 1. Quickly reduce costs by increasing model and computation
           | efficiency.
           | 
           | 2. Massively reduce prices while still maintaining some gross
           | margin.
           | 
           | 3. Massively increase market size and take the vast majority
           | of market share.
           | 
           | 4. End up with a higher gross profit due to a much larger
           | market size despite decreasing prices and gross margins.
           | 
           | 5. Profit.
        
           | theturtletalks wrote:
           | Step 3 also includes raising prices once people have
           | integrated the API. Google Maps was the "easy" and "cheap"
           | way of integrating maps into apps until they got almost all
           | the market share and raised prices through the roof.
        
         | swatcoder wrote:
         | The wildly successful public buzz draws internal and external
         | money towards the project. Outsiders now see Microsoft as
         | freshly repositioned against Google, and OpenAI as a rising
         | rocket; internal budget is likewise drawn to related endeavors
         | because everybody wants to claim a piece of whatever hits big.
         | 
         | Meanwhile, yes, the preview provides both training data for the
         | tooling, which has engineering value in AI, and usage data into
         | how users think about this technology and what they intuitively
         | want to do with it, which helps guide future product
         | development.
         | 
         | Both these reasons are also why they're (1) being so careful to
         | avoid scandal, and (2) being very slow to clear up public
         | misconceptions.
         | 
         | An safe, excited public that's fully engaged with the tool
         | (even if misusing and misunderstanding it) is worth a ton of
         | money to them right now and so has plenty of justification to
         | absorb investment. It won't last forever, but a new innovation
         | door seems to have opened and we'll probably see this pattern a
         | lot for a while.
        
         | sebzim4500 wrote:
         | It's also possible they have found a way to run the model
         | extremely cheaply. To be fair, there has been many improvements
         | to transformer inference since they initally set their prices
         | (most notably flash attention), so if they were barely making a
         | profit back then they could still be making a profit now.
         | 
         | That's a big if, however, and no one really will give you
         | figures on exactly what this costs at scale. Especially since
         | we don't know for a fact how big GPT-3.5-turbo actually is.
        
         | waboremo wrote:
         | Yes your second guess is accurate. They will be changing
         | pricing down the line when enough of the market is captured and
         | competitors have been deterred. Most notably, Microsoft's
         | largest competitor: Google.
        
         | sourcecodeplz wrote:
         | There is plenty of money from the 2012 - 2020 meteoric period
         | that has not been spent yet. If I had plenty of money I would
         | bet on Microsoft and OpenAI, as I am sure others are doing
         | already. Thus they have enough to sustain this growth.
        
         | bakugo wrote:
         | They'll hike up the price by 10x once enough companies are
         | relying on it to do business.
        
           | politician wrote:
           | We'll see AWS step in at that point with their own product
           | offering.
        
         | nico wrote:
         | This is a market grab. They are moving fast to capture the
         | market. Being cheap allows them to capture the market faster.
         | 
         | The main customers won't be end users of ChatGPT directly, but
         | instead companies with a lot of data and documents that are
         | already integrating the apis with their systems.
         | 
         | Once companies have integrated their services with OpenAIs
         | apis, they are unlikely to switch in the future. Unless of
         | course something revolutionary happens again.
        
           | riku_iki wrote:
           | > their services with OpenAIs apis, they are unlikely to
           | switch in the future.
           | 
           | why is that? If competitor release better or cheaper LLM, it
           | is not that hard to switch API calls..
        
             | nico wrote:
             | Sure, that's the case if all your software does is make a
             | couple of api calls and you have very few stake holders.
             | 
             | But when you have built a big service around an external
             | api, you have thousands or millions of users and thousands
             | of employees - replacing an api is not just a big technical
             | project, it's also a huge internal political issue for the
             | organization to rally the necessary teams to make the
             | changes.
             | 
             | People hate change, they actively resist it. The current
             | environment is forcing companies to adapt and adopt the new
             | technologies. But once they've done it, they'll need an
             | even bigger reason to switch apis.
        
           | potatolicious wrote:
           | > _" Being cheap allows them to capture the market faster."_
           | 
           | I think it's worth remarking that this is IMO a smarter way
           | of using price to capture market than what we've seen in the
           | post decade (see: Uber, DoorDash) - in OpenAI's case there's
           | every reasonable expectation that they can drop their
           | operating costs well below the low prices they're offering,
           | so if they are running in the red the expectation of
           | temporariness is reasonable.
           | 
           | What was unreasonable about the past tech cycle is that a lot
           | of the expectations of cost reduction a) never panned out,
           | and b) if subjected to even slight scrutiny would never have
           | reasonably panned out.
           | 
           | OpenAI has direct line-of-sight to getting these models
           | _dramatically_ cheaper to run than now, and that 's a huge
           | benefit.
           | 
           | That said I remain a bit skeptical about the market overall
           | here - I think the tech here is legitimately groundbreaking,
           | but there are a few forces working against this as a
           | profitable product:
           | 
           | - Open source models and weights are catching up very
           | rapidly. If the secret sauce is sheer scale, this will be
           | replicated quickly (and IMO is happening). Do users need
           | _ChatGPT_ or do they need _any decently-sized LLM_?
           | 
           | - Productization seems like it will largely benefit incumbent
           | large players (see: Microsoft, Google) who can afford to tank
           | the operating costs _and_ additional R &D required on top to
           | productize. Those players are also most able to train their
           | own LLMs _and_ operate them directly, removing the need for a
           | third party provider.
           | 
           | It seems likely to me that this will break in three
           | directions (and likely a mixture of them):
           | 
           | - Big players train their own LLMs and operate them directly
           | on their own hardware, and do not do business with OpenAI at
           | any significant volume.
           | 
           | - Small players lean towards undifferentiated LLMs that are
           | open source and run on standard cloud configurations.
           | 
           | - Small players lean towards proprietary, but non-OpenAI
           | LLMs. There's no particular reason why GCP and AWS cannot
           | offer a similar product and undercut OpenAI.
        
       | boringg wrote:
       | Until they raise prices. Classic venture playbook here - get
       | everyone hooked on the product then raise rates.
       | 
       | Also depends how you calculate cost. If its simply $ or if you
       | are counting the externalities as 0.
        
       | sourcecodeplz wrote:
       | I don't think were going to reach winter before we can run our
       | own ChatGPT locally with mundane hardware.
        
       | vkou wrote:
       | And yet, this is ChatGPT's attempt at generating a college essay:
       | 
       | https://acoup.blog/2023/02/17/collections-on-chatgpt/
       | 
       | Looking at the actual essay it produced, I don't need to know
       | anything about Roman history to know that the essay sucks.
       | Looking at the professor's markup of the essay, it becomes very
       | clear that for someone who knows a lot about Roman history, the
       | essay sucks - a _lot_.
       | 
       | And it's not like it was prompted to write about an _esoteric_
       | topic! According to the grader, the essay made 38 factual claims,
       | of which 7 were correct, 7 were badly distorted, and 24 were
       | outright bullshit. According to both myself, and the grader, way
       | too much heavy lifting is done by vague, unsubstantiated, overly
       | broad statements, that don 't really get expanded on further in
       | the composition.
       | 
       | But yes, if we're looking to generate vapid, low-quality, low-
       | value content spam, ChatGPT is great, it will produce billions of
       | dollars of value for advertisers, and probably net negative value
       | for the people reading that drivel.
        
         | photochemsyn wrote:
         | What is you sequentially fed ChatGPT with samples of the course
         | professor's own writing, and then asked it to write an essay on
         | the subject of interest? As the article notes, optimization is
         | possible:
         | 
         | > "For example, high school and college students have been
         | using ChatGPT to cheat on essay writing. Since current
         | recognition of AI generated content by humans involve
         | identifying ChatGPT's signature overly-academic voice, it
         | wouldn't surprise me if some kids on TikTok figure out a system
         | prompt that allow generation such that it doesn't obviously
         | sound like ChatGPT and also avoid plagiarism detectors."
         | 
         | A decent student might go to the trouble of checking all the
         | factual claims produced in the essay in other sources, thus
         | essentially using ChatGPT to write a rough draft then spending
         | the time saved on checking facts and personalizing the style. I
         | don't even know if that would count as serious cheating,
         | although the overall structure of such essays would probably be
         | similar. Running 'regenerate response' a few times might help
         | with that issue, maybe even, 'restructure the essay in a novel
         | manner' or similar.
        
       | specproc wrote:
       | I don't agree it's cheap. For generation at fairly small scale,
       | sure, but generation is just the party trick. The real power for
       | my use case lies in how much better it seems to do at traditional
       | NLP tasks than an out-of-the-box model, with no further fiddling
       | and faffing required.
       | 
       | Say I've got a corpus of ~1m documents, each of 10+ paragraphs
       | and I want to run quote extraction on them (it does this
       | beautifully), vectorise them for similarity search, whatever.
       | This gets pretty expensive pretty fast.
        
         | andix wrote:
         | What's the alternative? Hiring humans to do the job for you?
         | Probably much more expensive.
        
         | avibhu wrote:
         | Tangential: you can finetune something like flan-ul2 to do
         | quote extraction using examples generated from chatgpt. If you
         | have a good enough GPU, it should help cut down costs
         | significantly
        
           | winddude wrote:
           | Don't they have in the ToS you aren't allowed to use outputs
           | for training downstream? Which is a little ridiculous,
           | considering it's ToS.
           | 
           | But yea, they cheap cost and lack of training is making me a
           | take a long hard look at how I'm implementing more
           | traditional NLP solutions.
        
           | specproc wrote:
           | Nice, that sounds like it's worth exploring. Much
           | appreciated.
           | 
           | Again though, it's the zero-effort part that's appealing. I'm
           | on a very small team and getting that to close to the same
           | standard will take time for a ham-fisted clod like myself.
           | Worth giving a shot all the same though, thanks again.
        
             | pfdietz wrote:
             | It's interesting what you can do with ChatGPT with few shot
             | learning. It generalizes at the drop of a hat, often
             | correctly.
        
             | leobg wrote:
             | The zero shot ability is convenient. But for tasks that you
             | need to get done millions of times, I'd much rather spend
             | $10 on GPU compute and maybe a day of training data
             | generation to train a T5 which I then "own".
             | 
             | Also, running your own specialized model locally can be
             | much faster than using someone's API.
        
         | Ultimatt wrote:
         | I suspect the author doesnt realise one request with hardly
         | anything returned is many hundreds if not thousands of
         | "tokens". It adds up very fast. Just some debug effort on a
         | nonsense demo learning project cost $5 in a couple of hours.
         | For maybe a hundred or so requests.
        
           | carlosdp wrote:
           | That's straight up not true, unless that "demo learning
           | project" is feeding GPT the entire Bible or something.
           | 
           | I have a project that uses davinci-003 (not even the cheaper
           | ChatGPT API) like _crazy_ and I don 't come close to paying
           | more than $30-120/month. With the ChatGPT API, it'll be 10x
           | less...
        
           | pharke wrote:
           | You could have saved some money by writing tests. How much
           | text were you sending at a time? I've been summarizing
           | multiple 500 word chunks per query in my app as well as
           | generating embeddings and haven't broken $10 over the course
           | of a couple weeks.
        
           | sebzim4500 wrote:
           | It is not possible to pay anywhere close to $5 for a hundred
           | requests, even if you used the max payload size every time.
           | 
           | Is it possible you had a bug that caused you to send far more
           | requests than you were intending to send? Or maybe you used
           | the older models which are 10x more expensive?
        
           | KyeRussell wrote:
           | I can understand making a mistake on the Internet, but to say
           | it with such snarky gusto is inexcusable.
           | 
           | I've been playing with davinci pretty extensively and the
           | only reason I've actually given OpenAI my credit card was
           | because they won't let you do any fine-tuning with their free
           | trial credit, or something like that. You're off by orders of
           | magnitude, ESPECIALLY with the new 3.5 model.
        
           | manmal wrote:
           | I used it for dozens of requests yesterday and that amounted
           | to less than 7 cents. I used MacGPT for that.
        
         | celestialcheese wrote:
         | 1000x this. Entity extraction from unstructured text with
         | zero/few-shot is fantastic.
         | 
         | I've got a use case where I need to extract model numbers from
         | text - these LLMs are so good at it with very little work.
        
         | manmal wrote:
         | I'd wager it could cost anywhere between 1-10k to do that,
         | which is a considerable amount of money. Might still be worth
         | it though? If the alternative is mechanical turk, that would
         | probably cost x1000-10000? Are there any ML alternatives that
         | reliably produce useful results?
        
       | ApolIllo wrote:
       | When will they raise prices?
        
         | nateburke wrote:
         | When the PR gains from widespread adoption no longer cover the
         | costs.
        
         | kkielhofner wrote:
         | As soon as there's a reliable base of foundational
         | services/orgs/products/startups built on top of it.
         | 
         | Especially with this edge, for now, it's Hotel California.
        
       | lamontcg wrote:
       | You think Google Search is polluted with AI written SEO'd trash
       | already, well just wait for what it is in store when the chatbots
       | attack whatever value is still contained in reddit-as-a-search-
       | engine...
        
         | Velc wrote:
         | They already are. This week I uncovered a ChatGPT bot farm
         | operating on reddit.
        
         | sourcecodeplz wrote:
         | Who uses Google anyway nowadays besides getting the address of
         | a website you don't remember (assuming you know about ChatGPT).
        
           | Toutouxc wrote:
           | I can mostly tell when webpages from my search results are
           | trying to bullshit me (and Kagi gives me some nice tools to
           | suppress the bullshitting kind), but with ChatGPT I have no
           | idea.
        
             | darepublic wrote:
             | Yes true if you trust the website then generally that trust
             | can extend to all its content. You are putting your faith
             | into the competence and consistency of a human being which
             | is generally more trustworthy than the hit or miss results
             | of a word predictor
        
       | osigurdson wrote:
       | I largely agree but I don't see how ChatGPT hits the same use
       | cases as a fine-tuned model. Prompts can only have 8K tokens so
       | any "in-prompt" fine tuning would have to be pretty limited. I'm
       | not certain that the lack of ChatGPT fine tuning will be a
       | permanent limitation however.
        
         | riku_iki wrote:
         | but they have API for fine tuning?..
        
           | osigurdson wrote:
           | They do, but not for gpt3.5 turbo (the ChatGPT model). See
           | the following link for details:
           | https://platform.openai.com/docs/guides/chat/is-fine-
           | tuning-...
        
             | riku_iki wrote:
             | probably this may change soon..
        
       | al2o3cr wrote:
       | AI's main accomplishment so far is rendering its human shills
       | indistinguishable from Markov generators.
        
       | cs702 wrote:
       | It will also make a _lot_ of simple machine-learning models
       | obsolete. It 's just not that obvious yet.
       | 
       | Imagine feeding a query akin to the one below to GPT4 (expected
       | to have a 50,000-token context), and then, to GPT5, GPT6, etc.:
       | query = f"The guidelines for approving or denying a loan are:
       | {guidelines}.                 Here are sample application that
       | were approved: {sample_approvals}.                 Here are
       | sample applications that were denied: {sample_denials}.
       | Please approve or deny the following loans: {loan_applications}.
       | Write a short note explaining your decision for every
       | application."            decisions = LLM(query)
       | 
       | Whether you like it or not, this kind of use of LLMs looks almost
       | inevitable, because it will give nontechnical execs something
       | they have always wanted: the ability to "read and understand" the
       | machine's "reasoning." They machine will give them what they have
       | always wanted: an explanation in plain English.
        
         | logifail wrote:
         | You've seen the movie The Big Short?
         | 
         | Someone is likely coding:
         | 
         | query = f"The guidelines for approving or denying a loan are:
         | {guidelines}. Here are sample application that were approved:
         | {sample_approvals}. Here are sample applications that were
         | denied: {sample_denials}. Please write a loan application which
         | is very likely to be approved. Provide necessary supporting
         | details.
        
           | cs702 wrote:
           | Yeah, that sort of thing looks inevitable too.
        
         | RC_ITR wrote:
         | I imagined it and my theoretical supervised fine tuning bills
         | are through the roof!
        
         | vinni2 wrote:
         | This would be a privacy nightmare. Banks would get into trouble
         | if they send customer data to openAI. Unless they host their
         | own LLM this is not yet practical.
        
           | KyeRussell wrote:
           | This is an entirely immaterial detail that could and would
           | easily be addressed. I'm just going to assume that OpenAI's
           | arm can be twisted wrt terms and conditions for Big Clients,
           | as is standard practice. But even if it couldn't be, I've got
           | no doubt that OpenAI will accept the literal shipping
           | containers of money from a bank in exchange for an on-prem
           | GPT-3 appliance.
        
         | andix wrote:
         | No, that probably won't work well. For such a task you need to
         | train your model with thousands of samples, way too much for a
         | simple prompt. But also you can't teach knowledge to a language
         | model.
         | 
         | The language model is trained for answering/completing text.
         | You can do some additional training, but it will only pick up
         | new words or new grammar. But it won't be able to learn how to
         | calculate or how to draw conclusions.
        
           | cs702 wrote:
           | Your understanding is _very_ outdated. Go take a look at some
           | of the things people are doing with LangChain to get a sense
           | of what 's possible today and what will likely be possible in
           | the very near future. LLMs are normally used in a zero-shot
           | setting, without any kind of fine-tuning.
        
         | loxias wrote:
         | > Write a short note explaining your decision for every
         | application
         | 
         | Is there any evidence or reason to suspect that this would
         | result in the desired effect? (explanations that faithfully
         | correspond to the specifics of the input data resulting in the
         | generated output)
         | 
         | I suspect the above prompt _would_ produce _some_ explanations.
         | I just don 't see anything tethering the explanations to the
         | inner workings of the LLM. It would make some very convincing
         | text that would convince a human... that would only be
         | connected to the decisions by coincidence. Just like when
         | ChatGPT hallucinates facts, internet access, etc. They look
         | extremely convincing, but are hallucinations.
         | 
         | In my unscientific experience, to the LLM, the "explanation"
         | would be just more generation to fit a pattern.
        
           | micromacrofoot wrote:
           | A vast amount of the world is built on "close enough" and
           | this is no different
        
           | cs702 wrote:
           | > Is there any evidence or reason to suspect that this would
           | result in the desired effect?
           | 
           | Yes.
           | 
           | There's evidence that you can get these models to write
           | chain-of-thought explanations that are consistent with the
           | instructions in the given text.
           | 
           | For example, take a look at the ReAct paper:
           | https://arxiv.org/abs/2210.03629
           | 
           | and some of the LangChain tutorials that use it:
           | 
           | https://langchain.readthedocs.io/en/latest/modules/agents/ge.
           | ..
           | 
           | https://langchain.readthedocs.io/en/latest/modules/agents/im.
           | ..
        
           | Al-Khwarizmi wrote:
           | Not to refute what you said, but what you describe is quite
           | similar to what we humans call rationalization, and it has
           | been argued (e.g. by Robert Zajonc) that most of the time we
           | make decisions intuitively and then seek a rationalization to
           | explain them.
           | 
           | Also, good luck with human explanations in the presence of
           | bias. No human is going to say that they refused a loan due
           | to the race or sex of the applicant.
        
             | pixl97 wrote:
             | Well no smart humans, but it turns out there are plenty of
             | dumb ones.
        
         | roflyear wrote:
         | Awful use of the language model.
        
         | jcoc611 wrote:
         | Probably should not fully automate this, but if you omit the
         | "approve or deny" part then you got yourself a nice system that
         | can pre-screen and surface statistical concerns with
         | applications. You can still have a human making the final
         | decisions
        
           | cs702 wrote:
           | Yes. In fact, I think that's how it will likely be used at
           | first :-)
        
         | jsemrau wrote:
         | I have implemented an AI for credit decisioning in 13 countries
         | on a multi-billion dollar portfolio. Here are my concerns about
         | this elegant yet ineffective prompt:
         | 
         | 1. LLMs in general are not build for quantitative analysis.
         | Loan-to-value, income ratios, etc are not supposed to be
         | calculated by such a model. Possible solution would be to
         | calculate this beforehand and provide it to the model or train
         | a submodel using a supervised learning approach to identify
         | good/bad
         | 
         | 2. Lending models are governed quarterly yet see relevant
         | cohort changes only after a period of time after credit
         | decision which can be many years. This prompt above does not
         | take this performance of the cohort into consideration
         | 
         | 3. Based on the governance companies adjust parameters and
         | models regularly to adjust to changes in the environment. I.e.,
         | a new car models comes out or the company is accessing a new
         | customer segment. This process could not be covered well with
         | this prompt since there would be no approvals/ denies for this
         | segment.
         | 
         | 4. Since transfer of personal-identififation data needs to be
         | consented, it would likely be necessary to host an LLM like
         | this internally or find a way to ensure there is no data
         | leakage from the provider to other users on the platform.
         | 
         | 5. Credit approval limits are not necessarily covered by this
         | proceess. I.e., the credit decisions is unclear but would work
         | with 5-10% more downpayment. Or the customer would be asked to
         | lower the loan value or find someone in the company who can
         | underwrite that loan volume. This person then has usually a
         | bunch of additional questions (liquidity risk, interest risk
         | ,etc) to ensure that the company is well protected and the
         | necessary compliance checks are adhered to.
         | 
         | 6. The discussions about this with regulators and auditors will
         | be entertaining.
         | 
         | Yet, I think it IS an elegant prompt which might provide some
         | insights.
        
           | cs702 wrote:
           | There's evidence that you can get LLMs to write chain-of-
           | thought explanations that are consistent with the
           | instructions in the given text, including quantitative data,
           | cohort performance, governance imperatives, qualitative
           | considerations, etc. The models can even be given directions
           | to write conditional approvals if necessary.
           | 
           | To get a sense of what is and will be possible, take a look
           | at the ReAct paper: https://arxiv.org/abs/2210.03629
           | 
           | and some of the LangChain tutorials that use it:
           | 
           | https://langchain.readthedocs.io/en/latest/modules/agents/ge.
           | ..
           | 
           | https://langchain.readthedocs.io/en/latest/modules/agents/im.
           | ..
        
         | mtlmtlmtlmtl wrote:
         | Yes, let's put LLM in charge of loan applications. Definitely
         | no financially devastating 2008-like slippery slope there.
         | 
         | It'll be fine.
        
           | thatwasunusual wrote:
           | Yes, let's put people in charge of loan applications.
           | Definitely no financially devastating 2008-like slippery
           | slope there.
           | 
           | It'll be fine.
        
           | sebzim4500 wrote:
           | You could check a random sample of them with expert humans to
           | ensure there isn't a systematic issue causing you to issue
           | large loans that you shouldn't be issuing.
           | 
           | I doubt regulators would be happy with this though,
           | especially since regulations are often a jobs program for
           | former employees of regulators.
        
             | psychphysic wrote:
             | You mean you could ask a LLM to look at a sample of the
             | loans and decide if there was a bias.
        
             | NBJack wrote:
             | The fun part is when a LLM hits that small probabiliy where
             | they decide to go 'offscript'. It can result in a
             | beautifully terrifying cascade of grammatically acceptable
             | nonsense, and joe fun it would be in a legal document. We
             | go from a loan for a home to a generous offer that includes
             | a unicorn, a purple dishwasher, a unicorn, and a few
             | dangling participles at the going market rate, all for the
             | low low rate of 555.555.1212. [END TOKEN]--- WASHINGTON,
             | D.C. President Trump today met with lawmakers to
        
               | sebzim4500 wrote:
               | I think the hope is that as LLMs get larger these issues
               | will go away. Certainly there are issues with GPT-2 that
               | completely went away when moving to larger models.
               | 
               | Honestly, I haven't even seen GPT-3.5-turbo exhibit this
               | behavior myself, although I am willing to believe it
               | could happen. Llama 7B, however, goes off-script
               | constantly.
        
               | cs702 wrote:
               | I laughed really hard -- _after_ trying to make sense of
               | your comment!
               | 
               | Thank you for posting this :-)
        
             | rafram wrote:
             | I don't think regulator nepotism is the main reason that
             | the authorities would be uncomfortable with loan decisions
             | being made by a system that definitionally reinforces
             | existing biases and is incapable of thought. It's just a
             | bad idea!
        
             | credit_guy wrote:
             | You don't need to check a random sample. You can have a
             | policy where every single loan application is checked by a
             | human, and you can add whatever affirmation is needed. It
             | will still increase the productivity of those loan officers
             | by a factor of 5. (Put it differently, banks would be able
             | to lay off 80% of their loan officers).
        
             | corbulo wrote:
             | This is the exact kind of dystopic thinking that is feared
             | with the use of AI.
             | 
             | "We regularly take randomized samples and have not found
             | error, your appeal has been denied."
             | 
             | I mean come on, ethics anybody??
        
               | sebzim4500 wrote:
               | So long as it is better than the thing it replaces, I
               | don't get the big deal.
        
           | sangnoir wrote:
           | Yep, can't wait for loan "hacks" like randomly name-dropping
           | job titles and institutions in the loan application. "Our pet
           | hamsters 'Stanford University' and 'Quantitative Analyst' are
           | looking forward to having more room"
        
             | KyeRussell wrote:
             | We call him little quanty tables.
        
           | [deleted]
        
       | sposeray wrote:
       | [dead]
        
       | virtualjst wrote:
       | Where is juniper notebook code for the chatgpt_generate function?
        
       | virtualjst wrote:
       | Where is the jupyter notebook code with the example
       | chatgpt_generate function?
        
         | minimaxir wrote:
         | That function is just a wrapper over the base
         | openai.ChatCompletion.create call from the documentation with
         | no changes:
         | https://twitter.com/minimaxir/status/1631044069483749377
        
       | jonatron wrote:
       | I pasted the article into a Markov Chain:
       | 
       | exist generative text unfortunately with the current recognition
       | of its creation which uses the chatgpt api which can confirm the
       | media has weirdly hyped the upcoming surge of ai generated
       | content its hard to keep things similar results without any
       | chatgpt to do much better signalto-noise.
       | 
       | --
       | 
       | Is ChatGPT just an improved Markov Chain?
        
         | bitL wrote:
         | RLHF uses Markov chains as its backbone, at least theoretically
         | (deep NN function approximations inside might override any
         | theoretical Markov chain effect though).
        
         | TechBro8615 wrote:
         | I guess if you squint, it kinda is, in the sense that it
         | generates one token at a time.
        
           | theGnuMe wrote:
           | It is in the sense that it is Markov chain with an 8k token
           | memory and the "MCMC step" is the DNN.
        
           | nerdponx wrote:
           | It's not a Markov chain because by definition a Markov chain
           | only looks at the previous word. ChatGPT looks at a long
           | sequence of previous words. But the general idea is still
           | broadly the same.
        
             | PartiallyTyped wrote:
             | That's not correct. In a Markov chain, the current state is
             | a sufficient characteristic of the future. For all intents
             | and purposes you can create a state with sufficiently long
             | history to look at a long sequence of words.
        
               | nerdponx wrote:
               | Also fair, but then the "current" state would _also_ be a
               | long window /sequence. Maybe that interpretation is valid
               | if you look at the activations inside the network, but I
               | wouldn't know about that.
        
               | PartiallyTyped wrote:
               | Yes, the state for both is a long window / sequence.
               | Under this view, for the transformer we do not need to
               | compute anything for the previous tokens as due to the
               | causal nature of the model, the tokens at [0, ... N-1]
               | are oblivious to the token N. For token N we can use the
               | previous computations since they do not change.
        
             | Accujack wrote:
             | Actually, it looks at all the _meanings_ of the tokens
             | within its window.
        
             | xwdv wrote:
             | Hence why you have to squint.
        
         | jfengel wrote:
         | To a first approximation, yes.
         | 
         | The second approximation has significant differences, but
         | that's an ok first pass at it.
        
           | skybrian wrote:
           | As a mathematical model, it's almost completely unhelpful,
           | like saying that all computers are technically state machines
           | because they have a finite amount of memory.
           | 
           | Treating every combination of 4k tokens as a separate state
           | with independent probabilities is useless for making
           | probability estimates.
           | 
           | Better to say that it's a stateless function that computes
           | probabilities for the next token and leave Markov out of it.
        
         | PartiallyTyped wrote:
         | "just" is doing a lot of heavy lifting.
         | 
         | ChatGPT needs a language model and a selection model. The
         | language model is a predictive model that given a state
         | generates tokens. For chatGPT it's a decoder model (meaning
         | auto-regressive / causal transformer). The state for the
         | language model is the fixed length window.
         | 
         | For a Markov chain, you need to define what "state" means. In
         | the simplest case you have a unigram where each next token is
         | completely independent of all previously seen tokens. You can
         | have a bi-gram model, where the next state is dependent on the
         | last token, or an n-gram model that uses the last N-1 tokens.
         | 
         | The problem with creating a markov chain with n-token state is
         | that it simply doesn't generalize at all.
         | 
         | The chain may be missing states and can't produce a probability
         | distribution. e.g. since we use a fixed window for the state,
         | our training data can have a state like "AA" that transitions
         | to B, thus the sentence is "AAB". The model however may keep
         | producing stuff, thus we need to get the new state, which is
         | "AB". If "AB" is out of the dataset, well... tough luck, you
         | need to improvise on how to deal with this. Approaches exist
         | but nowhere near as good of a performance as a basic RNN let
         | alone LSTMs and transformers.
        
         | ar9av wrote:
         | ChatGPT and Markov Chain are both text-generating models, but
         | they use different approaches and technologies. Markov Chain
         | generates text based on probabilities of word sequences in a
         | given text corpus, while ChatGPT is a neural network-based
         | model.
         | 
         | Compared to Markov Chain, ChatGPT is more advanced and capable
         | of producing more coherent and contextually relevant text. It
         | has a better understanding of language structure, grammar, and
         | meaning, and can generate longer and more complex texts.
        
       | fancyfredbot wrote:
       | OpenAI could probably make money offering the API for free at
       | this point - the data they are getting is so valuable for them in
       | building a competitive advantage in this space.
       | 
       | Once they know use cases for the model they can make sure they
       | are very good at those, and then they can consider hiking the
       | price.
        
         | eeegnu wrote:
         | Charging a small amount is more optimal since it mitigates API
         | spam without having to set a low rate limit. It also ties your
         | users to a financial id, which is (probably) harder to get in
         | bulk for nefarious purposes than just requiring a phone number
         | to sign up.
        
       | 1f60c wrote:
       | I would love a breakdown of how they made the incredible header
       | image (https://minimaxir.com/2023/03/new-chatgpt-
       | overlord/featured....)
        
         | preommr wrote:
         | It says it in the caption: Stable diffusion + controlnet.
         | 
         | If you haven't seen it already, controlnet has allowed for
         | massive improvements in addint constraints to generated images.
         | 
         | Here's an example of using a vector logo to make it semalessly
         | integrate it in different environments:
         | https://www.reddit.com/r/StableDiffusion/comments/11ku886/co...
        
         | minimaxir wrote:
         | Here's a few bonus OpenAI charcuterie:
         | https://twitter.com/minimaxir/status/1633635144249774082
         | 
         | 1. I used a ControlNet Colab from here based on SD 1.5 and the
         | original ControlNet app:
         | https://github.com/camenduru/controlnet-colab
         | 
         | 2. Screenshotted a B/W OpenAI logo from their website.
         | 
         | 3. Used the Canny adapter and the prompt: charcuterie board,
         | professional food photography, 8k hdr, delicious and vibrant
         | 
         | Now that ControlNet is in diffusers, my next project will be
         | creating an end-to-end workflow for these types of images:
         | https://www.reddit.com/r/StableDiffusion/comments/11bp30o/te...
        
         | aqme28 wrote:
         | It says it's made with ControlNet and Stable Diffusion.
         | Probably SD'd "charcuterie board" over a drawing of the logo.
        
       | napier wrote:
       | Yes last week, but Llama 65B is as of today running on an M1 so
       | yeah difficult to predict how centralised AI APIs will play out
       | six months from now:
       | https://twitter.com/lawrencecchen/status/1634507648824676353
       | 
       | Still expecting OAI to be able to leverage a flywheel effect as
       | they plough their recent funding injection into new foundation
       | models and other systems innovations but there's also going to be
       | increasing competition from other platform providers and also the
       | open source community boosted by competitors open sourcing /
       | leaking expensive to train model tech with the second order
       | function of diffusing wind from sales.
        
       | zirgs wrote:
       | One of the biggest drawbacks of ChatGPT is that OpenAI knows
       | everything that its users are doing with it. Every prompt and its
       | answer are being logged. Hackers might breach OpenAI systems and
       | leak its data.
       | 
       | If you're Rockstar that's working on GTA 7 then you'll propbably
       | want to keep all the AI written mission scripts, story ideas,
       | concept art and other stuff like that on your own servers.
        
         | altdataseller wrote:
         | Isnt this the case for a lot of web products? Hackers can hack
         | into Adobe and steal my prototypes. They can hack into my
         | Dropbox and steal my files. They can hack into my Asana project
         | and steal my roadmap
        
         | BoiledCabbage wrote:
         | > OpenAI retains API data for 30 days for abuse and misuse
         | monitoring purposes.
         | 
         | They just changed this. It is now only 30 day retention -
         | https://openai.com/policies/api-data-usage-policies
        
           | corbulo wrote:
           | Data retention is kind of meaningless in this context since
           | there's so many ways it is laundered/absorbed/analyzed while
           | not technically violating whatever legalese they use this
           | month.
        
           | throwawayapples wrote:
           | Does that only apply to API usage and not ChatGPT the web
           | app?
           | 
           | It would seem to, because the web app doesn't seem to expire
           | your old chats.
        
         | sebzim4500 wrote:
         | I agree with you, but I do think that people are overstating
         | the problem. It's no worse than sticking your data on the
         | cloud, and a huge portion of companies are doing that willingly
         | already.
        
         | simonw wrote:
         | ... which is one of the strongest arguments for being able to
         | run a large language model on your own hardware instead!
         | https://til.simonwillison.net/llms/llama-7b-m2
        
           | Karrot_Kream wrote:
           | Don't want to be vapid, but these are some cool guides! I
           | know how to run these models but want to link my friends to
           | guides to get started. Thanks!
        
       | simonw wrote:
       | Does anyone have a good feel for how likely it is that OpenAI
       | might be running it at this price to get companies hooked, with
       | plans to then raise the price later on once everyone is locked
       | in?
       | 
       | I'm personally much more excited about the LLaMA + llama.cpp
       | combo that finally brings GPT-3 class language models to personal
       | hardware. I wrote about why I think that represents a "Stable
       | Diffusion" moment for language models here:
       | https://simonwillison.net/2023/Mar/11/llama/
        
         | pixl97 wrote:
         | Depends how much competition ends up in this market. If there
         | is plenty of competition that gives good results at a similar
         | costs rising prices will be difficult. Now if it actually costs
         | far more to run than the API cost is currently, we'll see it go
         | up.
        
         | minimaxir wrote:
         | I pointed that out in the caveats since that happened with
         | Google Maps, but in practice I don't think it'll happen (or if
         | it happens it will only be a slight increase) since that would
         | seriously upset its users. Especially since the low price was
         | likely due to competition anyways.
         | 
         | In the case of Google Maps it was effectively a monopoly.
        
           | iampims wrote:
           | Being a monopoly is what OpenAI is aiming for.
        
             | minimaxir wrote:
             | Specifically in the case of Google Maps it was a de facto
             | monopoly, and thus has full control of pricing, due to the
             | lack of _good_ competitors (OpenStreetMap doesn 't count).
             | 
             | For LLMs, instead competition is very fierce which will
             | pressure down prices such as here with the ChatGPT API.
        
         | mirekrusin wrote:
         | They want to take the widest possible share, which atm, without
         | competition means bringing on people/companies that wouldn't
         | otherwise consider it.
         | 
         | The price will only go down when competition appears. They can
         | only slow it down with the cheapest possible offering (to put
         | market entry bar higher for competitors). They don't know what
         | competition will do, but they know if they move fast they'll
         | have very low chance of catching up anytime soon and that's all
         | that matters.
         | 
         | Competition will be interesting because interface is as simple
         | as it can be (easy to switch to different provider).
         | 
         | Providers can hook people though pre-training but I don't know
         | if it's possible to do dedicated pre-training on large models
         | like this. They may need to come up with something special for
         | that.
        
         | fullshark wrote:
         | It's very likely, they are in a race and that's the tech
         | playbook to win the market.
        
       | fareesh wrote:
       | How does one go about minimizing costs for certain situations?
       | 
       | For example if I share a db schema and then ask it to generate
       | some sql, I need to share that entire db schema for every single
       | question that follows, is that right?
       | 
       | Or is it possible for me to somehow pay and have it "retain" that
       | schema knowledge for all subsequent queries without having to
       | send the schema along with every single question?
        
         | mirekrusin wrote:
         | It's called fine tuning [0] but it's not yet available for
         | chatgpt, only older models.
         | 
         | [0] https://platform.openai.com/docs/guides/fine-tuning
        
           | fareesh wrote:
           | Yeah that seems to be the trouble - the pricing is great with
           | ChatGPT but not so much with the older ones
        
             | mirekrusin wrote:
             | That may be partially because it's not possible to do pre-
             | training on such a large model.
        
               | fareesh wrote:
               | On the web version is it doing that under the hood? I.e.
               | if I have a 10+ follow up conversation does it start from
               | the beginning each time?
        
               | mirekrusin wrote:
               | It keeps context of 8k tokens.
               | 
               | In pre-training you're using much more examples and
               | network is tuned around them.
        
       ___________________________________________________________________
       (page generated 2023-03-11 23:00 UTC)