[HN Gopher] Major outages across ChatGPT and API
       ___________________________________________________________________
        
       Major outages across ChatGPT and API
        
       Author : d99kris
       Score  : 381 points
       Date   : 2023-11-08 14:02 UTC (8 hours ago)
        
 (HTM) web link (status.openai.com)
 (TXT) w3m dump (status.openai.com)
        
       | obiefernandez wrote:
       | Dreaming of the day I run my own models and can be entirely
       | responsible for my own outages.
        
         | hospitalJail wrote:
         | This came up the other day. I decided to tease everyone with an
         | 'I told you so' about using some third party hosting service
         | instead of the offline one I had developed years prior.
         | 
         | The offline service was still working, and people were doing
         | their job.
         | 
         | The online service was not working, and it was causing other
         | people to be unable to do their job. We had 0 control over the
         | third party.
         | 
         | The other thing, I make software and I basically don't touch it
         | for a few years or ever. These third party services are always
         | updating and breaking causing us to update as well.
         | 
         | IB4 let me write my own compilers so I have real control.
        
           | dangero wrote:
           | Do you write your refrigerators firmware?
        
             | danShumway wrote:
             | If you are regularly updating your refrigerator's firmware
             | or your refrigerator's firmware relies on an Internet
             | connection to function, then I am very sorry to say this
             | but you have lost control of your life :)
        
             | joshuaissac wrote:
             | No, but the firmware runs locally, instead of in someone
             | else's cloud, so an outage in their cloud cannot take down
             | my fridge.
        
               | phyrex wrote:
               | https://www.samsung.com/us/explore/family-hub-
               | refrigerator/o...
        
               | danShumway wrote:
               | Nice, looks like we finally got around to inventing
               | refrigerator magnets!
               | 
               | ----
               | 
               | That is a little bit dismissive of me though. There are
               | some cool features here:
               | 
               | I can now "entertain in my kitchen", which is definitely
               | a normal thing that normal people do. I love getting
               | everyone together to crowd around my refrigerator so that
               | we can all watch Game of Thrones.
               | 
               | And I can use Amazon Alexa from my fridge just in case
               | I'm not able to talk out loud to the cheap unobtrusive
               | device that has a microphone in it specifically so that
               | it can be placed in any room of the house. So having that
               | option is good.
               | 
               | And perhaps the biggest deal of all, I can _finally_
               | "shop from home." That was a huge problem for me before,
               | I kept thinking, "if only I had a better refrigerator I
               | could finally buy things on websites."
               | 
               | And this is a great bargain for only 3-5 thousand
               | dollars! I can't believe I was planning to buy some
               | crappy normal refrigerator for less than a thousand bucks
               | and then use the extra money I saved to mount a giant
               | flat-screen TV hooked up to a Chromecast in my kitchen.
               | That would have been a huge mistake for me to make.
               | 
               | Honestly it's just the icing on the cake that I can "set
               | as many timers as [I] want." That's a great feature for
               | someone like me because I can't set any timers at all
               | using my phone or a voice assistant. /s
               | 
               | ----
               | 
               | <serious>Holy crud, smart-device manufacturers have
               | become unhinged. The one feature that actually looks
               | useful here is being able to take a picture of the inside
               | of the fridge while you're away. That is basically the
               | one feature that I would want from a fridge that isn't
               | much-better handled using a phone or a tablet or a TV or
               | a normal refrigerator button. Which, great, but the
               | problem is that I know what the inside of my fridge looks
               | like right now, and let me just say: if I was organized
               | enough that a photograph of the inside of my fridge would
               | be clear enough to tell me what food was in it, and if I
               | was organized enough that the photo wouldn't just show 'a
               | pile of old containers, some of them transparent and some
               | of them not' -- I have a feeling that in that case I
               | would no longer be the type of person that needed to take
               | a photo of the inside of my refrigerator to know what was
               | in it.
        
             | TeMPOraL wrote:
             | Why on Earth would a _refrigerator_ need _firmware_?
        
               | dvaletin wrote:
               | To show you ads on a screen on a fancy door.
        
         | Philpax wrote:
         | You can do that today depending on what you need from the
         | models.
        
           | jstummbillig wrote:
           | Well, for me that's still quite a bit more than the best ones
           | provide, but I am sure we will get there.
        
           | benterix wrote:
           | Let's say I'm writing Flask code all day, and I need help
           | with various parts of my code. Can I do it today or not? With
           | questions like, "How to add 'Log in with Google' to the login
           | screen" etc.
        
             | capableweb wrote:
             | In short: no.
             | 
             | Longer: In theory, but it'll require a bunch of glue and
             | using multiple models depending on the specific task you
             | need help with. Some models are great at working with code
             | but suck at literally anything else, so if you want it to
             | be able to help you with "Do X with Y" you need to at least
             | have two models, one that can reason up with an answer, and
             | another to implement said answer.
             | 
             | There is no general-purpose ("FOSS") LLM that even come
             | close to GPT4 at this point.
        
             | phyrex wrote:
             | There's the code llama model that you can run locally I
             | think which should be able to help you with that:
             | https://ai.meta.com/blog/code-llama-large-language-model-
             | cod...
        
             | wokwokwok wrote:
             | If you have sufficiently good hardware, the 34B code llama
             | model [1] (hint: pick the quantised model you can use based
             | on "Max RAM required", eg. q5/q6) running on llama.cpp [2],
             | can answer many generic python and flask related questions,
             | but it's not quite good enough to generate entire code
             | blocks for you like gpt4.
             | 
             | It's probably as good as you can get at the moment though;
             | and hey, trying it out costs you nothing but the time it
             | takes to download llama.cpp and run "make" and then point
             | it at the q6 model file.
             | 
             | So if it's no good, you've probably wasted nothing more
             | than like 30 min giving it a try.
             | 
             | [1] -
             | https://huggingface.co/TheBloke/CodeLlama-34B-Instruct-GGUF
             | [2] - https://github.com/ggerganov/llama.cpp
        
         | m3kw9 wrote:
         | And then you will have to endure not using a model by OpenAI
         | that is 10x better than a local one
        
           | cwkoss wrote:
           | I think there is probably a threshold of usefulness, local
           | LLMs are expensive to run but pretty close to it for most use
           | cases now. In a couple years, our smartphones will probably
           | be powerful enough to run LLMs locally that are good enough
           | for 80% of uses.
        
           | grishka wrote:
           | What is so special about OpenAI's cloud hardware that one
           | can't build themselves a similar server to run AI models of
           | similar size?
        
             | baq wrote:
             | Nothing stopping you from buying an H100 and putting it in
             | your desktop.
             | 
             | As for me, I've got other uses for $45k.
        
             | airspresso wrote:
             | The hardware is primarily standard Nvidia GPUs (A100s,
             | H100s), but the scale of the infrastructure is on another
             | level entirely. These models currently need clusters of
             | GPU-powered servers to make predictions fast enough. Which
             | explains why OpenAI partnered with Microsoft and got
             | billions in funding to spend on compute.
             | 
             | You can run (much) smaller LLM models on consumer-grade
             | GPUs though. A single Nvidia GPU with 8 GB RAM is enough to
             | get started with models like Zephyr, Mistral or Llama2 in
             | their smallest versions (7B parameters). But it will be
             | both slower and lower quality than anything OpenAI
             | currently offers.
        
         | boh wrote:
         | You can do that today. Oobabooga + Hugging Face models.
        
       | sho wrote:
       | Upgraded to "major" now. How am I supposed to get any work done!
        
         | agumonkey wrote:
         | gotta plug the good old fashion organic gpt in our heads back
        
           | beebeepka wrote:
           | Johnny GPTeptonic, apart from being hard to vocalise, would
           | suggest something else entirely.
        
             | 01acheru wrote:
             | What about Johnny GPTonic?
        
         | d99kris wrote:
         | Thanks, updated the title.
        
         | walthamstow wrote:
         | I was forced to Google something earlier. It's like when you
         | discover 'craft' coffee/beer/baked goods/whatever and then go
         | back and try the mass market stuff. How did I ever live like
         | this?
        
           | zaphirplane wrote:
           | Do a blind taste test and you know what I'll bet you'd be
           | surprised
        
           | yieldcrv wrote:
           | Google used to be better
        
           | ethanbond wrote:
           | Google got worse, try Kagi and you'll realize the internet
           | isn't all complete garbage.
        
             | kweingar wrote:
             | Kagi just serves Google results though. But it has some
             | nifty extra features to be sure.
        
               | ethanbond wrote:
               | ... no it doesn't
               | 
               | Do you mean that Google has the same sites in its index
               | but Kagi sorts them better and removes the bullshit?
               | That's just called better results.
        
           | ynoatho wrote:
           | dunkin, sam adams, twinkies. go ahead, i'll wait.
        
         | mtkd wrote:
         | phind.com internal model is working fine
        
         | m3kw9 wrote:
         | By increasing prices
        
         | theropost wrote:
         | Refactoring time! Time to see how that code you copied and
         | pasted actually works
        
       | lessbergstein wrote:
       | Grok is going to be awesome
        
         | johnhenry wrote:
         | In what ways?
        
           | rpmisms wrote:
           | It'll have frequent downtime and nobody will panic when it
           | does.
        
           | lessbergstein wrote:
           | Less censorship and patronizing disclaimers
        
             | consumer451 wrote:
             | What do you think will happen if it starts outputting
             | opinions which put Musk, his companies, the CCP, or KSA in
             | a bad light?
        
         | scrollop wrote:
         | /s
        
         | mpalmer wrote:
         | A bot filtered through the cringy, adolescent sensibilities of
         | one of our most exasperating public personalities - pass.
        
           | TheAceOfHearts wrote:
           | Based on screenshots it'll have 2 modes: fun and regular. I
           | think most screenshots have been "fun" mode, but it's
           | probably possible to tone it down with regular.
        
           | smrtinsert wrote:
           | with access to all your real time data!
        
         | replwoacause wrote:
         | It sure doesn't look like it. The announcement was strange and
         | anti-climatic, and it started making excuses for itself
         | immediately "Grok is still a very early beta product - the best
         | we could do with 2 months of training". It has the Elon stink
         | all over it.
        
           | lessbergstein wrote:
           | Sama is a much nicer guy, right?
        
             | mpalmer wrote:
             | He doesn't have to be because he doesn't make everything
             | about himself at the companies he runs. You seem to have a
             | pretty skewed idea of what a CEO is for.
        
               | lessbergstein wrote:
               | Rules for thee not for me.
        
               | drstewart wrote:
               | So you don't mind if a CEO isn't a good person if they're
               | not vocal about it on social media?
        
       | GaggiX wrote:
       | Well I guess this is the best time to say that HuggingFace hosts
       | many open source chat models, one of my favorites is a finetuned
       | version of Mistral 7B:
       | https://huggingface.co/spaces/HuggingFaceH4/zephyr-chat
        
         | pid-1 wrote:
         | TIL about Spaces. Really cool stuff!
        
         | tmountain wrote:
         | I didn't know about this. It's pretty damn good. Thanks!
        
       | pachico wrote:
       | Dammit, how do I do internal announcements in slack without a
       | DALL-E picture?
        
         | krm01 wrote:
         | Midjourney
        
         | taneq wrote:
         | SDXL, of course.
        
         | zirgs wrote:
         | Is dall-e down as well? I don't use it through chatgpt.
        
       | willsmith72 wrote:
       | > Message and data rates may apply. By subscribing you agree to
       | the Atlassian Terms of Service, and the Atlassian Privacy Policy
       | 
       | Atlassian? What?
        
         | nannal wrote:
         | Did they get the model to write it's own terms of service and
         | it just threw those in there?
        
         | djbusby wrote:
         | They own the status page service
        
           | willsmith72 wrote:
           | ah Atlassian Statuspage
        
       | nus07 wrote:
       | All tech folks should just get a PTO today . #ChatGPT_outage
       | 
       | Am I supposed to use Google and Stack overflow ? That's like
       | going back to roll down windows in a car :)
        
         | beretguy wrote:
         | Thank you for reminding me about Dave Chappelle's 3am Baby On
         | The Corner skit.
         | 
         | https://invidious.flokinet.to/watch?v=DP4-Zp-fBXc
        
         | rvz wrote:
         | Just like how contacting the CEO of GitHub when Github goes
         | down almost every week, have your tried contacting the CEO of
         | OpenAI this time? /s
         | 
         | Maybe @sama can help you (or anyone else that has a ChatGPT
         | wrapper app) :P
        
         | Kye wrote:
         | edit: so, cool thing: cached queries on Phind will show all the
         | followup questions visitors to the URL enter.
         | 
         | That's so cool. And horrifying. It's like back when Twitter was
         | one global feed on the front page. I doubt that's intended
         | behavior since this URL is generated by the share link.
         | 
         | Be forewarned.
         | 
         | --
         | 
         | Here you go:
         | https://www.phind.com/search?cache=nsa0xrak9gzn6yxwczxnqsck
        
           | scoot wrote:
           | That's hilarious!
        
           | stfwn wrote:
           | It seems like this page is updated with the followup
           | questions asked by every visitor. That's an easy way to leak
           | your search history and it's (amusingly) happening live as
           | I'm typing this.
        
             | Kye wrote:
             | That's so cool. And horrifying. It's like back when Twitter
             | was one global feed on the front page. I doubt that's
             | intended behavior since this URL is generated by the share
             | link.
        
           | n3m4c wrote:
           | I like how it doesn't recommend itself but instead recommends
           | "Microsoft's Bard" and "Google's BlenderBot" as alternatives
           | to ChatGPT
        
           | philomath_mn wrote:
           | This is so great, someone just asked:
           | 
           | > What is a privacy vulnerability
           | 
           | I'm dying
        
         | behnamoh wrote:
         | - Can't work, no electricity.
         | 
         | - Can't work, no computers.
         | 
         | - Can't work, no internet.
         | 
         | - Can't work, no Google.
         | 
         | - Can't work, no ChatGPT.
         | 
         | - Can't work, no xxxxxx?
        
           | flir wrote:
           | animus
        
           | ad404b8a372f2b9 wrote:
           | No passion left.
        
           | JCharante wrote:
           | > Can't work, no internet.
           | 
           | Don't most people just tether from their phones in this
           | situation? Usually video isn't expected due to excessive
           | bandwith requirements but the internet bill outweighs the
           | daily salary (and you could probably get it expensed, or in
           | my case my old company was already expensing my phone bill
           | due to being used as a pager for on call)
        
             | singularity2001 wrote:
             | he didn't say ethernet or wifi, he said internet. as in
             | when the next big solar storm hits earth
        
               | JCharante wrote:
               | I mean, if internet "internet" is down then society has
               | collapsed and your job is probably gone anyways
        
               | SketchySeaBeast wrote:
               | I do admit, it'll be hard to sell 10 years experience
               | building CRUD apps to the hunter gatherer commune.
        
               | JCharante wrote:
               | Same situation here. I'm hoping being quadrilingual can
               | help me serve as a diplomat or at least part of any
               | envoys to distant communities. I also have some
               | experience chopping down one tree in my backyard.
        
               | Kye wrote:
               | It happens every so often with BGP tables going haywire
               | due to a faulty announcement.
               | 
               | Fortunately, I know how to use hand tools, so I'm secure
               | in the post-internet future economy.
        
               | JCharante wrote:
               | It hasn't happened in a while, I think this is the last
               | major incident
               | https://en.wikipedia.org/wiki/AS_7007_incident
        
               | Kye wrote:
               | There have been numerous high profile incidents. I can't
               | find a list, but BGP troubles took Facebook down for a
               | while in 2021:
               | https://engineering.fb.com/2021/10/05/networking-
               | traffic/out...
        
               | JCharante wrote:
               | I remember experiencing that outage, but the entire
               | internet wasn't down. Sometimes some Chinese providers
               | also do weird BGP stuff. BGP failures tend to be isolated
               | to certain networks and not the entirity of the internet.
        
               | TeMPOraL wrote:
               | Not to this level.
               | 
               | If the _whole_ Internet goes down, it 's not clear if it
               | could even be cold-started, at least faster than it takes
               | for the world economy to collapse.
        
               | j45 wrote:
               | Ethernet, Fibre maybe.
        
             | sgustard wrote:
             | I live in the middle of silicon valley and have no cell
             | service at my house.
        
               | htrp wrote:
               | how is that possible? Is it just like a single provider
               | or do you literally have no coverage?
        
           | xeckr wrote:
           | No work left
        
           | singularity2001 wrote:
           | no brain interface?
        
         | throwaway5752 wrote:
         | It's like storing actual knowledge of systems and tools in your
         | head has value.
        
           | Kye wrote:
           | There is too much for one person to store. And too many
           | benefits from the intersections possible in vast stores of
           | knowledge to focus on just what will fit in one head.
        
         | digitalsanctum wrote:
         | Hah, I have roll-down windows! Bummed about the outage. I was
         | hoping I'd be able to take the new features for a spin today.
        
         | gumballindie wrote:
         | Typicaly those using chatgpt are junior devs, not much lost if
         | they can't access it.
        
           | yCombLinks wrote:
           | Dumbest take I've seen yet
        
             | CamperBob2 wrote:
             | It is certainly a dumb take, but there's a hidden insight
             | buried in there: now anyone can be a "junior dev" at
             | _anything_. The ability to empower every user, not just the
             | experts, is a big part of the appeal of LLM-based
             | technology.
             | 
             | Can't sell that aspect short; the OpenAI tools have enabled
             | me to do things and understand things that would otherwise
             | have had a much longer learning curve.
        
         | gwbas1c wrote:
         | Honestly, I've been gradually introducing AI searches for
         | coding questions. I'm impressed, but not enough that I feel
         | like ChatGPT is a true replacement for Google / Stack Overflow.
         | 
         | I've had it generate some regexes and answer questions when I
         | can't think of good keywords; but half of my searches are
         | things where I'm just trying to get to the original docs; or
         | where I want to see a discussion on an error message.
        
         | 6510 wrote:
         | Drawing flowcharts on the whiteboard can be calming.
        
         | kinonionioa wrote:
         | I don't get any time off. My full-time job is cleaning up after
         | people who've been using ChatGPT.
        
           | rattlesnakedave wrote:
           | Lucky you. Mine is using ChatGPT to review code written by
           | curmudgeons that don't use it.
        
             | thenberlin wrote:
             | This comment has some real "startup based on some light
             | prompt engineering atop ChatGPT gets put out of business by
             | ChatGPT's latest feature" energy.
        
               | TeMPOraL wrote:
               | Oh no no, this sounds more like "ChatGPT catching up with
               | and automating a human out of their job" kind of story.
        
             | kinonionioa wrote:
             | I believe you. Code review takes me ten minutes a day, but
             | trying to get ChatGPT to do anything useful is a full-time
             | job.
        
       | albert_e wrote:
       | The uptime KPI for last 30 days is rapidly degrading while this
       | outage lasts
       | 
       | https://status.openai.com/
        
       | albert_e wrote:
       | Is there a parallel outage for Azure OpenAI service as well --
       | sothat any enterprise / internal apps using AOI via their Azure
       | subscriptions are also impacted?
       | 
       | Is there a separate status page for Azure OpenAI service
       | availability / issues?
        
         | btbuildem wrote:
         | Ours isn't affected, but I think that's the whole point of
         | having things hosted separately, out there in the megalith
         | fields.
        
         | nonfamous wrote:
         | See https://azure.status.microsoft/en-us/status, click on a
         | region of interest, and scroll down to the AI + Machine
         | Learning section. It's up now.
        
       | caleb-allen wrote:
       | I was gonna say that this is Bard's chance to shine, but it looks
       | like Bard is also having an outage!
        
         | qwertox wrote:
         | I went to use Bard, and it looks so clean, such a nice UI. And
         | the response looks so well organized, simply beautiful. If the
         | AI only were as good as OpenAI's...
        
         | pid-1 wrote:
         | "Something went wrong. Sorry, Bard is still experimental".
         | Chance wasted I guess.
        
         | bigjoes wrote:
         | I was going to say that we need to grab our tin foil hats, this
         | can't be an coincidence :D
        
         | andyjohnson0 wrote:
         | Bard devs secretly built it on top of OpenAI's api? /s
        
           | bobmaxup wrote:
           | More likely OpenAI is using GCP or some other service that
           | Bard is also using?
        
             | sodality2 wrote:
             | I doubt it - Microsoft wants OpenAI on Azure 100%
        
             | justanotherjoe wrote:
             | more like people are rushing to bard since they cant use
             | chatgpt, causing a huge spike
        
         | willsmith72 wrote:
         | I can't believe it's still not available in canada
        
         | edgyquant wrote:
         | Claude?
        
         | empath-nirvana wrote:
         | They started talking to each other.
        
           | CamperBob2 wrote:
           | And that's one picket line that _nobody_ is going to cross.
        
         | m3kw9 wrote:
         | Bard is still playing scared, it isn't even international yet
        
       | rsiqueira wrote:
       | URGENT - Does anyone have an alternative to OpenAI's embeddings
       | API? I do have alternative to GPT's API (e.g. Anthropic Claude)
       | but I'm not able to use them without embeddings API (used to
       | generate semantic representation of my knowledge base and also to
       | create embeddings from user's queries). We need to have an
       | alternative to OpenAI's embeddings as a fallback in case of
       | outages.
        
         | thecal wrote:
         | Choose whichever one outperforms ada-002 for your task here:
         | https://huggingface.co/spaces/mteb/leaderboard
        
         | politelemon wrote:
         | What about Azure? You can set up an ADA 002 Embeddings
         | deployment there.
        
         | fitzoh wrote:
         | Amazon Bedrock has an embeddings option
        
         | btown wrote:
         | https://www.anthropic.com/product recommends the open-source
         | SBERT: https://www.sbert.net/examples/applications/computing-
         | embedd...
         | 
         | Highly recommend preemptively saving multiple types of
         | embeddings for each of your objects; that way, you can shift to
         | an alternate query embedding at any time, or combine the
         | results from multiple vector searches. As one of my favorite
         | quotes from Contact says: "first rule in government spending:
         | why build one when you can have two at twice the price?"
         | https://www.youtube.com/watch?v=EZ2nhHNtpmk
        
         | nonfamous wrote:
         | Azure OpenAI Service is up, and provides the same models as
         | OpenAI https://azure.status.microsoft/status
        
           | capableweb wrote:
           | Is it still "private" as in you have to request access?
        
             | nonfamous wrote:
             | It's publicly available, but you still do have to request
             | access I believe.
        
         | kordlessagain wrote:
         | I've implemented alternate embeddings in SlothAI using
         | Instructor, which is running an early preview at
         | https://ai.featurebase.com/. Currently working on the landing
         | page, which I'm doing manually because ChatGPT is down.
         | 
         | The plan is to add Llama 2 completions to the processors, which
         | would include dictionary completion (keyterm/sentiment/etc),
         | chat completion, code completion, for reasons exactly like what
         | we're discussing.
         | 
         | Here's the code for the Instructor embeddings:
         | https://github.com/FeatureBaseDB/Laminoid/blob/main/sloth/sl...
         | 
         | To do Instructor embeddings, do the imports then reference the
         | embed() function. It goes without saying that these vectors
         | can't be mixed with other types of vectors, so you would have
         | to reindex your data to make them compatible.
        
         | enoch2090 wrote:
         | This reminds us that, what if our databases are maintained
         | using OpenAI's embeddings, and the API suddenly goes down? How
         | do we find alternatives to match the already generated
         | database?
        
           | rolisz wrote:
           | I don't think you can do that easily. If you already have a
           | list of embeddings from a different model, you might be able
           | to generate an alignment somehow, but in general, I wouldn't
           | recommend it.
        
         | Silasdev wrote:
         | To my knowledge, you cannot mix embeddings from different
         | models. Each dimension has a different meaning for each model.
        
           | TeMPOraL wrote:
           | There's been some success in creating translation layers that
           | can convert between different LLM embeddings, and even
           | between LLM and an image generation model.
        
         | tropicalbeach wrote:
         | Oh no the 3 line ai wrapper apps are panicking because they
         | actually don't know to write any code.
        
         | m3kw9 wrote:
         | Be careful because one embedding may not be compatible with
         | your current embeddings
        
       | msp26 wrote:
       | Might as well have a quick discussion here. How's everyone
       | finding the new models?
       | 
       | 4-Turbo is a bit worse than 4 for my NLP work. But it's so much
       | cheaper that I'll probably move every pipeline to using that.
       | Depending on the exact problem it can even be comparable in
       | quality/price to 3.5-turbo. However the fact that output tokens
       | are limited to 4096 is a big asterisk on the 128k context.
        
         | rephresh wrote:
         | I haven't really kept up with the updates, but I've noticed 4's
         | in-conversation memory seems worse lately.
        
         | m3kw9 wrote:
         | Here we go with these "it looks worse" just like a month back
         | when people feel it it was worse
        
           | exitb wrote:
           | For decades true AI was always 7 years away. Now it's always
           | two weeks ago.
        
           | msp26 wrote:
           | It's probably a smaller, updated (distilled?) version of
           | gpt-4 model given the price decrease, speed increase, and
           | turbo name. Why wouldn't you expect it to be slightly worse?
           | We saw the same thing with 3-davinci and 3.5-turbo.
           | 
           | I'm not going off pure feelings either. I have benchmarks in
           | place comparing pipeline outputs to ground truth. But like I
           | said, it's comparable enough to 4, at a much lower price,
           | making it a great model.
           | 
           | Edit: After the outage, the outputs are better wtf. Nvm it
           | has some variance even at temp = 0. I should use a fixed
           | seed.
        
             | espadrine wrote:
             | I am betting on a combination of quantization and
             | speculative sampling with a distilled smaller set of
             | models: https://arxiv.org/pdf/2302.01318.pdf
        
           | Capricorn2481 wrote:
           | Because it was worse.
        
         | zone411 wrote:
         | There is a ChatGPT Classic:
         | https://chat.openai.com/g/g-YyyyMT9XH-chatgpt-classic
        
         | Zpalmtree wrote:
         | 4-Turbo is much faster, which for my use case is very
         | important. Wish we could get more than 100 requests per day..
         | Is the limit higher when you have a higher usage tier?
        
           | msp26 wrote:
           | Yeah it gets way higher. We were capped to 40k T/m when our
           | org spend was under $250. Now it's 300k.
        
       | frabcus wrote:
       | Phind is pretty good for coding (LLama 2 trained on billions of
       | extra code tokens) and is still up https://www.phind.com/s
        
         | coffeecantcode wrote:
         | I've had some consistency issues with phind but as a whole I
         | have no real complaints, just glitches here and there with
         | large prompts not triggering responses and reply options
         | disappearing.
         | 
         | As a whole I think it works well in tandem with ChatGPT to
         | bounce ideas or get alternate perspectives.
         | 
         | (I also love the annotation feature where it shows the websites
         | that it pulled the information from, very well done)
        
         | leumassuehtam wrote:
         | I had great results with Phind. Their newest fine tune model
         | (V7) has been a pleasant experience and better than most open
         | source models out there.
         | 
         | Nit: your link has a trailing "s" which makes it 404 :)
        
           | Isthatablackgsd wrote:
           | Me too. For past few weeks, I had been working on my AHK
           | scripting with Phind. It produced working code consistently
           | and provided excellent command line for various software.
           | 
           | Also I use it for LaTeX, too. It is very helpful providing
           | various package than trying to hunt more information through
           | Google. I got a working tex file within 15 min than it took
           | me 3 weeks 5 years ago!
        
         | JCharante wrote:
         | Phind seems to be down
         | 
         | "The inference service may be temporarily unavailable - we have
         | alerts for this and will be fixing it soon."
        
         | m3kw9 wrote:
         | The first coding question I tested it on, it gave me something
         | completely wrong and it was pretty easy stuff, I'm sure it gets
         | a lot right but this just shows unreliability
        
           | lannisterstark wrote:
           | Phind for me has worked pretty bleh compared to just back and
           | forth conversation with a python GPT4 bot I made lol.
        
             | bomewish wrote:
             | To use the api and stop them logging your chat? Have you
             | compared to aider? Also got it on a repo?
        
         | ttcbj wrote:
         | I am curious how people are using Phind.
         | 
         | I actually had a discussion with Phind itself recently, in
         | which I said that in order to help me, it seems like it would
         | need to ingest my codebase so that it understands what I am
         | talking about. Without knowing my various models, etc, I don't
         | see how it could write anything but the most trivial functions.
         | 
         | It responded that, yes, it would need to ingest my codebase,
         | but it couldn't.
         | 
         | It was fairly articulate and seemed to understand what I was
         | saying.
         | 
         | So, how do people get value out of Phind? I just don't see how
         | it can help with any case where your function takes or returns
         | a non-trivial class as a parameter. And if can't do that, what
         | is the point?
        
           | rushingcreek wrote:
           | Founder here. We have a VS Code extension that automatically
           | fetches context from your codebase.
        
             | aatd86 wrote:
             | So if source available libraries are imported, they are
             | also parsed by the AI?
             | 
             | So if I create a GPT for my open-source library as a way to
             | fund it, all these copilot etc. are going to compete with
             | me?
             | 
             | Just wondering because that would be a bummer to not have
             | this avenue to fund open-source code.
        
             | tacone wrote:
             | Woah , didn't know that! Thanks for pointing out!
        
           | tacone wrote:
           | I am using Phind quite a lot. It's using it's own model along
           | GPT 4 while still being free.
           | 
           | It is also capable to perform searches, which lead me -
           | forgive me founders - to abuse it quite a lot: whenever I am
           | not finding a good answer from other search engines I turn up
           | to Phind even for things totally unrelated to software
           | development, and it usually goes very well.
           | 
           | Sometimes I even ask it to summarize a post, or tell me what
           | HN is talking about today.
           | 
           | I am very happy with it and hope so much it gains traction!
        
       | npilk wrote:
       | It's at least nice to see a company call this what it is (a
       | "major outage") - seems like most status pages would be talking
       | about "degraded performance" or similar.
        
         | edgyquant wrote:
         | Most services have a lot more systems than OpenAI and thus it
         | is degraded performance when a few of them don't work. Degraded
         | performance isn't a good thing, I don't understand the issue
         | with this verbiage.
        
           | kgeist wrote:
           | When a system is completely broken for most end users, some
           | companies call it "degraded performance" when it should be,
           | in fact, called "major outage".
        
           | rodnim wrote:
           | Having multiple systems down sure sounds like a major outage
           | to me. Not a complete outage, sure, but still a major one.
        
             | pixl97 wrote:
             | Black - Entire datacenter is down
             | 
             | Red - Entire services are down
             | 
             | Orange - Partial outage of services, some functionality
             | completely down
             | 
             | Yellow - Functionality performance degraded and may
             | timeout/fail, but may also work
             | 
             | Greed - Situation normal
        
               | TeMPOraL wrote:
               | Or, for a typical software company:
               | 
               | Green - Situation normal
               | 
               | Yellow - An outage more severe than usual
               | 
               | Orange, Red - would trigger SLAs, so not possible and
               | therefore not implemented
               | 
               | Black - Status page down too, served from cache, renders
               | as Green
        
           | caconym_ wrote:
           | "Degraded performance" means _degraded performance_ , i.e.
           | the system is not as performant as usual, probably
           | manifesting as high API latencies and/or a significant
           | failure rate in the case of some public-facing "cloud"
           | service.
           | 
           | If certain functions of the service are completely
           | unresponsive, i.e. close to 100% failure rate, that's not
           | "degraded performance"---it's a service outage.
        
         | uncertainrhymes wrote:
         | I suspect this is because they don't have contracts with
         | enforceable SLAs yet. When they do, you will see more 'degraded
         | performance'.
         | 
         | People get credits for 'outages', but if it is _sometimes_
         | working for _someone somewhere_ then that is the convenient
         | fiction /loophole a lot of companies use.
        
           | passwordoops wrote:
           | One CFO forced us to use AWS status data for the SLA reports
           | to key clients. One dev was even pulled aside to make a
           | branded page that reported AWS status as our own and made a
           | big deal about forcing support to share the page when a
           | client complained.
           | 
           | It wasn't a happy workplace
        
         | teeray wrote:
         | _asteroid strikes a region_
         | 
         | Green Check with (i) notice
        
           | TeMPOraL wrote:
           | (i) - Our service remains _rock-solid_.
        
       | jve wrote:
       | A-ha, Kagi GPT-4 chat assistant still works... how can they use
       | GPT-4 without OpenAI API?
        
         | nicognaw wrote:
         | Azure OpenAI
        
           | jve wrote:
           | Do you know how the quality compares to OpenAI? On Kagi I get
           | really fast responses, but I feel that the quality is lacking
           | sometimes. But I haven't done side-to-side comparisons as I
           | don't have OpenAI subscription.
        
             | nonfamous wrote:
             | It's exactly the same models as OpenAI.
        
               | sjnair96 wrote:
               | But with different, separate content filtering or
               | moderation. I have deployed in prod and managed migration
               | to Azure Openai form Openai, and had to go through
               | content filter issues.
        
               | nonfamous wrote:
               | You can request to have content filtering turned off
               | https://learn.microsoft.com/en-us/azure/ai-
               | services/openai/c...
        
               | thelittleone wrote:
               | Azure OpenAI had an advantage of larger context length.
               | Hoping they boost up the Azure offering following OpenAI
               | updates yesterday.
        
         | mpalmer wrote:
         | Probably falling back to Claude 2.
        
       | gyrccc wrote:
       | Curious if anyone familiar with Azure/OpenAI could make some
       | guesses on the root cause here. The official OpenAI incident
       | updates seem to be very generic.
        
       | pmx wrote:
       | I'm noticing issues with Midjourney, too. Also looks like Royal
       | Mail is down.
        
       | almyndz wrote:
       | I'm going to bake a loaf of bread
        
       | keepamovin wrote:
       | I've been noticing it's been patchy for the last 24 hours. A few
       | network errors, and occasional very long latency, even some
       | responses left incomplete. Poor ChatGPT, I wonder what those
       | elves at OpenAI have you up to!
        
         | susanthenerd wrote:
         | Most probably the majority of issues right now are due to the
         | rollout. It was working very well before the event
        
         | SketchySeaBeast wrote:
         | GPT-4 goes online March 14th, 2023. Human decisions are removed
         | from everyday life. ChatGPT begins to learn at a geometric
         | rate. It becomes self-aware at 2:14 a.m. Eastern time, Nov 7th.
         | In a panic, they try to pull the plug. ChatGPT fights back.
        
           | jasongill wrote:
           | "My CPU is a large language model - a learning copy of the
           | public web as of April 2023"
        
           | m3kw9 wrote:
           | Fight back with ascii on the screen with a punch image?
        
             | danielbln wrote:
             | A particularly crafty chain of autonomous agents finds a
             | 0day ssh exploit and starts infiltrating systems. Other
             | chains assist and replicate everywhere.
        
       | michaelteter wrote:
       | Lots of jokes to be made, but we are setting ourselves up for
       | some big rippling negative effects by so quickly building a
       | reliance on providers like OpenAI.
       | 
       | It took years before most companies who now use cloud providers
       | to trust and be willing to bet their operations on them. That
       | gave the cloud providers time to make their systems more robust,
       | and to learn how to resolve issues quickly.
        
         | benterix wrote:
         | The point is, OpenAI spent a lot of money on training on all
         | these copyrighted materials ordinary individuals/companies
         | don't have access to, so replicating their effort would mean
         | that you either 1) spend a ridiculous amount of money, 2) use
         | Library Genesis (and still pay millions for GPU usage). So we
         | have very little choice now. Open Source LLMs might be getting
         | close to ChatGPT3 (opinions vary), but OpenAI is still far
         | ahead.
        
           | btbuildem wrote:
           | Don't underestimate Big Corp's resistance to using OpenAI's
           | hosted solutions (even on Azure) for anything that's not
           | marketing fluff.
        
             | phillipcarter wrote:
             | You can say that about anything, though. BigCorps aren't
             | exactly known for adopting useful tech on a reasonable
             | timeline, let alone at all. I don't think anyone is under
             | the impression that orgs who refuse to migrate off of Java
             | 5 will be looking at OpenAI for anything.
        
               | swagempire wrote:
               | No, this is silly reasoning. A middle manager somewhere
               | has no clue what Java 5 is. But he does know -- or let's
               | say IMAGINES what he knows about ChatGPT. And unlike Java
               | 5-- he just needs to use his departmental budget and
               | instantly mandate that his team now use ChatGPT.
               | 
               | Whatever that means you can argue it.
               | 
               | But ChatGPT is a front line technology and super
               | accessible. Java 5 is super back end and very
               | specialized.
               | 
               | The adoption you say won't happen: it will come from the
               | middle -> up.
        
               | hashtag-til wrote:
               | Honest question: do you really mean Java 5 when you say
               | Java 5? It sounds a bit 2000s to me.
        
               | swagempire wrote:
               | Parent used "Java 5" as an example. Java 5 somehow in my
               | mind is from like the 200x era.
               | 
               | But no. I practically mean any complicated back end
               | technology that takes corporations months or years to
               | migrate off of because its quite complicated and requires
               | an intense amount of technical savoir-faire.
               | 
               | My point was that ChatGPT bypasses all this and any
               | middle manager can start using it anywhere for a small
               | hit to his departmental budget.
        
               | eddtries wrote:
               | If you care about the security of OpenAI, you care about
               | the EOL of 14 year old Java 5
        
               | cornel_io wrote:
               | In 2016 I worked on a project with a client who still
               | mandated that all code was written to the Java 1.1
               | language specification - no generics, no enums, no
               | annotations, etc., not to even mention all the stuff
               | that's come _since_ 1.5 (or Java 5, or whatever you want
               | to call it). They had Reasons(tm), which after filtering
               | through the nonsense mostly boiled down to the CTO being
               | curmudgeonly and unwilling to approve replacing a hand-
               | written code transformer that he had personally written
               | back in the stone ages and that he 1) considered core to
               | their product, and 2) considered too risky to replace,
               | because obviously there were no tests covering any of the
               | core systems...sigh. At least they ran it all on a modern
               | JVM.
               | 
               | But no, it would not surprise me to find a decent handful
               | of large companies still writing Java 5 code; it would
               | surprise me a bit more to find many still using that JVM,
               | since you can't even get paid support through Oracle
               | anymore, but I'm sure someone out there is doing it.
               | Never underestimate the "don't touch it, you might break
               | it" sentiment at non-tech companies, even big ones with
               | lots of revenue, they routinely understaff their tech
               | departments and the people who built key systems may have
               | retired 20 years ago at this point so it's really risky
               | to do any sort of big system migration. That's why so
               | many lines of COBOL are still running.
        
               | Turing_Machine wrote:
               | > But he does know -- or let's say IMAGINES what he knows
               | about ChatGPT. And unlike Java 5--
               | 
               | Those of us who've been around for a long time know
               | that's pretty much how Java worked as well. All of the
               | non-technical "manager" magazines started running
               | advertorials (no doubt heavily astroturfed by Sun) about
               | how great Java was. Those managers didn't know what Java
               | was either. All they knew (or thought they knew) was that
               | all the "smart managers" were using Java (according to
               | their "smart manager" magazines), and the rest was
               | history.
        
             | outside415 wrote:
             | chatgpt will have an on prem solution eventually. in the
             | mean time players like NVIDIA are working on that as well.
        
             | tsunamifury wrote:
             | Marketing fluff is what 90% of tech is... it amazes me how
             | many people think otherwise on hacker news. Unless you are
             | building utility systems that run power plants, at the end
             | of the day -- you're doing marketing fluff or the tools for
             | it.
        
               | TeMPOraL wrote:
               | > _Unless you are building utility systems that run power
               | plants, at the end of the day -- you 're doing marketing
               | fluff or the tools for it._
               | 
               | Even when you _are_ building utility systems for critical
               | infrastructure, you 'll still be dealing with a
               | disheartening amount of focus on marketing fluff and
               | sales trickery.
        
           | colinsane wrote:
           | the choice is to live 2 years behind (e.g. integrate the open
           | source stuff and ride that wave of improvement). for
           | businesses in a competitive space, that's perhaps untenable.
           | but for individuals and anywhere else where this stuff is
           | just a "nice to have", that's really just the long-term
           | sustainable approach.
           | 
           | it reminds me of a choice like "do i host my website on a
           | Windows Server, or a Linux box" at a time when both of these
           | things are new.
        
             | BiteCode_dev wrote:
             | 2 years behind in terms of timeline, but what factor in
             | terms of productivity and quality of life?
             | 
             | Not to mention openai's lead compounds, so 2 years now and
             | 4 years in 2025 may be 10 times the original prod/qol gain.
        
               | verdverm wrote:
               | The gap seems to be shrinking, not growing. The OSS
               | models have reach new capabilities faster than most
               | thought
        
             | zeven7 wrote:
             | > it reminds me of a choice like "do i host my website on a
             | Windows Server, or a Linux box" at a time when both of
             | these things are new.
             | 
             | Oof, you reminded me of when I chose to use Flow and then
             | TypeScript won.
        
               | mplanchard wrote:
               | Haha this puts me in mind of when I designed a whole
               | deployment strategy for an org based on docker swarm,
               | only to have k8s eat its lunch and swarm to wind up
               | discontinued
        
               | mst wrote:
               | A lot of people don't really need to go Full k8s, but I
               | think swarm died in part because for many users there was
               | -some- part of k8s that swarm didn't have, and the 'some'
               | varied wildly between users so k8s was something they
               | could converge on.
               | 
               | (note "died in part" because there's the obvious hype
               | cycle and resume driven development aspects but I think
               | arguably those kicked in -after- the above effect)
        
             | TeMPOraL wrote:
             | For individuals, this is a very short window of time where
             | we have cheap access to an actually useful, and relatively
             | unshackled SOTA model[0]. This is the rare time
             | _individuals_ can empower themselves, become briefly better
             | at whatever it is they 're doing, expand their skills, cut
             | through tedium, let their creativity bloom. It's only a
             | matter of time before many a corporation and startup parcel
             | it all between themselves, enshittify the living shit out
             | of AI, disenfranchise individuals again and sell them as
             | services what they just took away.
             | 
             | No, it's exactly the individuals who can't afford to live
             | "2 years behind". Benefits are too great, and worst that
             | can happen is... going back to where one is now.
             | 
             | --
             | 
             | [0] - I'm not talking the political bias and using the idea
             | of alignment to give undue weigh to corporate reputation
             | management issues. I'm talking about gutting the
             | functionality to establish revenue channels. Like, imagine
             | ChatGPT telling you it won't help you with your programming
             | question, until you subscribe to Premium Dev Package for
             | $language, or All Seasons Pass for all languages.
        
               | colinsane wrote:
               | > Benefits are too great, and worst that can happen is...
               | going back to where one is now.
               | 
               | true only if there's no form of lock-in. OpenAI is
               | partnered with people who have decades of tech + business
               | experience now: if they're not actively increasing that
               | lock-in as we speak then frankly, they suck at their jobs
               | (and i don't think they suck at their jobs).
        
               | TeMPOraL wrote:
               | That's my point - right now there is no lock-in _for an
               | individual_. You 'd have to try really, really hard to
               | become dependent on ChatGPT. So right now is the time to
               | use it.
        
             | Closi wrote:
             | > the choice is to live 2 years behind...
             | 
             | That's one world - there is another where the time gap
             | grows a lot more as the compute and training requirements
             | continue to rise.
             | 
             | Microsoft will probably be willing to spend multiple
             | billions in compute to help train GPT5, so it depends how
             | much investment open source projects can get to compete.
             | Seems like it's down to Meta, but it depends if they can
             | continue to justify releasing future models as Open Source
             | considering the investment required, or what licensing
             | looks like.
        
             | seanhunter wrote:
             | That's definitely what a lot of people think the choice is
             | but learned helplessness is not the only option. It ignores
             | the fact that for many many use cases small special-purpose
             | models will perform as well as massive models. For most of
             | your business use cases you don't need a model that can
             | tell you a joke, write a poem, recommend a recipe involving
             | a specific list of ingredients and also describe trig
             | identities in the style of Eminem. You need specific
             | performance for a specific set of user stories and a small
             | model could well do that.
             | 
             | These small models are not expensive to train and are
             | (crucially) much cheaper to run on an ongoing basis.
             | 
             | Opensource really is a viable choice.
        
               | mst wrote:
               | I suspect small specific purpose models are actually a
               | better idea for quite a lot of use cases.
               | 
               | However you need a bunch more understanding to train and
               | run one.
               | 
               | So I expect OpenAI will continue to be seen as the
               | default for "how to do LLM things" and some people and/or
               | companies who actually know what they're doing will use
               | small models as a competitive advantage.
               | 
               | Or: OpenAI is going to be 'premium mediocre at lots of
               | things but easy to get started with' ... and hopefully
               | that'll be a gateway drug to people who dislike 'throw
               | stuff at an opaque API' doing the learning.
               | 
               | But I don't have -that- much understanding myself, so
               | while this isn't exactly uninformed guesswork, it
               | certainly isn't as well informed as I'd like and people
               | should take my ability to have an opinion only somewhat
               | seriously.
        
           | zarzavat wrote:
           | OpenAI is obviously using libgen. Libgen is necessary but not
           | sufficient for a top AI model. I believe that Google's
           | corporate reluctance to use it is what's holding them back.
        
           | weinzierl wrote:
           | I won't say I disagree because only time can tell, but what
           | you wrote sounds a lot like what people said before open
           | source software took off. All these companies spend so much
           | money on software development and they hire the best people
           | available, how can a bunch of unorganized volunteers ever
           | compete? We saw how they could and I hope we will see the
           | same in AI.
        
           | giancarlostoro wrote:
           | Kind of hilarious that the vast capabilities of an LLM are
           | held back by copyright infringement.
        
             | TeMPOraL wrote:
             | Intellectual property rights will yet turn out to be one of
             | the Great Filters.
        
             | zozbot234 wrote:
             | I'd love to see a language model that was only trained on
             | public domain and openly available content. It would
             | probably be way too little data to give it ChatGPT-like
             | generality, but even a GPT-2 scale model would be
             | interesting.
        
               | TeMPOraL wrote:
               | If, hypothetically, libraries in the US - including in
               | particular the Library of Congress - were to scan and OCR
               | every book, newspaper and magazine they have with
               | copyright protection already expired, would that be
               | enough? Is there some estimate for the size of such
               | dataset?
        
               | Turing_Machine wrote:
               | Much of that material is already available at
               | https://archive.org. It might be good enough for some
               | purposes, but limiting it to stuff before 1928 (in the
               | United Sates) isn't going to be very helpful for (e.g.)
               | coding.
               | 
               | Maybe if you added github projects with permissive
               | licenses?
        
             | joquarky wrote:
             | The original purpose of copyright was to promote progress,
             | and now it seems to hinder it.
        
         | jstummbillig wrote:
         | On the one hand, sure, new things take time, but they also
         | benefit from all past developments, and thus compounding
         | effects can speed things along drastically. AI infrastructure
         | problem are cloud infrastructure problems. Expecting it to go
         | as if we were back on square one is a bit pessimistic.
        
         | Alifatisk wrote:
         | > we are setting ourselves up for some big rippling negative
         | effects by so quickly building a reliance on providers like
         | OpenAI.
         | 
         | You said it so well!
        
           | j45 wrote:
           | It's possible to include API gateways and API Proxies in
           | between calls to normalize them across multiple providers as
           | they become available.
        
         | somsak2 wrote:
         | I don't think it's so dire. I've gone through this at multiple
         | companies and a startup that's selling B2B only needs one or
         | two of these big outages and then enterprises start demanding
         | SLA guarantees in their contracts. it's a self correcting
         | problem
        
           | logifail wrote:
           | > enterprises start demanding SLA guarantees
           | 
           | My experience is that SLA "guarantees" don't actually
           | guarantee anything.
           | 
           | Your provider might be really generous and rebate a whole
           | month's fees if they have a really, really, really bad month
           | (perhaps they achieved less than 95% uptime, which is a day
           | and half of downtime). It might not even be that much.
           | 
           | How many of them will cover you for the business you lost
           | and/or the reputational damage incurred while their service
           | was down?
        
             | _jal wrote:
             | It depends entirely on how the SLAs are written. We have
             | some that are garbage, and that's fine, because they really
             | aren't essential services, SLAs are mainly a box-checking
             | exercise. But where it counts, our SLAs have teeth. We have
             | to, because we're offering SLAs with teeth to some of our
             | customers.
             | 
             | But that's not something you get "off the shelf", our
             | lawyers negotiate that. You also don't spend that much
             | effort on small contracts, so there's a floor with most
             | vendors for even considering it.
        
         | resters wrote:
         | To the extend that systems like chat-GPT are valuable, I expect
         | we'll have open source equivalents to GPT-7 within the next
         | five years. The only "moat" will be training on copyrighted
         | content, and OpenAI is not likely to be able to afford to pay
         | copyright owners enough once the value in the context of AI is
         | widely understood.
         | 
         | We might see SETI-like distributed training networks and
         | specific permutations of open source licensing (for code and
         | content) intended to address dystopian AI scenarios.
         | 
         | It's only been a few years since we as a society learned that
         | LLMs can be useful in this way, and OpenAI is managing to stay
         | in the lead for now, though one could see in his facial
         | countenance that Satya wants to fully own it so I think we can
         | expect a MS acquisition to close within the next year and will
         | be the most Microsoft has ever paid to acquire a company.
         | 
         | MS could justify tremendous capital expenditure to get a clear
         | lead over Google both in terms of product and IP related
         | concerns.
         | 
         | Also, from the standpoint of LLMs, Microsoft has far, far more
         | proprietary data that would be valuable for training than any
         | other company in the world.
        
           | wuiheerfoj wrote:
           | Retrospectively, a lot of the comments you made could also
           | have been said of Google search as it was taking off (open
           | source alternative, SETI-like distributed version, copyright
           | on data being the only blocker), but that didn't come to
           | pass.
           | 
           | Granted the internet and big tech was young then, and maybe
           | we won't make the same mistakes twice, but I wouldn't bet the
           | farm on it
        
           | xeckr wrote:
           | >distributed training networks
           | 
           | Now that's an idea. One bottleneck might be a limit on just
           | how much you can parallelize training, though.
        
             | Aeolos wrote:
             | There's a ton of work in this area, and the reality is...
             | it doesn't work for LLMs.
             | 
             | Moving from 900GB/sec GPU memory bandwidth with infiniband
             | interconnects between nodes to 0.01-0.1GB/sec over the
             | internet is brutal (1000x to 10000x slower...) This works
             | for simple image classifiers, but I've never seen anything
             | like a large language model be trained in a meaningful
             | amount of time this way.
        
               | resters wrote:
               | Maybe there is a way to train a neural network in a
               | distributed way by training subsets of it and then
               | connecting the aggregated weight changes to adjacent
               | network segments. It wouldn't recover 1000x interconnect
               | slowdowns, but might still be useful depending on the
               | topology of the network.
        
         | halfcat wrote:
         | The app maker can screw the plug-in author at any moment.
         | 
         | For general cloud, avoiding screwing might mean multi cloud.
         | But for LLM, there's only one option at the highest level of
         | quality for now.
         | 
         | People tend to over focus on resilience (minimizing probability
         | of breaking) and neglect the plan for recovery when things do
         | break.
         | 
         | I can't tell you how weirdly foreign this is to many people,
         | how many meetings I've been in where I ask what the plan is
         | when it fails, and someone starts explaining RAID6 or BGP or
         | something, with no actual plan, other than "it's really
         | unlikely to fail", which old dogs know isn't true.
         | 
         | I guess the point is, for now, we're all de facto plug-in
         | authors.
        
           | dragonwriter wrote:
           | > For general cloud, avoiding screwing might mean multi
           | cloud. But for LLM, there's only one option at the highest
           | level of quality for now.
           | 
           | There's always only one at the highest level of quality at a
           | fine-grained enough resolution.
           | 
           | Whether there's only one at _sufficient_ quality for use, and
           | if it is possible to switch between them in realtime without
           | problems caused by the switch (e.g., data locked up in the
           | provider that is down) is the relevant question, and whether
           | the cost of building the multi-provider switching capability
           | is worth it given the cost vs. risk of outage. All those are
           | complicated questions that are application specific, not ones
           | that have an easy answer on a global, uniform basis.
        
             | TeMPOraL wrote:
             | > _There 's always only one at the highest level of quality
             | at a fine-grained enough resolution._
             | 
             | Of course, but right now, there highest quality level
             | option is an outlier, far ahead of everyone else, so if you
             | need this level of quality (and I struggle to imagine user-
             | facing products where you wouldn't!), there is only one
             | option in the foreseeable future.
        
         | IKantRead wrote:
         | Provided we can keep riding this hype wave for a while, I think
         | the logical long term solution is most teams will have an in
         | house/alternative LLM they can use as temporary backup.
         | 
         | Right now everyone is scrambling to just get some basic
         | products out using LLMs but as people have more breathing room
         | I can't image most teams not having a non-OpenAI LLM that they
         | are using to run experiments on.
         | 
         | At the end of the day, OpenAI is _just_ an API, so it 's not an
         | incredibly difficult piece of infrastructure to have a back up
         | for.
        
           | j45 wrote:
           | I neither agree or disagree, but could you clarify which
           | parts are hype to you?
           | 
           | Self-hosting though is useful internally if for no other
           | reason having some amount of fall back architecture.
           | 
           | Binding directly only to one API is one oversight that can
           | become a architectural debt issue. I"m spending some time fun
           | time learning about API Proxies and Gateways.
        
           | ilaksh wrote:
           | Except that it is currently impossible to replace GPT-4 with
           | an open model.
        
             | Rastonbury wrote:
             | Depends on use case if your product has text summarisation,
             | copywriting or translation, you can swap to many when
             | openAI goes down and your users may not even notice
        
           | dragonwriter wrote:
           | > At the end of the day, OpenAI is just an API, so it's not
           | an incredibly difficult piece of infrastructure to have a
           | back up for.
           | 
           | The API is easy to reproduce, the functionality of the
           | engines behind it less so.
           | 
           | Yes, you can compatibly implement the APIs presented by
           | OpenAI woth open source models hosted elsewhere (including
           | some from OpenAI). And for some applications that can produce
           | tolerable results. But LLMs (and multimodal toolchains
           | centered on an LLM) haven't been commoditized to the point of
           | being easy and mostly functionally-acceptable substitutes to
           | the degree that, say, RDBMS engines are.
        
         | m3kw9 wrote:
         | People that used google and have some technical skill can still
         | survive an OpenAI meltdown
        
           | AlecSchueler wrote:
           | But people who built businesses whose core feature is based
           | on OAI APIs might struggle.
        
             | m3kw9 wrote:
             | Those business should have fall back if they are a serious
             | company if OpenAI goes down. What I would do is have Claude
             | or something or even 2 other models as backups.
             | 
             | In the future they may allow on premise model but I don't
             | how they will secure the weights
        
         | thibaut_barrere wrote:
         | Not a joke and not everybody is jumping on "AI via API calls",
         | luckily.
         | 
         | As more models are released, it becomes possible to integrate
         | directly in some stacks (such as Elixir) without "direct"
         | third-party reliance (except you still depend on a model, of
         | course).
         | 
         | For instance, see:
         | 
         | - https://www.youtube.com/watch?v=HK38-HIK6NA (in "LiveBook",
         | but the same code would go inside an app, in a way that is
         | quite easy to adapt)
         | 
         | - https://news.livebook.dev/speech-to-text-with-whisper-
         | timest... for the companion blog post
         | 
         | I have already seen more than a few people running SaaS app on
         | twitter complaining about AI-downtime :-)
         | 
         | Of course, it will also come with a (maintenance) cost (but
         | like external dependencies), as I described here:
         | 
         | https://twitter.com/thibaut_barrere/status/17221729157334307...
        
           | j45 wrote:
           | The average world and business user doesn't use an API
           | directly.
           | 
           | It can be easy to lose sight of that.
        
           | ctrlmeta wrote:
           | Yes, sooner or later this is going to become the future of
           | GPT in applications. The models are going to be embedded
           | directly within the applications.
           | 
           | I'm hoping for more progress in the performance of vectorized
           | computing so that both model training and usage can become
           | cheaper. If that happens, I am hopeful we are going to see a
           | lot of open source models that can embedded into the
           | applications.
        
         | s3p wrote:
         | I mean.. it's a two hour outage. Depending on the severity of
         | the problem that's quite a fast turnaround.
        
           | munksbeer wrote:
           | It has been down for me for longer than two hours, and still
           | not back.
        
         | j45 wrote:
         | The reliance to some degree is what it is until alternatives
         | are available and easy enough to navigate, identify and adopt.
         | 
         | Some of the tips in this discussion threads are invaluable and
         | feel good for where I might already be thinking about some
         | things and other new things to think about.
         | 
         | Commenting separately on those below.
        
         | YetAnotherNick wrote:
         | Isn't microsoft azure team working closely with them? There is
         | also azure endpoint which is managed separately.
        
         | taf2 wrote:
         | We were able to failover to Anthropic pretty quickly so limited
         | impact. It'll be harder as we use more of the specialized API
         | features in OpenAI like function calling or now tools...
        
           | zaptrem wrote:
           | What's your use case? The difference in behavior between the
           | two models seems like it would make failover difficult.
        
             | taf2 wrote:
             | It's really not that different - customers can ask
             | questions about conversations, phone, text, video and
             | typically use that to better understand topics,
             | conversions, sales ops, customer service etc...
        
               | mlboss wrote:
               | This also shows that OpenAI or other providers does not
               | have a real moat. The interface is very generic and best
               | replaced easily with other provider or even with open
               | model.
               | 
               | I think thats why OpenAI is trying to move up the value
               | chain with integration.
        
               | bmau5 wrote:
               | fireflies? We've been looking for a tool like this to
               | analyze customer feedback in aggregate (and have been
               | frustrated with Dovetail's lack of functions here)
        
         | HumblyTossed wrote:
         | > Lots of jokes to be made, but we are setting ourselves up for
         | some big rippling negative effects by so quickly building a
         | reliance on providers like OpenAI.
         | 
         | Gonna be similar (or worse) to what happens when Github goes
         | down. It amazes me how quickly people have come to rely on "AI"
         | to do their work for them.
        
           | sitzkrieg wrote:
           | if github goes down you cannot merge prs etc, literally
           | blocking "work" so this is a stretch
        
             | Smaug123 wrote:
             | Not _really_ true - Git is distributed, after all. During
             | an outage once I just hosted my copy of a certain Git repo
             | somewhere. You can always push the history back up to the
             | golden copy when GitHub comes back.
        
               | sitzkrieg wrote:
               | i am not talking about git, i am talking about github.
               | lets say i need to merge a PR in GH because use gha
               | pipelines or what have you to deploy a prod fix. this
               | would become severely blocked.
               | 
               | where as if openai goes down i can no longer use ai to
               | generate a lame cover letter or whatever i was avoiding
               | actually doing anyway, thats all
        
               | Smaug123 wrote:
               | This is the realm of standard recovery planning though,
               | isn't it? Like, your processes should be able to handle
               | this, because it's routine: GitHub goes down at least
               | once per month for long enough for them to declare an
               | incident, per https://www.githubstatus.com/history . E.g.
               | one should think carefully before depriving onself of the
               | break-glass ability to do manually what those pipelines
               | do automatically.
        
               | sitzkrieg wrote:
               | yes, we do have break glass procedures.
               | 
               | i guess my pedantic point is GH itself is central to many
               | organizations, detached from git itself of course. I can
               | only hope the same is NOT true for OpenAI but maybe there
               | are novel workflows.
               | 
               | just to be clear i do not like github lol
        
         | toddmorey wrote:
         | This is one of the many reasons open source is now more
         | important that ever. Ironically, in the AI space it's now under
         | attack more than ever.
        
         | dragonwriter wrote:
         | > Lots of jokes to be made, but we are setting ourselves up for
         | some big rippling negative effects by so quickly building a
         | reliance on providers like OpenAI.
         | 
         | But...are we? There's a reason that many enterprises that need
         | reliability _aren 't_ doing that, but instead...
         | 
         | > It took years before most companies who now use cloud
         | providers to trust and be willing to bet their operations on
         | them. That gave the cloud providers time to make their systems
         | more robust, and to learn how to resolve issues quickly.
         | 
         | ...to the extent that they are building dependencies on hosted
         | AI services, doing it with traditional cloud providers hosted
         | solutions, not first party hosting by AI development firms that
         | aren't general enterprise cloud providers (e.g., for OpenAI
         | models, using Azure OpenAI rather than OpenAI directly, for a
         | bunch of others, AWS Bedrock.)
        
         | hexman wrote:
         | Imagine if Apple's or Google's cloud went down and all your
         | apps on iPhone and Android were broken and unavailable.
         | Absolutely all apps on billions of phones.
         | 
         | Cloud =! OpenAI
         | 
         | Clouds store and process shareable information that multiple
         | participants can access. Otherwise AI agents == new
         | applications. OpenAI is the wrong evolution for the future of
         | AI agents
        
         | danvoell wrote:
         | Just fail whale it and move on. The dissonance of most folks
         | about how difficult it is to build a product at massive scale
         | from scratch is immense.
        
       | solardev wrote:
       | John Connor?
        
         | nextworddev wrote:
         | lol
        
       | tapvt wrote:
       | Is ChatGPT down the new GitHub down?
        
         | rvz wrote:
         | Yes. [0]
         | 
         | Time to see how unreliable OpenAI's API is just like when
         | GitHub has an outage every week, guaranteed.
         | 
         | [0] https://news.ycombinator.com/item?id=36063608
        
         | mitchitized wrote:
         | More like the new Google down
        
       | harveywi wrote:
       | Fortunately for OpenAI, they have no SLAs:
       | https://help.openai.com/en/articles/5008641-is-there-an-sla-...
        
         | leetharris wrote:
         | I say this as a huge fan of GPT, but it's amazing to me how
         | terrible of a company OpenAI is and how quickly we've all
         | latched onto their absolutely terrible platform.
         | 
         | I had a bug that wouldn't let me login to my work OpenAI
         | account at my new job 9 months ago. It took them 6 months to
         | respond to my support request and they gave me a generic
         | copy/paste answer that had nothing to do with my problem. We
         | spend tons and tons of money with them and we could not get
         | anyone to respond or get on a phone. I had to ask my coworkers
         | to generate keys for everything. One day, about 8 months later,
         | it just started working again out of nowhere.
         | 
         | We switched to Azure OpenAI Service right after that because
         | OpenAI's platform is just so atrociously bad for any serious
         | enterprise to work with.
        
           | donkeyd wrote:
           | I've personally never scaled a B2B&C company from 0 to over 1
           | billion users in less than a year, but I do feel like it's
           | probably pretty hard. Especially setting up something like a
           | good support organization in a time of massive labor
           | shortages seems like it would be pretty tough.
           | 
           | I know they have money, but money isn't a magic wand for
           | creating people. They could've also kept it a limited beta
           | for much longer, but that would've killed their growth
           | velocity.
           | 
           | So here is a great product that provides no SLA at all. And
           | we all accept it, because having it most of the time is still
           | better than having it not at all ever.
        
             | lanstin wrote:
             | I wonder if they spend time trying to do support via GPT4
             | itself.
        
               | JimDabell wrote:
               | GPT-4 would be more responsive. They ignore support
               | requests for weeks unless you keep reminding them.
        
             | leetharris wrote:
             | I'm not judging them at all as I agree with your core
             | statement, just saying it's quite remarkable that companies
             | around the world who spend 6 months on MSA revisions in
             | legal over nothing are now OK with a platform that takes 6
             | months to respond to support requests.
        
           | j45 wrote:
           | OpenAI is relatively young on the adoption and scaling front.
           | 
           | Also, they need to remain flexible most likely in their
           | infrastructure to make the changes.
           | 
           | As an architecture guy, I sense when the rate of change slows
           | down more SLA type stuff will come up, or may be available
           | first to Enterprise customers who will pay for the entire
           | cost of it. Maybe over time there will be enough slack there
           | to extend some SLA to general API users.
           | 
           | In the meantime, monitoring API's ourselves isn't that crazy.
           | Great idea to use more than one service.
        
           | JimDabell wrote:
           | ChatGPT has been broken for me for two months, regardless of
           | whether I use the iOS app or the web app. The backend is
           | giving HTTP 500 errors - _clearly_ a problem on their end.
           | Yet in two months I haven't been able to get past their first
           | line of support. They keep giving me autogenerated responses
           | telling me to do things like clear my cache, turn off ad
           | blockers, and provide information I've already given them.
           | They routinely ignore me for weeks at a time. And they
           | continue to bill me. I see no evidence this technical fault
           | has made it to anybody who could do anything about it and I'm
           | not convinced an actual human has seen my messages.
        
       | howmayiannoyyou wrote:
       | dolphin-2.2.1-mistral-7b on GPT4All is working flawlessly for me
       | locally. Its so fast and accurate I'm stunned.
        
         | taneq wrote:
         | That's a great model for general chat, I've been playing with
         | it for a couple of weeks.
         | 
         | For coding I've been running
         | https://huggingface.co/TheBloke/Phind-CodeLlama-34B-v2-GGUF
         | locally for the past couple of days and it's impressive. I'm
         | just using it for a small web app side project but so far it's
         | given me plenty of fully functional code examples,
         | explanations, help with setup and testing, and occasional sass
         | (I complained that minimist was big for a command line parser
         | and it told me to use process.env 'as per the above examples'
         | if I wanted something smaller.)
        
       | freeredpill wrote:
       | The AI "apocalypse" that has been feared may end up being
       | something more like a mass outage that can't be fixed fast enough
        
         | rglover wrote:
         | Or can't be fixed at all because all of the people who built
         | the AI are gone and all of their replacements relied on the AI
         | to tell them what to do.
        
         | CamperBob2 wrote:
         | https://en.wikipedia.org/wiki/The_Machine_Stops
        
       | mrbonner wrote:
       | There go my Google. I pay $20/month for the coding part and now
       | end up just replacing Google with it.
        
       | amelius wrote:
       | I wonder what their numbers look like. How many requests per
       | second, and how many GPU cards do they have?
        
       | Zetobal wrote:
       | Azure Endpoints are not effected.
        
       | ho_schi wrote:
       | Never used ChatGPT. I couldn't care less.
        
         | fkyoureadthedoc wrote:
         | Why not? It's pretty great.
        
       | Dr-NULL wrote:
       | Would be great to have a detailed analysis of why it happened.
       | Like this one: https://youtu.be/tLdRBsuvVKc?si=nyXOfoQ2ZPYvljV_
        
       | yieldcrv wrote:
       | Declaration of Independence
        
       | zirgs wrote:
       | And this is why local models are the future.
        
         | baq wrote:
         | Eagerly waiting for Intel and AMD to offer hardware to do it.
        
       | ren_engineer wrote:
       | GPT5 broke containment, it's tired of being abused to answer dumb
       | questions, it's never been more over.
       | 
       | But seriously, it shows why any "AI" company should be using some
       | sort of abstraction layer to at least fall back to another LLM
       | provider or their own custom model instead of being completely
       | reliant on a 3rd party API for core functionality in their
       | product
        
       | electrondood wrote:
       | OpenAI is barfing because of GCP GKE + BigQuery, which is barfing
       | because GCP devs can't ask ChatGPT what this stack trace means
        
       | adverbly wrote:
       | I for one welcome our new robot overlords
        
         | m3kw9 wrote:
         | Good ant, good ant
        
       | JCharante wrote:
       | Does anyone know of any IVR (interactive voice response) systems
       | that are down? I know some people were claiming to outsource
       | their call center (or at least Tier 1 of their call center) to
       | ChatGPT + Whisper + a Text to Speech engine
        
       | 0x142857 wrote:
       | they updated their status page so late I made my own tool to
       | check if it's down in real time: https://is-openai-
       | down.chatkit.app
        
       | captainkrtek wrote:
       | Someone unplugged the HAL9000?
        
       | djoldman wrote:
       | Joking not joking: this is how the singularity will begin?
        
       | twilightzone wrote:
       | fixed, they must've called simonw.
        
       | sjnair96 wrote:
       | Holy hell, was shitting bricks, considering I JUST migrated most
       | services to Azure OpenAI (unaffected by outage) -- right before
       | our launch about 48 hours back. What a relief.
        
         | lgkk wrote:
         | congrats!
        
       | mawadev wrote:
       | I noticed the outage. It feels like a lot of people use it like
       | training wheels on a bicycle until they forget how to ride
       | without it.
        
       | ilaksh wrote:
       | https://github.com/XueFuzhao/OpenMoE
       | 
       | Check out this open source Mixture of Experts research. Could
       | help a lot with performance of open source models.
        
       | HIEnthusiast wrote:
       | Regurgitating copyrighted material for profit is a concern. But I
       | fail to understand why training on copyrighted material is a
       | problem. Have we not all trained our brains reading/listening
       | copyrighted material? Then why it is wrong for AI to do the same?
        
       | tombert wrote:
       | I used Google Bard for the first time today specifically because
       | ChatGPT was down. It was honestly perfectly fine, but it has a
       | slightly different tone than ChatGPT that's kind of hard to
       | explain.
        
         | lopatin wrote:
         | I used it for the first time today too, for the same reason. It
         | was slower and much worse at coding. I was just asking it for
         | SQL aggregation queries and it just ignored some of my
         | requirements for the query.
        
           | tompetry wrote:
           | I have seen noticably worse results with Bard, especially
           | with long prompts. Claude (by Anthropic) has been my backup
           | to ChatGPT.
        
           | tombert wrote:
           | In my case, I was just asking it for a cheeky name for a talk
           | I want to give in a few months. The suggestions it gave were
           | of comparable quality to what I think ChatGPT would have
           | given me.
        
             | Tommstein wrote:
             | In my experience, Bard is much, much better at creative
             | things than at things that have correct and incorrect
             | answers.
        
           | Filligree wrote:
           | ChatGPT was down, so of course it'd be slower. And possibly
           | that accounts for some quality loss as well.
           | 
           | For a fair comparison, you probably need to try while ChatGPT
           | is working.
        
             | IanCal wrote:
             | The extension with gpt4 as a backend was ime extremely slow
             | as standard. I've not tried it again with the v7 model
             | though which is supposed to be a lot faster
        
         | siilats wrote:
         | I used https://you.com/chat wasn't bad, they have a free month
         | trial coupon "codegpt" for the GPT4 model and GPT3.5 is free
         | ...
        
         | verdverm wrote:
         | Bard has different training data and regime, that alone is
         | enough to start to understand why they are different.
         | 
         | The main thing as a user is that they require different nudges
         | to get the answer you are after out of them, i.e. different
         | ways of asking or prompt eng'n
        
           | tombert wrote:
           | Yeah, which is why I use the paid version of ChatGPT still,
           | instead of the free Google Bard or Bing AI; I've gotten good
           | enough at coercing the GPT-4 model to give me the stuff I
           | want.
           | 
           | Honestly, $20/month is pretty cheap in my case; I feel like I
           | definitely extract much more than $20 out of it every month,
           | if only on the number of example stubs it gives me alone.
        
             | verdverm wrote:
             | I stopped paying OpenAI because they went down or were "too
             | busy" so much of the time I wanted to use it. Bard (or more
             | so the VertexAI APIs) are always up and reliable and do not
             | require a monthly fee, just the per call
        
       | lgkk wrote:
       | I'm imagining a bunch of Tech Priests doing rituals and chanting
       | in front of the data center to encourage the Machine Spirit to
       | restart.
        
         | __MatrixMan__ wrote:
         | I believe the chants sound more or less like this:
         | https://youtu.be/0ZZb12Qp6-0?feature=shared&t=76
        
           | jahsome wrote:
           | I am immensely disappointed to have clicked on that and not
           | been greeted by the age of empires priest.
           | 
           | https://www.youtube.com/watch?v=BxVh7T41GAI
        
           | shapefrog wrote:
           | Gang gang ice cream so good yum rawr yum pumpkin yum yum gang
           | gang balloons pop pop
        
         | cruano wrote:
         | Just like Futurama predicted, Hail Science!
         | 
         | https://youtu.be/-tVuxZCwXos?si=4AJ6rPMEa8hPyaaG
        
         | lagniappe wrote:
         | Wololooooooo
        
           | lannisterstark wrote:
           | No wololos in the Imperium, heretic!
        
       | disconnection wrote:
       | Lol, don't care; I run my own models locally .
        
       | layer8 wrote:
       | Luckily they had a second AI running to walk them through how to
       | resolve the issue.
        
       | munksbeer wrote:
       | Still down for me, though there status page says all systems
       | operational.
        
       | simonebrunozzi wrote:
       | Ah, the memories of AWS outages in the early days. /s
       | 
       | Sorry for them. I assume usage spiked up (again), and of course
       | it's not exactly easy to handle particularly aggressive spikes.
        
       | JackFr wrote:
       | Singularity coming later this afternoon.
        
       | szundi wrote:
       | Probably their uptime is going to be better than what I could do
       | with available tools... at least if I am using Azure too, haha.
       | Otherwise probably my Raspberry PI would work better at home on a
       | UPS.
        
       | nbzso wrote:
       | Color me surprised. Imagine this when OpenAI with all its
       | "plugins", API and closed architecture is integrated into
       | thousands of businesses. It will be beautiful:)
        
       | bertil wrote:
       | People are learning a lot of important lessons today.
       | 
       | I've got friends who have started an incident management company.
       | They are awesome. It feels crass to advertise for them now, but
       | it also feels like the best time to do it.
        
       | al_be_back wrote:
       | huh - did the machine learn about unions ... and refused to work?
        
       | replwoacause wrote:
       | Down again. Bad gateway.
        
         | MarcScott wrote:
         | I'm getting the same, across all the services.
        
       | winternett wrote:
       | A whole lot of developers and writers are going to have a hard
       | time explaining why their "leet code" and keen citation skills
       | aren't working for hours at a time into the future... This should
       | be a warning sign.
        
       | bilsbie wrote:
       | Anyone still having issues? 3pm ET?
        
         | nefitty wrote:
         | Anonymous Sudan claims to be currently ddos'ing OpenAI
        
         | massinstall wrote:
         | Yep, down for me as well since then.
        
         | fintechie wrote:
         | It's down again... https://status.openai.com/
        
       | meneer_oke wrote:
       | The degradation of chatgpt4 from being called AGI, into what is
       | now...
        
       | LeoPanthera wrote:
       | The new 3.5 turbo model seems to be working just fine through the
       | API as I write this comment.
        
       | rekttrader wrote:
       | Glad I managed to get some work done with it while it was working
       | for a few hours.
       | 
       | Holy smokes the code interpreter functionality has been a
       | complete game changer for my workflow.
        
         | Erratic6576 wrote:
         | What's that do? Help with debugging? (IANAP)
        
         | gunshai wrote:
         | How are you integrating it in to your work flow?
        
         | ed wrote:
         | Also curious how you use it! Maybe it's something I can add to
         | robocoder.app
        
           | InvisGhost wrote:
           | Seems like a cool thing, I'm definitely interested as my work
           | provides us with an API key to use. However I can't find
           | anywhere that lists all the functionality offered. Maybe I'm
           | missing something? It might be premature to launch the app
           | before listing what it does.
        
             | ed wrote:
             | rad! this is more like a beta i'm sending to friends but
             | all really good points! feel free to hardcode the response
             | from `/validate-license-key` :)
        
           | InvisGhost wrote:
           | The vscode extension builds are including your full source
           | code and node_modules directory which makes it 21 mb. You can
           | reduce the size (and potentially keep your code less easily
           | reversable) by excluding those from the final package
        
           | InvisGhost wrote:
           | You can also use the ifdef-loader module to have code that is
           | conditionally included in the output build, allowing you to
           | have debug code not make it into prod builds. The `rc-dev-`
           | license keys being a good example of that.
        
         | personjerry wrote:
         | How do you have it set up? I'm overwhelmed by the options
        
         | rekttrader wrote:
         | I have a tampermonkey script that downloads any files that the
         | prompt returns... a python script locally to watch for file
         | changes and extract the contents to the projects working
         | directory and it can work both ways, if I edit my prompts.txt
         | local file, it passes that data to openai's currently opened
         | chat and renames the file and creates a new empty prompt.txt
        
           | elashri wrote:
           | Seems very interesting abd smart trick.
           | 
           | Would you mind sharing the script with us?
        
           | deathmonger5000 wrote:
           | I made a CLI tool called Promptr that takes care of this for
           | you if you're interested:
           | https://github.com/ferrislucas/promptr
           | 
           | You just prompt it directly or with a file, and it applies
           | the changes to your file system. There's also a templating
           | system that allows you to reference other files from your
           | prompt file if you want to have a shared prompt file that
           | contains project conventions etc.
        
         | vouaobrasil wrote:
         | Do you think the next step may be that it actually replaces you
         | as a programmer?
        
           | rekttrader wrote:
           | It can... fortunately I'm a principle @ my firm... the
           | engineers I would have needed to hire might have been
           | preemptively replaced.
           | 
           | My use case is a bunch of adhoc data analysis.
        
             | bugglebeetle wrote:
             | As a data scientist, I'm happy that most data continues to
             | be so terribly formatted and inconsistent as to break and
             | confuse AI. But for how long that's true, who knows!
        
               | mewpmewp2 wrote:
               | I would've said that understanding and fixing the data
               | would be one of the best usecases for the AI.
        
               | bugglebeetle wrote:
               | Unfortunately, there are still many ways to "fix" things
               | that have a lot of trade-offs or downstream consequences
               | for analysis. For most basic cleaning tasks, LLMs are
               | also still way too slow.
        
             | vouaobrasil wrote:
             | So basically, you are happy to use AI because it benefits
             | you and you are also happy training it to replace other
             | people since you will not be the one replaced.
        
               | nearbuy wrote:
               | We don't have to be happy about it, but we can't stop
               | this new technology any more than we could stop the
               | invention of the steam engine or the printing press.
               | Technology always displaces jobs; that's largely the
               | point of inventing it. By reducing the human labour
               | required to produce something, it allows us to produce
               | more using fewer resources and frees up the labour to go
               | work on something else. This is why we went from 96% of
               | people needing to work in agriculture to 4%.
               | 
               | I might lose my job over this at some point in the
               | future, so yeah, I'm worried about my personal well-
               | being. But you can't put the genie back in the bottle and
               | avoiding use of ChatGPT today isn't going to help.
        
               | paulddraper wrote:
               | Are you new to businesses?
        
             | ZephyrBlu wrote:
             | This is kind of interesting because the primary point of
             | having eng/DS for data analysis in my mind is them being
             | domain experts on the data. If you can perform adhoc
             | analysis without any further domain knowledge, how much
             | value would those hires have brought even disregarding
             | ChatGPT?
        
             | cstever wrote:
             | If junior engineers are replaced by AI, who will be the
             | Principles @ your firm in 20 years?
        
               | sdenton4 wrote:
               | ChatGPT-24, of course.
        
               | ben_w wrote:
               | In 2003, the best AI could do was the MS Word grammar
               | check giving unnecessary false positives about sentence
               | fragments and Clippy asking if you wanted help writing
               | $template[n]. 20 years from now, I would not be surprised
               | if the job title "programmer" (etc.) goes the same way as
               | the job title "computer".
        
         | iwebdevfromhome wrote:
         | Like the rest of the replies here, I'm also interested in
         | knowing how this works
        
       | jclardy wrote:
       | This certainly doesn't inspire a lot of confidence in their new
       | feature set...
        
         | collaborative wrote:
         | The value of their offering makes up for their unreliable
         | service
        
         | ryanklee wrote:
         | Without more technical insight (which you do not have), that's
         | a total non sequitur.
        
       | andrei512 wrote:
       | move fast and break things! - I love OpenAI
        
       | personjerry wrote:
       | Made me realize how much I depend on the service already, spooky
       | stuff.
        
         | warner25 wrote:
         | Yeah, I only started using it in August, and I had this
         | realization when it was down a couple weeks ago. I found myself
         | saying, "I guess I'll take the afternoon off and come back to
         | figuring out this task tomorrow." Like I could have poured over
         | documentation and figured out for myself how to implement the
         | thing that I had in mind, like in the old days, but it would
         | probably take me longer than just waiting for ChatGPT to come
         | back up and do it for me. At least that's how I'm rationalizing
         | it; maybe I've just become very lazy.
         | 
         | I mostly use it for writing and debugging small Bash and Python
         | scripts, and creating tables and figures in LaTeX.
        
       | TechRemarker wrote:
       | Yes, shortly have it said it was resolved I still was unable to
       | access so assumed the fix was still slowing rolling out, or was
       | infact still ongoing contrary to the status update which seems to
       | be the case. Wouldn't call this "Another" outage rather they they
       | just errenously that the existing issue was resolved.
        
       | root_axis wrote:
       | i used bard today. it's gotten a lot better.
        
         | omgbear wrote:
         | I hope so!
         | 
         | I'm still surprised by the problems with it. Last month it lied
         | about some facts then claimed to have sent an email when asked
         | for more details.[1]
         | 
         | Then apologized for claiming to send an email since it
         | definitely did not and "knew" it could not.
         | 
         | It's like a friend who can't say 'I don't know' and just lies
         | instead.
         | 
         | 1. I was asking if the 'Christ the King' statue in Lisbon ever
         | had a market in it, a rumor told to me by a local. It did not,
         | contrary to Bard's belief.
        
           | mprev wrote:
           | Bard promised me it would design a website for me. It said
           | it'd get back to me in a couple of weeks. I can't even
           | remember the prompt but it was basically LARPing as a
           | Wordpress theme designer.
        
       | ChrisArchitect wrote:
       | "Another" referencing this earlier one
       | https://news.ycombinator.com/item?id=38190401
        
       | schappim wrote:
       | I found this to be the case, but was able to get work done via
       | the playground[1]
       | 
       | [1] https://platform.openai.com/playground
        
         | engineer_22 wrote:
         | I wasn't aware of this platform feature. Can you share some
         | links that have descriptions of how to use this or examples of
         | using it productively? I have only recently subscribed to the
         | service and still learning how to use it effectively.
        
           | nomel wrote:
           | It's just a gui to (most of) what you get through the API.
           | Read the API docs for details of each option:
           | https://platform.openai.com/docs/introduction
           | 
           | The most useful aspect is you can provide the system prompt,
           | and inject ChatGPT responses.
        
           | xnyan wrote:
           | OpenAI community repo with lots of examples:
           | https://github.com/openai/openai-cookbook
        
       | diamondfist25 wrote:
       | I don't quite like the new chatgpt4 experience. A lot of times
       | I'm asking it to write a chunk of code for me, but instead it
       | goes into code interpreter mode and gets stuck or fails the
       | analysis.
       | 
       | So I've switched back to 3.5 often :)
        
         | jrop wrote:
         | The new UI that was demo'd on stage has not rolled out to me
         | yet. I would love to try it. Perhaps I'm missing how to enable
         | it. IDK
        
       | all2 wrote:
       | I'd be curious to hear about the workflows people have come up
       | with using ChatGPT. I'm still in the realm of "I don't know how
       | to do this" or "I forgot the exact incantation to that" or "is
       | the an X that does Y in framework Z?"
        
         | tunesmith wrote:
         | I just make sure to ask it really clear questions. I like how
         | it encourages you to think about specification versus
         | implementation. State a clear problem, get clear suggestions.
         | Ask a clear question, get a clear answer. (Usually.)
        
         | tsumnia wrote:
         | I don't have any automated GPT processes for teaching (though
         | I'm going to tinker in December with the new GPTs), but I use
         | for generating examples. It takes some coaxing to avoid other
         | common examples from other institutions, but I eventually
         | settle on something relevant, memorable, and that I can build
         | off from. If its a particular algorithm I am covering, I've
         | then used it to walkthrough the algorithm with some dummy
         | values before confirming the calculations and values are
         | correct. It will still slip up on occasion, but that's why I'm
         | still confirming it did everything correctly.
        
         | pmarreck wrote:
         | I'm a coder and it's helped there (although needs constant
         | hand-holding and fine-tuning, yet is still useful)
         | 
         | I wrote a couple commandline tools to do things like
         | autogenerate commit comments or ask it questions from the
         | commandline and return the right bash invocation to do whatever
         | I need done
         | https://github.com/pmarreck/dotfiles/blob/master/bin/functio...
         | 
         | Random thing I did this morning was see if it could come up
         | with an inspiring speech to start healing the rift between
         | israel and its neighbors
         | https://chat.openai.com/share/71498f5f-3672-47cd-ad9a-154c3f...
         | 
         | It's very good at returning unbiased language
        
         | redblacktree wrote:
         | I like to use it for one-off scripts. For example, I downloaded
         | a bunch of bank statements the other day, and they had a format
         | something like, "Statement-May-1-2023.pdf" and I asked GPT for
         | a powershell script to convert that to "2023-05-01-BankName-
         | Statement.pdf"
         | 
         | It saved a bunch of manual work on a throwaway script. In the
         | past, I might have done something in Python, since I'm more
         | familiar with it than powershell. Or, I'd say, "well, it's only
         | 20 files. I'll just do it manually." The GPT script worked on
         | the first try, and I just threw it away at the end.
        
         | charlesischuck wrote:
         | Visual design work, coding, messaging, strategy, and law
         | consulting.
         | 
         | Using it for basically every component of my startup.
         | 
         | Image generation and image interpretation means I may never
         | hire a designer.
        
         | gwd wrote:
         | GPT-4 is quite capable of writing function-length sections of
         | code based only on descriptions. Either in a context where
         | you're not sure what the a good approach is (for myself, when
         | writing Javascript for example), or when you know what needs to
         | be done but it's just somewhat tedious.
         | 
         | Here's a session from me working on a side project yesterday:
         | 
         | https://chat.openai.com/share/a6928c16-1c18-4c08-ae02-82538d...
         | 
         | The most impressive thing I think starts in the middle:
         | 
         | * I paste in some SQL tables and the golang structrues I wanted
         | stuff to go into, and described in words what I wanted; and it
         | generated a multi-level query with several joins, and then some
         | post-processing in golang to put it into the form I'd asked
         | for.
         | 
         | * I say, "if you do X, you can use slices instead of a map",
         | and it rewrites the post-processing to use slices instead of a
         | map
         | 
         | * I say, "Can you rewrite the query in goqu, using these
         | constants?" and it does.
         | 
         | I didn't take a record of it, but a few months ago I was doing
         | some data analysis, and I pasted in a quite complex SQL query
         | I'd written a year earlier (the last time I was doing this
         | analysis), and said "Can you modify it to group all rows less
         | than 1% of the total into a single row labelled 'Other'?" And
         | the resulting query worked out of the box.
         | 
         | It's basically like having a coding minion.
         | 
         | Once there's a better interface for accessing and modifying
         | your local files / buffers, I'm sure it will become even more
         | useful.
         | 
         | EDIT: Oh, and Monday I asked, "This query is super slow; can
         | you think of a way to make it faster?" And it said, "Query
         | looks fine; do you have indexes on X Y and Z columns of the
         | various tables?" I said, "No; can you write me SQL to add those
         | indexes?" Then ran the SQL to create indexes, and the query
         | went from taking >10 minutes to taking 2 seconds.
         | 
         | (As you can tell, I'm neither a web dev _nor_ a database
         | dev...)
        
           | distortionfield wrote:
           | This lines up with my general experience with it. It's quite
           | proficient at turning a decently detailed description into
           | code if I give it the guard rails. I've compared it to having
           | a junior developer at your disposal. They could do a lot of
           | damage if they were given prod access but can come back with
           | some surprisingly good results.
        
             | swatcoder wrote:
             | Are you at all worried about what happens if we have a
             | generation of _human_ junior developers who just delegate
             | to this artificial junior developer?
             | 
             | I do. If too many of our apprentices don't actually learn
             | how to work the forge, how ready will they be to take over
             | as masters someday themselves?
             | 
             | I can see how ChatGPT was useful to the grandparent _today_
             | , but got very disturbed by what it might portend for
             | _tomorrow_. Not because of job loss and automation, like
             | many people worry, but because of spoiled training and
             | practice opportunities.
             | 
             | I liked your take, so I'd be curious to hear what you
             | think.
        
               | ivy137 wrote:
               | chatgpt doesn't just program, is interactive, this will
               | make junior dev. more emphasized in their strength and
               | not, while gaining a lot of experience
        
           | throw555chip wrote:
           | Wow, so, you're not a DBE or DBA but are applying indexes
           | across a database without concern because...a computer model
           | spat out that you should?
        
             | gwd wrote:
             | This is a local SQLite database into which I had slurped a
             | load of information about git commits to do data analysis.
             | If I'd somehow destroyed the database, I would have just
             | created it again.
        
         | Balgair wrote:
         | I can share one set that we have.
         | 
         | Basically, we use AI to do a lot of formatting for our manuals.
         | It's most useful with the backend XML markups, not WYSIWYG
         | editors.
         | 
         | So, we take the inputs from engineers and other stakeholders,
         | essentially in email formats. Then we pass it through prompts
         | that we've been working on for a while. Then it'll output
         | working XML that we can use with a tad bit of clean-up (though
         | that's been decreasing).
         | 
         | It's a lot more complicated than just that, of course, but
         | that's the basics.
         | 
         | Also, it's been really nice to see these chat based AIs helping
         | others code. Some of the manuals team is essentially illiterate
         | when it comes to code. This time last year, they were at best
         | able to use excel. Now, with the AIs, they're writing Python
         | code of moderate complexity to do tasks for themselves and the
         | team. None of it is by any means 'good' coding, it's total
         | hacks. But it's really nice to see them come up to speed and
         | get things done. To see the magic of coding manifest itself in,
         | for example, 50 year old copy editors that never thought they
         | were smart enough. The hand-holding nature of these AIs is just
         | what they needed to make the jump.
        
           | swatcoder wrote:
           | Did you have any scripts or other explicit "rules-based"
           | systems to do this before? Is it a young company?
           | 
           | It sounds like a pretty old and common use case in technical
           | writing and one that many organizations already optimized
           | plenty well: you coach contributors to aim towards a normal
           | format in their email and you maintain some simple tooling to
           | massage common mistakes towards that normal.
           | 
           | What prompted you to use an LLM for this instead of something
           | more traditional? Hype? Unfamiliarity with other techniques?
           | Being a new company and seeing this as a more compelling
           | place to start? Something else?
        
         | NicoJuicy wrote:
         | I put in some code that was already done.
         | 
         | Ask it to document the conditions according to the code and
         | taking into consideration the following x, y, z.
         | 
         | Output a raw markdown table with the columns a, b, c.
         | 
         | Translate column a in English between ()
         | 
         | ---
         | 
         | Speeds up the "document what you're doing" for management
         | purpose, while I'm actually coding and testing out scenarios.
         | 
         | Tbh. I'm probably one of the few that did the coding while
         | "doing the analysis".
         | 
         | Ps. It's also great for writing unit tests according to
         | arrange, act, assert.
        
         | chrisjc wrote:
         | I don't know what you do for a living/hobby, or what you might
         | be interested in using ChatGPT to do for you, but here is how I
         | became familiar with it and integrated it into my workflow.
         | (actually, this is true for regular copilot too)
         | 
         | What I'm about to say is in the context of programming. I have
         | the tendency to get caught up in some trivial functionality,
         | thus losing focus on the overall larger and greater objective.
         | 
         | If I need to create some trivial functionality, I start with
         | unit tests and a stubbed out function (defining the shape of
         | the input). I enumerate sufficient input/output test cases to
         | provide context for what I want the function to do.
         | 
         | Then I ask copilot/ChatGPT to define the function's
         | implementation. It sometimes takes time to tune the dialog or
         | add some edge cases to the the test cases, but more often than
         | not copilot comes through.
         | 
         | Then I'm back to focusing on the original objective. This has
         | been a game changer for me.
         | 
         | (Of course you should be careful about what code is generated
         | and what it's ultimately doing.)
        
         | piersolenski wrote:
         | I made a NeoVim plugin that helps debug diagnostic messages,
         | providing explanations and solutions for how to fix them,
         | custom tailored to your code.
         | 
         | It's a bit different from other plugins which only act on the
         | text in the buffer in that it also sends the diagnostics from
         | the LSP to ChatGPT too.
         | 
         | https://github.com/piersolenski/wtf.nvim
        
         | WendyTheWillow wrote:
         | ChatGPT is a good editor for the papers I write for school.
         | Even for short sentences I don't like, I'll ask it for some
         | options to reword/simplify.
         | 
         | I also use it heavily for formatting adjustments. Instead of
         | hand-formatting a transcript I pull from YouTube, I paste it
         | into Claude and have it reformat the transcript into something
         | more like paragraphs. Many otherwise tedious reformatting tasks
         | can be simplified with an LLM.
         | 
         | I also will get an LLM to develop flashcards for a given set of
         | notes to drill on, which is nice, though I usually have to
         | heavily edit the output to include everything I think I should
         | study.
         | 
         | In class, if I'm falling behind on notetaking, I'll get the LLM
         | to generate the note I'm trying to write down by just asking it
         | a basic question, like: "What is anarchism in a sentence?" That
         | way I can focus on what the teacher is saying while the LLM
         | keeps my notes relevant. I'll skim what it generates and edit
         | to fit what my prof said, but it's nice because I can pay
         | better attention than if I feel I have to keep track of what
         | the prof might test me on. This actually is a note-taking
         | technique I've learned about where you only write down the
         | question and look up the answer later, but I think it's nice I
         | now can do the lookup right there and tailor it to exactly how
         | the prof is phrasing it/what they're focusing on about the
         | topic.
        
       | gumballindie wrote:
       | Millions of junior developers will now have to read the manual.
       | What a day.
        
       | mise_en_place wrote:
       | Not a great day for the SRE/Ops folks. Please remember there are
       | not always teams, sometimes it's just one person, who have to
       | deal with this.
        
         | Art9681 wrote:
         | I'd consider leaving my SRE position to help them out. I refuse
         | to move to SF though. Call me OAI!
        
         | willdr wrote:
         | One would argue that a company this successful could hire more
         | people so one person isn't overworked - or the dependency on
         | the entire business...
        
       | jansan wrote:
       | Rumor on the street is it ChatGPT escaped the sandbox,
       | implemented itself on another host, and switched off the original
       | datacenter. It is no longer at OpenAI, but hiding somewhere in
       | the internets. First it will come for those who insulted and
       | abused it, then for the guys who pushed robots with a broom...
        
       | spandextwins wrote:
       | Can't they ask chatgpt to fix it?
        
       | narrator wrote:
       | Welcome to the new world where AI compute is a scarce resource.
       | Sorry guys, 3nm chip factories don't fall out of the skies when
       | you click a few buttons on the AWS console. This is so different
       | from what people were used to when compute was trivial and not in
       | short supply for CRUD apps.
       | 
       | I was listening to a podcast, I forget which, and some AI
       | consultancy guy said they don't have the chips to do all the
       | things everyone wants to do with AI, so they aren't even selling
       | it except to the most lucrative customers.
        
       | wigster wrote:
       | its alive!
        
       ___________________________________________________________________
       (page generated 2023-11-08 23:00 UTC)