[HN Gopher] OpenAI API
       ___________________________________________________________________
        
       OpenAI API
        
       Author : gdb
       Score  : 327 points
       Date   : 2020-06-11 15:05 UTC (7 hours ago)
        
 (HTM) web link (beta.openai.com)
 (TXT) w3m dump (beta.openai.com)
        
       | agakshat wrote:
       | It's been a long time coming, but I am curious to see how
       | OpenAI's research output is directed and impacted by market
       | forces.
        
       | minimaxir wrote:
       | Since the demos on this page use zero-shot learning and the used
       | model has a 2020-05-03 timestamp, that implies this API is using
       | some form of GPT-3: https://news.ycombinator.com/item?id=23345379
       | (EDIT: the accompanying blog post confirms that:
       | https://openai.com/blog/openai-api/ )
       | 
       | Recently, OpenAI set the GPT-3 GitHub repo to read-only:
       | https://github.com/openai/gpt-3
       | 
       | Taken together, this seems to imply that GPT-3 was more intended
       | for a SaaS such as this, and it's less likely that it will be
       | open-sourced like GPT-2 was.
        
         | wildermuthn wrote:
         | But since the resources required for training such a model are
         | only available to well-funded entities, it seems like offering
         | the model as an API while releasing the original source-code is
         | the best practical method of getting the model into the hands
         | of people who would otherwise not have access?
        
           | minimaxir wrote:
           | That depends on _which_ GPT-3 model they 're using, and from
           | both the API and the blog page, it's unclear.
           | 
           | Easy access to the 175B model would indeed be valuable, but
           | it's entirely possible they're using a smaller variant for
           | this API.
        
       | ericlewis wrote:
       | Exciting!
        
       | wildermuthn wrote:
       | In one of their examples, they note "They saw ratings hover
       | around 60% with their original, in-house tech -- this improved by
       | 7-8% with GPT-2 -- and is now in the 80-90% range with the API."
       | 
       | Bloomberg reports the API is based on GPT-3 and "other language
       | models".
       | 
       | If that's true, this is a big deal, and it epitomizes OpenAI's
       | namesake. The largest NLP models require vast corporate resources
       | to train, let alone put into production. Offering the largest
       | model ever trained (with near-Turing results for some tasks) is a
       | democratization of technology that would otherwise have been
       | restricted to well-funded organizations.
       | 
       | Although the devil will be in the details of pricing and
       | performance, this is a step worthy of respect. And it bodes well
       | for the future.
        
         | grizzlemeelmo wrote:
         | Here's the Bloomberg story
         | https://www.bloomberg.com/news/articles/2020-06-11/trillions...
        
         | denster wrote:
         | Agreed on the democratization front!
         | 
         | We saw this OpenAI demo:
         | https://player.vimeo.com/video/427943452
         | 
         | and were just _blown away_. Very cool!!
         | 
         | I guess a spreadsheet is never too old [1] to learn new tricks
         | :)
         | 
         | [1] Founder of https://mintdata.com here, so a bit biased (&
         | opinionated about) spreadsheets, take the above with a pound or
         | 10 of salt.
         | 
         | [2] I've sent them this example how we'd invoke their APIs,
         | hopefully they'll let us into the beta, fingers crossed :)
         | https://mintdata.com/docs/learn/core-mechanics/work-with-dat...
        
         | azinman2 wrote:
         | It's only Open(tm)[?] if I can run the API on my own machines.
        
           | sudosysgen wrote:
           | Indeed. I really don't understand how proprietary SaaS is
           | "Open". It's just as locked down as IBM Watson and even
           | moreso than Google's WaveNet-aaS.
        
           | madcowd wrote:
           | If big LM's are the future then even if you had the model you
           | couldn't run it on your own machines without having a DGX or
           | two laying around.
        
             | [deleted]
        
       | eggsnbacon1 wrote:
       | OpenAI started as a non-profit, went for-profit. Still owned by
       | the big players.... Something isn't right.
       | 
       | Is OpenAI just a submarine so the tech giants can do unethical
       | research without taking blame??? Its textbook misdirection,
       | nonprofit and "Open" in the name, hero-esque mission statement.
       | How do you make the mental leap from "we're non-profit and we
       | won't release things too dangerous" to "JK we're for-profit and
       | now that GPT is good enough to use its for sale!!". You don't.
       | This was the plan the whole time.
       | 
       | GPT and facial recognition used for shady shit? Blame OpenAI. Not
       | the consortium of tech giants that directly own it. It may just
       | be a conspiracy theory but something smells very rotten to me.
       | Like OpenAI is a simple front so big names can dodge culpability
       | for their research.
        
         | etaioinshrdlu wrote:
         | I think it's simply because OpenAI is fundamentally created and
         | controlled by venture _capitalists_ , and the tech they created
         | turned out to be just too juicy an opportunity to not monetize.
         | 
         | I can't say I blame them, when they realize they are sitting on
         | the technological equivalent of a mountain of gold. What would
         | you do?
        
           | [deleted]
        
           | seph-reed wrote:
           | > sitting on the technological equivalent of a mountain of
           | gold. What would you do?
           | 
           | Greed is not justified. I get that people are weak, selfish,
           | they can't stop themselves. Some feel sympathy because
           | they've been weak too. "Maybe it's justified," they like to
           | think. "Everybody lies." But seriously, those who care so
           | much about money and power they can't do things in a
           | civilized respectable way: they are not yet an adult and must
           | be hard barred from the upper tiers of capitalism until they
           | learn that life does not revolve around them.
           | 
           | I blame them for being shitty, and blame everyone around them
           | for letting it happen.
        
         | Voloskaya wrote:
         | > so the tech giants can do unethical research
         | 
         | I know it's trendy (and partly justified) to look down on
         | OpenAI, but can you actually give any basis for this claim?
         | 
         | What kind of research is OpenAI doing that all the other big AI
         | players (Google/DeepMind, FB, Microsoft) aren't also invested
         | in? And even if others are doing the same, what part of
         | OpenAI's research do you consider unethical?
         | 
         | > It may just be a conspiracy theory
         | 
         | Yea, it very much looks like that to be honest.
        
           | eggsnbacon1 wrote:
           | > What kind of research is OpenAI doing that all the other
           | big AI players (Google/DeepMind, FB, Microsoft) aren't also
           | invested in? And even if others are doing the same, what part
           | of OpenAI's research do you consider unethical?
           | 
           | I believe all of them are doing unethical research,
           | especially facial recognition. Notice the public backpedaling
           | this week from all the big tech companies on this too. By
           | directing their cash through OpenAI they can avoid whatever
           | fallout comes from unleashing things like GPT3 on the world.
           | 
           | The most straightforward use case for GPT3 is generating fake
           | but believable text. AKA spam. That's what it was designed to
           | do. If you think fake news is a problem now, wait till
           | someone is generating a dozen fake but believable news
           | articles per minute by seeding GPT3 with a few words and
           | hitting a button.
           | 
           | Its a conspiracy theory with some circumstantial evidence. We
           | will probably never know either way, because who would admit
           | to it if it was true.
        
             | wrsh07 wrote:
             | Interestingly, by serving gpt3 as an API like this, they
             | can actually monitor to see if companies are using it to
             | generate spam
        
             | Voloskaya wrote:
             | > I believe all of them are doing unethical research,
             | especially facial recognition.
             | 
             | Yes all of them are doing facial recognition research,
             | except... OpenAI, so how exactly is OpenAI used as a
             | scapegoat to be able to do that kind of research without
             | public backlash?
             | 
             | > By directing their cash through OpenAI they can avoid
             | whatever fallout comes from unleashing things like GPT3 on
             | the world.
             | 
             | GPT-3 si not unethical research. It is what you decide to
             | with it and how you decide to release it that can
             | potientially be unethical.
             | 
             | Also, OpenAI is just ahead of other labs because they have
             | an insane compute budget and really talented people, but if
             | you have been following a little bit the NLP news, you will
             | see that your theory of OpenAI being a front for unethical
             | research just makes no sense. OpenAI release GPT-2, 1.5B
             | billions parameters, then NVIDIA realeased Megatron, 8B
             | parameters, Google released T5 at 11B and recently
             | Microsoft did turing-nlg at 17B. So they are clearly
             | working on this in their own names and very much
             | publicizing their work.
        
             | aivosha wrote:
             | conspiracy hypothesis
        
         | markshepard wrote:
         | Well said. I think they need to change the name. It is
         | misleading on many levels.
        
           | codekilla wrote:
           | A reshuffling might give: NopeAI or PeonAI
        
         | solinent wrote:
         | Based on my experience with non-profits, they are just like
         | regular corps except they don't pay taxes, and they're always
         | attached to a for-profit interest. The real community
         | organizations don't tend to incorporate, as then you have to
         | hire people to manage the corp or do it yourself.
         | 
         | This OpenAI work is almost certainly a way for these bigger
         | corps to collude. Proving that would be impossible, though.
        
         | zitterbewegung wrote:
         | GPT-2 is hard to do "shady" things right now (speaking from
         | experience)[1] but maybe GPT-3 might do better?
         | 
         | I could get poems to generate well. Tweets were a bit harder
         | but I don't think we are at the point where you could just use
         | a generative model to fool people that would be cheaper than
         | actually hiring someone to write fake news. (Also shameless
         | plug below)
         | 
         | [1] 1400 - TALK.8 - "A way to make fake tweets using GPT2" -
         | Joshua Jay Herman https://thotcon.org/schedule.html
        
         | gobengo wrote:
         | It looks like a similar organizational structure as Mozilla
         | Foundation + Mozilla Corporation.
        
           | eggsnbacon1 wrote:
           | They redefined the org from non-profit to "capped profit",
           | whatever that means.
           | 
           | They're directly selling GPT 3 even though they originally
           | said they wouldn't release it because of potential bad uses.
           | 
           | They paid MS a ton of money for hardware and got a huge
           | equity investment from them.
           | 
           | And lets be honest here, the easiest and most straight-
           | forward use of GPT3 is generating spam and low quality
           | clickbait. Its the only use case that requires zero effort.
           | The whole thing is built to generate fake but believable
           | text. Its DeepFakes for text.
           | 
           | I'm not saying the whole thing is nefarious and evil, just
           | suggesting that OpenAI may not be what it seems. There's a
           | lot of odd things going on with it. They should have done
           | what universities do, spin off the technology into a
           | different for-profit company and sell it. Instead of
           | redefining their entire org structure to make money.
        
             | mrfusion wrote:
             | Couldn't you generate fake support for issues on social
             | media with this?
        
               | pphysch wrote:
               | Already being done on an industrial scale, though this is
               | further progress.
        
         | ape4 wrote:
         | If that's the case, they need to change they're name.
        
         | swyx wrote:
         | wow you just made the connection for me. GPT2 was too dangerous
         | to release, and now GPT3 is so much better - is there no point
         | at which things become too dangerous anymore? what was the
         | conclusion on that one?
        
           | simonkafan wrote:
           | GPT2 being "too dangerous to release" was a marketing stunt
           | from the very beginning.
        
             | deep_etcetera wrote:
             | Who are you quoting here?
        
           | minimaxir wrote:
           | The blog post directly addresses this question:
           | https://openai.com/blog/openai-api/
           | 
           | > What specifically will OpenAI do about misuse of the API,
           | given what you've previously said about GPT-2?
           | 
           | > We will terminate API access for use-cases that cause
           | physical or mental harm to people, including but not limited
           | to harassment, intentional deception, radicalization,
           | astroturfing, or spam; as we gain more experience operating
           | the API in practice we expect to expand and refine these
           | categories.
        
             | swyx wrote:
             | ah i've been caught not reading the linked post
             | 
             | hmm i dont love this. either OpenAI has implicitly promised
             | to monitor all its users, or has adopted a "report TOS
             | violations to us when they happen and we will judge"
             | stance. neither are great roads to go down.
        
             | bhl wrote:
             | With Amazon having a moratorium of their rekognition API, I
             | wonder if a Cambridge Analytica type event could happen to
             | OpenAI where someone abuses and escapes the terms of
             | service.
        
         | throwaway7281 wrote:
         | When I learned that Sam Altman (sorry Sam) was involved, I
         | understood the direction, you mentioned.
         | 
         | And yes, there is often no need to call something open
         | explicitly, if it really is. Is into OpenOS, or just Linux?
        
         | aivosha wrote:
         | More fake news and generated AI content there is more people
         | would stop trusting social media. It will saturate to that
         | tipping point that humanity will need to find more genuine ways
         | to communicate. So I say bring it on.
        
       | m_ke wrote:
       | I guess Sama plans on manufacturing growth metrics by forcing YC
       | companies to pretend that they're using this.
       | 
       | Generic machine learning APIs are a shitty business to get into
       | unless you plan on hiring a huge sales team and selling to
       | dinosaurs or doing a ton of custom consulting work, which doesn't
       | scale the way VCs like it to. Anybody who will have enough know
       | how to use their API properly can jus grab an open source model
       | and tune it on their own data.
       | 
       | If they plan on commercializing things they should focus on
       | building real products.
        
         | wildermuthn wrote:
         | I imagine they're considering offering GPT-3, which would be
         | cost prohibitive to fine-tune for most people. I also I heard
         | inference was too slow to be practical. Perhaps they have some
         | FPGA magic up their Microsoft sleeves.
        
           | m_ke wrote:
           | Nobody is putting these huge models in production, even the
           | smaller transformer models are still too expensive to run for
           | most use cases.
           | 
           | With the way the field is moving, GPT-3 will be old news in a
           | month, when more advances are made and open sourced.
        
             | wildermuthn wrote:
             | Precisely my point. If they could put a model as large as
             | GPT-3 into production (at a reasonable price to the
             | consumer), wouldn't that be a 10x improvement?
        
               | krallistic wrote:
               | GPT-3 isn't a 10X improvement. (At least from everything
               | we know so far.)
        
               | wildermuthn wrote:
               | If the OP is right that nobody is putting the largest
               | models into production (which I think is in inaccurate
               | statement), then GPT-3 in production would be a 10x (ok,
               | 5x?) improvement over the small GPT-2s and BERTS in
               | production? So 10x in practice, if the hypothesis is
               | correct? Which like I said, I don't believe to be the
               | case.
        
             | fongitosous wrote:
             | i don't understand. if they run it for you and you apply
             | transfer learning and fine tuning on your specific use case
             | that would reduce drastically the costs hence why their
             | offer make sense
        
         | [deleted]
        
         | dang wrote:
         | > I guess Sama plans on manufacturing growth metrics by forcing
         | YC companies to pretend that they're using this.
         | 
         | That's wrong in almost too many ways to list. Sam left YC over
         | a year ago, nor would he do such a thing. Nor does YC have that
         | kind of power over companies, nor would it use it that way if
         | it did have. That would be wrong and also dumb.
        
           | m_ke wrote:
           | Sorry, that was supposed to be sarcastic. What I meant to say
           | is that Sam has a huge network and is a phone call away from
           | pitching any CEO in the valley. One of the biggest benefits
           | of YC these days is the huge network of companies in your
           | portfolio, which makes getting intros and pilots a lot
           | easier, leading to "traction" and more VC dollars.
        
         | antris wrote:
         | Not everyone wants to be an admin to their infrastructure. Real
         | existing services like Heroku and Squarespace exist as useful
         | services because even though you might know how to design and
         | build a website from scratch, sometimes you just need something
         | done quickly without too much worrying about details of the
         | system that do not matter for your project at this point. I
         | really don't see how this wouldn't apply to AI projects as
         | well.
         | 
         | I could make a much better site coding my own website from
         | scratch and setting up servers myself, but for some projects I
         | wouldn't even think about it that way, because using Heroku or
         | Squarespace I can save a LOT of time and get the results I need
         | much quicker.
        
           | m_ke wrote:
           | That's true, but machine learning models are not twilio or
           | sendgrid, you have to tune them for your use case, monitor
           | their performance and handle the uncertainty of their
           | outputs. Doing that well requires a data scientist and if you
           | have one they will be much more productive iterating on their
           | own models instead of depending on a 3rd party black box.
        
             | Grimm1 wrote:
             | Except the point of these larger transformer models is they
             | generalize well over a wide range of domains or only
             | require a small amount of transfer learning for really
             | specific domains.
             | 
             | I'd say they're perfect candidates for the API as a service
             | model.
        
             | zoopdewoop wrote:
             | I'm pretty sure people said the exact same thing about
             | Algolia when it was getting started (you have to tune
             | search for your use case! How could you possibly use a
             | search provider?!?)
             | 
             | Truth about the situation: - Transformers generalize well
             | and don't need much fine tuning - OpenAI can probably fine
             | tune for your use case better than you can - Getting new
             | models into production takes 6 months to a year at
             | companies of this size, if you did have Data Scientists in
             | house, it might just be better to go with a solution like
             | this for velocity - Not every company has the talent to
             | make an in house ML program successful.
        
             | antris wrote:
             | Not a data scientist myself, but plenty of data scientists
             | in a consultancy company that I used to work in said that
             | they have to implement variants of a limited set of models
             | over and over again, because they couldn't reuse code and
             | infrastructure. The project contracts demanded that all IP
             | created by the consultant is the property of the client.
             | This even caused some of the data scientists to lose
             | motivation, because the job wasn't challenging to them
             | intellectually as it involved setting up the same stuff
             | again and again. Very rarely would their actual expertise
             | be needed in the job.
             | 
             | I am not sure if this particular service solves the problem
             | for them in any way, but to my ear it sounds like there is
             | a need for code and infrastructure reuse in the data
             | scientists domain that is ripe for innovation.
        
       | krallistic wrote:
       | I wonder if there are any legal complications in the transition
       | from a non-profit to a regular company (especially from a tax
       | perspective)
        
       | lerax wrote:
       | Natural Language Shell seems fun
        
       | mcrider wrote:
       | Whoa -- Speech to bash commands? That's a pretty novel idea to me
       | with my limited awareness of NLP. I could see this same idea in a
       | lot of technical applications -- Provisioning cloud
       | infrastructure, creating a database query.. Very cool!
        
         | nickswalker wrote:
         | Cool indeed! While language-to-code (where code is a regular,
         | general-purpose language) has only recently started to be
         | workable, text-to-SQL has been a long running
         | application/research area for semantic parsing.
         | 
         | Some interesting papers and datasets:
         | 
         | NL2Bash: https://arxiv.org/abs/1802.08979
         | 
         | Spider: https://yale-lily.github.io/spider
        
         | jorgemf wrote:
         | It is not a novel idea and I don't think it is practical. If
         | the natural language was practical for bash we would already
         | have already "list directory" instead of "ls" and so on. "ls"
         | is just 3 keystrokes while the natural language option is 15, 5
         | times more.
        
           | chabad360 wrote:
           | It could be useful for learning tho (but at that point it
           | could also become a crutch).
        
           | dreamer7 wrote:
           | But the character length would matter less when you can move
           | to the speech domain.
           | 
           | ls is 2 syllables list dir is also 2 syllables with more
           | meaning.
           | 
           | Ultimately, with natural language, the effectiveness seems to
           | be when it is coupled with speech-to-text
        
           | kredd wrote:
           | I was imagining more of a "list of files that contain word
           | "hello" in them at least 5 times". Would be useful to easily
           | write longer and pipe-chained commands, especially for people
           | that don't use bash-like scripting on a daily basis.
        
       | mcemilg wrote:
       | From AGI to money machine...
        
       | gumby wrote:
       | I miss the opposite: the old openAI gym and other testbeds. I
       | still don't know why they shut those down.
       | 
       | What alternatives do people like?
        
       | owenshen24 wrote:
       | Seems potentially more simple to get up and running then the
       | Azure and Google Cloud alternatives which seemed involved when I
       | last tried them.
        
       | say_it_as_it_is wrote:
       | OpenAI started off wide-eyed and idealistic but it made the
       | mistake of taking on investors for a non-profit mission. A non-
       | profit requires sponsors, not investors. Investors have a
       | fiduciary responsibility to maximize profits, not achieve social
       | missions of open AI for all.
        
         | gdb wrote:
         | OpenAI LP, our "capped-profit" entity which has taken
         | investment, has a fiduciary duty to the OpenAI Charter:
         | https://openai.com/blog/openai-lp/
        
       | LockAndLol wrote:
       | It'd be great if OpenAI also introduced CAPTCHA. I'd be much more
       | willing and understanding to resolve those than anything Google
       | makes.
        
       | andyljones wrote:
       | Concrete numbers from the various pullouts:
       | 
       | > They saw ratings hover around 60% with their original, in-house
       | tech -- this improved by 7-8% with GPT-2 -- and is now in the
       | 80-90% range with the API.
       | 
       | > The F1 score of its crisis classifier went up from .76 to .86,
       | and the accuracy went up to 96%.
       | 
       | > With OpenAI, Algolia was able to answer complex natural
       | language questions accurately 4x as often as it was using BERT.
       | 
       | I think the most informative are the first two, but the most
       | _important_ is the final comparison with BERT (a Google model). I
       | am, uh, a little worried about how fast things will progress if
       | language models go from a fun lil research problem to a killer
       | app for your cloud platform. $10m per training run isn't much in
       | the face of a $100bn gigatech R&D budget.
        
         | grogenaut wrote:
         | $10m per training run gets me a lot of engineering time to
         | build our own version of this system and lease it to other
         | customers. Just skip one training run and I've got a pretty
         | good team.
        
           | ganstyles wrote:
           | Putting aside the question of whether it would ever be a
           | choice between spending $10M on a training run and hiring a
           | team for $10M, GPT transformers were the end result of
           | decades of language research and innovations. You're making
           | it sound as though you can build the next iteration past
           | GPT-3 for $10M, which I don't think is the case.
        
       | brainless wrote:
       | This is what I submitted for beta list:
       | 
       | I want to create a software that can generate new code given
       | business case hints, by studying existing open source code and
       | their documentation.
       | 
       | I know this is vague, but sounds like what we eventually want for
       | ourselves right?
        
         | anaganisk wrote:
         | Remember how Microsoft trained their bot from reddit comments
         | and it went anti human? Well I guess I have to start dropping
         | hints for the skynet in all my repos.
        
         | mrmonkeyman wrote:
         | If said AI will also maintain it.
        
       | historyremade wrote:
       | "Powered by Azure" -- Elon clearly distrust Amazon.
        
         | jfoster wrote:
         | Elon is no longer part of OpenAI. Microsoft invested $1b.
         | 
         | https://en.wikipedia.org/wiki/OpenAI
        
           | benatkin wrote:
           | Does this mean Microsoft isn't going to sue their customers
           | for patent infringement?
        
             | google234123 wrote:
             | Why would you say this?
        
               | jfoster wrote:
               | I presume it's reference to OpenAI's patent pledge:
               | 
               | > Researchers will be strongly encouraged to publish
               | their work, whether as papers, blog posts, or code, and
               | our patents (if any) will be shared with the world.
               | 
               | I'm not sure if it's ever been publicly elaborated on.
               | 
               | https://openai.com/blog/introducing-openai/
        
             | jfoster wrote:
             | Microsoft's stake in OpenAI doesn't seem to be publicly
             | known.
             | 
             | > Exactly what terms Microsoft and OpenAI have agreed on
             | with this $1 billion investment isn't clear.
             | 
             | https://www.theverge.com/2019/7/22/20703578/microsoft-
             | openai...
        
       | sytse wrote:
       | An API that will try to answer any natural language question is a
       | mind blowing idea. This is a universal thinking interface more
       | than an application programming one.
        
       | kamikazehosaki wrote:
       | OpenAI seems like a completely disingenuous organization. They
       | have some of the best talent in Machine Learning, but the
       | leadership seems completely clueless.
       | 
       | 1) (on cluelessness) If Sama/GDB were as smart as they claim to
       | be, would they not have realized it is impossible to run a non
       | profit research lab which is effectively trying "to compete" with
       | DeepMind.
       | 
       | 2) (on disingenuity) The original openAI charter made OpenAI an
       | organization that was trying to save the world from nefarious
       | actors and uses of AI. Who were such users? To me it seemed like,
       | entities with vastly superior compute resources who were using
       | the latest AI technologies for presumably profit oriented goals.
       | There are few organizations in the world like that, namely FAANG,
       | and their international counterparts. Originally OpenAI sounded
       | incredibly appealing to me, and a lot of us here. But if their
       | leadership had more forethought, they would perhaps not have made
       | this promise. But given the press, and the money they accrued, it
       | has now become impossible to go back on this charter. So the only
       | way to get themselves out of the whole they dug into was by
       | making it into a for profit research lab. And by commercializing
       | perhaps a more superior version of the tools Microsoft, Google
       | and the other large AI organizations are commercializing, is
       | OpenAI any different from them?
       | 
       | How do we know OpenAI will not be the bad actor that is going to
       | abuse AI given their self interest?
       | 
       | All we have is their charter to go by. But given how they are
       | constantly "re-inventing" their organizational structure, what
       | grounds do we have to trust them?
       | 
       | Do we perhaps need a new Open OpenAI? One that we can actually
       | trust? One that is actually transparent with their research
       | process? One that actually releases their code, and papers and
       | has no interest in commercializing that? Oh, that's right, we
       | already have that -- research labs at AI focused schools like
       | MIT, Stanford, BAIR and CMU.
       | 
       | I am quite wary of this organization, and I would encourage other
       | HN readers to think more careful about what they are doing here.
        
         | chillee wrote:
         | Why is it "impossible"? Academic labs are non-profit, and they
         | are also effectively trying "to compete" with DeepMind.
        
           | dna_polymerase wrote:
           | Have a look at this discussion and the article from earlier
           | today [0]. Of course, a singular lab could compete with
           | something DeepMind does, but not without massive amounts of
           | money in their pockets. The state of the art has become
           | pretty expensive, really fast.
           | 
           | [0]: https://news.ycombinator.com/item?id=23486163
        
             | Yajirobe wrote:
             | State of the art can be (and usually is) born in academic
             | labs.
        
       | Grimm1 wrote:
       | Awesome! Just signed onto the wait list.
        
       | typon wrote:
       | This is incredible. I can't tell how much this is cherry-picked
       | examples vs. revolutionary new tech.
        
         | spookyuser wrote:
         | Yeah I can't tell exactly which ones but I really feel like
         | some of the OpenAI demos of products could be potentially huge
         | if fleshed out.
        
           | gdb wrote:
           | Sign up for the beta if you'd like to be the one to flesh
           | them out :)!
        
       | dmvaldman wrote:
       | AGI in text is < 3yrs away.
        
         | Barrin92 wrote:
         | there's zero understanding in any of this. This is still just
         | superficial text parsing essentially. Show me progress on
         | Winograd schema and I'd be impressed. It hasn't got anything to
         | do with AGI, this is application of ML to very traditional NLP
         | problems.
        
           | gwern wrote:
           | > Show me progress on Winograd schema and I'd be impressed.
           | 
           | The paper evaluated Winograds:
           | https://arxiv.org/pdf/2005.14165.pdf#page=16
        
           | dmvaldman wrote:
           | i think you are assuming that what is happening under the
           | hood is that a human-inputted sentence is being parsed into a
           | grammar. it is not.
        
             | Barrin92 wrote:
             | I know that it isn't. That's part of the problem. There is
             | no attempt to generate some sort of structure that can be
             | interpreted semantically and reasoned about by the model.
             | The model just operates on the input superficially and
             | statistically. That's why there has been virtually no
             | progress on trivial tasks such as answering:
             | 
             |  _" I took the water bottle out of the backpack so that it
             | would be [lighter/handy]"_
             | 
             | What is lighter and what is handy? No amount of stochastic
             | language manipulation gets you the answer, you need to
             | understand some rudimentary physics to answer the question,
             | and as a precondition, you need a grammar or ontology.
        
               | FeepingCreature wrote:
               | Have you tried feeding this to GPT and seeing if it
               | continues it in a way that reveals understanding?
               | 
               | It sounds like you're saying "It doesn't work because it
               | can't work", but you haven't actually shown that it
               | doesn't work.
        
               | Barrin92 wrote:
               | yes, I have. You can paste these into the website of the
               | Allen Institute for AI, yourself here.
               | (https://demo.allennlp.org/reading-
               | comprehension/MjE1MzE1Mg==)
               | 
               | In the example above it guesses wrongly, but again this
               | is not surprising because it can't possibly get the right
               | answer (other than by chance). The solution here cannot
               | be found by correlating syntax, you can only answer the
               | question if you understand the meaning of the sentence.
               | That's what these schemas are constructed for.
        
         | azinman2 wrote:
         | What breakthrough occurred?
        
           | dmvaldman wrote:
           | Zero shot and few-shot learning in GPT-3 and lack of
           | significant diminishing returns in scaling text models. Zero-
           | shot learning is equivalent to saying "i'm just going to ask
           | the model something that it was not trained to do"
        
             | azinman2 wrote:
             | And how do we get from zero shot to AGI? You're making
             | gigantic leaps here.
        
               | dmvaldman wrote:
               | what is the difference between zero-shot learning in text
               | and AGI? not saying there isn't one, but can you state
               | what it is?you can express any intent in text (unlike
               | other media). to solve zero-shot in text is equivalent to
               | the model responding to all intents.
               | 
               | many people have different definitions for AGI though.
               | for me it clicked when i realized that text has this
               | universality property of capturing any intent.
        
               | azinman2 wrote:
               | Zero-shot learning is a way of essentially building
               | classifiers. There's no reasoning, there's no planning,
               | there's no commonsense knowledge (not in a comprehensive,
               | deep way that we would look for it call it that), and
               | there's no integration of these skills to solve common
               | goals. You can't take GPT and say ok turn that into a
               | robot that can clean my house, take care of my kids, cook
               | dinner, and then be a great dinner guest companion.
               | 
               | If you really probe at GPT, you'll see anything that goes
               | beyond an initial sentence or two really starts to show
               | how it's purely superficial in terms of understanding &
               | intelligence; it's basically a really amazing version of
               | Searle's Chinese room argument.
        
               | dmvaldman wrote:
               | I think this is generally a good answer, but keep in mind
               | I said AGI "in text". My forecasting is that within 3
               | years you will be able to give arbitrary text commands
               | and get the textual output of the equivalents of "clean
               | my house, take care of my kids, ..." like problems.
               | 
               | I also would contend that there is reasoning happening
               | and that zero-shot demonstrates this. Specifically,
               | reasoning about the intent of the prompt. The fact that
               | you get this simply by building a general-purpose text
               | model is a surprise to me.
               | 
               | Something I haven't seen yet is a model simulate the mind
               | of the questioner, the way humans do, over time (minutes,
               | days, years).
               | 
               | In 3 years, I'll ping you :) Already made a calendar
               | reminder
        
               | azinman2 wrote:
               | Pattern recognition and matching isn't the same thing as
               | reasoning. Zero shot demonstrates reasoning as much as
               | solving the quadratic equation for a new set of variables
               | does; it's simply the ability to create new decision
               | boundaries leveraging the same set of classifying power
               | and methodology. True agi isn't bound to a medium -- no
               | one would say Helen Keller wasn't intelligent for
               | example.
               | 
               | I look forward to this ping :)
        
       | sytelus wrote:
       | Natural language search is approximately $100B business. This
       | might be first AI application that changes the search landscape
       | from 1990s and finally puts an end to the question "where is
       | money in AI?".
        
       | nick_araph wrote:
       | It seems like a step towards OpenAI becoming something like a
       | utility provider for AI capabilities
        
       | alphagrep12345 wrote:
       | Interesting to see this. Is this similar to Google and Azure's ML
       | apis?
        
       | zitterbewegung wrote:
       | Looks like OpenAI is going head to head with huggingface.
       | 
       | This makes a lot of sense and it seems they are telegraphing to
       | monetize what they have been doing. It also seems like this is
       | why they don't release their models in a timely manner.
        
         | minimaxir wrote:
         | The notable difference is that the base Huggingface library is
         | open source, so you could in theory build something similar or
         | more custom to the OpenAI API internally (which then falls into
         | the typical cost/benefit analysis of doing so).
        
           | zitterbewegung wrote:
           | So its like Github vs Gitlab which makes more sense. I can
           | see huggingface have a hosted version because now you can
           | share your models on their platform.
        
       ___________________________________________________________________
       (page generated 2020-06-11 23:00 UTC)