[HN Gopher] OpenAI disbands its robotics research team ___________________________________________________________________ OpenAI disbands its robotics research team Author : morty_s Score : 76 points Date : 2021-07-16 21:03 UTC (1 days ago) (HTM) web link (venturebeat.com) (TXT) w3m dump (venturebeat.com) | Jack000 wrote: | Makes sense I guess, integrating robot hardware requires an | entirely different set of skills to ML research and has a much | slower dev cycle. | | I think OpenAI has progressively narrowed down its core | competency - for a company like 3M it would be something like | "applying coatings to substrates", and for OpenAI it's more like | "applying transformers to different domains". | | It seems like most of their high-impact stuff is basically a big | transformer: GPT-x, copilot, image gpt, DALL-E, CLIP, jukebox, | musenet | | their RL and gan/diffusion stuff bucks the trend, but I'm sure | we'll see transformers show up in those domains as well. | varelse wrote: | Fascinating in the wake of Fei Fei Li's lab publishing | significant work on embodied intelligence... | | https://arxiv.org/abs/2102.02202 | | Not to mention a bunch of relatively inexpensive reinforcement | learning research relying on consumer knockoffs of Spot from | Boston Dynamics... | | Really does seem like they are following the money and while | there's nothing wrong with that it's also nothing like their | original mission. | [deleted] | coolspot wrote: | The team was probably replaced by GPT-4. No need for humans to | slow down great mind. | wly_cdgr wrote: | This feels like a strong sign that AGI is quite close now | amerine wrote: | Why do you think that? | wly_cdgr wrote: | They smell the urgency in the air, and they're close enough | to the center to get a good and accurate whiff | abeppu wrote: | How on earth would you know if a whiff was accurate, when | we're talking about something which has never before been | created? | | I think even if you have intuitions about an approach, and | have promising results, if you're trying to arrive at | something new, it's really hard to know how far away you | are. | wly_cdgr wrote: | It's just a hunch, no need to get your boxers in a bunch | fartcannon wrote: | This is lunacy. The first country/company to replace human labour | with general bipedal robots, will reap wealth beyond imagination. | The short sitedness is astonishing, if you ask me. | | I genuinely believe how we as a society act once human labour is | replaced is first aspect of the great filter. | tejtm wrote: | There are no mechanisms in place for the generated wealth to | benefit the replaced people, the wealth will go mainly to | vanishingly few persons self selected to be okay with gross | economic inequality. | | We have been at this since at least the dawn of the industrial | revolution and do not have it right yet. Backing off and taking | it slow now to let some cultural adjustments happen is a | responsible step. | | My cultural norms are repulsed by the thought of me not working | as much as possible, it is how I expect my value to society to | be gauged (and rewarded). | | This line of reasoning will be (is) obsolete and we need | another in its place globally. | | I hope some may have better ideas of what these new cultural | norms should look like than I with my too much traditional | indoctrination. | | I only know what I will not have it look like; humanity as | vassals of non corporeal entities or elites. | joe_the_user wrote: | _There are no mechanisms in place for the generated wealth to | benefit the replaced people, the wealth will go mainly to | vanishingly few persons self selected to be okay with gross | economic inequality._ | | That hasn't stopped the march of progress so far. | Conveniently (or not), humanoid robots do not appear likely | for the foreseeable future. But keep worrying, the problem | you list are appearing in other fashions anyway. | ragebol wrote: | > replace human labour with general bipedal robots | | No need for bipeds, car factories employ dumb robot arms, no | humans needed. Not very general purpose robots though. | | The first country/company to create robots that can be | instructed similar to a humans to do any job will indeed have | great benefits, but how long until that happens? Not within any | amount of time that an investor wants to see. I'm unsure if I | will ever see that in my life (counting on ~60 years to go | still maybe?) | TaylorAlexander wrote: | One thing that struck me recently was that the famous | imagenet competition that was won by a neural net took place | in 2012. So we have made fantastic advances in ten years. But | I'd still say at best robots like you describe are 20 years | away, and that's a long time horizon for a small | organization. | ragebol wrote: | Has robotics had such an 'ImageNet moment'? Nothing springs | to mind, just slow advancement over decades. | | If suddenly robot manipulators could grasp any object, | operate any knob/switch, tie knots, manipulate cloth, with | the same manipulator, on first sight, that would be quite a | feat. | | But then there's still task planning which is a very | different topic. And ... and .... So much still to develop | for generally useful robots. | TaylorAlexander wrote: | Not yet. I have a four wheel drive robot I designed with | four 4k cameras feeding in to an Nvidia Jetson Xavier. | [1] | | Just getting it to navigate itself using vision would | mean building a complex system with a lot of pieces | (beyond the most basic demo anyway). You need separate | neural nets doing all kinds of different tasks and you | need a massive training system for it all. You can see | how much work Tesla has had to do to get a robot to | safely drive on public roads. [2] | | From where I am sitting now, I think we are making good | inroads on something like an "Imagenet moment" for | robots. (Well, I should note that I am a robotics | engineer but I mostly work on driver level software and | hardware, not AI. Though I follow the research from the | outside.) | | It seems like a combination of transformers plus scale | plus cross domain reasoning like CLIP [3] could begin to | build a system that could mimic humans. I guess as good | as transformers are we still haven't solved how to get | them to learn for themselves, and that's probably a hard | requirement for really being useful in the real world. | Good work in RL happening there though. | | Gosh, yeah, this is gonna take decades lol. Maybe we will | have a spark that unites all this in one efficient | system. Improving transformer efficiency and achieving | big jumps in scale are a combo that will probably get | interesting stuff solved. All the groundwork is a real | slog. | | [1] https://reboot.love/t/new-cameras-on-rover/277 | | [2] https://www.youtube.com/watch?v=hx7BXih7zx8 | | [3] https://openai.com/blog/clip/ | brutus1213 wrote: | I am a researcher on the AI/Systems side and I wanted to | chime in. Transformers are amazing for language, and have | broken all the SOTA is many areas (at the start of the | year, some people may have wondered if CNNs are dead | [they are not as I see it]). The issue with Transformer | models is the insane amount of data they need. There is | some amazing progress on using unsupervised methods to | help, but that just saves you on data costs. You still | need an insane about of GPU horsepower to train these | things. I think this will be a bottleneck to progress. | The average university researcher (unless from tier 1 | school with large funding/donors) are going to pretty | much get locked out. That basically leaves the 5-6 key | corporate labs to take things forward on the transformer | front. | | RL, which I think this particular story is about, is an | odd-duck. I have papers on this and I personally have | mixed feelings. I am a very applications/solutions- | oriented researcher and I am a bit skeptical about how | pragmatic the state of the field is (e.g. reward function | specification). The argument made by the OpenAI founder | on RL not being amenable to taking advantage of large | datasets is a pretty valid point. | | Finally, you raise interesting points on running multiple | complex DNNs. Have you tried hooking things to ROS and | using that as a scaffolding (I'm not a robotics guy .. | just dabble in that as a hobby so curious what the | solutions are). Google has something called MediaPipe, | which is intriguing but maybe not what you need. I've | seen some NVIDIA frameworks but they basically do pub-sub | in a sub-optimal way. Curious what your thoughts are on | what makes existing solutions insufficient (I feel they | are too!) | TaylorAlexander wrote: | Great comment thank you. | | Yes unless the industry sees value in a step change in | the scale on offer to regular devs, progress on massive | nets will be slow. | | Hooking things together is pretty much my job. I have | used ROS extensively in the past but now I just hook | things together using python. | | But I consider what Tesla is doing to be pretty | promising, and they are layering neural nets together | where the output of three special purpose networks feed | in to one big one etc. They call that a hydra net. No | framework like ROS is required because each net was | trained in situ with the other nets on the output of | those nets, so I believe all compute logic is handled | within the neural network processor (at some point they | integrate standard logic too but a lot happens before | that). Definitely watch some Karpathy talks on that. | | And currently I am simply not skilled enough to compose | multiple networks like that. So I _could_ use multiple | standalone networks, process them separately, and link | them together using IPC of some kind, but it would be | very slow compared to what 's possible. That's why I say | we're "not there yet". Something like Tesla's system | available as an open source project would be a boon, but | the method is still very labor intensive compared to a | self-learning system. It does have the advantage of being | modular and testable though. | | I probably will hand compose a few networks (using IPC) | eventually. I mean right now I am working on two networks | - an RL trained trail following network trained in | simulation on segmentation-like data (perhaps using | Dreamer V2), and a semantic segmentation net that is | trained on my hand labeled dataset with "trail/not-trail" | segmentation. So far my segmentation net works okay. And | a first step will actually be to hand-write an algorithm | to go from segmentation data to steering. My simulation | stuff is almost working. I built up a training | environment using Godot video game engine and hacked the | shared memory neural net training add on to accept image | data, but when I run the sim in training on DreamerV2, | something in the shared memory interface crashes and I | have not resolved it. [1] | | But all of this is a hobby and I have a huge work project | [2] I am managing myself that is important to me, so the | self driving off road stuff has been on pause. But I | don't stress about it too much because the longer I wait, | the better my options get on the neural network side. | Currently my off road rover is getting some mechanical | repairs, but I do want to bring it back up soon. | | [1] https://github.com/lupoglaz/GodotAIGym/issues/15 | | [2] https://community.twistedfields.com/t/a-closer-look- | at-acorn... | brutus1213 wrote: | First off, amazing farm-bot project! I am looking forward | to reading the details on your site. | | Thx for the pointers on Tesla. Had not seen the Hydranet | stuff. There was a Karpathy talk about 2 weeks back at a | CVPR workshop .. he revealed the scale of Tesla's current | generation deep learning cluster [1]. It is insane! | Despite being in industrial research, I don't foresee | ever being able to touch a cluster like that. | | A lot of our current research involves end-to-end | training (some complex stuff with transformers and other | networks stitched together). There was a CVPR tutorial on | autonomous driving [2], where they pretty much said | autonomy 2.0 is all about end-to-end. I've spoken to a | few people who actually do commercial autonomy, and they | seemed more skeptical on whether end2end is the answer in | the near-term. | | One idea we toy with is to use existing frozen | architectures (OpenAI releases some and so do other big | players) and do a small bit of fine-tuning. | | [1] https://www.youtube.com/watch?v=NSDTZQdo6H8 [2] | https://www.self-driving-cars.org/ | toxik wrote: | Imagine that there only needs to be ten people to "run the | world". What is the population size going to be then? Ten? As | large as possible? Somehow it seems that the way we're headed, | it'll be ten plus some administrative overhead. | kadoban wrote: | The way we're headed it'll be billions in misery and dozens | in luxury. | Zababa wrote: | > The first country/company to replace human labour with | general bipedal robots, will reap wealth beyond imagination. | | Humans ARE genral bipedal robots. The price of these robots is | determined by the minimum wage. | nradov wrote: | We are decades away from being able to build a general bipedal | robot that can snake out a plugged toilet or dig a trench or | nail shingles to a roof. It's just not a rational goal yet. Aim | lower. | TaylorAlexander wrote: | This is correct. Right now our best and brightest can only | build demos that fall apart the moment something is out of | place. Humanoid or even partial humanoid (wheeled base) | robots are far from ready for general purpose deployment. | Animats wrote: | And we're further away since nobody bought Schaft from | Google, and Schaft was shut down. They had the best humanoid. | | But so many of the little problems have been solved. | Batteries are much better. Radio data links are totally | solved. Cameras are small and cheap. 3-phase brushless motors | are small and somewhat. Power electronics for 3-phase | brushless motors is cheap. 3D printing for making parts is | cheap. | | I used to work on this stuff in the 1990s. All those things | were problems back then. Way too much time spent on low-level | mechanics. | | You can now get a good legged dog-type robot for US$12K, and | a good robot arm for US$4K. This is progress. | joe_the_user wrote: | I basically agree. | | I'd just note that "decades away" means "an unforeseeable | number of true advances away" - which could mean ten years or | could mean centuries. | | And private companies can't throw money indefinitely at | problems others have been trying to solve and failing at. | They can it once and a while but that's it. | throwaway_45 wrote: | If robots are doing all the work how will people make money to | buy the stuff the robots make? Is Jeff Bezos going to own the | whole world or are we going to have another French revolution? | TaylorAlexander wrote: | We should really endeavor to build collectively owned | institutions that can purchase and operate the robots (and | physical space) we depend on. | | EDIT: Imagine the "credit unions" I mention in the following | linked comment, but holding homes and manufacturing space to | be used by members. | https://news.ycombinator.com/item?id=27860696 | xnx wrote: | Interesting contrast to another story today: | https://ai.googleblog.com/2021/07/speeding-up-reinforcement-... | ansk wrote: | Is the prevailing opinion that progress in reinforcement learning | is dependent on algorithmic advances, as opposed to simply | scaling existing algorithms? If that is the case, I could see | this decision as an acknowledgement that they are not well | positioned to push the frontier of reinforcement learning - at | least not compared to any other academic or industry lab. Where | they have seen success, and the direction it seems they are | consolidating their focus, is in scaling up existing algorithms | with larger networks and larger datasets. Generative modeling and | self supervised learning seem more amenable to this engineering- | first approach, so it seems prudent for them to concentrate their | efforts in these areas. | abeppu wrote: | I think the premise of your question actually points to the | real problem. In RL, b/c your current policy and actions | determine what data you see next, you can't really just "scale | existing algorithms" in the sense of shoving more of the same | data through them on more powerful processors. There's a | sequential process of acting/observing/learning which is | bottlenecked on your ability to act in your environment (ie | through your robot). Off-policy learning exists, but scaling up | the amount of data you process from a bad initial policy | doesn't really lead anywhere good. | andyxor wrote: | Reinforcement learning itself is a dead-end on a road to AI. | They seem to slowly starting to realize it, probably ahead of | academia. | TylerLives wrote: | What's the alternative? | nrmn wrote: | Why do you believe this to be the case? | andyxor wrote: | In a nutshell it's too wasteful in energy spent and it | doesn't even try to mimic natural cognition. As physicists | say about theories hopelessly detached from reality - "it's | not even wrong". | | The achievements of RL are so dramatically oversold that it | can probably be called the new snake oil. | vladTheInhaler wrote: | I'm going to need you to unpack that a bit. Isn't | interacting with an environment and observing the result | exactly what natural cognition does? What area of machine | learning do you feel is closer to how natural cognition | works? | kirill5pol wrote: | Maybe true if you consider policy gradient methods and Q | learning the only things that exist in RL... it's a pretty | wide field that encompasses a lot more than the stuff OpenAI | puts out. | nrmn wrote: | Yes, it feels like we have squeezed most of the performance out | of current algorithms and architectures. OpenAI and deepmind | have thrown tremendous compute against the problem with little | overall progress (overall, alpha go is special). There was a | big improvement in performance by bringing in function | approximators in the form of deep networks. Which as you said | can scale upwards nicely with more data and compute. In my | opinion as an academic in the deep RL, it feels like we are | missing some fundamental pieces to get another leap forward. I | am uncertain what exactly the solution is but any improvement | in areas like sample efficiency, stability, or task transfer | could be quite significant. Personally I'm quite excited about | the vein of learning to learn. | an_opabinia wrote: | > alpha go is special | | The VC community is in denial about how much Go resembled a | problem purpose built to be solved by deep neural networks. | dougSF70 wrote: | Designing robots to pick fruit and make coffee / pizzas cannot | have a positive ROI until labor laws make the bsuiness-case for | them. Majority of use cases where we can use robots for | activities humans cannot perform (fast spot welding on production | line, moving nuclear fuel rods, etc) have been solved. It is | smart to focus on language and information processing, given that | we are producing so much more of it, everyday. | cweill wrote: | I think the comments are confounding shutting down the robotics | research team with eliminating all RL research. Most robotics | teams don't use data-hungry RL algorithms because the cost of | interacting with the environment is so expensive. And even if the | team has a simulator that can approximate the real world to | produce infinite data, there is still the issue of the | "simulator-gap" with the real world. | | I don't work for openAI but I would guess they are going to keep | working on RL (e.g hide and seek, gym, DoTA style Research) to | push the algorithmic SoTA. But translating that into a physical | robot interacting with the physical world is extremely difficult | and a ways away. | samstave wrote: | Curious idea: | | With the mentioning that they can shift their focus to domains | with extensive data that they can build models of action with | etc... Why not try the following (If easily possible) | | --- | | Take all the objects on the various 3D warehouses (thingiverse, | and all the other 3d modeling repos out there) -- and have a | system whereby an OpenAI 'Robotics' platform can virtually learn | to manipulate and control a 3D model | (solidworks/blender/whatever) and learn how to operate it. | | It would be amazing to have an AI robotics platform where you | feed it various 3D files of real/planned/designed machines, and | have it understand the actual constituancy of the components | involved, then learn its degrees of motion limits, or servo | inputs etc... and then learn to drive the device. | | Then, give it various other machines which share component types, | built into any multitude of devices - and have it eval the model | for familiar gears, worm-screws, servos, motors, etc... and have | it figure out how to output the controller code to run an actual | physically built out device. | | Let it go through thousands of 3D models of things and build a | library of common code that can be used to run those components | when found in any design.... | | Then you couple that code with Copilot and allow for people to | have a codebase for controlling such based on what OpenAI has | already learned.... | | As Copilot is already built using a partnership with OpenAI... | marcinzm wrote: | I suspect it's because at a certain point detailed physics | matters and simulating things well enough is really hard. A | robotic arm might flex just a bit, a gear may not mesh quite | correctly, signals may take just a bit longer to get somewhere, | a grip might slip, a plastic object might break from too much | force, etc, etc. | robotresearcher wrote: | Sounds like a perfect domain to explore robust methods that | can't overfit to silly details. | verall wrote: | NVIDIA Isaac sounds very close to what you're describing. | adenozine wrote: | I'm sure the overhead and upkeep of a robotics lab far outweighs | that of a computer lab for software research. | | Are there any Open* organizations for robotics that could perhaps | fill the void here? I think robotics is really important and I | think the software is a big deal also, but it's important that | actual physical trials of these AIs are pursued. I would think | that seeing something in real space like that offers an | unparalleled insight for expert observers. | | I remember the first time I ever orchestrated a DB failover | routine, my boss took me into the server room when it was | scheduled on the testing cluster. Hearing all the machines spin | up and the hard drives start humming, that was a powerful and | visceral moment for me and really crystallized what seemed like | importance about my job. | spiritplumber wrote: | www.robots-everywhere.com we have a bunch of free stuff hereif | it helps any | minimaxir wrote: | The cynical-but-likely-accurate take is that researching language | modeling has a higher ROI and lower risk than researching | robotics. | madisonmay wrote: | Wojciech stated this pretty explicitly on his Gradient Dissent | podcast a few months back. | texasbigdata wrote: | After a bit of Googling are you referring to Wojciech, the | head of YouTube? | ingenieros wrote: | https://open.spotify.com/episode/0f9Ht2vtdCYuHvKjMGf0al?si= | K... | kirill5pol wrote: | http://wojzaremba.com/ | Animats wrote: | Yes. | | Also, regular ML researchers sit at tables with laptops. | Robotics people need electronics labs and electronics | technicians, machine shops and machinists, test tracks and test | track staff... | | If you have to build stuff, and you're not in a place that | builds stuff on a regular basis, it takes way too long to get | stuff built. | [deleted] | amelius wrote: | Order picking in e-commerce warehouses seems a potentially | profitable market. | johnmoberg wrote: | Definitely! Pieter Abbeel (who was working with OpenAI at | some point) and others realized this and founded | https://covariant.ai/. | [deleted] | high_derivative wrote: | I dont think this is cynical and I don't think it's a bad | thing. OpenAI is not a huge org. The truth in 2021 is that not | only is robotics 'just not there yet' in terms of being a | useful vehicle for general intelligence research (obviously | robotics research itself is still valuable), there is also | nothing really pointing at this going to be the case in the | next 5-10 years. | | Given that, unless they want to commercialise fruit picking or | warehouse robots, it seems sensible. | BigBubbleButt wrote: | > Given that, unless they want to commercialise fruit picking | or warehouse robots, it seems sensible. | | How successful do you think attempts to monetize this will | be? Apart from Kiva at Amazon, I'm not even sure most shelf- | moving robots are profitable enterprises (GreyOrange, | Berkshire Grey, etcetera). I'm very skeptical of more general | purpose warehouse robots such as you see from Covariance, | Fetch, etcetera. I don't really know too much about fruit- | picking other than grokking how hard it would be and how | little it would pay. | | To be clear, I'm not saying these companies make no money or | have no customers. But it's not clear to me that any of them | are profitable or likely will be soon, and robots are very | expensive. I'm happy to learn why I'm wrong and these | companies/technologies are further ahead than I realize. | zitterbewegung wrote: | I was wondering why OpenAI's gym was archived on GitHub this | pivot seems more sense. | Syntonicles wrote: | Can you explain what that means? I'm familiar with OpenAI | Gym, I've been away from Github for a long time. | the-dude wrote: | Read only | jablongo wrote: | GYM is not exclusively for robotics - it's for reinforcement | learning in simulated environments, which I'm sure they will | keep doing. Also it looks like it is still being maintained, | so not really sure what you mean. | fxtentacle wrote: | My prediction is that dropping the real-world interactions will | severely slow down their progress in other areas. But then | again, I'm super biased because my current work is to make AI | training easier by building specialized hardware. | | Reinforcement learning can work quite well if you produce the | hardware, so that your simulation model perfectly matches the | real-world deployment system. On the other hand, training | purely on virtual data has never really worked for us because | the real world is always messier/dirtier than even your most | realistic CGI simulations. And nobody wants an AI that cannot | deal with everyday stuff like fog, water, shiny floors, rain, | and dust. | | In my opinion, most recent AI breakthroughs have come from | restating the problem in a way that you can brute-force it with | ever-increasing compute power and ever-larger data sets. "end | to end trainable" is the magic keyword here. That means the | keys to the future are in better data set creation. And the | cheapest way to collect lots of data about how the world works | is to send a robot and let it play, just like how kids learn. ___________________________________________________________________ (page generated 2021-07-17 23:00 UTC)