[HN Gopher] Powerful AI Can Now Be Trained on a Single Computer
       ___________________________________________________________________
        
       Powerful AI Can Now Be Trained on a Single Computer
        
       Author : MindGods
       Score  : 61 points
       Date   : 2020-07-17 20:48 UTC (2 hours ago)
        
 (HTM) web link (spectrum.ieee.org)
 (TXT) w3m dump (spectrum.ieee.org)
        
       | rgovostes wrote:
       | Is SLIDE being used anywhere, or were flaws discovered? It was
       | supposed to massively accelerate training on CPUs.
       | 
       | https://www.hpcwire.com/off-the-wire/rice-researchers-algori...
        
       | RcouF1uZ4gsC wrote:
       | > His group took advantage of working on a single machine by
       | simply cramming all the data to shared memory where all processes
       | can access it instantaneously.
       | 
       | If you can get all your data into RAM on a single computer, you
       | can have a huge speedup, even over a cluster that has in
       | aggregate more resources.
       | 
       | Frank McSherry has some more about this, though not directly
       | about ML training.
       | 
       | http://www.frankmcsherry.org/graph/scalability/cost/2015/01/...
        
       | ladberg wrote:
       | So this basically boils down to keeping your training data in
       | memory? Is there something else I missed?
        
         | dan-robertson wrote:
         | It looks obvious when you write it like that but I think many
         | people are surprised by just how much slower distributed
         | computations can be compared to non distributed systems. Eg the
         | COST paper [1]
         | 
         | [1]
         | https://www.usenix.org/system/files/conference/hotos15/hotos...
        
       | datameta wrote:
       | The machine used is a 36-core + single gpu. So not quite a home
       | computer yet but this is some serious progress!
       | 
       | Paper: https://arxiv.org/abs/2006.11751
       | 
       | Source: https://github.com/alex-petrenko/sample-factory
        
         | cheez wrote:
         | That's my machine!
        
         | cbozeman wrote:
         | I dunno... $8000 builds a 64c/128t 256 GB RAM workstation with
         | the same GPU these researchers used
         | (https://pcpartpicker.com/list/P6WTL2). That's arguably in the
         | realm of home computer for just about anyone making $90,000 and
         | above, I would think; I would also think anyone working in
         | those fields could command at least that salary or greater,
         | unless they're truly entry level positions. Seems it would be a
         | reasonable investment for someone actively working in the area
         | of machine learning / artificial intelligence.
        
           | eanzenberg wrote:
           | Anyone that can afford a car can afford this.
        
           | nqzero wrote:
           | what's the per-hour cost spot price of this machine on AWS ?
        
       | rbanffy wrote:
       | It always could be trained on a single computer. It was just a
       | matter of physical size versus time.
        
       | dan-robertson wrote:
       | Lots of people are focusing on this being done on a particularly
       | powerful workstation, but the computer described seems to have
       | power at a similar order of magnitude to the many servers which
       | would be clustered together in a more traditional large ML
       | computation. Either those industrial research departments could
       | massively cut costs/increase output by just "magically keeping
       | things in ram," or these researchers have actually found a way to
       | reduce the computational power that is necessary.
       | 
       | I find the efforts of modern academics to do ML research on
       | relatively underpowered hardware by being more clever about it to
       | be reminiscent of soviet researchers who, lacking anything like
       | the access to computation of their American counterparts, were
       | forced to be much more thorough and clever in their analysis of
       | problems in the hope of making them tractable.
        
       | fxtentacle wrote:
       | I'm surprised that this is IEEE worthy and not just common sense.
       | Of course there'll be huge speedups if, and only if, your dataset
       | fits into main RAM and your model fits into the GPU RAM.
       | 
       | But for most state of the art models (think gpt with billions of
       | parameters) that is far from being the case.
        
       | softwrdethknell wrote:
       | Apple's in-house SOC is the future.
       | 
       | I suspect the cloud has a decade, maybe less, of hype to grift
       | on.
       | 
       | Huge data sets on a personal computer and opt-in data sharing
       | with business and healthcare, etc will be the new norm.
       | 
       | Further out, software as we know it will cease to exist as
       | entirely custom chips per application are the norm. IN TIME.
       | 
       | New hardware wars to capture consumer attention incoming.
        
         | rbanffy wrote:
         | I don't need to own a fast workstation unless I want to
         | continuously train my models. I can, however, quickly get a
         | cloud instance that's much larger than that and train the model
         | at a fraction of the time and cost of a desktop workstation.
        
       | vz8 wrote:
       | How much RAM did their test workstation have? I can't seem to
       | spot it.
        
         | neatze wrote:
         | According to article System 1 had 128 GB DDR4 and System 2 has
         | 256 GB DDR4.
        
         | genpfault wrote:
         | Paper lists 3 configs, one with 128GB and two with 256GB.
        
         | m463 wrote:
         | I'm guessing most of the perf comes from the gpu memory size.
        
           | Enginerrrd wrote:
           | While that's likely true, I generally find that it's quite
           | rare that enough attention is paid to how information is
           | moved between disk, RAM, CPU and GPU. And paying close
           | attention to that can be extremely helpful. Taking the RAM up
           | to 11 can eliminate a lot of the art to it, which is a good
           | thing.
        
             | rbanffy wrote:
             | A machine in this class can easily have a terabyte of RAM.
             | Add a couple Optane DC sticks and you have enormous storage
             | at exceedingly high bandwidth.
        
       ___________________________________________________________________
       (page generated 2020-07-17 23:00 UTC)