[HN Gopher] The Computational Limits of Deep Learning
       ___________________________________________________________________
        
       The Computational Limits of Deep Learning
        
       Author : ozdave
       Score  : 39 points
       Date   : 2020-08-17 21:07 UTC (1 hours ago)
        
 (HTM) web link (arxiv.org)
 (TXT) w3m dump (arxiv.org)
        
       | xiphias2 wrote:
       | ,,This article reports on the computational demands of Deep
       | Learning applications in five prominent application areas and
       | shows that progress in all five is strongly reliant on increases
       | in computing power.''
       | 
       | I don't agree with the conclusion of the paper. The computing
       | architectures have been improving dramatically over the last few
       | years, and almost any task that was achievable 5 years ago with
       | deep learning is orders of magnitudes cheaper to train.
       | 
       | The energy resources taken by deep learning is increasing because
       | of the huge ROI for companies, but it will probably slow down as
       | the compute cost gets close to the cost of software engineers (or
       | the profit of a company), because at that point researching
       | improvements to the models gets relatively cheaper again.
        
       | cs702 wrote:
       | We can look Deep Learning's growing demands for computation and
       | despair, or view those growing demands as an economic incentive
       | to develop more powerful hardware that uses energy more
       | efficiently at a lower marginal cost and in a more sustainable
       | manner.
       | 
       | In other words, Deep Learning's growing need for computing power
       | seems to have reached a point at which it is now motivating
       | fundamental research to find greener, cheaper, more energy-
       | efficient hardware.
       | 
       | The economic incentives are _very_ powerful: Whichever companies
       | (or organizations, or countries) find ways to harness the most
       | computing power at the lowest marginal cost will win the race in
       | this market.
       | 
       | --
       | 
       | PS. The same could be said for Bitcoin mining: it is also
       | motivating fundamental research to develop greener, cheaper, more
       | energy-efficient, more powerful hardware. Whoever finds ways to
       | harness the most computing power at the lowest marginal cost will
       | make the most money processing transactions on the network.
        
         | rocqua wrote:
         | I think the case here is a lot easier than the case for bitcoin
         | mining. Bitcoin miners are so stupidly single purpose that
         | development there doesn't help much. Maybe in general it helps
         | create an industry for designing and manufacturing ASICs. I
         | suppose that might go into making ASICs for deep learning at
         | some point.
        
           | saddlerustle wrote:
           | Bitmain is one of TSMC's largest customers, and that
           | absolutely has been reinvested by TSMC in developing more
           | advanced fabrication techniques.
           | 
           | Also bitcoin mining chips are actually a lot like deep
           | learning chips in that it's a lot of simple operations scaled
           | out. And indeed, Bitmain now produces deep learning chips
           | too.
        
         | peterthehacker wrote:
         | Aren't companies like google[0] and nvidia[1] already doing
         | this?
         | 
         | The paper's point is that eventually we will reach computing
         | power limits and then we will have to improve the deep learning
         | algorithm's efficiency to continue to improve. From the
         | abstract:
         | 
         | > Continued progress in these applications will require
         | dramatically more computationally-efficient methods, which will
         | either have to come from changes to deep learning or from
         | moving to other machine learning methods.
         | 
         | [0] https://cloud.google.com/tpu [1] https://www.nvidia.com/en-
         | us/data-center/v100/
        
       | 256lie wrote:
       | I wonder how long we can continue overfitting these benchmark
       | datasets as a community of researchers? How much is ImageNet is
       | labeled incorrectly/subotimally?
        
       | freeone3000 wrote:
       | Will it? Or, like capital-intensive industries of the past, will
       | deep learning funnel its profits into bigger and bigger
       | computers, as has been done in the past and will be done again?
        
         | rabidrat wrote:
         | These are order-of-magnitude increases. If it costs $5m to
         | train GPT-3, which is 100x more compute than GPT-2, then it may
         | cost $500m to train GPT-4, and $50b to train GPT-5. This is
         | what is meant by economically (not to mention environmentally)
         | unsustainable.
        
           | freeone3000 wrote:
           | Environment aside (I don't even think the CURRENT rate of
           | training is environmentally safe) -- I see no inherent
           | problem with a 100x increase in cost per step holding anybody
           | back. Once it's trained, you can run it much cheaper. Who's
           | to say $50 billion of value can't be extracted from GPT-5?
        
             | AnimalMuppet wrote:
             | > Who's to say $50 billion of value can't be extracted from
             | GPT-5?
             | 
             | Perhaps it can. (Though the number of companies that can
             | afford to train it is rather small.) But can $5 trillion be
             | extracted from GPT-6? Even if it can, who can afford to
             | train it?
        
             | saddlerustle wrote:
             | Using electricity is not inherently damaging to the
             | environment. Very low cost and high power zero-carbon
             | generation sources exist (hydro and nuclear). For scale
             | also keep in mind that all the datacenters in the world
             | still use much less power than is used for smelting
             | aluminium.
        
       ___________________________________________________________________
       (page generated 2020-08-17 23:00 UTC)