[HN Gopher] Explainable Deep Learning: A Field Guide for the Uni... ___________________________________________________________________ Explainable Deep Learning: A Field Guide for the Uninitiated Author : BERTHart Score : 100 points Date : 2020-05-03 17:53 UTC (5 hours ago) (HTM) web link (arxiv.org) (TXT) w3m dump (arxiv.org) | mindgam3 wrote: | Lost me at the first sentence. | | > Deep neural network (DNN) is an indispensable machine learning | tool for achieving human-level performance on many learning | tasks. | | Not to be pedantic, but words matter. Is anyone actually claiming | that deep learning achieves true "human-level performance" on any | real world open-ended learning task? | | Even the most state of the art computer vision/object | classification algorithms still don't generalize to weird input, | like familiar objects presented at odd angles. | | I get that the author is trying to write something motivating and | inspirational, but it feels like claiming "near" or "quasi"-human | performance, with disclaimers, would be a more intellectually | honest way to introduce the subject. | kahnjw wrote: | Words matter, concretely define "open ended"? Did you just add | that phrase to preemptively nullify evidence to the contrary? | | Deep learning has surpassed human level performance on many | tasks [1][2]... (could add more you get the point). | | [1] | https://www.sciencedirect.com/science/article/pii/S2215017X1... | [2] https://arxiv.org/pdf/1502.01852v1.pdf | rwilson4 wrote: | "Many learning tasks" is a wiggle term. Sure, edge cases exist, | but the methods do work impressively well in many cases. | svara wrote: | > Is anyone actually claiming that deep learning achieves true | "human-level performance" on any real world open-ended learning | task? | | No, but the text you quoted doesn't say that. | | Human level performance in this context means humans perform no | better than some algorithm on some specific dataset. | | Incidentally, that's also how you get to claim superhuman | performance on classification tasks. Just include some classes | that aren't commonly known in your dataset, e.g. dog breeds, | plant species, or something like that. ;) | mindgam3 wrote: | > No, but the text you quoted doesn't say that [deep learning | achieves human-level performance | | Uh, it says DNNs are indispensable for achieving human level | performance. That clearly implies that this level of | performance is achievable, despite all evidence to the | contrary. | jdminhbg wrote: | This is a weird interpretation of that sentence. There are | lots of fields where human-level performance has been | achieved. See Go, for example. | asah wrote: | Maybe you need an RNN to help parse that sentence!! :-) | | Buffalo buffalo Buffalo buffalo buffalo buffalo Buffalo | buffalo | | etc | troelsSteegin wrote: | This is a review paper. It's long. Past the first sentence I find | it readable and well organized. It mentions some of the work on | interpretability I'd expect to see, Finale-Velez, Rudin, Wallach, | and LIME, but does not appear to mention Shapley. The bottom line | conclusion is "In the end, the important thing is to explain the | right thing to the right person in the right way at the right | time." That's both an obvious truth and a differentiating mindset | in research-first space. It's worth a skim. | cs702 wrote: | I'm surprised there is no mention of capsules and capsule-routing | algorithms. | | Capsules are groups of neurons that represent _discrete entities_ | in different contexts. For example, a 4x4 pose matrix is a | capsule representing a particular object in different | orientations seen from different viewpoints. Similarly, a subword | embedding can be seen as a capsule with vector shape representing | a particular subword in different natural language contexts. More | generally, a capsule can have any shape, but it always represents | only one entity in some context. | | In certain new capsule-routing algorithms -- e.g., EM routing[a], | Heinsen routing[b], dynamic routing[c], to name a few off the top | of my head[d] -- each capsule can activate or not _depending on | whether the entity it represents is detected or not in the | context of input data_. | | Models using these algorithms therefore make it possible for | human beings to interpret model behavior _in terms of capsule | activations_ -- e.g., "the final layer predicts label 2 because | capsules 7, 23, and 41 activated the most in the last hidden | routing layer." | | While these new routing algorithms are not yet widely used, in my | humble opinion they present a promising avenue of research for | building models that are explainable and/or enable assignment of | causality at high levels of function composition. | | -- | | [a] https://research.google/pubs/pub46653/ | | [b] https://arxiv.org/abs/1911.00792 | | [c] https://arxiv.org/abs/1710.09829 | | [d] If you're aware of other routing algorithms that can | similarly activate/deactivate capsules, please post a link to the | paper or code here. | Eridrus wrote: | It should be pretty obvious why, they don't work as well as | what is standard. This is about ways to explain the models we | get good performance with. | orionr wrote: | For those interested in this area for PyTorch models, take a look | at Captum (https://captum.ai/). Still a lot of work to do, but | we've provided a number of algorithms described in this field | guide in the library. Always looking for collaborators and | contribution of others. | | Disclosure: I support the team that developed Captum. ___________________________________________________________________ (page generated 2020-05-03 23:00 UTC)