IEEE Transactions on Neural Networks and Learning Systems, ( ISI ), Volume (34), No (11), Year (2022-1) , Pages (9016-9028)

Title : ( Prototype-Based Interpretation of the Functionality of Neurons in Winner-Take-All Neural Networks )

Authors: Ramin Zarei Sabzevar , Kamaledin Ghiasi Shirazi , Ahad Harati ,

Access to full-text not allowed by authors

Citation: BibTeX | EndNote

Abstract

Prototype-based learning (PbL) using a winner-take-all (WTA) network based on minimum Euclidean distance (ED-WTA) is an intuitive approach to multiclass classification. By constructing meaningful class centers, PbL provides higher interpretability and generalization than hyperplane-based learning (HbL) methods based on maximum inner product (IP-WTA) and can efficiently detect and reject samples that do not belong to any classes. In this article, we first prove the equivalence of IP-WTA and ED-WTA from a representational power perspective. Then, we show that naively using this equivalence leads to unintuitive ED-WTA networks in which the centers have high distances to data that they represent. We propose ±ED-WTA that models each neuron with two prototypes: one positive prototype, representing samples modeled by that neuron, and a negative prototype, representing the samples erroneously won by that neuron during training. We propose a novel training algorithm for the ±ED-WTA network, which cleverly switches between updating the positive and negative prototypes and is essential to the emergence of interpretable prototypes. Unexpectedly, we observed that the negative prototype of each neuron is indistinguishably similar to the positive one. The rationale behind this observation is that the training data that are mistaken for a prototype are indeed similar to it. The main finding of this article is this interpretation of the functionality of neurons as computing the difference between the distances to a positive and a negative prototype, which is in agreement with the BCM theory. Our experiments show that the proposed ±ED-WTA method constructs highly interpretable prototypes that can be successfully used for explaining the functionality of deep neural networks (DNNs), and detecting outlier and adversarial examples.

Keywords

, Adversarial examples, BCM theory, interpretability, neuron, prototype-based learning (PbL), winner-take-all.
برای دانلود از شناسه و رمز عبور پرتال پویا استفاده کنید.

@article{paperid:1089326,
author = {Zarei Sabzevar, Ramin and Ghiasi Shirazi, Kamaledin and Harati, Ahad},
title = {Prototype-Based Interpretation of the Functionality of Neurons in Winner-Take-All Neural Networks},
journal = {IEEE Transactions on Neural Networks and Learning Systems},
year = {2022},
volume = {34},
number = {11},
month = {January},
issn = {2162-237X},
pages = {9016--9028},
numpages = {12},
keywords = {Adversarial examples; BCM theory; interpretability; neuron; prototype-based learning (PbL); winner-take-all.},
}

[Download]

%0 Journal Article
%T Prototype-Based Interpretation of the Functionality of Neurons in Winner-Take-All Neural Networks
%A Zarei Sabzevar, Ramin
%A Ghiasi Shirazi, Kamaledin
%A Harati, Ahad
%J IEEE Transactions on Neural Networks and Learning Systems
%@ 2162-237X
%D 2022

[Download]