|
||||
|
||||
ANN vs SNN The main difference
between ANN and SNN operation is the notion of time. While ANNs inputs are static, SNNs operate
based on dynamic binary spiking inputs as a function of time. In ANNs the information is transmitted
on the artificial synapses through numerical values expressed in digital
form. In SNNs, information is transmitted on artificial synapses through
impulses whose timing is somewhat equivalent to the numerical values of the
ANNs. Wolfgang Maass has shown that SNNs,
also called third generation neural networks, have a greater computational
power and can approximate any ANN, also called second generation neural
networks. In any case, ANNs are universal approximators and therefore there
is no reasonable motivation for using SNN from a practical and applicative
point of view. The main motivation that led to
develop SNN-based neural chips and related sensory devices is the greater
biological plausibility of the SNN model compared to the ANN model. We
believe that research in the field of SNNs is very important and some
applications are particularly promising in some sectors. Nevertheless we see
two criticisms both in the applicative field and in the biological
plausibility field. Criticality in the application field:
the learning algorithms based on STDP (Spike Timing Dependent Plasticity) are
more difficult to apply. Efficiency is intrinsically linked to a HW
realization. The realization of complex applications is always much more
expensive than using ANN. From a computational power point of view there is
no problem that can be solved with SNN that cannot be solved with ANN. Chips
based on communication with spikes typically use a protocol called AER
(Address Event Representation) or in any case a proprietary protocol that
allows communication via events (digital pulses in the time domain) between
NPUs (Neural Processing Units). Multiple NPUs can be interconnected
(configuration) to simulate different neural network architectures. From a
practical point of view Event-Based Processing and Event-Based Learning are
efficient only if the sensors are also designed to communicate the input data
directly with an Event Based Communication method. This factor restricts the
number of sensors that can be used effectively. Almost all commercial sensors
communicate data with different digital protocols and, therefore, it is
necessary to convert the data towards an Event Based Communication protocol.
For example a traditional digital camera needs a "pixel to event"
conversion pre-processing stage. This pre-processing stage is mandatory if
the chip is to accept input from traditional sensors. About the property of “one shot
learning”, this is strictly bound to the neural network model that is
configured. When the one-shot learning capability on CNN is mentioned, it
means that only the final classification layer is able to learn new examples
but all the parameters of the previous layers (features extraction) are
configured through a transfer learning procedure. Scalability (ability to connect many
chips to get a bigger neural network), when present, is bound to the configured
neural network architecture. Chips based on spiking neurons are
very energy efficient. However, energy efficiency comparisons are almost
always done with GPUs and, in this context, there is a clear advantage. If
you compare the power consumption of a spiking neural chip with that of a
non-spiking digital neural chip operating with a low clock on byte-type data,
the differences are almost irrelevant. Despite these criticisms, we are
carefully observing and evaluating the new generations of spiking chips and
intend to integrate them into devices for aerospace applications in the short
term. Criticality in the field of biological
plausibility: although SNNs are much closer to the biological model than
ANNs, we must point out that there is an impassable threshold between an
implementation on silicon and the countless variables involved in the
chemical processes of a real biological neuron. Beyond all the opinions related to the
research aspect, we have chosen to use ANN also as an on-chip HW
implementation. The key technology of our HW devices is, in fact, the
Neuromem® chip which offers maximum scalability, extremely low clock
frequency and ease of use. This chip implements a fixed neural
architecture of the RBF (Radial Basis Function) type with RCE (Restricted
Coulomb Energy) learning algorithm and is therefore a classifier. Scalability
is assured and virtually unlimited. One Shot Learning is inherently guaranteed
by RCE. The possibility of obtaining degrees of confidence on multiple
classifications allows to transform the classifier into a GRNN (General
Regression Neural Network) through weighted interpolations on LUTs (Look Up
Table). All Anomaly Detection problems can be solved in an extremely simple
way. |
LUCA MARCHESE Aerospace_&_Defence_Machine_Learning_Company VAT:_IT0267070992 Email:_luca.marchese@synaptics.org |
Contacts_and_Social_Media |
Copyright_©_2025_LUCA_MARCHESE |