SYN@PTICS® Lab: twenty-fifth birthday of the website (Neural Networks / Cognitive Systems / Lifelong Deep Learning Systems "on Chip")
My research, software development, and teaching activities in the field of Neural Networks, Adaptive and Cognitive Systems began in 1984 but this year the Syn @ ptics website which describes part of the history of these activities is "only" 25 years old (1993 ).
This article does not want to be a self-commemoration or even an advertisement but just a thanks to all those who believed in my research work in the field of neural networks and, above all, in the applicability of neuronal chips even in the years when the Artificial Intelligence has seen a long winter, remaining relegated to the academic world and very poorly considered in industry. During that winter, many of us, although moved by passion for research in the field of Neural Networks, Fuzzy Logic and Soft-Computing in general, have also had to devote themselves to projects and developments not particularly related to the way of Artificial Intelligence.
In 1993 I have developed "NeurFuzz" (Freeware Distributed OS:DOS), an Error Back Propagation MLP Neural Network trainer and C source code generator. The program included: “Elliptic” Fuzzy Interfaces based on Fuzzy Logic (Lofti Zadeh, Bart Kosko), Genetic Algorithm Learning, Learning with Simulated Annealing & Thermoshok (local minima overcoming), Multiple (up to 100) Hidden Layers (very long training time with many hidden layers on Intel 386/486). The limited RAM imposed, with many layers, the use of mass storage of synapses and fast access through hash algorithms.
Soon I realized that the Multilayer Perceptron with Error Back Propagation was a neural network suitable for being programmed through data, but not suitable for learning new data in real time. The need to cycle on all the data already learned through learning "epochs" was a not negligible limit:
My real interest was aimed at neural networks able to learn new data without having to cycle on previously learned data. The problem of plasticity vs stability has been solved by Stephen Grossberg and Gail Carpenter with Adaptive Resonance Theory (ART). From this unsupervised learning model, supervised models have been derived, among which I want to remember the simple and intuitive SFAM (Simplified Fuzzy ARTmap).
Other Neural Network models based on prototype neurons like the RBF (Radial Basis Function) with supervised RCE (Restricted Coulomb Energy) learning algorithm were able to learn in an adaptive and non-competitive way by adding neural resources based on "novelty detection" in the input data. The RCE learning algorithm was invented by the Nobel Prize Leon Cooper. At that time I realized that SFAM and RBF / RCE were classifiers suited to my main objective: to build machines able to learn in a continuous way while maintaining the stability of the acquired knowledge.
At the same time I learned that the RBF / RCE model had been implemented in some chips with on-chip learning capability. I have, therefore, chosen to work on these chips because I thought they would be the future in the applications of Artificial Intelligence with the need for continuous learning in embedded contexts. The learning algorithm is executed in few steps and new data are learned without cycling on previously learned data avoiding catastrophic forgetting.
Despite the difficulties of those years, I believed in these technologies and I presented projects in the 90's that described SERVERS (today called CLOUD) based on neuronal chips, meeting the distrust of those who saw the neural networks closed in some University laboratories. These projects proposed pattern recognition "on the cloud" and "Context based Internet Search Engines", not based on GPU (Graphic Processing Units) but on neural chips such as ZISC (Zero Instruction Set Computer) which implemented a RBF (Radial Basis Function) neural network with RCE learning algorithm (Restricted Coulomb Energy) and had on-chip learning capabilities. Today this chip still exists and has 1000 neurons (instead of 36) and continues to be the only neural chip on the market with on-chip learning capabilities ... it has changed its name to NeuroMem®.
The early version of the RBF/RCE chip ZISC036 (Zero Instruction Set Computer). The name was derived from the evolution of CISC processors (Complex Instruction Set Computer) to RISC (Restricted Instruction Set Computer): ZISC was obviously the third generation and the concept, even today, represents the future of "neural devices", since the current GPUs (very useful for algorithmic research and cloud services) still process instructions flows in high energy consumption SIMD / MIMD architectures.
The new NeuroMem® chip by General-Vision implementing a RBF neural network with on-chip RCE (Restricted Coulomb Energy) learning algorithm and 1000 prototypes.
In the 80s and 90s powerful GPUs (SIMD + MIMD) were not available but there were accelerators boards (ISA/PCI) based on SIMD (Single Instruction Multiple Data) or MIMD (Multiple Instruction Multiple Data) processors with limited capacity of parallelism. Also the ISA / PCI cards based on DSP (Digital Signal Processor) were widely used to accelerate software implementations of MLP (Multilayer Perceptron) thanks to the ability to carry out many multiplication-accumulate operations (MAC) used in Fast Fourier Transform (FFT) but very useful to multiply input data with weights (synaptic values) and accumulate in the neuron.
.