Preskoči na glavno vsebino

Optimization for Multi Layer Perceptron: Without the Gradient


  These days, the publishing house Nova Publishers published the book, entitled Advances in Machine Learning Research. In it is a chapter entitled OPTIMIZATION FOR MULTI LAYER PERCEPTRON: WITHOUT THE GRADIENT where I describe two new algorithms for neural networks learning (Bipropagation and Border Pairs Method ). Both of them are much more powerful than their predecessors - Backpropagation algorithm. The second algorithm is among other things constructive.

Abstract of the book chapter


During the last twenty years, gradient-based methods have been primarily focused on the Feed Forward Artificial Neural Network learning field. They are the derivatives of Backpropagation method with various deficiencies. Some of these include an inability to: cluster and reduce noise, quantify data quality information, redundant learning data elimination. Other potential areas for improvement have been identified; including, random initialization of values of free parameters, dynamic learning from new data as it becomes available, explanation of states and settings in hidden layers of learned ANN, among other.
This chapter deals with the contemporary, non-gradient approach to ANN learning, which is not based on gradual reduction of remaining learning error anymore and it tries to eliminate most of the mentioned deficiencies. Introduction includes a chronological description of some methods, which deal with solutions of mentioned problems: Initializing Neural Networks using Decision Trees (Arunava Banerjee, 1994), DistAl: An inter-pattern distance-based constructive learning algorithm (Jihoon Yang, 1998), Geometrical synthesis of multilayer feedforward neural networks or Multi-Layer Perceptron (Rita Delogu, 2006) and Bipropagation - a new way of MLP learning (Bojan Ploj, 2009).We continue with the description of a new learning method - Border Pairs Method (BPM), which in comparison with the gradient methods carries numerous advantages or eliminates most of the predecessor’s deficiencies. The BMP implements and uses border pairs – learning patterns pairs in the input space, which are located close to the class border. 
The number of boundary pairs gives us some information about the complexity of the learning process. Boundary pairs are also the perfect basis for the noise reduction. We determine that performing a noise reduction of the border pairs is sufficient.
By dividing the input space, the homogenous areas (clusters) are established. For every linear segment of border we assign one neuron in the first layer. The MLP learning begins in the first layer by adapting individual neurons. Neurons on the first layers are saturated, so we get a binary code on the output of the first layer - the code is unified for all members of the same cluster. Logical operations based on the data from the first layer are executed in the following layers. Testing showed that such learning is reliable, it is not subject to overfitting, and is appropriate for on-line learning and susceptible to concept drift in the process of learning (forgetting and additional learning).

Komentarji

Priljubljene objave iz tega spletnega dnevnika

Artificial intelligence

          Artificial intelligence (AI) is a relatively young branch of science that stirs the imagination of many. Even movie directors from hollywood are not exceptions. Development in AI area is very fast and there is no indication that this will change soon. I still remember my first contact with learning devices. This happend at the end of the last millennium when I realized neural networks (NN). They have immediately attracted my attention, because such devices were not known till then.           NN are made along the lines of mammalian brain. During the learning NN extract the essence from the data. After the learning we can ask NN questions. It gives us the right answers even to questions that during learning did not participate.  NN learns autonomously and therefore may exceed the teacher's (poeple's) knowledge. Here are some important achievements of artificial intelligence: A couple of years ago the co...

Umetna inteligenca

Umetna inteligenca (UI) je razmeroma mlada veja znanosti, ki mnogim buri domišljijo. Tudi Hollywoodski režiserji niso izjema, saj skorajda ni več znanstveno-fantastičnega filma brez pametnih naprav. Razvoj na  področju UI je zelo hiter in nič ne kaže, da se bo to kmalu spremenilo.  Dobro se še spominjam svojega prvega srečanja z napravami, ki imajo sposobnost učenja. Bilo je to konec minulega tisočletja, ko sem se seznanil z nevronskimi mrežami (NM) V hipu so pritegnile mojo pozornost navkljub dejstvu, da inteligentne naprave takrat še niso bile splošno znane in so le redki verjeli v njihovo svetlo prihodnost. NM  so narejene po vzoru možganov kakršne imamo ljudje in živali. Med samostojnim učenjem so sposobne iz kopice učnih podatkov izluščiti bistvo, zato  lahko uspešno odgovarjajo tudi na vprašanja, ki jih med učenjem niso srečale. Lahko presežejo znanje svojega  učitelja ali tudi celotnega  človeštva.  Omenimo nekatere pomembne dosežke ...

A new Deep Learning Algorithm: One-Step Method

We are living in the AI era where progress is faster and faster each and every single day. Here is another one discovery in this field: One Step Method , a new machine learning algorithm which can do many things, amongst other can replace digital circuits with neurons, can find the even better construction of neural network than Border Pairs Method. More you can find in the 3rd chapter of our book:  Machine Learning: Advances in Research and Applications  from Nova Science Publishers . This new algorithm is also suitable for Deep Learning in combination with other methods like convolutional learning, bipropagation, border pairs method, autoencoder and others.