Skip to main content
Computational Intelligence and Neuroscience logoLink to Computational Intelligence and Neuroscience
. 2015 Feb 1;2015:947098. doi: 10.1155/2015/947098

Training Spiking Neural Models Using Artificial Bee Colony

Roberto A Vazquez 1,*, Beatriz A Garro 2
PMCID: PMC4331474  PMID: 25709644

Abstract

Spiking neurons are models designed to simulate, in a realistic manner, the behavior of biological neurons. Recently, it has been proven that this type of neurons can be applied to solve pattern recognition problems with great efficiency. However, the lack of learning strategies for training these models do not allow to use them in several pattern recognition problems. On the other hand, several bioinspired algorithms have been proposed in the last years for solving a broad range of optimization problems, including those related to the field of artificial neural networks (ANNs). Artificial bee colony (ABC) is a novel algorithm based on the behavior of bees in the task of exploring their environment to find a food source. In this paper, we describe how the ABC algorithm can be used as a learning strategy to train a spiking neuron aiming to solve pattern recognition problems. Finally, the proposed approach is tested on several pattern recognition problems. It is important to remark that to realize the powerfulness of this type of model only one neuron will be used. In addition, we analyze how the performance of these models is improved using this kind of learning strategy.

1. Introduction

Artificial neural networks (ANNs) are applied in a broad range of problems. Among the most popular tasks using ANN, we could mention pattern recognition, forecasting, and regression problems. However, the accuracy of these models could drastically diminish if the topology is not well-designed and if the training algorithm is not selected carefully. One interesting alternative to designing the topology and training and exploiting the capabilities of an ANN is to adopt a learning strategy based on evolutionary and swarm intelligence algorithms. It is well-known that designing and training tasks can be stated as optimization problems; for that reason, it is possible to apply different types of evolutionary and swarm intelligence algorithms. For example, particle swarm optimization [1] and differential evolution [2] have been used to design and train ANNs automatically.

Several swarm intelligence algorithms based on the collective behavior of self-organizing systems have been proposed in the last years. Among the most popular, we could mention ant colony system (ACO) [3] and particle swarm optimization (PSO) [1] and artificial bee colony (ABC) [4]. Most of the studies related to honey bee swarm are focused on the dance and communication, task allocation, collective decision, nest site selection, mating, marriage, reproduction, foraging, floral and pheromone laying, and navigation behaviours of the swarm [5]. ABC is a novel algorithm that tries to mimic the behavior of the bees in nature, which tasks consist in exploring their environment to find a food source.

The ABC algorithm has been used in a broad range of optimization problems. This algorithm is relatively simple, and its implementation is straightforward for solving optimization problems, being able to produce acceptable results at a low computational cost. The different studies performed in the literature compare its efficiency against other traditional strategies such as genetic algorithm (GA), differential evolution (DE), particle swarm optimization (PSO), ant colony optimization (ACO), and their variants. The efficiency obtained on numerical problems using numerical test functions, multivariable functions, constrained and unconstrained optimization problems, and even multiobjective problems, suggests that the ABC algorithm is a serious candidate for training ANN [5].

Some of the first works that use the ABC algorithm to adjust the synaptic weight of an ANN are described in [6, 7]. In [8] the authors train a feed-forward ANN using ABC applied to solve the XOR, 3-bit parity, and 4-bit encoder-decoder problem, and some signal processing applications. In [9], the ANN is trained using ABC algorithm to solve a medical pattern classification problem. In [10], the authors train an ANN to classify different dataset utilized in the machine learning community. In [11], the ANN is trained for the classification of the acoustic emission signal to their respective source. Another interesting paper for designing and training an ANN is presented in [12], where the authors described a methodology for maximizing its accuracy and minimizing its connections by evolving the weights, the architecture, and the transfer function of each neuron. In the context of forecasting, [13] used ABC to train an ANN for bottom hole pressure prediction in underbalanced drilling. Whereas in [14] the authors train a recurrent ANN for stock price forecasting, in [15] the author uses ABC for training an ANN for earthquake time series data. Reference [16] presents an ANN trained with ABC for S-models of biomedical networks approximation. From all these papers the authors conclude that ABC algorithm is capable of training ANN with an acceptable accuracy and, in some cases, the results are better than those obtained with other traditional techniques.

Moreover, there exist many algorithms based on the bees' behavior such as bee algorithm, honey-bee mating algorithm, and bee colony optimization (BCO), among others [17]. Hence, there are investigations about training ANNs that use different kinds of algorithms related to the ABC. For example, in [18] the authors apply the bee colony algorithm to train an ANN, which later is applied to the wood's defect problem. In [19], the authors estimate the state variables in distribution networks, including distributed generators using a honey-bee mating algorithm. Furthermore, in [20], the authors use this algorithm in combination with a self-organizing map (SOM) in the market segmentation problem.

Swarm intelligence algorithms have contributed and gained popularity in the field of ANN as a learning strategy. However, the intrinsic limitations of ANN do not allow applying them in complex pattern recognition problems, even using different learning strategies. These limitations motivate to explore other alternatives to model and generate neural models to make possible their application in several pattern recognition problems.

Although ANNs were inspired by the behavior of the human brain, the fact is that they do not mimic the behavior of a biological neuron. In that sense, the development and application of more realistic neural models could improve the accuracy of an ANN during several pattern recognition tasks.

Spiking neuron models are called the 3rd generation of artificial neural networks [21]. These neurons increase the level of realism in a neural simulation and integrate the concept of time. These types of models have been used in a broad range of areas, mainly from the field of computational neurosciences [22], brain region modeling [23], auditory processing [24, 25], visual processing [2628], robotics [29, 30], and so on.

Several spiking models have been proposed in the last years. One of the most realistic and complex models was proposed in [31]. Nonetheless, there are simplified versions of this model that reduce its computational complexity. Among these models, we could mention the well-known integrate-and-fire model [32], Izhikevich model [33], FitzHugh-Nagumo model [34], and Hindmarsh-Rose [35].

Theoretically, these types of models could simulate the behavior of any perceptron type neural network. However, their application in computer vision and pattern recognition has not been widely explored. Although there are some works related to image segmentation [3638] and pattern recognition [3943], there still are several issues to research related to the learning process, design, and implementation.

The process of learning of these models is conducted with different techniques. In [44], the authors present the Spike-Prop, an adaptation of the well-known backpropagation algorithm to train a spiking neural model. Furthermore, several variants to improve the efficiency of spike-prop have been proposed [4547]. However, these algorithms require a careful tuning of the network to obtain acceptable results.

Another approach to train these models is based on probabilistic models [4749], information bottleneck learning [5052], and reinforcement learning [5355].

On the other hand, nongradient based methods like evolutionary strategies (such as GA, PSO, and DE) have emerged as an alternative to traditional methods for training spiking neural models. Although this approach is computationally more expensive compared with traditional methods, it has several advantages that make possible its application in real pattern recognition problems [42, 56, 57].

Recently, it has been proven that only one spiking neuron model can solve nonlinear pattern recognition problems, showing a clear advantage against the traditional perceptron [5860]. One alternative to simulate the learning process of this type of model is to use swarm intelligence algorithms. For example, in [58] the authors describe an approach to applying a leaky-integrate-and-fire spiking neuron in various linear and nonlinear pattern recognition problems. In that work, the authors use the differential evolution algorithm as a learning strategy. In other researches, the authors use the Izhikevich spiking model to the same set of problems using a differential evolution strategy [60]. In [61, 62], the authors use the Izhikevich spiking model to the same set of problem using cuckoo search and particle swarm optimization algorithms, respectively. In general, the methodology described in those papers can be stated as follows: given a set of input patterns belonging to K classes, first of all, each input pattern is transformed into an input signal. Then the spiking neuron is stimulated during T ms and the firing rate is computed. After adjusting the synaptic weights of the spiking model by means of a swarm intelligence algorithm, we expect that input patterns from the same class produce similar firing rates, and input patterns from different classes generate firing rates different enough to discriminate among the categories.

Despite the results presented in those papers, it is still necessary to explore and develop strategies that allow these models to learn from their environment and improve their accuracy solving complex pattern recognition problems. Due to the capabilities of producing acceptable results at a low computational cost in a broad range of optimization problems, including ANN field, the ABC algorithm could be an excellent tool for simulating the learning process of a spiking neural model.

In this paper, we proposed to use the ABC algorithm as a learning strategy to training a spiking neuron model aiming to perform various linear and nonlinear pattern recognition problems. Based on the methodology described in [60, 61], we present a comparison of the results presented in [58, 60, 61] to determine which learning strategy provides the best accuracy and how it affect the accuracy of the spiking neuron. In order to test the accuracy of the learning strategy combined with the spiking neuron model, we perform several experiments using different pattern recognition problems, including an odor recognition problem and cancer classification based on DNA microarrays.

The outline of this paper is divided into five sections. A brief introduction to the ABC algorithm is presented in Section 2. The concepts related to the third generation of neural networks knowing as spiking neural networks are presented in Section 3. Section 4 presents the proposed methodology for training spiking neural networks using the ABC algorithm. The experimental results, as well as the discussion of the results, are presented in Section 5. Finally, the conclusions of this work are presented in Section 6.

2. Basics on Artificial Bee Colony

The artificial bee colony (ABC) algorithm is a novel approach in the area of the swarm optimization proposed by Karaboga and Akay [6]. The ABC algorithm is based on the behavior of bees in nature, whose task consists of exploring their environment to find a food source, picking up the flower's nectar and returning to the hive in order to evaluate the quality and the amount of the food, and then call the other bees of the community to fly towards the food source. Communication among bees is done by a particular dance.

This algorithm can find the optimum values in the search space of a given optimization problem. A global optimization problem can be defined as finding the parameter vector x that minimizes the objective function f(x):

minimizef(x),x=(x1,x2,,xi,,xn1,xn)Rn (1)

which is constrained by the following inequalities and/or equalities:

lixiui,i=1,,n. (2)

f(x) is defined on a search space, S, which is n dimensional rectangle in R n. The variable domains are limited by their lower and upper bounds (2).

This algorithm represents the solutions of a given problem by means of the position of different food sources visited by a bee. Furthermore, it works with three kinds of bees in order to explore and exploit the search space: employed, onlookers, and scouts bees.

Following the work described in [7], the employed bee has to modify the position (solution) in its memory based on the local information (visual information) and test the nectar amount (fitness value) of the new source (new solution). If the quantity of nectar in the new position is better than the old one, the bee memorizes it and forgets the old one. Contrarily, it keeps the previous one in its memory. After the entire employed bees complete the search process, they share the nectar information about the food sources and their position, with the onlooker bees in the dance area.

An onlooker bee checks the nectar information obtained from all employed bees and selects a food source with a probability in terms of its nectar amount. The employed bee modifies the position in its memory and checks the quantity of nectar obtained from the candidate source. If its nectar is higher than that of the previous one, then the bee memorizes the new position and forgets the old one.

An artificial onlooker bee chooses a food source depending on the probability p i associated with that food source. This probability is calculated with the following expression:

pi=fitin=1SNfitn, (3)

where fiti is the fitness value of the solution i of dimension D which is proportional to the nectar amount of the food source in the position i and SN is the number of food sources that is equal to the number of employed bees.

In order to produce a candidate food position from the old one in memory, the ABC algorithm uses the following expression:

vij=xij+ϕijxijxkj, (4)

where k ∈ {1,2,…, SN} and j ∈ {1,2,…, D} are indexes randomly chosen. Although k is determined randomly, it has to be different from i. ϕ ij is a random number between [−1,1] that controls the production of neighbor food sources around x ij.

The food source whose nectar is discarded by the bees is changed with a new food source by the scouts. In the ABC algorithm, this is simulated by producing a random position and replacing it with the abandoned one. If that position cannot be enhanced after a number of trials, then the solution is discarded. This parameter is called the “limit” for abandonment. Assume that the abandoned source is x i and j ∈ {1,2,…, D}; then the scout bee discovers a new food source to be replaced with x i. This operation can be defined as

xij=lj+rand0,1ujlj, (5)

where l j and u j are the lower and upper bounds of the parameter x ij, respectively.

After each candidate source position v ij is generated and then evaluated by the artificial bee, its performance is compared to the performance associated with the previous position. If the new food source has equal or better nectar than the old one, then it is substituted with a new one in the memory. Otherwise, the old one is retained in the memory.

The pseudocode of the ABC algorithm is shown in Pseudocode 1.

Pseudocode 1.

Pseudocode 1

The ABC algorithm randomly initializes a population of solutions that represent the position of food sources within the lower and upper bounds. The size of the population is a parameter that corresponds to the number of food sources that is equal to the number of employed or onlooker bees. The stop criterion adopted in this algorithm is the maximum cycle number. The solutions are limited by their lower and upper bounds (2). If they are out of the boundaries, they are set to the lower or upper bounds.

One advantage of this algorithm is that only three control parameters are needed: population size (SN), the maximum cycle number (MCN), and the value of the “limit.”

3. Spiking Neural Models

The main distinctive elements that compose a typical spiking neuron are: dendrites (input device), soma (central processing unit), and axon (output device). The dendrites collect signals from other neurons and transmit them to the soma. The soma performs an important nonlinear processing step where an output signal is generated if the total input exceeds a certain threshold. Finally, the output signal is taken over by the axon, which delivers the signal (short electrical pulses called action potentials or spikes) to other neurons. Typically, the spikes have an amplitude of about 100 mV and a duration of 1-2 ms [63]. Although the same elements exist in a linear perceptron, the main difference between a linear perceptron and a spiking model is the action potential generated during the stimulation time. Furthermore, the activation function used in spiking models is a differential equation that tries to model the dynamic properties of a biological neuron in terms of spikes.

According to [63], a spike train is a sequence of stereotyped events generated at regular or irregular intervals. The form of the spike does not carry any information, and what is important is the number and the timing of spikes. The shortest distance between two spikes defines the absolute refractory period of the neuron that is followed by a phase of relative refractoriness where it is difficult to generate a spike.

Several spiking models have been proposed in the last century aiming to model different neurodynamic properties of neurons [64]. Among these models, we could mention the well-known integrate-and-fire model, resonate-and-fire, Izhikevich model, FitzHugh-Nagumo model, and Hodgkin-Huxley model.

One of the most simple and versatile models is the one proposed by Izhikevich. This model has only nine dimensionless parameters, and it is described with the next equation:

Cv˙=kvvrvvtu+Iif  vvpeak  thenu˙=abvvruvc,uu+d. (6)

Depending on the values of a and b, the spiking model can act as an integrator or a resonator. Whereas a is the recovery time constant, the sign of b force to u behaves as an amplifying (b < 0) or a resonant (b > 0) variable. The parameters c and d consider that the action of high-threshold voltage-gated currents activated during the spike affects only the after-spike transient behavior. In addition, c define the voltage reset value, and d describes the total amount of outward minus inward currents activated during the spike. On the other hand, the variable v represents the membrane potential, u is the recovery current, C is the membrane capacitance, v r is the resting membrane potential, v t is the instantaneous threshold potential, and v peak is the spike cutoff value [65].

According to [33], the spiking model can reproduce various intrinsic firing patterns based on different values of the parameters. Regular spiking (RS) neurons are the most typical neurons in the cortex; see Figure 1. For the intrinsically bursting (IB) behavior, the neurons fire a stereotypical burst of spikes followed by repetitive single spikes. Whereas the neurons with a chattering (CH) behavior can fire stereotypical bursts of closely spaced spikes, neurons with a fast-spiking (FS) behavior can fire periodic trains of action potentials with extremely high frequency practically without any adaptation (slowing down). Finally, for the low-threshold spiking (LTS), the neurons can also fire high-frequency spike trains but with a notable spike frequency adaptation.

Figure 1.

Figure 1

Simulation of the Izhikevich neuron model 100v˙=0.7v+60v+40-u+I, u˙=0.03-2v+60-u, if v ≥ 35 then v ← −50 and uu + 100. (a) Injection of the step of DC current I = 70 pA. (b) Injection of the step of DC current I = 100 pA.

A profound description of the Izhikevich model can be found in [65]. In this paper, we will concentrate on the parameters that produce regular spiking patterns. However, we do not discard the use in the near future of other firing patterns, useful for solving pattern recognition tasks. The Izhikevich model can produce a class 1 neural excitability behavior: action potentials can be generated with arbitrary low firing rate, depending on the strength of the applied current [65]. The response of the Izhikevich neuron changes if the input current changes generating different firing rates; see Figure 1.

The neuron is stimulated during T ms with an input signal, and it fires when its membrane potential reaches a specific value and thus generating an action potential (spike) or a train of spikes.

4. Proposed Method

Based on the hypothesis “patterns from the same class produce similar firing rates in the output of the spiking neuron and patterns from other classes produce firing rates different enough to discriminate among the classes,” the authors in [58, 60] proposed a methodology which describes the way that a spiking neuron can be applied to solve pattern recognition problems.

Following the same approach, let D = {x i, k}i=1 p be a set of associations composed of p input patterns, where k = 1,…, K is the class to which x iR n belongs. The learning process adjusts the synaptic values of the spiking model in such way that the output generates a different firing rate for each class k, reproducing the behavior described in the hypothesis.

4.1. Classifying with Firing Rates

This subsection describes how a spiking neural model can be applied in a pattern classification task based on [58, 60].

In order to use a spiking neural model to solve a pattern classification problem, it is necessary to compute the input current I that stimulates the model. In other words, the spiking neuron model is not directly stimulated with the input pattern x iR n but with the input current I. If we assume that each feature of the input pattern x i corresponds to the presynaptic potential of different receptive fields, then we can calculate the input current I that stimulates the spiking neuron as

I=x·w, (7)

where w iR n is the set of synaptic weights of the neuron model. This input current is used in the methodology to stimulate the spiking model during T ms.

Instead of using the spike train generated by the spiking model to perform the pattern classification tasks, we compute the firing rate of the neuron defined as

fr=NspT, (8)

where N sp is the number of spikes that occur within the time window of lenght T.

It is necessary to calculate the average firing rate AFR ∈ R K of each class, by using the firing rates produced by each input pattern. In this sense, the learning process consists of finding the synaptic values of the spiking model in such way that it generates a different average firing rate for each class k.

Suppose that the spiking neuron is already trained using a learning strategy. To determine the class to which an unknown input pattern x belongs, it is necessary to compute the firing rate generated by the trained spiking neuron. After that, the firing rate is compared against the average firing rate of each class. The minimum difference between the firing rate and the average firing rates determines the class of an unknown pattern. This is expressed with the following equation:

cl=argmink=1KAFRkfr, (9)

where fr is the firing rate generated by the neuron model stimulated with the unknown input pattern x~.

4.2. Adjusting Synapses of the Spiking Neuron Model

In order to achieve the desired behavior at the output of the spiking neuron, it is necessary to adjust its synaptic weights. This step corresponds to the learning (training) phase of the spiking neuron. Several swarm intelligence algorithms can be used in the learning phase, but in this research we focus on the artificial bee colony (ABC) algorithm as a learning strategy.

The synapses of the neuron model w are adjusted by means of the ABC algorithm. The food source in the ABC algorithm represents the synaptic weights of the spiking model. In order maximize the accuracy of the spiking neuron model during a pattern recognition task, the best set of synaptic weights must be found using the ABC algorithm. However, if we state the problem as a minimization problem, the classification error must be minimized. The fitness function that uses the classification error to find the set of synaptic weights is defined as

fw,D=1performancew,D, (10)

where w are the synapses of the spiking model, D is the set of input patterns, and performance(w, D) is a function which computes the classification accuracy, in terms of (9), given by

performancew,D=PccPt, (11)

where P cc denotes the number of patterns correctly classified and P t denotes the number of tested patterns.

5. Experimental Results

In this section, we analyze and discuss the results generated with the proposed methodology. In order to evaluate the accuracy of the proposed methodology, we perform several experiments using different datasets. The section is divided into three subsections. The first section presents a comparison of the proposed methodology against different swarm intelligence algorithms. In the last two subsections, we present preliminary results applying spiking neural models using the proposed methodology for solving an odor classification problem and cancer classification based on DNA microarrays.

5.1. Analysis and Comparison of Experimental Results

In order to evaluate the classification accuracy, when the spiking neuron is trained using the ABC algorithm, six sets of experiments were performed. Each set of experiments was performed with a dataset that corresponds to a specific pattern recognition problem. The iris plant, glass, diabetes, liver-bupa, and wine datasets were taken from the UCI machine learning benchmark repository [66].

The iris plant dataset is a three-class problem composed of input patterns with four features. The wine dataset also is a three-class problem composed input patterns with 13 features. The glass dataset is a six-class problem composed of patterns with nine features; however, in this experiment we used the two most representative classes. The diabetes dataset is a two-class problem composed of input patterns with eight features. Finally, the liver dataset also is a two-class problem composed of input patterns with six features.

Furthermore, an object recognition problem, composed of five different objects, was used to evaluate the accuracy of the proposal; see Figure 2 [67]. A dataset was generated from a set of 100 images in different positions, rotations, and scale changes. Instead of recognizing objects directly from their images, an invariant description of each object was calculated applying the next process over each image: a standard threshold [68] was applied to get its binary version; small spurious regions were eliminated from each image by means of a size filter [69]; finally, the seven well-known Hu moments invariant to translations, rotations, and scale changes [70] were computed over each image. As a result of this process, we obtained a five-class dataset problem composed of patterns with seven features.

Figure 2.

Figure 2

(a)–(d) Some of the images used to train the proposed method. (f)–(i) Some of the images were used to test the proposed method.

To reproduce the regular spiking behavior, the parameters for the Izhikevich neuron model were set as C = 100, v r = −60, v t = −40, v peak = 35, k = 0.7, a = 0.03, b = −2, c = −50, and d = 100. To solve the differential equation, Euler method with dt = 1 with a simulation time of T = 1000. Finally, the parameters for the artificial bee colony algorithm were set to SN = 40, MAXGEN = 1000, LIM = 100, and the boundaries for the solution to XMAX = 10 and XMIN = −10.

The classification rate of the model was computed in terms of the number of input patterns correctly classified divided by the total number of tested input patterns. 30 experiments over each dataset were performed to validate the accuracy of the spiking neuron when the ABC algorithm is used as a learning strategy. In order to train the spiking neural model, 80% of the samples (training subset) were randomly selected to do so, the remaining samples (testing subset) were used during the testing phase.

At the beginning of the evolutionary learning process, we observed that the learning error rapidly converges to a stable error and after a certain number of generations the learning error changes at a slower rate. Although this behavior was observed in whole experiments, the learning error achieved with the ABC algorithm was not good enough for all the problems. For some problems, the achieved learning error was highly acceptable. In Figure 3 is shown how the learning error evolves through each generation of the ABC algorithm.

Figure 3.

Figure 3

Training error achieved by the spiking model during the learning phase using ABC algorithm. (a) 30 experiments using the wine dataset are shown. (b) 30 experiments using the iris dataset are shown. (c) 30 experiments using the glass dataset are shown. (d) 30 experiments using the diabetes dataset are shown. (e) 30 experiments using the liver dataset are shown. (f) 30 experiments using the object recognition dataset are shown.

After the spiking neuron was trained, we proceed to evaluate the accuracy of the neuron using the remaining 20% samples. From these experiments, we can observe that the performance of the spiking neuron trained with the ABC algorithm was highly acceptable for the six pattern recognition problems. The average percentage of classification computed from all experiments is shown in Table 1. Something that should be remarked is that these methods only used one Izhikevich neuron model. Furthermore, we present the results using the t-distribution test to construct the confidence intervals around the mean.

Table 1.

Statistical results obtained with the proposed methodology.

Dataset Mean Confidence intervals
Tr. cr. Te. cr. TR TE
Wine 0.963 ± 0.015 0.878 ± 0.038 [0.957–0.968] [0.864–0.892]
Iris plant 0.996 ± 0.006 0.957 ± 0.025 [0.994–0.999] [0.948–0.967]
Glass 0.832 ± 0.029 0.703 ± 0.064 [0.822–0.843] [0.679–0.727]
Diabetes 0.800 ± 0.016 0.743 ± 0.024 [0.794–0.806] [0.734–0.752]
Liver 0.749 ± 0.024 0.688 ± 0.034 [0.740–0.757] [0.675–0.700]
Object recognition 1.000 ± 0.000 0.990 ± 0.021 [1.000-1.000] [0.982–0.998]

Tr. cr = training classification rate, Te. cr. = testing classification rate.

We observed that the classification rate achieved for the iris, wine, and object recognition problems using the spiking neuron trained with the ABC algorithm during the training phase was greater than 96%. However, the classification rate achieved during the testing phase most of the times diminished. Nonetheless, the classification rate was higher than 87%.

On the contrary, for the case of the glass and diabetes problems, we observed that whereas the classification rate achieved using the spiking neuron trained with the ABC algorithm during the training phase was not greater than 84%, the classification rate for the liver problem was not greater than 75%. As a consequence, the accuracy achieved with the spiking neuron drastically diminished during the testing phase. Nonetheless, the results obtained with the spiking neuron trained with the ABC algorithm were acceptable.

Figures 4(a), 4(b), and 4(c) show the experimental results obtained with the wine, iris plant, and object recognition datasets. Each dot represents the time when the neuron produces a spike. As the reader can observe, the synaptic weights found with the ABC algorithm provoke that the Izhikevich neuron generates almost the same firing rate when it is stimulated with patterns from the same class. Furthermore, the Izhikevich neuron generates firing rates different enough to discriminate patterns that belong to other classes.

Figure 4.

Figure 4

Experimental results obtained with proposed methodology for different datasets. Notice that different firing rates which correspond to different classes can be observed. Each dot represents the time that the neuron generate an spike. (a) Wine dataset. (b) Iris plant dataset. (c) Object recognition dataset. (d) Glass dataset. (e) Diabetes dataset. (f) Liver dataset.

On the other hand, Figures 4(d), 4(e), and 4(f) show the experimental results obtained with the glass, diabetes, and liver datasets, respectively. For this pattern recognition problems, the set of synaptic weights found with the ABC algorithm during the training phase were not good enough for the spiking neuron to generate a similar firing rate when it is stimulated with patterns from the same class; as a result, some patterns belonging to the same class were classified in a different class. Although the Izhikevich neuron does not generate firing rates different enough to discriminate among patterns from different classes, the results achieved with the proposed method were quite acceptable.

In addition, in Table 2, we compare the behavior of the methodology using three different swarm intelligence algorithms as a learning strategy: differential evolution (DE) [58, 60], cuckoo search (CS) [61], and particle swarm optimization (PSO) [62].

Table 2.

Average accuracy provided by the methods using different databases.

Dataset Method using DE Method using PSO Method using CS
Tr. cr. Te. cr. Tr. cr. Te. cr. Tr. cr. Te. cr.
Wine 0.9796 0.8744 0.9782 0.8879 0.9831 0.9078
Iris plant 0.9933 0.9833 0.9933 0.97 0.9942 0.9467
Glass 0.8158 0.7411 0.8178 0.7457 0.8080 0.7646
Diabetes 0.8038 0.7371 0.7990 0.7619 0.8051 0.7477
Liver 0.7620 0.6870 0.7591 0.6754 0.7609 0.6536
Object recognition 1 0.9850 1 0.9950 1 1

Tr. cr = training classification rate, Te. cr. = testing classification rate.

Furthermore, we perform a comparison against the well-known feedforward neural network (FNN). Two different topologies (1 and 2 hidden layers) and two different training algorithms (descendant gradient and Levenberg-Marquardt) [71] were used to design the FNN. The parameters for training algorithms were set as 0.1 for the learning rate, 2000 epochs for the stop criterion. For each combination (topology-learning algorithm), 30 experiments were done to prove the obtained results statistically. For each experiment, we randomly select the 70% and 10% of the samples for both, training and validation phases. The remaining 20% of the samples were chosen for the testing phase.

The number of neurons that compose the hidden layers depends on the classification problem. The next two topologies were used in combination with each dataset.

  1. Wine dataset: 1 HL composes 8 neurons. 2 HL compose 5 and 3 neurons, respectively.

  2. Iris plant dataset: 1 HL composes 4 neurons. 2 HL compose 2 and 2 neurons, respectively.

  3. Glass dataset: 1 HL composes 8 neurons. 2 HL compose 5 and 3 neurons, respectively.

  4. Diabetes dataset: 1 HL composes 5 neurons. 2 HL compose 3 and 2 neurons, respectively.

  5. Object recognition dataset: 1 HL composes 6 neurons. 2 HL compose 4 and 2 neurons, respectively.

Table 3 shows the classification efficiencies using feedforward neural networks (FNN). Each column contains the obtained results using a combination of topology and learning algorithms during training and testing phases. Label DG-1 means that the descendant gradient method and one hidden layer (HL) were used to design the FNN, and label LM-2 indicates that Levenberg-Marquardt method and two hidden layers were used to create the FNN, and so on.

Table 3.

Average accuracy provided by the FNN using different databases.

FNN FNN FNN FNN
Dataset DG-1 DG-2 LM-1 LM-2
Tr. cr. Te. cr. Tr. cr. Te. cr. Tr. cr. Te. cr. Tr. cr. Te. cr.
Wine 0.9867 0.9800 0.9497 0.9171 0.9986 0.9714 0.9797 0.9629
Iris plant 0.9133 0.8867 0.6625 0.6367 0.9942 0.9767 0.7667 0.7733
Glass 0.6907 0.6738 0.6390 0.6310 0.8453 0.7714 0.7413 0.7119
Diabetes 0.7663 0.7732 0.7153 0.7144 0.7993 0.7320 0.7850 0.7588
Liver 0.5924 0.6145 0.5928 0.5623 0.7243 0.6652 0.7181 0.6812
Object recognition 0.7413 0.6800 0.7413 0.6800 0.7413 0.6800 0.7413 0.6800

Tr. cr = training classification rate, Te. cr. = testing classification rate.

The results obtained with the spiking neuron model trained with the ABC algorithm are similar to those obtained in [58, 6062]. Although with some dataset the ABC algorithm provides better results than DE and PSO algorithm and vice versa, we could not say that one learning strategy is better than the other.

In fact, swarm intelligence algorithms can be considered as a learning strategy to adjust the synaptic weights of a third-generation neural model. After comparing the different algorithms, we note that they present the same behavior and are serious candidates to be considered as a learning strategy. These results suggest that swarm intelligence algorithms combined with spiking neurons can be regarded as an alternative way to perform various pattern recognition tasks. After several experiments, we observed that if we want to improve the efficiency of the methodology, it is not enough to change the learning strategies because they provide similar efficiencies. Nonetheless, the results obtained are highly acceptable.

On the other hand, the reader can observe in Table 2 that the proposed methodology provides better results than traditional FNN composed of several neurons. These results, obtained with the described methodology, show a visible advantage of a spiking model against FNN.

We can also conjecture that if only one neuron is capable of solving pattern recognition problems, perhaps several spiking neurons working together can improve the experimental results obtained in this research.

5.2. Application of the Proposed Methodology in a Odor Recognition Problem

In this subsection, we present some preliminary results obtained with the proposed methodology in a problem related to odor recognition. The dataset used was built in Bermak and Martinez [72] where the authors developed an automated gas delivery experimental setup for extracting volatile compounds at given concentrations from liquids, composed of two pumps, two mass flow controllers (MFCs), one bubbler, a gas chamber, and a data acquisition system. This dataset contains patterns obtained from ethanol and butanol vapors injected into the gas chamber at a flow rate determined by the mass flow controllers. The authors also used sensor arrays composed of five commercial TGS Figaro gas sensors (TGS 2600, 2602, 2610, 2611, and 2620). The potential differences across the sensor resistances were measured using a voltage divider with 2.2 k load resistors while keeping the heating voltage constant to 5 V. Finally, the sensors output voltages were sampled at a rate of 10 Hz and quantized with an 11 bit analog to digital converter to build a dataset composed of 124 patterns with 5 features each pattern.

The parameters for the Izhikevich neuron and for the artificial bee colony algorithm and the number of experiments as well as the samples used for training and testing phases were set as equal as in the previous subsection.

Table 4 presents the results achieved with the proposed methodology applied to an odor recognition problem. According to the results presented in Table 4, the reader can observe that the accuracy of the spiking neuron is highly acceptable.

Table 4.

Statistical results obtained with the proposed methodology.

Dataset Mean Confidence intervals
Tr. cr. Te. cr. TR TE
Odor 0.952 ± 0.024 0.898 ± 0.064 [0.943–0.961] [0.874–0.922]

Tr. cr = training classification rate, Te. cr. = testing classification rate.

These results obtained with the proposed methodology confirm that spiking neurons can be considered as a pattern recognition technique useful in a wide range of applications, including odor recognition.

5.3. Application of the Proposed Methodology in Cancer Classification Based on DNA Microarrays

In this subsection, we present some preliminary results obtained with the proposed methodology for a problem related to cancer classification based on DNA microarrays.

DNA microarrays are a powerful technique in genetic science due to the possibility to analyze the gene expression level of millions of genes at the same time. However, the main problem that arises when DNA microarrays are analyzed with computational intelligent techniques is that the number of genes is too big, and the samples are too few [73].

In that sense, we evaluate the capabilities of a spiking neuron for predict a disease based on the DNA microarrays dataset. It is important to mention that in order to capture the real behavior of the spiking neuron, we do not apply any dimensional reduction technique over the dataset.

The dataset used in this paper was built in [74] where the authors described a generic approach to cancer classification based on gene expression monitoring by DNA microarrays that discover the distinction between acute myeloid leukemia (AML) and acute lymphoblastic leukemia (ALL). The dataset consists of 38 bone marrow samples for training (27 ALL and 11 AML), over 7129 probes from 6817 human genes. Also, 34 samples testing data are provided, with 20 ALL and 14 AML.

The parameters for the Izhikevich neuron and for the artificial bee colony algorithm and the number of experiments were set as equal as in the previous subsection.

Table 5 present the results achieved with the proposed methodology applied to a cancer recognition problem using DNA microarrays. According to the results presented in Table 5, the reader can observe that the accuracy of the spiking neuron is highly acceptable during training phases; however during the testing phase, the accuracy drastically diminished.

Table 5.

Statistical results obtained with the proposed methodology.

Dataset Mean Confidence intervals
Tr. cr. Te. cr. TR TE
Cancer DNA 0.883 ± 0.024 0.625 ± 0.097 [0.875–0.892] [0.588–0.661]

Tr. cr = training classification rate, Te. cr. = testing classification rate.

These results obtained with the proposed methodology are preliminary and suggest that only one spiking neuron is not enough to solve this problem. Nonetheless, if we combine several spiking neurons to build a network or apply a dimensional reduction stage before training the spiking model, the accuracy could increases. It is important to remark that few highly dimensional samples were used to train the spiking neural model.

6. Conclusion

In this paper, we described how the artificial bee colony algorithm can be used as a learning strategy to train a spiking neural model. The results obtained with this approach suggest that ABC algorithm is a useful alternative to adjusting the synaptic weights of the spiking neuron model. Furthermore, we observed that the spiking neuron model provides acceptable results during the pattern recognition task, regardless of the swarm intelligence algorithms used as a learning strategy, the spiking neuron model provides acceptable results during the pattern recognition task.

Although we found that the spiking neuron did not behave as good as we desired with the six datasets, we could say that, in general, the behavior achieved with the spiking neuron provokes that patterns belonging to the same class generate almost the same firing rate in the output of the spiking neuron, and patterns belonging to different classes produce firing rates different enough to discriminate among the different classes. Thanks to the swarm intelligence algorithm used during the evolutionary learning process, the spiking model behaves like we previously described in our starting hypothesis.

Furthermore, a comparison among the ABC, DE, CS, and PSO algorithms was performed. In general, we observe that the accuracy obtained with the spiking neuron model trained with the ABC algorithm is comparable to the accuracy obtained with the methods described in [58, 6062]. We could not say that one learning strategy is better than the other because, with some datasets, the ABC algorithm provides better results than DE and PSO algorithm and vice versa.

This research provides a clear idea of how powerful a spiking neuron model is in a pattern classification task. Furthermore, the results obtained with the odor recognition and cancer classification problem based on DNA microarrays were highly encouraged.

Nowadays, we are testing different spiking neuron models. Despite the encouraging results achieved with this methodology, we are developing a new method to evolve the synaptic weights at the same time not only for one spiking neuron, but for several spiking neurons as well as to adjust the parameters of the models.

Acknowledgments

The authors would like to thank Universidad La Salle for the economic support under Grant no. ULSA I-061/12. Beatriz Garro thanks CONACYT for the posdoctoral scholarship.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

References

  • 1.Garro B. A., Sossa H., Vazquez R. A. Design of artificial neural networks using a modified particle swarm optimization algorithm. Proceedings of the International Joint Conference on Neural Networks (IJCNN '09); June 2009; pp. 938–945. [DOI] [Google Scholar]
  • 2.Garro B. A., Sossa H., Vázquez R. A. Proceedings of the 17th International Conference on Neural Information Processing: Models and Applications—Part II (ICONIP '10) Vol. 6444. Berlin, Germany: Springer; 2010. Design of artificial neural networks using differential evolution algorithm; pp. 201–208. (Lecture Notes in Computer Science). [Google Scholar]
  • 3.Dorigo M., Blum C. Ant colony optimization theory: a survey. Theoretical Computer Science. 2005;344(2-3):243–278. doi: 10.1016/j.tcs.2005.05.020. [DOI] [Google Scholar]
  • 4.Karaboga D., Basturk B. A powerful and efficient algorithm for numerical function optimization: artificial bee colony (ABC) algorithm. Journal of Global Optimization. 2007;39(3):459–471. doi: 10.1007/s10898-007-9149-x. [DOI] [Google Scholar]
  • 5.Karaboga D., Gorkemli B., Ozturk C., Karaboga N. A comprehensive survey: artificial bee colony (ABC) algorithm and applications. Artificial Intelligence Review. 2014;42(1):21–57. doi: 10.1007/s10462-012-9328-0. [DOI] [Google Scholar]
  • 6.Karaboga D., Akay B. Artificial Bee Colony (ABC) algorithm on training artificial neural networks. Proceedings of the IEEE 15th Signal Processing and Communications Applications (SIU '07); June 2007; Eskişehir, Turkey. IEEE; pp. 1–4. [DOI] [Google Scholar]
  • 7.Karaboga D., Basturk B. Artificial bee colony (abc) optimization algorithm for solving constrained optimization problems. Proceedings of the 12th International Fuzzy Systems Association World Congress on Foundations of Fuzzy Logic and Soft Computing (IFSA '07); 2007; Cancun, Mexico. Springer; pp. 789–798. [Google Scholar]
  • 8.Karaboga D., Akay B., Ozturk C. Artificial bee colony (ABC) optimization algorithm for training feed-forward neural networks. Proceedings of the 4th International Conference on Modeling Decisions for Artificial Intelligence (MDAI '07); 2007; Springer; pp. 318–329. [Google Scholar]
  • 9.Karaboga D., Ozturk C., Akay B. Training neural networks with abc optimization algorithm on medical pattern classification. Proceedings of the International Conference on Multivariate Statistical Modelling and High Dimensional Data Mining; 2008. [Google Scholar]
  • 10.Karaboga D., Ozturk C. Neural networks training by artificial bee colony algorithm on pattern classification. Neural Network World. 2009;19(3):21–57. [Google Scholar]
  • 11.Omkar S. N., Senthilnath J. Artificial bee colony for classification of acoustic emission signal source. International Journal of Aerospace Innovations. 2009;1(3):129–143. doi: 10.1260/175722509789685865. [DOI] [Google Scholar]
  • 12.Garro B. A., Sossa H., Vazquez R. A. Artificial neural network synthesis by means of artificial bee colony (ABC) algorithm. Proceedings of the IEEE Congress of Evolutionary Computation (CEC '11); June 2011; New Orleans, La, USA. pp. 331–338. [DOI] [Google Scholar]
  • 13.Irani R., Nasimi R. Application of artificial bee colony-based neural network in bottom hole pressure prediction in underbalanced drilling. Journal of Petroleum Science and Engineering. 2011;78(1):6–12. doi: 10.1016/j.petrol.2011.05.006. [DOI] [Google Scholar]
  • 14.Hsieh T. J., Hsiao H. F., Yeh W. C. Forecasting stock markets using wavelet transforms and recurrent neural networks: an integrated system based on artificial bee colony algorithm. Applied Soft Computing. 2011;11(2):2510–2525. doi: 10.1016/j.asoc.2010.09.007. [DOI] [Google Scholar]
  • 15.Shah H., Ghazali R., Nawi N. M. Using artificial bee colony algorithm for MLP training on earthquake time series data prediction. Journal of Computing. 2011;3(6):135–142. [Google Scholar]
  • 16.Yeh W.-C., Hsieh T.-J. Artificial bee colony algorithm-neural networks for S-system models of biochemical networks approximation. Neural Computing and Applications. 2012;21(2):365–375. doi: 10.1007/s00521-010-0435-z. [DOI] [Google Scholar]
  • 17.Karaboga D., Akay B. A survey: algorithms simulating bee swarm intelligence. Artificial Intelligence Review. 2009;31(1–4):61–85. doi: 10.1007/s10462-009-9127-4. [DOI] [Google Scholar]
  • 18.Pham D. T., Soroka A. J., Ghanbarzadeh A., Koc E., Otri S., Packianather M. Optimising neural networks for identification of wood defects using the bees algorithm. Proceedings of the IEEE International Conference on Industrial Informatics (INDIN '06); August 2006; Singapore. IEEE; pp. 1346–1351. [DOI] [Google Scholar]
  • 19.Niknam T. Application of honey-bee mating optimization on state estimation of a power distribution system including distributed generators. Journal of Zhejiang University: Science A. 2008;9(12):1753–1764. doi: 10.1631/jzus.A0820047. [DOI] [Google Scholar]
  • 20.Amiri B., Fathian M. Integration of self organizing feature maps and honey bee mating optimization algorithm for market segmentation. Journal of Theoretical and Applied Information Technology. 2007;3:70–86. [Google Scholar]
  • 21.Maass W. Networks of spiking neurons: the third generation of neural network models. Neural Networks. 1997;10(9):1659–1671. doi: 10.1016/s0893-6080(97)00011-7. [DOI] [Google Scholar]
  • 22.Rieke F., Warland D., Robert, Bialek W. Spikes—Exploring the Neural Code. 1997. [Google Scholar]
  • 23.Hasselmo M. E., Bodelón C., Wyble B. P. A proposed function for hippocampal theta rhythm: separate phases of encoding and retrieval enhance reversal of prior learning. Neural Computation. 2002;14(4):793–817. doi: 10.1162/089976602317318965. [DOI] [PubMed] [Google Scholar]
  • 24.Hopfield J. J., Brody C. D. What is a moment? ‘Cortical’ sensory integration over a brief interval. Proceedings of the National Academy of Sciences of the United States of America. 2000;97(25):13919–13924. doi: 10.1073/pnas.250483697. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Loiselle S., Rouat J., Pressnitzer D., Thorpe S. Exploration of rank order coding with spiking neural networks for speech recognition. Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN '05); July-August 2005; Québec, Canada. pp. 2076–2080. [DOI] [Google Scholar]
  • 26.Azhar H., Iftekharuddin K., Kozma R. A chaos synchronization-based dynamic vision model for image segmentation. Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN '05); 2005; pp. 3075–3080. [DOI] [Google Scholar]
  • 27.Thorpe S. J., Guyonneau R., Guilbaud N., Allegraud J.-M., VanRullen R. SpikeNet: real-time visual processing with one spike per neuron. Neurocomputing. 2004;58–60:857–864. doi: 10.1016/j.neucom.2004.01.138. [DOI] [Google Scholar]
  • 28.Vazquez R. A., Girau B., Quinton J. C. Visual attention using spiking neural maps. Proceedings of the International Joint Conference on Neural Network (IJCNN '11); August 2011; pp. 2164–2171. [DOI] [Google Scholar]
  • 29.di Paolo E. A. Spike-timing dependent plasticity for evolved robots. Adaptive Behavior. 2003;10(3-4):243–263. [Google Scholar]
  • 30.Floreano D., Zufferey J.-C., Nicoud J.-D. From wheels to wings with evolutionary spiking circuits. Artificial Life. 2005;11(1-2):121–138. doi: 10.1162/1064546053278900. [DOI] [PubMed] [Google Scholar]
  • 31.Hodgkin A. L. The local electric changes associated with repetitive action in a non-medullated axon. The Journal of Physiology. 1948;107(2):165–181. doi: 10.1113/jphysiol.1948.sp004260. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Abbott L. F. Lapicque's introduction of the integrate-and-fire model neuron (1907) Brain Research Bulletin. 1999;50(5-6):303–304. doi: 10.1016/s0361-9230(99)00161-6. [DOI] [PubMed] [Google Scholar]
  • 33.Izhikevich E. M. Simple model of spiking neurons. IEEE Transactions on Neural Networks. 2003;14(6):1569–1572. doi: 10.1109/TNN.2003.820440. [DOI] [PubMed] [Google Scholar]
  • 34.FitzHugh R. Impulses and physiological states in theoretical models of nerve membrane. Biophysical Journal. 1961;1(6):445–466. doi: 10.1016/s0006-3495(61)86902-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Rose R. M., Hindmarsh J. L. The assembly of ionic currents in a thalamic neuron. I. The three-dimensional model. Proceedings of the Royal Society B: Biological Sciences. 1989;237(1288):267–288. doi: 10.1098/rspb.1989.0049. [DOI] [PubMed] [Google Scholar]
  • 36.Rowcliffe P., Feng J., Buxton H. Clustering within integrate-and-fire neu rons for image segmentation. In: Dorronsoro J., editor. Artificial Neural Networks—ICANN 2002. Vol. 2415. Berlin, Germany: Springer; 2002. pp. 69–74. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
  • 37.Meftah B., Lezoray O., Benyettou A. Segmentation and edge detection based on spiking neural network model. Neural Processing Letters. 2010;32(2):131–146. doi: 10.1007/s11063-010-9149-6. [DOI] [Google Scholar]
  • 38.Wu Q. X., McGinnity T. M., Maguire L. P., Belatreche A., Glackin B. Processing visual stimuli using hierarchical spiking neural networks. Neurocomputing. 2008;71(10–12):2055–2068. doi: 10.1016/j.neucom.2007.10.020. [DOI] [Google Scholar]
  • 39.Belatreche A., Maguire L. P., McGinnity T. M. Pattern recognition with spiking neural networks and dynamic synapse. Proceedings of the International FLINS Conference on Applied computational intelligence; 2004; pp. 205–210. [Google Scholar]
  • 40.Wade J. J., McDaid L. J., Santos J. A., Sayers H. M. SWAT: a spiking neural network training algorithm for classification problems. IEEE Transactions on Neural Networks. 2010;21(11):1817–1830. doi: 10.1109/tnn.2010.2074212. [DOI] [PubMed] [Google Scholar]
  • 41.Yu Q., Tan K., Tang H. Pattern recognition computation in a spiking neural network with temporal encoding and learning. Proceedings of the International Joint Conference on Neural Networks (IJCNN '12); June 2012; Brisbane, Australia. IEEE; pp. 1–7. [DOI] [Google Scholar]
  • 42.Abdull Hamed H. N., Kasabov N., Michlovský Z., Shamsuddin S. M. Neural Information Processing: 16th International Conference, ICONIP 2009, Bangkok, Thailand, December 1–5, 2009, Proceedings, Part II. Vol. 5864. Berlin, Germany: Springer; 2009. String pattern recognition using evolving spiking neural networks and quantum inspired particle swarm optimization; pp. 611–619. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
  • 43.Kasabov N. Evolving spiking neural networks for spatio-and spectro-temporal pattern recognition. Proceedings of the 6th IEEE International Conference Intelligent Systems (IS '12); September 2012; pp. 27–32. [DOI] [Google Scholar]
  • 44.Bohte S. M., Kok J. N., la Poutré H. Error-backpropagation in temporally encoded networks of spiking neurons. Neurocomputing. 2002;48(1–4):17–37. doi: 10.1016/s0925-2312(01)00658-0. [DOI] [Google Scholar]
  • 45.Schrauwen B., van Campenhout J. Extending SpikeProp. Proceedings of the IEEE International Joint Conference on Neural Networks; July 2004; pp. 471–475. [Google Scholar]
  • 46.Xin J., Embrechts M. J. Supervised learning with spiking neural networks. Proceedings of the International Joint Conference on Neural Networks (IJCNN '01); July 2001; pp. 1772–1777. [Google Scholar]
  • 47.Rao R. P. N. Advances in Neural Information Processing Systems. Vol. 17. MIT Press; 2005. Hierarchical bayesian infere nce in networks of spiking neurons. [Google Scholar]
  • 48.Deneve S. Bayesian spiking neurons. I: inference. Neural Computation. 2008;20(1):91–117. doi: 10.1162/neco.2008.20.1.91. [DOI] [PubMed] [Google Scholar]
  • 49.Kasabov N. To spike or not to spike: a probabilistic spiking neuron model. Neural Networks. 2010;23(1):16–19. doi: 10.1016/j.neunet.2009.08.010. [DOI] [PubMed] [Google Scholar]
  • 50.Barber D. Learning in spiking neural assemblies. Proceedings of the 16th Annual Neural Information Processing Systems Conference (NIPS '02); December 2002; [Google Scholar]
  • 51.Chechik G. Spike-timing-dependent plasticity and relevant mutual information maximization. Neural Computation. 2003;15(7):1481–1510. doi: 10.1162/089976603321891774. [DOI] [PubMed] [Google Scholar]
  • 52.Pfister J. P., Toyoizumi T., Barber D., Gerstner W. Optimal spike-timing-dependent plasticity for precise action potential firing in supervised learning. Neural Computation. 2006;18(6):1318–1348. doi: 10.1162/neco.2006.18.6.1318. [DOI] [PubMed] [Google Scholar]
  • 53.Seung H. S. Learning in spiking neural networks by reinforcement of stochastic synaptic transmission. Neuron. 2003;40(6):1063–1073. doi: 10.1016/S0896-6273(03)00761-X. [DOI] [PubMed] [Google Scholar]
  • 54.Izhikevich E. M. Solving the distal reward problem through linkage of STDP and dopamine signaling. Cerebral Cortex. 2007;17(10):2443–2452. doi: 10.1093/cercor/bhl152. [DOI] [PubMed] [Google Scholar]
  • 55.Legenstein R., Pecevski D., Maass W. A learning theory for reward-modulated spike-timing-dependent plasticity with application to biofeedback. PLoS Computational Biology. 2008;4(10) doi: 10.1371/journal.pcbi.1000180.e1000180 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Schliebs S., Defoin-Platel M., Worner S., Kasabov N. Integrated feature and parameter optimization for an evolving spiking neural network: exploring heterogeneous probabilistic models. Neural Networks. 2009;22(56):623–632. doi: 10.1016/j.neunet.2009.06.038. Advances in Neural Networks Research: International Joint Conference on Neural Networks (IJCNN'09) [DOI] [PubMed] [Google Scholar]
  • 57.Cachón A., Vázquez R. A. Tuning the parameters of an integrate and fire neuron via a genetic algorithm for solving pattern recognition problems. Neurocomputing. 2015;148:187–197. doi: 10.1016/j.neucom.2012.11.059. [DOI] [Google Scholar]
  • 58.Vazquez R. A., Cachón A. Integrate and Fire neurons and their application in pattern recognition. Proceedings of the 7th International Conference on Electrical Engineering, Computing Science and Automatic Control (CCE '10); September 2010; pp. 424–428. [DOI] [Google Scholar]
  • 59.Vazquez R. Izhikevich neuron model and its application in pattern recognition. Australian Journal of Intelligent Information Processing Systems. 2010;11(1):35–40. [Google Scholar]
  • 60.Vázquez R. A. Advances in Artificial Intelligence—IBERAMIA 2010: Proceedings of the 12th Ibero-American Conference on AI, Bahía Blanca, Argentina, November 1–5, 2010. Vol. 6433. Berlin, Germany: Springer; 2010. Pattern recognition using spiking neurons and firing rates; pp. 423–432. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
  • 61.Vazquez R. A. Training spiking neural models using cuckoo search algorithm. Proceedings of the IEEE Congress of Evolutionary Computation (CEC '11); June 2011; pp. 679–686. [DOI] [Google Scholar]
  • 62.Vázquez R. A., Garro B. A. Advances in Swarm Intelligence: 1st International Conference, ICSI 2010, Beijing, China, June 12–15, 2010, Proceedings. Part I. Vol. 6728. Berlin, Germany: Springer; 2011. Training spiking neurons by means of particle swarm optimization; pp. 242–249. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
  • 63.Gerstner W., Kistler W. M. Spiking Neuron Models: Single Neurons, Populations, Plasticity. Cambridge, UK: Cambridge University Press; 2002. [DOI] [Google Scholar]
  • 64.Izhikevich E. M. Which model to use for cortical spiking neurons? IEEE Transactions on Neural Networks. 2004;15(5):1063–1070. doi: 10.1109/TNN.2004.832719. [DOI] [PubMed] [Google Scholar]
  • 65.Izhikevich E. M. Dynamical Systems in Neuroscience: the Geometry of Excitability and Bursting. MIT Press; 2007. (Computational Neuroscience). [Google Scholar]
  • 66.Murphy P. M., Aha D. W. Irvine, Calif, USA: Department of Information and Computer Science, University of California; 1994. UCI repository of machine learning databases. [Google Scholar]
  • 67.de Los Monteros R. A. V. E., Azuela J. H. S. A new associative model with dynamical synapses. Neural Processing Letters. 2008;28(3):189–207. doi: 10.1007/s11063-008-9089-6. [DOI] [Google Scholar]
  • 68.Otsu N. A threshold selection method from gray-level histograms. IEEE Transactions on Systems, Man and Cybernetics. 1979;9(1):62–66. doi: 10.1109/tsmc.1979.4310076. [DOI] [Google Scholar]
  • 69.Jain R., Schunck R. K. B. G. Machine Vision. McGraw-Hill; 1995. [Google Scholar]
  • 70.Hu M.-K. Visual pattern recognition by moment invariants. IEEE Transactions on Information Theory. 1962;8(2):179–187. doi: 10.1109/tit.1962.1057692. [DOI] [Google Scholar]
  • 71.Lourakis M. I. A. A Brief Description of the Levenberg-Marquardt Algorithm Implemented by Levmar. Crete, GREECE: Foundation for Research and Technology-Hellas, Vassilika Vouton; 2005. [Google Scholar]
  • 72.Bermak A., Martinez D. A compact 3D VLSI classifier using bagging threshold network ensembles. IEEE Transactions on Neural Networks. 2003;14(5):1097–1109. doi: 10.1109/tnn.2003.816362. [DOI] [PubMed] [Google Scholar]
  • 73.Garro B. A., Vazquez R. A., Rodríguez K. Classification of DNA microarrays using artificial bee colony (ABC) algorithm. In: Tan Y., Shi Y., Coello C., editors. Advances in Swarm Intelligence. Vol. 8794. Springer; 2014. pp. 207–214. (Lecture Notes in Computer Science). [DOI] [Google Scholar]
  • 74.Golub T. R., Slonim D. K., Tamayo P., et al. Molecular classification of cancer: class discovery and class prediction by gene expression monitoring. Science. 1999;286(5439):531–527. doi: 10.1126/science.286.5439.531. [DOI] [PubMed] [Google Scholar]

Articles from Computational Intelligence and Neuroscience are provided here courtesy of Wiley

RESOURCES