 FACTOID # 19: Cheap sloppy joes: Looking for reduced-price lunches for schoolchildren? Head for Oklahoma!

 Home Encyclopedia Statistics States A-Z Flags Maps FAQ About

 WHAT'S NEW

SEARCH ALL

FACTS & STATISTICS    Advanced view

Search encyclopedia, statistics and forums:

(* = Graphable)

Encyclopedia > Neural nets

A neural network is an interconnected group of neurons. The prime examples are biological neural networks, especially the human brain. In modern usage the term most often refers to artificial neural networks (ANN), or neural nets for short, and this is the sense that is used in the rest of this article.

An artificial neural network is a mathematical or computational model for information processing based on a connectionist approach to computation. There is no precise agreed definition amongst researchers as to what a neural network is, but most would agree that it involves a network of relatively simple processing elements, where the global behaviour is determined by the connections between the processing elements and element parameters. The original inspiration for the technique was from examination of bioelectrical networks in the brain formed by neurons and their synapses. In a neural network model, simple nodes (or "neurons", or "units") are connected together to form a network of nodes — hence the term "neural network".

A neural network is an interconnected groups of nodes, akin to the vast network of neurons in the human brain.

Artificial neural networks are quite different from the brain in terms of structure. Like the brain, however, a neural net is a massively parallel collection of small and simple processing units where the interconnections form a large part of the network's intelligence; however, in terms of scale, a brain is massively larger than a neural network, and the units used in a neural network are typically far simpler than neurons. Nevertheless, certain functions that seem exclusive to the brain such as learning, have been replicated on a simpler scale, with neural networks.

### Models

A typical feedforward neural network is a set of nodes. Some of these are designated input nodes, some output nodes, and in-between are hidden nodes. Each connection between neurons has a numerical weight. When the network is in operation, a value will be applied to each input node -- the values being fed in by a human operator, from environmental sensors, or from some external program. Each node then passes its given value to the connections leading out from it, and on each connection the value is multiplied by the weight associated with that connection. Each node in the next layer then receives a value which is the sum of the values produced by the connections leading into it, and in each node a simple computation is performed on the value -- a sigmoid function is typical. This process is then repeated, with the results being passed through subsequent layers of nodes until the output nodes are reached. Early models (circa 1970) had a fixed number of layers. More recently, genetic algorithms are used to evolve the neural structure (see neuroevolution).

### Calculations

The sigmoid curve is often used as a transfer function because it introduces non-linearity into the network's calculations by "squashing" the neuron's activation level into the range [0,1]. The sigmoid function has the additional benefit of having an extremely simple derivative function, as required for back-propagating errors through a feed-forward neural network. Other functions with similar features can be used, most commonly tanh which squashes activations into the range of [-1,1] instead, or occasionally a piece-wise linear function that simply clips the activation rather than squashing it.

If no non-linearity is introduced by squashing or clipping, the network loses much of its computational power, becoming a simple matrix multiplication operation from linear algebra.

Alternative calculation models in neural networks include models with loops, where some kind of time delay process must be used, and "winner takes all" models, where the neuron with the highest value from the calculation fires and takes a value 1, and all other neurons take the value 0.

Typically the weights in a neural network are initially set to small random values. This represents the network knowing nothing; its output is essentially a random function of its input. As the training process proceeds, the connection weights are gradually modified according to computational rules specific to the learning algorithm being used. Ideally the weights eventually converge to values allowing them to perform a useful computation. Thus it can be said that the neural network commences knowing nothing and moves on to gain some real knowledge, though the knowledge is sub-symbolic.

### Usefulness

Neural networks are particularly useful for dealing with bounded real-valued data, where a real-valued output is desired; in this way neural networks will perform classification by degrees, and are capable of expressing values equivalent to "not sure". If the neural network is trained using the cross-entropy error function (see Bishop's book) and if the neural network output is sigmoidal, then the outputs will be estimates of the true posterior probability of a class.

## Real life applications

In real life applications, neural networks perform particularly well on the following common tasks:

Other kinds of neural networks, in particular continuous-time recurrent neural networks (CTRNN), are used in conjunction with genetic algorithms (GAs) to produce robot controllers. The genome is then constituted of the networks parameters and the fitness of a network is the adequacy of the behaviour exhibited by the controlled robot (or often by a simulation of this behaviour).

## Types of neural networks

### Single-layer perceptron

The earliest kind of neural network is a single-layer perceptron network, which consists of a single layer of output nodes; the inputs are fed directly to the outputs via a series of weights. In this way it can be considered the simplest kind of feed-forward network. The sum of the products of the weights and the inputs is calculated in each node, and if the value is above some threshold (typically 0) the neuron fires and takes the activated value (typically 1); otherwise it takes the deactivated value (typically -1). Neurons with this kind of activation function are also called McCulloch-Pitts neurons or threshold neurons. In the literature the term perceptron often refers to networks consisting of just one of these units. They were described by Warren McCulloch and Walter Pitts in the 1940s.

A perceptron can be created using any values for the activated and deactivated states as long as the threshold value lies between the two. Most perceptrons have outputs of 1 or -1 with a threshold of 0 and there is some evidence that such networks can be trained more quickly than networks created from nodes with different activation and deactivation values.

Perceptrons can be trained by a simple learning algorithm that is usually called the delta rule. It calculates the errors between calculated output and sample output data, and uses this to create an adjustment to the weights, thus implementing a form of gradient descent.

Single-unit perceptrons are only capable of learning linearly separable patterns; in 1969 in a famous monograph entitled Perceptrons by Marvin Minsky and Seymour Papert showed that it was impossible for a single-layer perceptron network to learn an XOR function. They conjectured (incorrectly) that a similar result would hold for a multi-layer perceptron network. Although a single threshold unit is quite limited in its computational power, it has been shown that networks of parallel threshold units can approximate any continuous function from a compact interval of the real numbers into the interval [-1,1]. This very recent result can be found in [Auer, Burgsteiner, Maass: The p-delta learning rule for parallel perceptrons, 2001 (state Jan 2003: submitted for publication)].

A single-layer neural network can compute a continuous output instead of a step function. A common choice is the so-called logistic function:

With this choice, the single-layer network is identical to the logistic regression model, widely used in statistical modeling.

### Multi-layer perceptron

A two-layer neural network capable of calculating XOR. The numbers within the neurons represent each neuron's explicit threshhold (which can be factored out so that all neurons have the same threshold, usually 1). The numbers that annotate arrows represent the weight of the inputs. This net assumes that if the treshhold is not reached, zero (not -1) is output. Note that the bottom layer of inputs is not always considered a real neural network layer

This class of networks consists of multiple layers of computational units, usually interconnected in a feed-forward way. Each neuron in one layer has directed connections to the neurons of the subsequent layer. In many applications the units of these networks apply a sigmoid function as an activation function.

The universal approximation theorem for neural networks states that every continuous function that maps intervals of real numbers to some output interval of real numbers can be approximated arbitrarily closely by a multi-layer perceptron with just one hidden layer. This result holds only for restricted classes of activation functions, e.g. for the sigmoidal functions.

Multi-layer networks use a variety of learning techniques, the most popular being back-propagation. Here the output values are compared with the correct answer to compute the value of some predefined error-function. By various techniques the error is then fed back through the network. Using this information, the algorithm adjusts the weights of each connection in order to reduce the value of the error function by some small amount. After repeating this process for a sufficiently large number of training cycles the network will usually converge to some state where the error of the calculations is small. In this case one says that the network has learned a certain target function. To adjust weights properly one applies a general method for non-linear optimization task that is called gradient descent. For this, the derivation of the error function with respect to the network weights is calculated and the weights are then changed such that the error decreases (thus going downhill on the surface of the error function). For this reason back-propagation can only be applied on networks with differentiable activation functions.

In general the problem of reaching a network that performs well, even on samples that were not used as training samples, is a quite subtle issue that requires additional techniques. This is especially important for cases where only very limited numbers of training samples are available. The danger is that the network overfits the training data and fails to capture the true statistical process generating the data. Computational learning theory is concerned with training classifiers on a limited amount of data. In the context of neural networks a simple heuristic, called early stopping, often ensures that the network will generalize well to examples not in the training set.

Other typical problems of the back-propagation algorithm are the speed of convergence and the possibility to end up in a local minimum of the error function. Today there are practical solutions that make backpropagation in multi-layer perceptrons the solution of choice for many machine learning tasks.

### Recurrent network

Recurrent networks (RNs) are models with bi-directional data flow. While a feed-forward network propagates data linearly from input to output, RNs also propagate data from later processing stages to earlier stages.

A simple recurrent network (SRN) is a variation on the multi-layer perceptron, sometimes called an "Elman network" due to its invention by Jeff Elman. A three-layer network is used, with the addition of a set of "context units" in the input layer. There are connections from the middle (hidden) layer to these context units fixed with a weight of one. At each time step, the input is propagated in a standard feed-forward fashion, and then a learning rule (usually back-propagation) is applied. The fixed back connections result in the context units always maintaining a copy of the previous values of the hidden units (since they propagate over the connections before the learning rule is applied). Thus the network can maintain a sort of state, allowing it to perform such tasks as sequence-prediction that are beyond the power of a standard multi-layer perceptron.

In a fully recurrent network, every neuron receives inputs from every other neuron in the network. These networks are not arranged in layers. Usually only a subset of the neurons receive external inputs in addition to the inputs from all the other neurons, and another disjunct subset of neurons report their output externally as well as sending it to all the neurons. These distinctive inputs and outputs perform the function of the input and output layers of a feed-forword or simple recurrent network, and also join all the other neurons in the recurrent processing.

### Hopfield network

The Hopfield network is a recurrent neural network in which all connections are symmetric. Invented by John Hopfield in 1982, this network guarantees that its dynamics will converge. If the connections are trained using Hebbian learning then the Hopfield network can perform robust content-addressable memory, robust to connection alteration.

### Boltzmann machine

The Boltzmann machine can be thought of as a noisy Hopfield network. Invented by Geoff Hinton and Terry Sejnowski in 1985, the Boltzmann machine is important because it is one of the first neural networks to demonstrate learning of latent variables (hidden units). Boltzmann machine learning was slow to simulate, but the contrastive divergence algorithm of Geoff Hinton (circa 2000) allows models including Boltzmann machines and product of experts to be trained much faster.

### Committee of machines

A committee of machines (CoM) is a collection of different neural networks that together "vote" on a given example. This generally gives a much better result compared to other neural network models. In fact in many cases, starting with the same architecture and training but different initial random weights gives vastly different networks. A CoM tends to stabilize the result.

The CoM is similar to the general machine learning bagging method, except that the necessary variety of machines in the committee is obtained by training from different random starting weights rather than training on different randomly selected subsets of the training data.

### Instantaneously trained networks

Instantaneously trained neural networks (ITNNs) are also called "Kak networks" after their inventor Subhash Kak. They were inspired by the phenomenon of short-term learning that seems to occur instantaneously. In these networks the weights of the hidden and the output layers are mapped directly from the training vector data. Ordinarily, they work on binary data but versions for continuous data that require small additional processing are also available.

### Spiking neural networks

Spiking (or pulsed) neural networks (SNNs) are models which explicitly take into account the timing of inputs. The network input and output are usually represented as series of spikes (delta function or more complex shapes). SNNs have an advantage of being able to continuously process information. They are often implemented as recurrent networks.

Networks of spiking neurons -- and the temporal correlations of neural assemblies in such networks -- have been used to model figure/ground separation and region linking in the visual system (see e.g. Reitboeck et.al.in Haken and Stadler: Synergetics of the Brain. Berlin, 1989).

## Relation to optimization techniques

Analysis of many neural network techniques reveals their close relationship to mathematical optimization techniques. For instance, multi-layer perceptron back-propagation can be substituted with more general global optimization techniques. The objective in training an ANN is, given some set of pairs of data and output, { (d0, o0) , (d1,o1), ... } to minimize some error function ||E||2, where E(xi) = F(w,xi) - oi. Here F is the neural network function which given a vector of weights w and an input vector produces an output vector for the network. Thus as well as using back-propagation to train the network, it is also possible to use global optimization techniques to produce a weight vector w.

For very large data sets, using more advanced optimization techniques is often slower than using gradient descent, if the weights of the network are updated by gradient descent after each training example. This is because one sweep of gradient descent through the training set can make a large amount of progress, while the same amount of computational effort can only compute a true gradient at one setting of the parameter vector.

## Related topics Results from FactBites:

 LTM - Why Use Neural Nets (537 words) Neural nets, because they are pattern recognition tools, differ from other quantitative modeling tools such as deterministic models and statistical models. Neural nets begin training by assigning randow weights between nodes in the input layer and nodes in the hidden layer and between nodes in the hidden layer and the nodes of the output layer. For example, we have tested whether a neural net model that is built to simulate land use change in the Detroit area can be used to predict historical changes in the Twin Cities metropolitan area, and visa versa.
 Neural network - Wikipedia, the free encyclopedia (2911 words) Neural network is sometimes used to refer to a branch of computational science that uses neural networks as models to either simulate or analyze complex phenomena and/or study the principles of operation of neural networks analytically. Neural networks, as used in artificial intelligence, have traditionally been viewed as simplified models of neural processing in the brain, even though the relation between this model and brain biological architecture is very much debated. In some of these systems neural networks, or parts of neural networks (such as artificial neurons) are used as components in larger systems that combine both adaptive and non-adaptive elements.
More results at FactBites »

Share your thoughts, questions and commentary here
Press Releases | Feeds | Contact